var/home/core/zuul-output/0000755000175000017500000000000015136151671014533 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015136176350015500 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000413773115136176256020302 0ustar corecorexikubelet.log_o[;r)Br'o b-n(!9t%Cs7}g/غIs,r.k9GfD p ?K"mv?_eGbuuțx{w7ݭ7֫g^|1Fr_?c^*߶E٬:rv筼ح_y~̎+\/_p/Bj^ֻ]Eo^O/(_/V?,<']_kmN:`S.ww}(o./WY<͉#5O H 'wo6C9yg|O~ €'} S[q?,!yq%a:y<\tunL h%$Ǥ].v y[W_` \r/Ɛ%aޗ' B.-^ mQYd'xP2ewEڊL|^ͣrZg7n͐AG%ʷr<>; 2W>h?y|(G>ClsXT(VIx$(J:&~CQpkۗgVKx*lJ3o|s`<՛=JPBUGߩnX#;4ٻO2{Fݫr~AreFj?wQC9yO|$UvވkZoIfzC|]|[>ӸUKҳt17ä$ ֈm maUNvSJ d3N^Z)~5ɑUIU"$`SoFKa"j[Hp'yfȼ-vE,4IZkL!~k0ߐNPJ|U }]=UD m}}O-5UOnOA~HXwh@'ڷMVwdOoe:[t.ٗ*AqΝ6 lv?Qw):yt٘8c'8Ai؋%\78:ZBBk`E\Ƹ#¿Øp*vxyPLSMY 9J}t/A`*t) O5]/* @.yhi-cS4 6"KaFٗt<>vRڡc0SAA\c}or|MKrO] g"tta[I!;c%6$V<[+*J:AI \:-rR b B"~?4 W4B3lLRD|@Kfځ9g ? j럚Sř>]uw`C}-{C):fUr6v`mSΟ1c/n߭!'Y|7#RI)X)yCBoX^P\Ja 79clw/H tBFKskޒ1,%$BվCh,xɦS7PKi0>,A==lM9Ɍm4ެ˧jOC d-saܺCY "D^&M){ߘ>:i V4nQi1h$Zb)ŠȃAݢCj|<~cQ7Q!q/pCTSqQyN,QEFKBmw&X(q8e&щu##Ct9Btka7v Ө⸇N~AE6xd~?D ^`wC4na~Uc)(l fJw>]cNdusmUSTYh>Eeք DKiPo`3 aezH5^n(}+~hX(d#iI@YUXPKL:3LVY~,nbW;W8QufiŒSq3<uqMQhiae̱F+,~Mn3 09WAu@>4Cr+N\9fǶy{0$Swwu,4iL%8nFВFL2#h5+C:D6A@5D!p=T,ښVcX㯡`2\fIԖ{[R:+I:6&&{Ldrǒ*!;[tʡP=_RFZx[|mi ǿ/&GioWiO[BdG.*)Ym<`-RAJLڈ}D1ykd7"/6sF%%´ƭ*( :xB_2YKoSrm_7dPΣ|ͣn/𚃚p9w#z A7yTJ$KOL-aP+;;%+_6'Sr|@2nQ{aK|bjܒ^o(מO80$QxBcXE ء\G=~j{Mܚ: hLT!uP_T{G7C]Ch',ެJG~Jc{xt zܳ'鮱iX%x/QOݸ}S^vv^2M!.xR0I(P 'fΑQ)ۢWP Pe>F=>l |fͨ3|'_iMcĚIdo阊;md^6%rd9#_v2:Y`&US tDkQ;>" ء:9_))wF|;~(XA PLjy*#etĨB$"xㄡʪMc~)j 1駭~բ>XiN .U轋RQ'Vt3,F3,#Y3,kJ3,LhVnKauomˠ_>2h-/ ђ(9Uq EmFjq1jX]DןR24d ;嶑, }t&&\5u17\I@ 5O? ʴ(aPqPϟ'Xa>EE衢^}p/:F?}bi0>Oh%\x(bdF"F 'u Qx`j#(g6zƯRo(lџŤnE7^k(|(4s\9#.\r= (mO(f=rWmd'rDZ~;o\mkmB`s ~7!GdјCyEߖs|n|zu0VhI|/{}BC6q>HĜ]Xgy G[Ŷ.|37xo=N4wjDH>:&EOΆ<䧊1v@b&툒f!yO){~%gq~.LK78F#E01g.u7^Ew_lv۠M0}qk:Lx%` urJp)>I(>z`{|puB"8#YkrZ .`h(eek[?̱ՒOOc&!dVzMEHH*V"MC Qؽ1Omsz/v0vȌJBIG,CNˆ-L{L #cNqgVR2r뭲⭊ڰ08uirP qNUӛ<|߈$m뫷dùB Z^-_dsz=F8jH˽&DUh+9k̈́W^̤F˖.kL5̻wS"!5<@&] WE\wMc%={_bD&k 5:lb69OBCC*Fn) u{Hk|v;tCl2m s]-$zQpɡr~]Si!ڣZmʢ鉗phw j8\c4>0` R?da,ȍ/ءfQ 2ؐfc}l 2窾ۉ1k;A@z>T+DE 6Хm<쉶K`'#NC5CL]5ݶI5XK.N)Q!>zt?zpPC ¶.vBTcm"Bsp rjﺧK]0/k<'dzM2dk–flE]_vE P / څZg`9r| 5W;`.4&XkĴp 6l0Cз5O[{B-bC\/`m(9A< f`mPіpNЦXn6g5m 7aTcTA,} q:|CBp_uFȆx6ڮܷnZ8dsMS^HэUlq 8\C[n膗:68DkM\7"Ǻzfbx]ۮC=1ÓOv$sY6eX%]Y{⦁# &SlM'iMJ았 t% ~@1c@K?k^rEXws zz.8`hiPܮbC7~n b?`CtjT6l>X+,Qb5ȳp`FMeXÅ0+!86{V5y8 M`_Uw ȗkU]a[.D}"\I5/1o٩|U戻,6t錳"EFk:ZM/!ݛ@pRu Iヵvyne 0=HH3n@.>C@{GP 9::3(6e™nvOσ =?6ͪ)Bppًu_w/m/0}T>CUX\!xl=ZVM\aٟ6h㗶E۶{O#X26.Fٱq1M k'JE%"2.*""]8yܑ4> >X1 smD) ̙TީXfnOFg㧤[Lo)[fLPBRB+x7{{? ףro_nն-2n6 Ym^]IL'M+;U t>x]U5g B(, qA9r;$IN&CM(F+ hGI~Q<웰[, qnriY]3_P${,<\V}7T g6Zapto}PhS/b&X0$Ba{a`W%ATevoYFF"4En.O8ϵq\FOXƀf qbTLhlw?8p@{]oOtsϑ`94t1!F PI;i`ޮMLX7sTGP7^s08p15w q o(uLYQB_dWoc0a#K1P,8]P)\wEZ(VҠQBT^e^0F;)CtT+{`Bh"% !.bBQPnT4ƈRa[F=3}+BVE~8R{3,>0|:,5j358W]>!Q1"6oT[ҟ^T;725Xa+wqlR)<#!9!籈K*:!@NI^S"H=ofLx _lp ꖚӜ3C 4dM @x>ۙZh _uoֺip&1ڙʪ4\RF_04H8@>fXmpLJ5jRS}_D U4x[c) ,`̔Dvckk5Ťã0le۞]o~oW(91ݧ$uxp/Cq6Un9%ZxðvGL qG $ X:w06 E=oWlzN7st˪C:?*|kިfc]| &ب^[%F%LI<0(씖;4A\`TQ.b0NH;ݹ/n -3!: _Jq#Bh^4p|-G7|ڸ=Bx)kre_f |Nm8p5H!jR@Aiߒ߈ۥLFTk"5l9O'ϓl5x|_®&&n]#r̥jOڧK)lsXg\{Md-% >~Ӈ/( [ycy`ðSmn_O;3=Av3LA׊onxlM?~n Θ5 ӂxzPMcVQ@ӤomY42nrQ\'"P؝J7g+#!k{paqTԫ?o?VU}aK q;T0zqaj0"2p؋9~bޏt>$AZLk;3qUlWU Ry==qޕ6ql?N/e1N2i ۓ,j|z6OSu;BKŨʐPqO K\{jDiy@}b|Z79ߜih(+PKO;!o\戔-QB EM;oH$$]?4~YrXY%Ο@oHwlXiW\ΡbN}l4VX|"0]! YcVi)@kF;'ta%*xU㔸,A|@WJfVP6`ڼ3qY.[U BTR0u$$hG$0NpF]\ݗe$?# #:001w<{{B\rhGg JGIެE.:zYrY{*2lVǻXEB6;5NE#eb3aīNLd&@yz\?))H;h\ߍ5S&(w9Z,K44|<#EkqTkOtW]﮶f=.*LD6%#-tңx%>MZ'0-bB$ !)6@I<#`L8턻r\Kuz*]}%b<$$^LJ<\HGbIqܢcZW {jfѐ6 QڣPt[:GfCN ILhbB.*IH7xʹǙMVA*J'W)@9 Ѷ6jىY* 85{pMX+]o$h{KrҎl 5sÁbNW\: "HK<bdYL_Dd)VpA@A i"j<鮗 qwc&dXV0e[g#B4x╙✑3'-i{SEȢbK6}{Ⱥi!ma0o xI0&" 9cT)0ߢ5ڦ==!LgdJΆmΉO]T"DĊKٙ@qP,i Nl:6'5R.j,&tK*iOFsk6[E__0pw=͠qj@o5iX0v\fk= ;H J/,t%Rwó^;n1z"8 P޿[V!ye]VZRԾ|“qNpѓVZD2"VN-m2do9 'H*IM}J ZaG%qn*WE^k1v3ڣjm7>ƽl' ,Τ9)%@ wl42iG.y3bBA{pR A ?IEY ?|-nz#}~f ‰dŷ=ɀ,m7VyIwGHέ 2tޞߛM{FL\#a s.3\}*=#uL#]  GE|FKi3&,ۓxmF͉lG$mN$!;ߑl5O$}D~5| 01 S?tq6cl]M[I5'ոfiҞ:Z YՑ"jyKWk^dd@U_a4/vvV qHMI{+']1m]<$*YP7g# s!8!ߐ>'4k7/KwΦθW'?~>x0_>9Hhs%y{#iUI[Gzďx7OnuKRv'm;/~n-KI`5-'YݦD-!+Y򼤙&m^YAKC˴vҢ]+X`iDf?U7_nMBLϸY&0Ro6Qžl+nݷ" 㬙g|ӱFB@qNx^eCSW3\ZSA !c/!b"'9k I S2=bgj쯏W?=`}H0--VV#YmKW^[?R$+ +cU )?wW@!j-gw2ŝl1!iaI%~`{Tռl>~,?5D K\gd(ZH8@x~5w.4\h(`dc)}1Kqi4~'p!;_V>&M!s}FDͳ֧0O*Vr/tdQu!4YhdqT nXeb|Ivż7>! &ĊL:}3*8&6f5 %>~R݄}WgѨ@OĹCtWai4AY!XH _pw騋[b[%/d>. !Df~;)(Oy )r#.<]]i-*ػ-f24qlT1  jL>1qY|\䛧\|r>Ch}Ϊ=jnk?p ^C8"M#Eޑ-5@f,|Ά(Շ*(XCK*"pXR[كrq IH!6=Ocnи%G"|ڔ^kПy׏<:n:!d#[7>^.hd/}ӾP'k2MؤYy/{!ca /^wT j˚ب|MLE7Ee/I lu//j8MoGqdDt^_Y\-8!ד|$@D.ݮl`p48io^.š{_f>O)J=iwwӑ؇n-i3,1׿5'odۆ3(h>1UW蚍R$W/{&Ά+4*Iqt~L4Ykja?BH 8yݪkIf-8>V#ہll/ؽnA(ȱbAj>C9O n6HNe">0]8@*0)QsUN8t^N+mXU q2EDö0^R) hCt{d}ܜFnԴ.2w⠪R/r| w,?VMqܙ7;qpUۚ5Tnj ۝jlN$q:w$U>tL)NC*<` `)ĉJآS2 z]gQ)Bی:D`W&jDk\7XD&?Y\9ȢG:${1`+i n8=%Ml%İȖb7AޗuV3A7ำqE*\qb'YpuHƩҬV nm=Ɂ-2=|5ʹ zi ' ׹U>8bK0%V\ t!Lku`+]c0h&)IVC)p| QUA:]XL/2La[Xѓ F;/-rtx-rei0hE˝ݸDt#{I} `v;jUvK S x1Q2XU&6k&lE"} Q\E)+u>.,SzbQ!g:l0r5aI`"Ǒm O\B!,ZDbjKM%q%Em(>Hm 2z=Eh^&hBk X%t>g:Y #)#vǷOV't d1 =_SEp+%L1OUaY쎹aZNnDZ6fV{r&ȑ|X!|i*FJT+gj׾,$'qg%HWc\4@'@—>9V*E :lw)e6;KK{s`>3X: P/%d1ؑHͦ4;W\hx锎vgqcU!}xF^jc5?7Ua,X nʬ^Cv'A$ƝKA`d;_/EZ~'*"ȜH*Duƽ˳bKg^raͭ̍*tPu*9bJ_ ;3It+v;3O'CX}k:U{⧘pvzz0V Y3'Dco\:^dnJF7a)AH v_§gbȩ<+S%EasUNfB7™:%GY \LXg3۾4\.?}f kj· dM[CaVۿ$XD'QǛU>UݸoRR?x^TE.1߬VwխmLaF݄",Uy%ífz,/o/Z^]ݖF\\UR7򱺹...m/~q[ /7n!7xB[)9nI [GۿsH\ow!>66}եl?|i [%۾s& Z&el-ɬeb.E)բA l1O,dE>-KjLOgeΏe|Bf".ax)֒t0E)J\8ʁ,Gulʂ+lh)6tqd!eó5d ¢ku|M"kP-&ђ5h ^pN0[|B>+q"/[ڲ&6!%<@fpѻKQ31pxFP>TU?!$VQ`Rc1wM "U8V15> =҆#xɮ}U`۸ہt=|X!~Pu(UeS@%Nb:.SZ1d!~\<}LY aBRJ@ѥuȑz.# 3tl7 ]وb Xnݔ[TN1|ttc‡-5=VrPhE0Ǐ}Wd|\aD;(;Ha.]1-{s1`HbKV$n}Z+sz'ʀ*E%N3o2c06JZW?V g>ed\)g.C]pj|4逜*@ nBID f"!!*7kS4޷V+8弔*A19`RI/Hй qPq3TY'퀜+/Ĥ'cp2\1: 0mtH,.7>\hSؗ΀ѩ آSNEYdEcaLF&"FhQ|![gIK v~,Jc%+8[dI368fp*CDrc3k.2WM:UbX[cO;R`RA]d+w!e rr솜[/V`+@;Τ`5d0ϕ_Lع`C"cK>JG.}Ε00e>& 2䯫vNj31c$ i '2Sn-51Y}rE~b>|Ď6Oj~ebIapul9| 3QtUqSCxTD7U9/nq.JYCtuc nrCtVDƖϧ;INOKx%'t+sFUJq:ǫf!NRT1D(3.8Q;І?O+JL0SU%jfˬ1lމZ|VA/.ȍȱh M-r ~[0AG꠭y*8D*-Rz_z{/S[*"꫒?`a;N6uilLn<Yllmb rY״͆jqTI!j.Pٱh s!:W_´KxA|Hk1nE6=W|$O -{]1Ak$ ѫQ6Plp;3F$RveL l5`:~@c>q,7}VE-Q8W70up˳ A¦g/OEU:غA>?=CۣPqȅlW11/$f*0@б 2Dݘrt +qrx!8 J&[V =͋A,z`S,J|L/vrʑ=}IhM4fG(Ȋ1{TT%41Oa'$ cJ{F_e#u9o<h׼CoIɒce0fD]OUʸb$i,2=%2VIE; jSddsǪ9 +iXu U! RwDQ)vwM9X1ͣqrAX% d|SRmSRbtSRũFƗzZiKck2Z{a eqeiNh8yY\ٺfqlNz, Kfhh/@ u 6msq۰! W }k1nYT5%ϊj檶jvpWߓ/^fez?#:.{ ?y?=Iq_۱U\2 O7_^(tC O |I⦶Rd1>̳mKeq񩩜B &P5R]O1@W\ ط0l:"0',-MEUڗJ ^ Mp%`ݦ>9]UΏl:UݺVkݼVkջ6j VI6_E5jAU#<0MCmplp~ ~bI<y\[՗XoOMqoP+(Q@fA5zj1LTbZfpB Z2**x=]/Y{Q(~d3-4f뺆%L-t54 PpS ôl빚, ]OsQ3?Ǚ$<# ®k {>Lz7GA []qxGi2-o `<"K! _l=Z܉1Xfk st@kEٰQ"<@ "_tELJW?Q9YPbR.: tIš(c-拀e:b|={< Yʛr֬%oU{#$̊Ik1RjM-gKi%u%^{\B~v+lUfa5Q dK(!uU%Do`N[~ryީ 7qVs p3˯WgENc^8?'|wm {3:{wb%?ꊥ]cj"D<Yqi'.78$U\T4pO;4xH(4p"U4(# RXO3 ˙0@ [z "<8\0 M=%%n[ѐQ.;O/*Ob,R#pʹ $+7OQMXp\1&*zbt!vy9&uJY[슓-ZRmxw`E hpِ͋[ļEz]ˌZhP梽ھUwNf>{cU$k_4X?+Qa|0d,*.{l2SP^L/O!<+^P#[ZMw WӅn ˷'+ds F[/uV,Aȅ%Z/Pi[=ehU/"D:.74 N/ɏ!zOJBCqx$|n{ :/|F_ k>8 ?Dmb ]wc`ΫjP7> _/͊wo|@9NY)X&ϧC3)ݮ6:w$gaJo bסN1V*0M\It ä6KIIuKb4aj7oK]'N>0Ke^M[,&_PZPoAq)WP"vm`x~ _U2/X%KvWKYSu[0ќnM,9O y1ܱ$i ^> \߾ZDh9n9#kSv4((-tJF:eSp ~ !U.'lF,o"l^"ugۓ^Oq:z|NQەGqv+R% n$ ]MF7UեG(꺄bǝ"+IVcPlg4!BW">>E*npSw<+Xʳdߺ8 ˡ scG *d>ƕinQ>;muisgdq|Ծ {-`HU݇ToQ>eeٮ*>J2}Ө]e4Zk呩2jM58' gQY3eŁ~ h'ޟ}Y鰎eTf )W NK]0F{6/F4#ƾH\= < iY @C|?{YQR2gȍ'huMN#ffh]]Z Rk 8x%0ÕZY {k.WG5RUZ-kSn`׶N,)ںj^aUɆՇ9^(e梏uͽc#ϚS"?Oxx6J uS>F+h'++%~]=V~_*DcΞ0Ó|ZY5}waYP6=X TK˨}\,kOxog[{L,ӱR0ț92 CcVyb#~x<, Yw5  ʖ=Ot$IVU1 (஄Tcjf.lUw).ō9FvВ$T\mvHג jl1EH]bEͱ?J}~Mheu5qin4m!⼇ILֆ9`žR_Z -ќ&%ԍQG[uCQ->灲 Zsb:\FXO*E%FUO$CNZm6aEI=G=e@ݓ`O**ӵaLt,P*됆Kև!Y5Gn/%#j&s;ĺhJC6]]~“h.QkTe[7{<7,Y=p91Ω ͕uLK{era\Xlžކ Cpmh'r%y5uOʽu+ (he[vֶ"l.߿jc;_[ay2X7@$o .RWX$2a%#F(+.bMmZ+IlLL,9m-1@4*)lj4eVf ͶK +fd곞;bKԗ/>>%.3Cpu:u5jӧN5ݞj6Tk]% E;nѿWO}-a{|>duďEdhhȰlsjLFT̐eGt뜳Jl!]eR`PyG$(3 S!GGjjtkhJwBr4p[Xw`e6 <μ0d3}0S=]7Af88j-d2dD֐mڂ-Ap;8ÖڛpLZO}]I%T"\Q#0dM97V YSSVUm2d =:diZEZTmnC4))+nB:nU<;]v$%n:d=ܓ^[s_OZI4e{6ڈHRxٴ Dy?Y Fzxtm^%+wtݎ_čuA/ZKٮov˴Tt[f& _duCbhɚ,ׁE/rA^!d8㸏%kԯ;au>'otd=|Q;ڳތ㣦|M< @6]h`i]YKɶ #s(w1'l u=mYm`s^C,MPO-* vj$dxOז}03 lH^몚B$z´ [?|ېx):b[uDFӳ$ ϕ Wrh!μ>W^*PUgq6_~ HXm5}F 2r)K˜L"|7ZKwԲL2,$$1gJчr2`q2Nc0lӆ1<F*gX&K3J=Q 0 uTC5 ~L$,gbpיswgQ Nꨖ1K0ҥJ(uo SS3wsL~ #dYG%b} r4uX%8f bf;AMs4cxSP"WJq4,k0ʐbt^ٻt\L|=l$ v&uveF@9;GAꊥ]h]Bb?MG7 ;wtNijeЖCpětMᨄwsv?v3 3ԎN^є! i5z\4i'Iow-mZ`8s;{E A䬷:z'vLP$dD-)= )ʒU`"m&pΑV8[ [L2tU$͗cGQp: 6ieVK?PGH%!`mX| "$hI0DeZR;P@k.wߖO$aEe>.MJUbC)"hV U'L$Tqnlsu_Ne{ 7.obߚ"gY(^"_}Izi^wO((;y8]+QQ; Xկ<~&Fڤ$=3٪'ޙ `CD f&ustņyrtG:0X7JеsKP0dDv} ݸ (+B  NWo\ @ѸѦa^Gkp2ZPG(nvә"8 1]0 ޮ#J 7@X׈*ط(X4Dp׏`1.-GW`z] 9@[e) XG<K7W.:db  `F*be`4-Ag{8wG~ j{nnCTlUG_wP,CU_V~be^N]#)jPLZ,p$Q([^njDu۔vX[)[Q!"xXs۱;~W.͜.l ة5x-_;dZ^1әC!ӡYw~G.uJ9X;em{r)/7PSn;s֬* xC;GP~︶w쎋;wr;wv=gw x1Ý#:W#w8bWsf9rזwpNLݎn8Ӡ&Yr @^\K9ožP˙׹h _84qm@ixK|(#ܵE )v#w -;wm>Aj0ۋ 0D>e_+pg UG eЀA5)3uL҂d^K[̡t\@ \D#׮#E4op4 ^q7Kmٞi٦͈冮Z>Gpq:eפԴ!I`A8~Aa?8a!Q {TҪFwdVoKwݭϺC{٥߸ @r%}P:[L W]HVN:?h\,>. hD. QbP=W̤Fqƺuj dqn -%QdƶKlsDO'lD n 5mnR`iY*A[Ĺy |hɐ }re3׍9;Doϔg8Zp+דaYv%B?<$§ʝ|2d;I] `$\1`UANUŰb`V J[k5?uW E<19M`c2'L~b \80Ht´|?=rb<0}.F09b8r86 JJΩ]ꨵΨT:BF-r҃z*tM`Ȗ5y-UYR [V{G_ŧzmGN>͚ (oYڰ`a`AKmh N l!1jz|ZmM0[z_O, zcs֪'z_N 6>< AFNUEbWs0y|-,^i6+ꠄI7Mq0I7E08C1 QH"==d!u-4UV*y9]˖ya:pvդkBV1k]=g{>tFnЬB}?LFzjn1 m/?mo$!џv6B> 9ߒ{~MWCn\Vd#g7S|dAl0˥b#$82 bg1ezsʬ\&װ}NW ^$|#s;6J~䝸B 3_ ЋHYGR4hPCxCF2u Lζ_*`?UׂUA/>SPҦJ_تJwj S>i_ڶn]Z^AX{wY헉j?SR{AUAuvy3uv v.({왂uWuw]PeAPoUPoA^&LA嫂 _&(|AUAw3w4X4A`wA  <8zs.0 XU_?җg€<4yYyE9UIE L7cˤXNS"aLUtpooZxR2٫rZ/EuBi*5/7<^ͣ>!ozQ{u୚=U-h],7r9f5܏ng(?KUOG;SֈG & !,COmDY17[ ۃ4z`x)DkOÓqq,E$6tCؖhzXͣrD3=H(;:K$Zۃ5f-k͡9x[_%?n 0JYwT+7>K2yz]ZLa<9:?p93VŠk:e~ w {T!pC0BA.;cWf4nm)px9Zi0 .B_ ճ,<졶<nPN:6IwyMuN=˦mzͯ >Sz O?ǡ‰ydv}~q!|NqCY7%'{ NnbOQ>"&Uat3qRe*s4ߏ3_0X>{4Z4N Bp5%!Ե2rd\%JZIٕQ?NT,Nv}ow$8*=ڌVs뙮Qdd=p G ͐`K:x&:Oʲud* %]q њb-XI 1Kw1E_?#&daIr,v8i#MWy o8H[0, mF g=ʄ"mIq@0OB3 l ,$(y./Il$M:[*ps,x!k-a8H8_HNpPcOVBX`+K 6'?>g#mܜ7VMڛ(X^`CM I#S211lNd8x8E&DcU҂ՠs<"i`GBٖ N%1ءZr5'P\I3bb LZlzHעIpUF, 6>y5&֪6ƈaa'ЁtyWr+Msaİ|QYlIt]]P8o!hC=23x6`fSB"#N^5\#`I i\$Y[ɟTG$ڱa. 0pӳ$-2Èks*W kEaY0Í&(]I t&E.gQg7oGdE#c8ô,y9S\!j\K{_ ^up6>݄5aİ8P@g̢20GPjg7k}'Np8 "@಍ꇙ!yY>˞:mV~zɬ]T 0Mwc-9G ?FG*0ޟn迾O6gkE4f^A)2(Wpj+Q)mlZsQRuٲ^}[oS\OɮԬRbl?&A|/>R@-_TScE.#sv ~#ʕ93|sr>G(pf& g ~j~u l$o?.D#M|xZGvGSKԄ;V"Mo[1eGf0X$l١ώS;5DXݢ:f9 1BXcշ}. 6aUI.@;܃KA['f^BhCCk:0i3Ӛe8=f;*6^3{~} 3s!t8^{Dal:%4n,^ qKȸ!鲊(~c[u>K>dRYjU1xZFDHbuU[S?E7C ~,zw,,^сA9n2G16#%ՉxVAT3g,iB#tVM+be?q'`^WueI8&E˹g.F4;7ع,kV2LhPKG<٢D`Ǟco.p&e=IZc S3ORa\lS<Ɋ Mf{)2obd*d `ސ`I n{߮ZBbBZ[κfg{7BҀQ/tj&1Q1U(9Y :{)6T#M k瓣 JMrD1Y*mwnuHIˊQ#$]x2{ Qx?.!*30?4fb~c*aCb4år5,Z+TPY0OPW,ʩ9f_C)_G3ѐ7/jy3tq=;Gu5xpe15r݊ ٤2D!51.e ࡉcN΍cc»\ԱEհßK/.É WޙD`wgTfԓX;$xu"d5Ռ%P gZ`i 5`kCJ\A \y `FY^ }Q o$8 YӐwcѯ$nQг3}1k,_IpTS`Y`rpk7%3Deyv[L_F #Llf?Yt~Q>O04ׇ[^8G*Yن,Jl19+}0n鷗yQL44#b:tz owdDefϹٔcWB2Cz0?Ŀʫ9fVwNo 4xk"is^l@H;p vQGp%{/TP:H1؅gf`JA,eP-z~.:b0۵C,}nQNg4fE-T/$8(N N}򶫎,Y `C=IH o`lוܤH~,n|;:##E^m7$H-w{|I2[IOempFXU2xa}F u4l%]ey]8:p%i矬Sbԝ uTd=_&zE4OaPO\-#"$jMB$&C9vU0Z2vWD~nE¸.=>VaUiΰk7 .R*R1F]M9-`;7hZD7#Kt7}l ĮUQ!m)ebT $W֕ϻzGDR*]ksqp֏A{sLY4U讲aeϳY( 4G>w9H$IݔDD(Mه˺Nllus1_GMD Q]^A7̵֨l45׆ )sy~lI>)>?@qtm>xgcݸQt;_t# .- "~_e]|Y22%*2P}5<@}wMnKC_?˿왢F?xTtOOk Hzo.CԓIU|Zh}9G!Inݿ@ uQ$QD &LY( [O 4F51W2 _MD|WZ{] ۆ};qhfph5 j3f43? (pJ4zKTg;fM33/t=l% ^ BoZ^@оU1ς~Pg72K{ GKLLZ[VgEi5W2jr쁙ܙ$Jؔh1f>x{9V00f5S;kBuQ\2SMI+u9y]]]Gxհo]Il(rk/3gڙr>837K̻*2Z~]c42RӲjFpu/ l"]rY0z 0hK]o+^ݛ!kN}©N+,Z8ٕh^;>'߫3~T_T1GZtӮ {}lFBGF/^De2W,͗N`~T2 •Ds:w@EZ|Y wnKXŮzLS-ׇ9JL[93ޚșf\h!OBa,|BNI>!G+( z.rfŸ8>f^;RfG@i)늒= ǡ {k90*Sdwad(B(>E(mEN/Zs^5 e +و0|U*ȢŠӤɒ.ŪJj)Aca(׋gΆeva<0Oѥ ىO˳-ujcԽU43RD2= kzsmw:LO8ǫ7Pc,PʓHD9U2wJ-LgY&k|#)FКq 6h/=\A]#wÆyVcA?iߖxKtWyTu /As;+뱳Jɖnl_W!%prPA"hvŷs5%O{з\8+7($0FA~΍ ٖ.*ЧaCam}0m9~^1;~ޠµԎ}QUp|T=?ᦠXIХS9q-ZigTxh ۻoUcp o -y~6hC'Ԝ-SLt&6:z OJpvKtr:=''ʂm:<~\KZZfZ 0c\7?lQ6gcWm?2~oBQ^~ȝ'f;$.54Tp7jo^?'tƦ r9IFiUu{,P (U}3sKj号.uX -ML_~lG^/ݣ)v^rJ= WM_N:QxѻPJ7R?nu?cu "R!e 4C7tKy@!/}7⣟oFw×bsy.ы'ў |֍Yz- >^b+R?A#SpMX=zX%XaWY^8wLZa0BZCV,Xm9f .{F 5:}vpAjq 6ۡjWSyG E "1O^f<ϒ^)7(cɘ01)C Sq O$8d+kIbM91|kbTQEQ >M+ z?/jq<{`:q 4CҌeu@`hMHZy☍]+9)ŀq(Ta+kFӊiF{׌%afDDʡsbhJG N(Z)ǔ1 Wĉ5iE4i 4֏޷f,W&%b 8˄i$:I]*si&$nJ4FsFӊY7 k7[3EPFHe}d`>Z@b^}^A nA-k:3ǴrofZ^x6é\5X3+\z,(h֙n@gۛ_}3ݰ =Ěn^{&"DK4;+!>;=ά'.!I˶:J}W=;zpEϡ5 ׏Z/]ƲRn6昕X4RKyHI(ջ7ėa_fFzkIґʧ|ؿMszf ^OB|qUh .SSz}]__#"zf4jެP~qReNf WU_- Jr[a)AL%|/$0S%VV'i,dBPǏ/==&OniLUV(5s.-rA`ETke/.mqBv<_[:'6|(FIkF'IR[SȊkR3I4U$a>mVƠ~n%VOB-D !k1&1MqJ0Kir1%q0q1cXʼn,8ISmi4bFÚb~uMT&RO1Sqi&A*6Zh wլݩ_L/c Lɩ# <ǵ3#RPΉ_%IN¨EsCoIRa\>8^¨1lq4PU] J9A6.)-2RïX-6dܳ4j!Se ̜ SQjwic`Ljr#&=0#$ohܣkWB)MD.VHlD×kxI@0IBQ@YьQ+d8AjRH w i{h)&ڬ5!dA`"F5[cq#JZ 1"mXaF(L[0-RZ~V(οe2|u_Q|y.wQy`В.;t ,9e;ʞ ׀L`jX9oRY 8Bh`<$0'=R< iV/HKA,R/{W俊? b"(j, Em` hOϨsێÚB ˡ16zG/u'$v=kTHH*-UBdz_"QGxf0o]kGE/\~L@>A V'džw] M o+Lu9IjNZ,gԃKe5M`.L;8!Q8gZ;o1`mεr)Bwe^h% ZMcB#*ׅb*5[((qQoiShBZ[sڲ"W<ꠕi!=] Q p`*S%R0`= T :9 ^g` hrLP?B SM$np=ue+vђa쪝 1R\K$ݒ_$\T;ǪZ0h`oYMׂ174P$ v8a;08;kR{6v6kUn{汬,i].V 1NОn\gf=Pzևka= ):0uR-+zplWV`Iy*P{,G)7DˊYoI*D?.$7c%~k3+L8Yˠ؝bh8;$l\I F'JwH{J (pjW{=2g=b0Fs-%*yN[iT.ӥ7UOzSdRňr/-"gl9[@Ct{S^~ ,iBeO>]n/HN]_[YH!.aҎ{VvuԫoGwVmnM VW@%Эk%<$F39AZ`?6ZHw[W>zPnY([20]ՕKցrLowd3w97fwqB땄*w=(W˟71%S?..dWuB+YVW.p^jȅh_-U0?rQfBX^[/ED%j6Z_o% +|w40q?qpM#2i`Ǩ+Q8|FOs{)z袕 O5'eTt1GW~6,N+c`]BWi.XE3;ӿ;' ix3UP@1l?H K9ϏQ~^d䗭/7p*n q^fj2,pKgӛQxDfd%|6Œ@\^N܌3;s8D8K^rP‰Ic2{*iBFS"u6}#b* $u{I*4ҏ%j+&G]yND6p1-g[ y4? one/FC#zڅ97X%{eJ,Q2qxdVYn .&%K|:>@5tz%P6ڔwadv/< =[i*o3ɱ6 ēQZr:^fxN['^s`p1 )sH[ ΉuXU^f Y~ճ &u@)Ԧhj6!s1`hu:cMaMEKsEbr;.LҒlMY*v܊aϦ sl7qNj˪ԙ6b:#>*~|NnAFՌ;,y+g?e8Le"QZd\m#l եS+c9_uĺvܲw{nu,{P<GC%`[mTʤ`|p0,rƿe#GzioL0ZSeEݍ_5jڛIqkuV^rcЌ"E^[l1+l>Zx'A#L8/j!DY˿m??Esv6ꛕ=qaC/aݖKz˖vn =/Prxg쒠t.|UHU%S_ڦOMzz(DWp1C!f3_L 14Xf슞̋+8"x`^ǀrC¬,, {οhgwgM棬pOFWL98A,yd]` z>{_o~O> OaIJX) )]< lW[*o2}~Ųq2sl*A2\Uٮ |8U42$Ϧ%n>͇uu|Gw5>8bxvx3]ƹegojyb'R 4xNÐߧww1rx/b ،Udg|ΌbS۫ mvqPkYV1H^TSChh>CN*QN::ݩNn窱&5pBʬcڤ>,6hVk'WdqHO4E?;3oOeI7XG*RTeeVګ6Z`3AF0680׿+i TrZy4 gt޺\[ʬۊӚߵ̣θQ,l"EF˩U^\8gx\d^{Qb< cA8 g&}ES @[hu,+*c^ELs40yŗTv1ekonOv|~|ҏ*QphmXJX%2t)L"W F*_'pNswo3f<Z_?Ӭ';a[QGoqKv3Ȫ3eysSHݰf j[dW6[[,%~+,=&Y[: Q^F HO >pB : ݛms؆5 %?%(b\p>'Z tEK"欵 ύ6Tf nv~ɏ48ml5nKXhG.pBZ^Ўoq;L%!ч:,fԏ/zf_vo.{M4/~tafuBqdg|~}MVtY@? W+:7%띰:5Ig>ze]Qp\E`d K D.gKvYI࿼[G_1X+iz1ZKK(NDdV$QdVn[cдukz'6;Y;,^RX"$,wgҚ 8F7ísxrm'406gzo,0D)/Z`%h9S,kG68,ŌiޗB gI(Y {hX:)?BpjLXbԁJj+ p}ez:zS; j1ͻL84SX>Y^5H i%#,:VkY頱e9G8,lfKn[1JdQ%Ms"LXw]wq}+;g=R 9|0D %-l6#bPpDeN+3?* (q@*×Hb sJ р}fq}٬kwVtBQ' Er .S( M&5_䍑62P}lՔ.R;σ!%hMԱb0 'w΄0͞fy#F(I.Bi%J֛b)cwZ̸kq}Yh7n hv%^aj16&P=Y.)\~[j\}*Sv%g-XB|/vԞAcb4we>́9 ފb)'3~DօkD^:8FE0AwiR+WG;Sz~N ? )X\:rā䃷)qK:h_DZǀU"G:Pzt*Q瀭!K:g2>㠹Gq/JBr\V-},m<{" M5J[ kZi1`] X':R֜ Ȉ/Қ\J=[bƶ_5 Z, *8+H%ھYZGi#23\x_ B(/ F#ȧ</_.}Px[Gd lī`HWwtزrxO:܍ă㔝>F`nƳ3ʊO&.DgZ9aA%u;IG,ғk>)_&UpK;*U猚k&LJ!ا L5(E})6Mg1 jCWT}+;s`D40ƼCnw(~ï^T  W?hKt(A]wИݽfMɈ&k~j(K0~3g1]7] x_3/!/f**m hNd oOLbe1`0y}10G)"L+eMr=໠.Uw.{E|Zz_gl*wNBR2 O0hC28V /9^?ŹJ4;FkՌzPYp:`Flx_6 {h|]c1_7:A1h1`ƑIGT:u"x!/fnuF jƔoZTK0QDT;YJMI3]K]=tW_&k**a@fef";LWh*8(,7c< 7 NCI͉g˧IL9LՑF{lcZ2Xpx}Z3D)SĘOf:֯h,\9)g1%tsIJg KdÁ$6r14,hޗ#;{F KHa .ZRRpVkf} C-]"h.A'WKxQ3b+a)1.&yUR N"BP>K:t743m{ܲo|yC\eZ&3V3;,[1aF-"pEx_XrOuVf=k<~#(dm.zK>x_r>llf¸^QF‡EFbDmw:ft;qAFC}У?t|cX<ѱNm_C`yI%7x;C<zO&1 H΋âSۈ3F(.6lZZ#]ĸUWБP@X=SRX>ߡuw ,^#X0VwӸVvKDKt Ưk`DӸDZOx#X$4vd֛b)c%'Db5Ec%-$)gIgn:E!B#Cp Dx^@ka Bz#, FX(.amRf< .ވC! gq 3֤ë+urn/n99Uyd9ɽ!7$2 j"]\vt5\qn8%Ηjk4,Vx*{CONUK)FؙϠ)L p>@Pԉabf|*&ub>7K7_ꗊizw b;ҧD|񵥥X6繏v 3geόat (8p76aKIBTTHIBUWmσ!%hMtr鴜aWm5nM͙C+Uz? `Uz70/3R9+ 8]yYLމVFk[ a07CwɮE&w܉],Tb0zXQ3@dwIW\6D чu0E()Bh֣w N_G&!pM Wa3j\K6~(N ;琭gqi1S#$1aŵӻAfWkN]M&i9`E-[-\0G.)"RMS0G޳,9+yߏEm**R$=={N}(?$ٖljQ;]@7<MԢӠ\oHuM&B4YQO梕eϊc]L  }RHzBC($%Fv=Qp F`ɘS]R;w; r+l-6Hj-2""d _Ȗ91G**P]-vېppNbΕjЇ1cY _ 56YF4; Z|]޸t>T*j1:涢'TeLq(ۀ ) )3\$}nUhwz+2nhgEcLoœ;wA]I<j6 u@sWR]ª)gx 9kq5'vsd#b:[ǫ{=zʬ "f0pEQP&dN)8Ѩp4Â\gۢF9.% +m2e–a%6xQ8!P|çwN;!uA4fRkv, rP\j2Gs2|(G(E+{Hfj8S>wT'@:|CrzMP-e i.eƫ&2xw#*o`whbboC+kX1(MqO7<P6K9倊,ʁ<ܶj-UNoߚhglQgu*}}>;Y@q;|;7=AֺwИwzPmVj,zY'8';tzmOFćh A rs 5d6+cG7@<R$#7+E: L8g\T:_ ik&#V* @pW[-|7Al5]qۂn5Jtߤ֦ǕT cFU ,eqkyk3HU{ʁd:h9{n_nvk3cr@}{ kpD^*G $Xr脳84=P8g:|ɽ`Ҷ,:db(똱 BJW09 d-GL"X &ZZ|Yf=h;3ٗ4;LZ<-&нގ`o5 `.?^u%V9\ F6-^V.ix~G P"E_o΃2zā.7j<-IqIeטΦoic]g`~wH +qm`sD@s9h'y"D TX{S{62?m OzOZ'.L(h\ILA +%a'E3mTiN2BK:˚;K/[#v7-|վ*p)h%Cj"yE'~%Ǚ'˨z/PF |FUKU酗F&/5%h :/jS`5IZ@]ҪI _8{cb})Z*{Y1.R\fw}fWo3C1&D'QЪ|*ЫB09c q1h*M vMY<[9-RLcދ|Aa +-6ߕdn!605H9l~>;jz:jcLQsǁlfUa/ۙԳ߈T424ۼz؀?#tA1&%W$*#Ic9dD8ebDΕsӿ}~V_o$QJ `DǘDR}ҳ= p=3GQ3A{u.bG˅U ;P3pZ0-3&h*ݵ?w?`y>B'6ǘBm2)>͋)A^[TϦq~ 8M`C#F9tN^[cZn ^ؠݻwnks~@ΏBo#䗟6>ދ}{ k<@7e\57 c] &Hʬ8# @.CC1p_ ?Ы`DǘH:W O/[Φ6a :ij_{Bl*W2o\`Mf (iiɖ잶Gcdy}w|U }#^%F0;cOPĕptd:l&ưRJdv wШe!]b>@llKЬ5@3d,w;Bgwڏ1Ʌdz%@x < OEz7ޢz>IB" 1'dF4GŸDN7 g, mޮ6oO]ǘL:~a8g^.]o߇  m~G*27gtIǘmn]˛~O\c`xf3/أz(wJ2\џRrNs6w} ~tr=TM{L1Mٹw73@Cn͚O=B WI շcLrj1W{mu^1~$^: =v?mA5Fp4G+y{&:24-GF ξoa[ |ֈjFwFON~,[2R牁L8Qm\U**q,2"}{l2,IL5W<~b|^3Ec_%7 8A;hw|bRcnh &vwQ>Fr>~Һđ6}`GA&+rə R1t@u\$^i5& ͍@.`λ8֥p஻&}ﮭ~7_i6A7-̓}xcoO5EXOKEf|=+<N0l )\Ȥx\H~!1]ϵۄЩ@L!JA'kHgMμĚW>Xu ֖̬7Um= s஫y "pRXƾ7G<bzI#-&zWo#|>@n>cjzi}k WBE3IjW9]Mʧ6Yh2l(;rr($9-dr9< Wg!SM*SKZE\vm{foh;DvslI.PK78$?zv({3S"c Q(k];t՘^1cF-Ң|rB!qsY+#ݿ}L&@>,|.ɶtJP!I#O]3"3ցjtfǘ!YAI|p%Gˈ1HO|n#s]×~ ^@1&(#;cXfabT{Lhluq퇎\gs*] ϞOr*Ԕ^^Vr5ec_6*d y'j;7&ƷR{\o>'b(D&WV#U ڪ1!{U"xY"T$ͫo6 ՙjTM"uic$}L&[jUPe_%|'ϽTAraX.0W'>'gBl0>fψx\Aџ5cmE  sgP2ea*X횩Lea4y˲xkԠ%N'hj^@-YMVKhPs1u}: 5 ƒQ,\<`aZڤ-Ȕf̄Ɉ3+gBmp>dkVD(k Pƙ Bob/8⢔6 buJ M*dPjwGsw|(xq[kmNL\"[ѧ9D.@'[FrHvL#;պl,gH.UqDZv$ㅊ/ L䯇T LL+~.KR&bXX#}A9WFǖLS'&aRRcİHf@nef iX]Vo_$AϔԣؔVIw.qU#'O%{+37~#<얺'Nf۩:$t)Gw:!K+˘vRC#1eo^WHHM\x6=2 ytCD2͛#'Zᒈ԰$}8*)Lx]#^Kȶ3J.uuPC؞Ued0+Zy4љC/3zGTM.@=L9ef0nmBu'h-9"̷kضŹL(.#LݼL"eJmNDcd<}9;'_{ Z12svv™љcd cs7O~Sf(P&^=EDAH. u93h_Zq ̡,PP@HoTް#Y{7+\w~.r0\ؾ/w>I6g_5)I9axRW]]^ hO~NF˖m_[5CzqR|^f1~{ o&e 9WtxYN'{eTz^Ջ)'j3}:EϳJm-TT&$&<Ώ64 2qׇExaj3q.&MZ4Qz͞^^/SջrD]gjuYzǟ#ӀgA\!]tX_e_MԀ_),E`Ә䜁c$PŒ\E\\{&G*gf7qxO-kЁ rPQ̴EJHTMQVF*Hzgt92E~0NL#ս:͓h=4o3q4"H00&թ!<ˆ*Y Wp[[9!%9_*Is_a]u/X^2khB3sP3xgH傸8Ք*R=g1x^(Kb%,.)Zd}%9KZډ2 6/MG`hn|\ˊ:ùpk=GŝGY RG {#ڻlNbVvQC*H]2vOd`5Y̧tҦ_grNg-(ӉX  4D|^EqL2..ɶnjk _RtE%t-G}3"Ib\3άpLTK[8 νCA $|Yto /AtRܢ k3q yՠfu,356wo@1=ǝY2x#j,^ۆ^ d?=G/jv XX:\}PcSpL}R{A%3JW2JúAxBF,/ ä L7 2TtP_nmhe M{h 58 !-XD3?ܙ$O"A/Ot,J*lOAA+A.4PwH Z:v58=d/[>ctGZ/L^`)E" *J8w_x>A 8z&P=ޔ#3S]홎QfT:֎t>1/j?RX 8z& "_Aa@xFl3q:xNЁލ"\޸Sbt*zQLUR0n[gbfϋ6wq~:! ^Uk2$ L.81qB8 ^HeL'1#PĤYb4g9I 8z&NϫOE*Y&Q4.81\ 6:8z&N5_hۀgAL+tբ K4UGXf'I ^ G("fա/qbzK>M|uYlqtuo׌hWրgJA[xUH[i3q885ġP\E(|\ZsM3f4' 8z&.\%l5L3A;4ǐ"r.$sANNvpsDž+rA <$ TZ⩶6@2=%꾸oC5Um3qXGpmi1 ?Rg.n-7шUiL:]P9rCT\tP8y}aL7LLLlj@7q6!N~U -*i+L3L9j:SAfyYBTʩ7\v#Q8$T Su4yO0;DYů>CqEوs >&Q1I>* 7FneFZfbzZNG~O0?W-Fwvy7$w~ST]C{XVN?,l>,A|ޱ%vnV0[챞+'g_f~Gs۷yp $WA1ko4ZhK^oK0ү1 G 'WQ OiE2nc!1] "`9 x@Q V<_h.=69"$ReHˌhC.YwƗ[G;8s51{ٲj̦›+:DiGMTq,rGߴd tjQ /`Vq"oқWP4efN7zȇaLl:nPsh*50Cc5jFEy0&ui#NΘ't z7"'E(*84w8"gpI6ͭyj.1oy+P8ܹΝ)֙ D!0ԁ( pq@} CFI2ex|I -'Q}?nzg%?ImZ.9kԃ>ChP_uշ CUwݤ·5md3AIltj)LZ`bq[tT!bT~\7̊[o/M1.^lSg,(&ŝw ixd`:I7b-!1MLZIso#(Ho`NC KM7P\csBGy;9>aj8O*b \ӰAM+jо^i1 zZGC`l̡1Q$ڹO{5As QL' o_9ZA W,,{ O_o q7磻2Va⏮yqςpోyp6&I,X)?gDv8 3ư|@|9fŤB_?We=`c ~ͱ_1[}9 pDP<KǷ1a \ e`0AR8Y̎$,>-s 5Co8Ex5oc<GQJU9ZA9( jpTLZC0ggoUKoc|uߍ!=ਤзsB5XuW/O_Ҕ8dsH }13k|yRM!zߙc>R,SڙZSoj^;k^0'1 2L:N|ih2" Ў`%:85O7>`쁮1ȁV<`>jhH0xjhe"H^UtZ\FTGFJ[?t'>9ԑ$w1aZ;RVb0<N:xSdaU_..Iĺ2Lx(TZ5*K~2[v:-wAO)~}!V/g7o%tSS]ص"Be}:|P@8RGBDYt<?|7 ץDEfn7|l*tƚ&W ןf@"M񍍆8$ 50,S."cFq -OdQ+bnY^>ßL S# e &H )74L|9H4㨮a Q@+eTo 8'~ǡ,Ƿ1ubj )K}Q*zwPpx]Th/FoE3čv/瞅g Jl(yQQ@#%sGXp0'IvG*'Asm7S X)2 z'(|xƱS~lbgU1IWes]zxw/+&5{|{|pK̄.meq܂nA&^(Ӕ ?OKxT!$>]LEc*4㾓 Y& əc4#F bg :ø=طqc.&8ߥH;s}~.?|7qH ɂn7Qyrދ|.؄jSg5KjR&'Q8H.oYv`o: R$~c2>#@7Xy/s0{;赫mĢ.vktAif)5"Z"Upi.YګD^І:\¤p:kCST PTdI7$0$Z,c"=9Si&Xab2*\X9] tsw΀27)d*h3A(mnr|Yp|.P#KLҟsP>o߿޵N|k|jq0> {p+sܚ\e̷RVI$%O@* $]d2F@Z8k^: Ps[<~`g(T XRQog1.IGK$HB"j\N1in2D*  صqXtݿZqcvK] dUˢ=c0# U: @ځm#'IV8W/uڀϮxbo>麫~#1G5!w`8o8ϛ섳1ħ^SOlo] lA6 BGaB{@/s# [aYHݟYϫ@KAxk>phw|>_/z7yh}rs4og|s..LV46|y7ziWwhu~m=k m(s3jLZNfkƏ㩟M w}x=dn:"'Zv4~{}-&ZO)H.IǠRĠrE/ F1De{ۦdLeFʹW.ҿ`֡35?fi=Uj5R(gs>?4 g9lGn:[5{OY^v ̫=x<сGdظ&_O( `3Shvwz&0ozwGyS\vac2q| >2$(rOZbh*5" QQ/ۚ.yy|k+ZJ')!]nbֺ`cbXZ62It@Z3e*)a`lk![35T|M*j9IǨ`owakXrt 嬑/@ڏ0j K%"lJ)JP(90,pcMHH@86chS_ u>7zE8S3)vJ3V`-)lLiXD3.P5sb}΋fޏ|~Z YYLyf`4Xy8b .iV(1Đ R =)hi.QnrHvXuExb"6E:Ksx M^@q:~l*q8 TXKZ|Q;pgèd^Dqq<"0i#X_k,rɖO΅@<|`y<7@gK^01 Î1DJע †wVƨga X`J\# ;cx愔 4-qPEh"QAw=3*[5̒3ͲçrEjayL]arp ʙ~NONI/¶#G=͝Yl^zՄg>;9ix* ׼ui}K&0YteAWէdmE-0Xo6VQûk AσS[ӏ)0^qJW_},!1Qi_? _o O()l=\dcZRƴ[ 3ӂ0G2 % #wG?2>>66Vպc)LFmx6۫_6~sv%aYޤ|1vqqm"q2 񘻠 6KsQq ygLXqh;|a+֖3L;>A|wu[pshhwjpu7ԄOTwyNIu{vX|wa]o&4bm;|9X.; /ޤPanvd_? d_N(Oc GȵEzlyNy^T06c}Qҷl\ g+K<[C"s&AS"'47HъpAGos DcI&d:{JG0o/r4ɶ v ۍĄLu'tU|]/vm2ѷxod1!eLXό'!\N<1x[Yx=qh9URa.'MFƱyQ(=VBYPs}; I GdގGM{gǻ? wR0aV)\Mca7e7 obo2w@3 S:E v:1O,WϩvȠ!7iW`fJZ+9>fM?|idػM?IOdDT7#MCX=,Ddj 2#.]cApPR#k8A O(OSaQ'`93t(EDz ӕGt 2"8'a&'v1Qh]o~+׸:bRs0Q2B ˍsFgO'`Tv8l|Yhr7cPGB}~MZb" U 1$ .q-E-c'g#'#v)23FQ7Fv)ž1{ !=nṬî` -fhML^(s?SP.eV6_F^%ç ]=V9kvL"gTB3;' y/iԛuOe|[ )A08yS[9@3FivQOg` 4MiZNDD6NMNJ\H3F  *8*eDqon4WB@]x/;&< lMH"F\vE$$ Z+ӺsS9W\1qiC68J<#.tԷtcXE& `G $ ئsr/~QM6ZLg3pD9b|몐;&Z!]{TJs8r$bZ*;& c5 9#7Ρ:SksczI!t+tv!á:ҽB_+)r'}S(1qnCo4ϗ0$#9 j伭Pa s1mZv0(E[MG[v^`fK&@L=%Ȁfd`y.:pIQE{8|7EzBe"SێB lǓ/d81ؒH󐐈RcAz ؽۏE2cp9q^{8ꬉE>KcWd҂ؠ"WIA[D|z2dhGqYS@V,iFۻ 0 HQӎvprC'֔S*R0=;< iP$=g<> "/^ O`R:1@]XPg1 mi>UCAO#D K#gX<jj)a̭v*jň)Ri'7|VkDdG`,O0u̱Fq硅w=بlcyԨRwqaY$i0C:(\pȽ'z?ܰ ;Jl(+wWnYW^fUrJaFӀ Dcp#峺|ooq\mN!f+68R$#/@ =Z4}2k|pIJGh?ric; 6hl#|( UUYA:Rb&y$ViǬQFYJyqp UX/5^#Zֺs@}K ,xJO{AƘh ZZ.HY;gZhV%%I + sZPA9* p 9 XEL! +ƤBD4V[hzyo*#\52D A  xC 1W¸`av4ZeAHҾeSZAY4+5`:S:x ˷Vbɣ-ߙ>I] .h[Wkk ա :D6G4?LΌzх-0`mu9 {n4ić\k{j}UD=oj>eT,fH7sz[ 7*R %#)-B1!^ex8ն[,+- 0w.N ,D[И'tIeDJn1)B=<2wT H;IpԊ2a@wZ B' LG] ӌ,1WC_R>6a"(j,L#<7d5 ]4 s)I2\I;Rei82cDHΣJQH(5* yh4B]AK"ܚ#"b4hIX tpʈ( ,'#( ZK,*sSf1R٧E%[Grʜ3w#"\d')LR[xMBL0ŢgFetX#|mftA."5k`Ijg,N)XT_ef5,NG\JS㲙K_@T1rdѰΥ!bO"JnYViEE{2E$99%Yfï)stbُq^Jg22.xjlbCg&hљss~s9T/網O wBo-&?1(1QK͔Ȑz2Ԓ2{ZB? YLo8_'z;M NO#A۩YGn"唖h{.#p]jM6YR]Y:EniT?sg`@97O?7OS8=_^C:<^T2uAޅvڡ7H]cҪ,i7'MkB9ل7+r ݷ.B(=]\{2sxx>JN;`@A &&s+s =Rc1I2Ty-Ȅ$^q2"NxPeA8 H5Â!ر깺 }W?BR\ ,H+5%F$8o-:j)-a)]GrЋ,UHrHNݔH=XM;jN"F:ϪΜJ:P;P;q9m[UG{3YJWMػwjz@cl5!X\D{fCJo-UX _q6`+cT ^_a61oWep@K~ y% IK-1i@=GMA@mDAXŻT rQ2gJZ-PUBn;t{8&e`%D1ʄ:#xJKԑȜ@ldMV:d ႉh m!M$i >g!)Ni;ܞ t>%rV",{|d4 TTx8()q3&cY]L$91XHN)ƹs$a In>{~|.FFybaAcZ)ZZqH ZXFǭʙAPI(I}#i& FT80\1 Am"irZwe<ݲ,0*<ޗ3DM9cR`n%> zk"#D-WA3E@O?5z2CQ#%%RQV-IA1 *z#ăr>:IOioSb !5b  s '1V|8ZQ{uШ5x&!@7122)e[#G_K,aDPeY$Z v-QMГ}X'Zy&Ey1OLNb)gy鸵!KI MoQ<`,wHGkr_P-KqYk{ +yJ>{9-DB'$R\O eQ% Tf2S^YH0ٿߤ>\~:*8qL=b<)Xj$8%JSf=ƒSO+L=5̑N8өVfS &`H&[\2幔Q AOi4 9=B9ʊ1'aϹ+3bJŖ|iAH-]- ׸7PJ.3b ʸ/d%`7W]LAPLAZ$].\"ͼ8O%rm˃w8(?z_.BcYLt^OqNGKe{9&* & oKkQD0]{Q_f//ǛH.=xu"#S&d>TbB;ٯ$!^dZqzli"Ya?H'?O"CGQN_E el[;<:k 6f5}Hql#k_fGٍ/f |^O+0Uc>~6+KւI.A%ԑ6emT:~,0VkF';6I4ޮVuۻ4 3uA$ɺ/z2i}V&wy LqtaQҹVpbT~2OqXt.**I V:#mN. J|^)yYWiZ+hn_-+*)v5P쬅 PKr1<WK54VpVzG`9Oǣ~,^[ӞtivK?nv &uD985`Ԍ`-}y4#ɼ浃N9˯&`MswŠ-%4)eMXt-yTϲht6iz{E,Kz.-Næb12.F>qܭYP%hVK.f!wX0nw*>L?616;Єz었K /p{ $]Wqb oqg;w{ϻ$pMfС20] Sްe׿@5o{˧wp] m|Vf'hFf\ dܽ:b-Fy ofI-gJx8_!&RUQhM]r,x'~cj\5xnnH tCNVnh~SkwgG{(a ݏǻ~FW٤gpma~ɤ`;yQSQFcڐ/Om.NԿ~wt1s}9STB˝h x'9=.ՓPB3UWesk6{8(pBiW.]!sJ,0rG,QS!h hf|so=Ť >m(j|+D*&:Dv%+wCoPI0-B؋3 ڸmcNrsbԔxeb`ښ,mxϫc7*2p՝ƺ=侹|TSü3p;Mjw|^҄3f\4[M㼰Q:q M?MB{/%A/a8:`G+iW\Nss_[~כ^ ҍ?~zqbD_9xu#Cc v8-9AҦL+ L(йS`D .ssSi7D%mB41 Ra\R$HY jg{iN%eg:Pmu([!XAfG;}=;UG^nN$ӆ>E' 1K\b(2j.Mqφmt>ص~#/ Ͷ~c"87MfuI)Z0%fmsG&듦^D+f}s@}|+n{.){ik $yUcr{5Ǿ!5{'CffO] BO&6Racj (|<ʆHY '45Q=d6Rfe(mF{ ,$S%EiE# nQhUGO-T*Tg(*JI#p J-H. MxV4yHhP!xSÔʙnvEӔ ZQ|1He!wwr3 +@EP&ɉH'HN.$ŒƸѨs$'PS,X2"+U["PCKe 8,:&gT&8B f3%Չp\Xr\NL"ܢsI9Y*bKhf:<$Tvϴ 'xzPaE2=M)(Dg2kݪ{ѹc-P(`ɖ~ѩ"6B9~kuB F,8\;(F}L<׹1p@*Ew\+J,PiEu(R [Q~)y/(S- $%J{ F2W|~F-d.Q!=V=rr$zfi%4u?T+L V_u t>ƫ:|yvPRLM?n^~l!| 8v>z,$ё$V?'JgOtzln19oO _L"|ӣև@IslAod Z=RsX5"ɬ wBE4B99Rh\ptN0_kVHFI6quw5iwNdϟE$[8S|RW|nkGQQדKl{_Gu 7{\\6 Ԡ@6DA~[+>OyOט;'}, l.Ky !FucmfK ettnlvlt2'Q\anDqK;ܤG\{O%ɣLO:|D]iHؖ8D'F=8摡53]rpyc;֘NBC00ݝg7hM #ΜٯA>6sa6sq`8emQۏׂu- %!Wt&࿲yayJ`Xۧas"$5ѬYͭNuc5Jٔ6h{#6~dl6!ϦP<>T o6A;?Eq 6?ɛNpprw?E_p.<w:`v?y m5MFlҴls ߤ]״vgJ["Z87]os o/Q'μ[^jWۨgӼz?~~nzDX3.ڑ`F>24/|-bS 2_kurW%Ln-j/ 2\3c0py5h.бl5H&i|n>[^~.=!wlwHjHh1[@ +Xi^.{4:x] 2: C*X·w~/(R Tt3q*m) {WĦ9v^e3?erD#hiPDR8a4#Ćs}]|r7/4N]\vHm{2 =W56;;,cB\zg@$:^I2K%)KҨEs[jy~\HgM` uJ*Q.X BA;Fpb`&V̒df)&]lXi3Qk \ytbQow e=k;n9s7*^ޗonV!LQAJR+*f(Ux`!$Ě&uWdQ6UsN?#zEJp%g4qd _\-i): &DF(jeߔu>6(x.Ig֐hH"D-0Tb%#^pb8Z1{ul5x&FCyIh)Zd42! qP˜˘U^d}-#}s_h_5VT7݄b5DL*T9u ]`R*8')\^G{1c_v&!jiY;K4hX\oAxKURoyG1k O& %䑒M2-A=ry0h~/x sގyrw"ܻ\H _%.@Eݒ;FvotR=2h:F~sQՔtB$/$TGL%SFZZRK*' ĒyKh zj`=]tꐞβ1sYKJS a3 BDNJx0US(8kGVZO'-|~X>6}~f^-\7y/Pq}+ۤ^zu- Ӆsm)8.%eK9He2&H$gSG?L+Mx*'wEq˂8j/W׻mv,o-P'HTl]~z6M/XSBC"!C6*S<Bi'J`Mwֈ] .a'~RmAZЊj*d/֫j7+odOE "rKL "!$:!q(J$OCJyr٢c{c)#]#<"+ִ͠Hہ@E k k k kAZk k k x-x-*x-x-V%5 ` 1p!.QB \ 1pT@*PB \0U!.bB \ SB=uHFsY +bB \ 1#=u@Kz,Ѥ6n4^W$ß@EX,u9QClzB%k 5J"B)'BJ`@[JB%)k#)Qnv+4l%L<!fQIUSm;ٻ6cWb$ *=ȔvNEvJ eɧӳ+BjX6NL_nyrE3JӍ$}E"-m)ym3J=[%|ؘxIKgwd"D~"'k)8qVrP(m?9 n˕5CH 2Ņ M J! '-PA(3P"4 iCB"0DԈ +TT!H6Pd4mX#rQW~}.W/80cR`n%> zk"#D-Wx+MSE]CO=jJiY(@+𖤠 IkaAKj5ܤNZ?wzrioc:ڊy)1 h1IB+A>I*F8LnoTG_+hݑosˊcF//CI, yP<X"dA#,#F8"*`㼕""@(>sXdW r pG'c& GF 3ϼtZ␥T"E&Vjن9˹HGctC~Aq2v)]{EH^(`u2Dm6`(h$8ɱRIPs{.)Pwk%#RNcRɜRafXTqL!ʓH$w*emQ D(3- wrdz#LFQcd幡 \Bt jP0R -ϯ4KeZZ1")=Z(R;oi1J  Hx; 5+H>Õ'̚WJS!E5o /I'uwջAa0Ic,ǡgV"Jd7P"Ŷ4}_..G9a+NIq!g75:EF-؉`>F&f+s>#X^/o'\.5hMМk7rm^R( ?N~ PHLT!HR92a0,`ZdŇBc{MSHQgjJHXqBf2>\z8dØ% ;KN// _gwPG^_uw߽|&w|,7*$}??Z!0lhXy.xqeS5.pf{nnr m7.kώ@tരJz3? b~]xM_QP;TiĐS}Bybs=R ﯊ N'q:]ɲSQ.^KX-ƘavxiRS}yWྍW z#(wR畁q8([_CZ!^Ɉdj"Fg^fX0icGFf ͢#g,,H+5%F$82ZDy P@UrcrA峇M`d^~\9:j[tttS#áM44m+$CsKFD} ZptZp aI{yNs0>`=r f؀ !OucGXHI@_{R{[RR%n@:$# @'甿(|Z4ښhciPrN-4zWcc1o l? 0k⣓H3qd6`< hpP%i6"SDl #[NZep,gŮx9kmߜƿ $MJ?L >yeҩ~իgYyAi/~0ېN)'7qeCzW?XE4Ej/ LmNiLAH^eCQ(; Fy޾GL ʛ(KޫKva J.5jD/9vRВ)E 0@hw"w3 _.|{ee#u2^2+Ꮝ ("`# մ}u~G0)np![n5@,NcʠZL40U%c=^t2!%k.PHb̤kCV{tHbaJ['Ϊ'smw"N0N:C 3>_ -z'zLa}p0EN93X[-[\q]&zp<G9DŽRn&x9)[GEd}rbTjbHZ'hUNggW 4pq7-; /Jr0Tz8(p XD?gC_l F(%7m' R\8Osm˃w4 _gۋdq`OQN=K|ʪ* Ӓik CmA.j?\c <7u^ь):3!RnLfĊ1vxxrJLLE} -b*_K`t$}G=j:R ޠLt Dc2cDq?{MҞhG7cyvIE7?UFiꁽ19CFtLTT"?g#ݱ*I.c7ݜeC㹚ϖ}R '~1c/%ET^o>`'tZ?]WmyEC($b+QZ-[}E׻+tIa\ZiQXj)e.eKB;Dg\9K$':mf7ue7*Xtvx)puIMtwzxW[3R.o۵%T֞u<*G0Sܱ`XP!GZ{.Zv+>O [R3 79.\GmrJ,i6jQ *6-\[`q-h>}c#mDvkq*Z)!kj⹠\CR~'$Cdޙ$laRןt-aR[A)abdOp#g{ǀYwa,)sj,h{sIJGD 24w^fm*؈SqGAPVr#+ 9I %6<2J^GP1f* րn]lpPߣZ"Y$.M@uT\JFRr5ZJͅb^6\`g-=X?Vlj[0r_N ,mcBc$'91e 𳌡N)ׅBOkmH#QCVn$:A!"L(q_ _E"e9ǯrp- hMqO!iТ&Xyl@W]MSY5M؟qi:+ȄI&C҂eIF)J$Ū9ž5_yl=i+΄Qecu*D/3,t*1Edvԭ J|gx+גT}N-!*(:t!GǛJRO4R{'ReW pRaN_~|SQ'>UxVj bxܠmr`(dHWWVGRrd.Qb=l,9 =wFP{e36?]Lͧ?Ѻ YM9zwV8(B\cSv.}T]F4W4cB'oJ 9*rvHpi]v~!;qN96Q4Eꪽ&T>q `\?<;oVbmDh<ə=%/0^cFrrム.V #Vca mLqbǣBOc.gW/`:#rըhڈŒNJJpHX}|ʭo&T,jj k7Uj_ۛ FU\;~Cw߿.?|{|7g\سӿ[8 \WM?b ӟ624lhiaUTO|qM${]m6$꧛?v_t|qYNU4;?eia5ƒM+4m `R-YάE[:e#f2#msU1䊃vٿx?ڦG.]VZ=ιhϘQPl@Mn=6(O C"$w[($6ԗ} ztuVɵ"-d}V7vCԻE C 謭Wuנd$}m ],'=e]nMӊ??Г.l3\{vVddgx]Su\k{t't\5:P}ppIL8JN^ Wk΄W UQ /7"UT& v1q# iAfM9:#n KbC+X*1#2 ub";.QgYH'^)y P4p?B è s_|ܘf"4w=sA hM(b+<{A2 Q&b^ N[wqԁq o';Eo_C3qgK[(s=7Fj7sll/bvp9U?Poqd%E/ZUzJq!rAKe|a3d `{ڐ Y<=~\;=6VG§ͪ;-/K.pA^@%)(µs\gCٮj[m8]nK}-qz.YHٴŸ hih?o)6!yNW4;X鞍iq!WwY1% U\Zt!1N 0}2tvc/7rEB1:R ! Ő$ d&l<}D"B'# ۙ4/O{9;Vӌ %II:Sh-cL0$=yw*ge'BOJXFYdR26%C'ʎ嬳l(gw;LJRB,ӄsҘ2+o9'(atsD2䲍{)+.E×SE7zr''A R,E2EüyKFKNBS qF:ZzxHo{?iKx%?+p%Xc$I2 @{щ88Rg5XFi!q *xVɪL a̓{$Z,NYMꥭͭO+hAOE:c Xfِ4'1sQ@Arg$T!'/y))C(AӇ;9ӡHGgTK7d]R7Kr (߳>{?ފ+szl׍_h-u7ߧٯ^`wV#(ܻiA+UGċW. )b4,hآQPkS 8՟ݦV(XvЉBCt{K9)mu siUcӨ9H'#8ߦEn}.]ghq(seV="Eqօ2g%}Pkʑ,[H4G'2+H_$iV}9y7Vy{LZx]ztGXJ›W.B0JzS^r@T!:eLp1Ld- ZIDkd=g/q&n v(yT˭f - ,| ?2.6ݐ%]0p]4A@] ke ;/jׁeC^>D/uAI_WGqkQ]F{m%9]O0j92!F;1AWg{r~0zǻt]QtQBz 3!?t1ץ1/*wzwN{,iDowS}8zaƶGZ{BZBۥRK^[Rm&-}O%zZ2A3:Q~4[N祥wx+Ǝh|-9D!y\M(s4 )LI>FTxv>xevgꠘY,sqmW>/ NX$6 H3M,֞t\˭RBJJYI =+)g);> Q؞CbH!٠gK^P#XVyӤ;^7: y,*mafjn *2.7$qQnl,b փ?5H;fGyB~愽KKJftV$xȱd(l{i c5&ԣNufwč;8n颟۷oUL?߆QMӪ]\LZ8ӟڊ7SU Vt8a4(}ɮ"OtB$ r5O[ ݫB [=Y?-S2fOyįm}XA}m#4}o׼Y_8)vzGBӥ}ׯ.~݆'`Ӻ}q/G+,zgLg*h- z-1K"y]*qe<$БTqB$1( բtMp<{%H5&SPb$yQHtk]ΉO9(#HA uaƕ2QuįCėpZ{egO5s@:du.ul: E9j%_iICYm~J"_'7N`B3gů>+~ٲ_w2fA(4RR(5)Zc "ܰ8@[|  8$R;AƌEcRaYV&wB/HT3骞xu*2 j!3ƌ&XǍ$Q:zw*; L %5!JvJc fJLjFd3Cꞁe8 HS yZxdo/l:#Lax3)ӗcB !_pݿz-釒Qz<@":>M>c1d#I@Pp,J0@zrK[$E2G>VV7!XOB7CP@vM|Ԝg!kl -{uiJ K(՚ M vM`3kAP;N2XƂV뇉]o,ZQs ^puLXCwxk'F4e&@("Yc{ vg0n %XWԒe}uKlђ촑v,*VifLUX2sllq~$58B lrn(_OZECDUئ*)>̋,APŖ$BJ`٨ rn@肼oި J,5jWщRbAőA~kz < 9sF%+P @D{|o9l#g ȷD҈ l։1-vOyelZ+M oX 6 Vg7S @MYhsHtX? 33@Y 쬱"YӈJEBT\۵w("gij?1=Gz:WGRh]C+e *듌ڌ)z oa-lT/ i`DLEV2 r@聼crꌼmb*m'&*-AK܍=N&{ˆy!-'o=>$c1'@<9[ dlVɨ[PBF7dz DZ15C(:ZͰ ZZ<U^#0!}>1?ݼaZh6 E*!{_ >Bjm@|wcq(TAP[Vij=eca\C}|ܯ6SW1w(vhPHB V_(hn_4Xr&w*Ua1'_3OBY!sG_rPO{tXn8ڙ?xn}qU80I 'm|Vɂ{*FWF*@a# FDelPƎ!?`ewcޱ5Rh/ Vz$Z8BI9.UH5['fn)H6tdi$=2B3#dDTb?Иp z ƪZecC`5٫,F~ Sd#6`+^}-*VOyvG>NNPMQy)̙,<#Oy}X5|{WkɇPKj@聼 F9X0Ѳ:0ʪMƐKg,LMA@TQ _A{9&7$|F sF«]#ݏ@肼7;)h*B '-B)HYWU+}l14(qAG'jɻȸՀLyǎBo6!`우nJL*#i#}@[+6l'D4}#%¼r|ZTNW"f$6rQf٢Y +F 5nT5lUfpJmVxMXZ 6bСiT{RTHsvs|,K!< ,j'&g΅b.o~ySssO앚CT p0tIT3٢ ;hCޔ a@6I+¼Iȯ$/͐޸/eOJ*'Cqjj,lĪRykkE`$Y/,%s5IWHVE4Ԫd߁.dslZeP1P,>H)צA^_^ }JݜK7g/۵,.ubG''N{}y;?{z8CNo3+:W˘LP24v\qm`_f{AɄs@vT䢖6$Iyc5gՕς5x/ldլIIcj }8tnۙg_@"T1j Z,Պ*N3+f@fR+u*U4W,Г>|-*'ǦqjtVd׆6Rkn+ |r1}KSfjD~'=C{DEr-9TrF/2)RFņ$sW;jCtMZ .c^)6&Ύ^8Df.6\X[&Oޞ=)|'Y^6-)jUpD(*Z6H1Uxl~rRFfHLaښl/؄P5XEDxoLJEUSan$g;3Y*v{B֐۬dQ0VbAȐ]ًC-ȭBipG7 TXSو{⯈++gsX{Hw1f %\v?.kV;T!|:}aa^1AZ1>wߝ//W|SJϧ\JʆB5YTM0/n(+޽un>_88o6)wF*>׃j֛|;>8?ݨujѩ(ylmVI@M#-}_Vj}?b˫٩A6r&Cϖz~_aڼ6IÿW16 ,3fH ¹ br^7(i ݻO%TIYLl0{@U ^}f]7/K"V, Nl^PrTY{>fc6Z*Zc X9{'ll5:B)Jk 8>80S,gc\t2CD'SJٿlI%r31+ߏ +b_5EvI'Ui!R&ymQX!$%ZixN5kMɳesnj{'DޱcQ3B8G9Ilcq #%=%f}xMP>c0R@R!1C!jV(?jH </#}J^ȜLM-բV!o;+G_#CjU<f8Ґf\lg3梍̳/?m7N#[{AA9CCfx{'/@F /9g>uhyk*;gQ&Ɲ C{+K(ؽH]Η۷Iv v<(-"rɳ'x_Nfa"7k{~c=8=2IJh@h4|=M/yj -ˢ΀aryaF- Nߞ-ޒ,!T^mˆ-!|3nkq.Rh% 岕.] v-/;2 3m.{|lشA/> 9{3$ Wo8y Px-L>?2ۯ&66Ժ/Z@2Mkl{o,ml19뗯l*7__QI֗m X?M^Kki^liaue tr2?^W rКpipɳ|#mfSnt b?$H:U̓mKr`HW6YLxH}T!˕:;kYj.溽tsmdjIFEVt&_Ht !Ba ( @X9t9OYt䉿ְXI%C *|FiD BQJ3P]POF7J^k|Jꋯ,t1\Uv6uk_vho.(nW' KDd|drs~~t#p=ʶ5.*IA# ~oƈk3~wG&Nm/ Qa|XDn~Kk$rUҫ4aIÔh'YeGM R iS\q II0}^7}m4}ҟ:6X8dtlY]8x޿B08QLylpkB7uo0ng4kj!齽c}6gBuCԡ~wpݏ*8갳n@5; x+>3Vd$1HFnmb񓷦Z|ȜHHI'%^zK2^ddpN"oisɢF\iuI<å&!CtR+e"A6BL(93 YD2%F'/>oȳ} > li8ݶŦ;hևs~+.?ԛ^;q64/iHۓw-ຨ`zf׫@=؊-3¹6VR[U-3څٸh2 C|o3*Bʑ2݁_]{k=S [+;Ӡi>-Ew讶Se6;n#Ctk3%#fUy.dtGp (NY)~"xw~{3,.|6 5ч}vnKxTxQ*=5Jy0c츏=Nͽ%)YspmZ=Dw}e&ߺwHf+%}{`-km#SS8e7:')#_w[&n= )U x+nWK-xR *C 9A(FWXx+v2eЯ[M)q)xaEe[։Ʋd?Gko)][@U:E|buIɴHgj{k‡ -S:\~nnS[DŽH1G@ъқP]݅x AIR.)H BBPNV"5>Q|ªsFb+rPHBNjɅ^R(Ȇ˖貌L;oe cru?6rZYd I8j ohhI+(UJT9 RUHJU de,.<̺m㭫Ax"w`ix\~H㛄*m4Z @&C$>G F&-ZZՙLrԀIr&Y6qPV9IQF"O&cl^E=O<.w2f}Z_lt;Ý-^ZSs/0^ůx0>e~QkrDH n!H[%0'FYIzNDd59g*:Peu|yj(E]f ڂe )B Kf)}я6Srbyf1!s5Zi*2o;n@#Q֜ݱC@pJ}8"ӱ:&hB8*վAsm]sjnۮfQU2T9qJES2Ecc4fxEK'F^e`)DJ7IB$R˹ ,cK opm 58LSQtrZ{P/k7rM#ߔ[ F Ő@ɤ:=~q^:` H-(YB$L}'Une_F8ZIrpG'Y9U.Pg[偛Z3G8O䛤BT5٤Q$e9km:;U(UE t q`N,bk|P:sf[zq'M ?UӃzs %5Vh-i%* M#)j ZZa='7N:-~;``cIQ@QpghRfY̞!D)0L$P?xߴh8`ĕ[.jZg:yH+9e22#LSQEJ ,N [qY=έL =:kE+|eᛅbwicf"@!NUA0oQ=Zq aq2DeVWW1c$E:ZJiY:} 'Ln}A|Lr 6 QNf 5T&J:O6)F{WZ{6)%d|ø?G(] wTOfzQAӱc_~u<ߐd|=5A9.0I.6f) rv>Vx!85VBf-Y mBf-Y mVɭ6Bf-Y mBf-Y mBf-Y mBfZhڬ6kZhڬ6k͝MX(NOnB^4 x_y R~v2db1JxL@taLەiDo!fcˎ(;憬΋9}fd|W{+gh2mLrF,WLJ3ց'V(&IQF"O&c0@GcM1$|UFEM1@FEāLĝ7etlsfF|M5}l?~팜8xvp[ieM{FC-a|lbU9V{" .kh !HI$'Fto:GzNDd59/9,"5e}j̉s7NT 3Sx13y5y*?/Jy!Ew[xdd'0;PE" Wݼ&)]"[fޫzW2`OaPLt~߇R|U"W|wݰٴ:0~47P#=7>.[@9C-zv\-+wt[ N$c;+&l NMs{J-c1x10 .fdrbqb#V|`Fƺ\t.g `*8ʝ7\ LŘ8P=d:FY^1Tk%ig9 @\&3i;Xx\JI L˕d[Fy{iaw?I6.MF%m KnӢ$JђdKk,)QV8dk.׭3¹6f Dqvq:[']>r~-?=#?S=wf}i\z ]>J|&&~M} (s!ʫN'rrgw]qZ{?]-UЮ |nCB8VBre5wvz21혿sery&^4-#^Q:.JVr9MY5dY\0ŋ"0:u-WݸT|G̲K> nbH;Væk?}2ť^by]<&]n.. rvk3kK!jdw&&2 2B"+]povl&}NRZʮvP3k dƔuY6i52y[/$gm$h}3fCsbtճYn%h:)֢FO^#-:n">yvD׳:ŹXdsS˂dI⪜Nw'Ǻ3LLZ ܚPFXɜ—ab 3]_B|}%2RNq,z̝vX-"ro0CJfGc$\YLt,> <+ߜ6~(j*0UDEeiD"rA 4xac9 3ǏXG>dDTEZ'|DQ(R;oi1J  Hx;Y#GYÕB&en*dz5!j$^kRœ~{%@zu xýʬWHuUOo[3B1Jo /bX>+t5c`WX".pR!ruKA)ʌB_g@0ƁzeM@ZX.*l҈χn=wS]$ I>ݜ? M2kЧCx,MToTi{.{QIS?r{]25̮EF8'nbg0+g]2Ezb_O~~N?x 6 &sg>{[Ӗm.pä221zb'7t iFn-a?)>g/&&H^gliJH 8+LFR/ÑKM?C1E6.A_ߞ6Z'_ohTOZ7ԏ ׻v|7o^uonNo*UCaP I}]+uMۧk欧ߧ_ۆ~B[)=7?r+]}o}*YVY~iqj6LAp&~6])'k6Q/k҈!=Ȼdf|P"p_t'C5TI۬_eb~&gXZj0:Ui3Yg@L`O ~ӄ+2K;\ *0wHsIMxQ?2۟?7Ư[ ЀQVp@+kRektyj<,8@H5ÂLV}32۟J8ۑ#k9yBT Dd8: G5*XM$T"ʃEL 9!i=V P0y?}ϭ^+"h+fr eKҿ 5u5% ZaYͭߍnVkNnV0`:k?&,WEt~\-Ո8e17덟Gn[I 3-dKK)FD¿(x9Lf K(E/ho?M>oe۠"3%RZQ:@ 4G\Dp r^)FZUd: y JُKI3N,s$iY}hm8Gs\3o7#lxP9`18ξ }EVpWI$Ҍc- :O0btడK !̳"SK݉Φ˯͛AK1czF[GU5۬H戮0޺~sg-S_}>n?})53^jZege`t('*mXR z_P,X) ^7|R, 7%1:VM {04P9lZUkb||[\+&h u0 圀zBRF|% QCJb s1Dt@Nxfl>km83 !Xy.Z4"IrHg.g)zn<vS$߇Ik%IQ זhQjc|nMi=ua]{=2&F%>*y]S1逰Y@SJx-NE @g;x4ųc*7o>}u*AGŧ0WigUUQ~\uҴ:0~4oJBͥ}7{r!mAyH'HJv6oR@r@#t^3#>LF鴐?+ s:=*»T4xƝa<#Hph/:aa{ȴPgs4.d .qx?J0Z.[,raK]TϢ̶҃xVxje?=kRahrjNC!D!e!x0k5f,h@k13Dk45 ih-d+O\GM3\740$5Iyܱ~s!,$S{+%Z y7c;FUZF@k5|-%p+NQ]9=I>Nwlٯ7U&! HdRl}'Fl]]{u\Z$޿b[6 +MCۛN~P+byOc(ޮ.뷷7w4:g*Ԭ7Kxsv˻O;D?Lhy767{=s1}x~/3ڊxN4ZnYf%uΚ^7P ĕSy˨(rtY? |&՛0A~8pu)ᨴwÏ%)Vf@>~Jd8`TEmQE P7tT_ĥG|6E M Qs=M;gY̕M5#Rg3j'AcfiGs +32? #!?{ƑlH~0^Ż1xX!SҚ"_ I4R+g؏ztmE?FBj] ^V'n|};Muq ف}AMz  K9=lN9IXN9=lN]9=lNu-aQJQr՜l5'[Vs՜l5'[V3jiUy#Q@Ln|F77aVmo^7tj j:Mg3}V227׿ޅ*EkQ TN%D0J*MR_FoBh۹(~S.M 0]œ[fr}`pܨh<Dp275798D쮺X[5_:/$t'y|c CAn|i \!ݎG)0xGya0Rڢ^sDPkVBbI~cnwcZ۽jb[f̦1ua_ {khׁ hUC{͗C[ʫsm||tĞY}tTc.[}=A=S]{~@<%IwzEoUtcQ3W0=J'hv?9D۩r)pSoꙹzfg4zfgJ(Jgf3s\=3y3e3eM˚gꙹzfL/f|Ox_zP]`xyV/ͷkБ]χ 8=C/wWGUB w0)HiO7 \54.䪡jh檡jh檡jhfCǞ.5ovO[ʦ\9YeKKeDC}U6Ol*㫲hm+!bP&!1"hi0:ĤňQN()ya`"ZA۠FH"I`ci9I1δ]i ks*á[)u7{TxF>trsCiS|TUdIKgw?@3 "J \ƋDzb{͂8H\4wg\Y<1DʰS\DpBXKK0i:D ˈU=3(* %BC6XrJ[3D)5VC*$6Pd4mYZΎzV_U:&)I3)GM$1#s+ 93m^ A &*oM1ށghS>~-?[wX- PF`TH4xTXKRP Ik%]?FPk妃$*IoǿmX;mT#c{JBA >Z @@a9 G+ $Ln\6c6;LTCjm+9ڥNꀵZ|?w٦edR)7G4YPpAˆˈH q=n\gRFO&"OU>Ѳ%yT(ۨar晗[ Vj Rd^hbߢz,XShGkTP-k#:A&icuf+^p2H&G 6D-P^{-}Q11CoG-x׻?)~5*(L1 0b^a@0a0Lu%gC |f2Q9VgKE|Ǡщe^GxY-EN͞5U/f1<"C։38MQͮE]_O@?C>)cE޽ %2@JF/tJÃ&=,!;+k_'O?((_Nc%ЃXjT%@SLЅL_ {fO̲ ubj1' +D`MQmZ:Utx" J1֮Gvl aodǝ8}:o]0s%YIYαCFÏZ4eb\qsH^[GG~(01$ R?rnTO(Y4dȒrAU :-mnA-x7)Pʍ"6:2,"`:sGkY)$q b~7 I"bv*_Lb%VAq@֤|tVKWUzTNMfUX;(\kSSM!e/7kK25uB4kפ״˥ڟ_g&HyJVMuD4jss~sMSe:@M a&~*;|;'8hk7>4kȮe%&-^k Pčoɦ r}fB#V}y7Xk CsaFLE˽&>]SסEf]ә+c^zfJ,#^5 TW[s5e7gkv}W^30?ncCAQǦ,at7fڼr~gA0CN% w.O$@eS.Eiʧ^JC$(o!],)K,|zqցgWVkĺhہe&{w'B]}? |  i{ ʡ(;~)J3{^-Tk 1n^Ô3vurCdW0X e _xfTt/ yϙC3eq(NQR)29DtMy?HxTnJ_RS7* EJKZ>]a1ҾL 'Em rw7ͻӰ!tmHz&gKyB[ʓsL4h7W}>@-e@vmt%֭49R> vmFPwh渶ʄYT։K-c1KA\ny Lq,!]Zn5`]Ncʐ~%Q s[R IUp;oB!1qV{tHbaJ36JW}/7I+&OtJD8~$/g0rs3-v9cw_vmR%K6p0EI%3ȖXR?p֕2?[¹6fg [lw;J[Jԗ/['bhTc󖟁ޑbk[@Za+LǟbΔ21i&;VΓ—a0،rLwx;2a` U(sVA-m=3{9AFɎ{^czs;GAKsY60׾ɉo~lr~RcCFFFF窔T{THq2\b]%U_IaF Il? =q|Eiu:d b'0h:cth{RB)|HS\L3ʭDŕ O(AIh)5ym[-MeS'lB O1jvhRC]K߭GX-=k`rkyK)X_,|FrJZ:i)D2q]T{sM'>]6eEsF霳8 rLĥD:$2"J%1VC"S#@|I &_Νv%(#RNcRYKVlɨ43SI%hD0AP4@2Ag1c%P 1Ѓx4]CmM(QYj aDQ(R;oi1J  Hx;YU8p%M9I.*k܇H:5.|{V5L0H3=~ħoMN]* ~yghiǷŕzPY3Vu {w="X3$kzNC 9S!rrN.Swb&pcO&SXoN@g+a Uiʚ\:$U<\_ ];H4|t9>U^sqkEzboU=$s)t:ނ.15NХ.uYMm~} *IFU|_ivQ\k{e`a0Zufܘ߮qp}yUӖ{.9hݴ:d8[w/BM=1{i4w#Ic7ևZi 焞y}s=69LF:dSMsUB{NyEHꘃe4vI(FH)>Ũ>_6ZWkhTL֏ w}{'X.^凋W^`.|~W (0~4 A?7?ܵ957*K׼YK)7fCJ[nV D⧻^ }>ܦ'8XHGA'~6Me)7k_FCz1*w%D fXGAu&tپPVh{xuZ9(qj4-S9}}ؗhBpDf `gAF=iR;ɷ\8F?1 qiBV!NIǘ3X/T3, dО\Vs \`e&$QتQj"6{Q,bR͙Iy>jjw{oXHk2A;D@wOk($|P@*mK'kCf8n Zt ڀsqb4KJ۾”Ӎǿj[ 3-dJK)FD¿|`˞KĞ4{]l÷mK$JiGr`T z"[llY+H0 KydcOL^a 1Y,-,xGI4 GZI akx"tBxyq3XMS خ'v\IG'GK@g«6`#aZ8yVVw޾ݝl*f_~V^ uyŮ$܀o}sF!}ّYSl~.ۏY>&5_Nyh^|.-<{ze PwmM$;8ej:/[yٳqNLRʆ70gUm0_6tu ,.JILɢQ;[kPARLW"1 b?2E; !N6뾇߿}_#Y4a'vh;!\ ΂^DJ(с-ʥ?CALu$;*`l-Luyqs򢃳o/} 2@G! W ]bƣJé'w>p\IIE\:qx9YwT߀AKރx3Z8W۔"_Id0ԑY\4Q5R#?r oc) ;y)ŝ)5rn `&*ȢuIM%~n+-oh7 Egh0[V0'+w..f;|8t[OV-ZF_,Zm bn2Ox3ٜM⬟iFKsa?ųygXb"@'#z)Z0Os"i 35w^Ox  Uh?Nx-#av q&xH)PVo !yU[7wJfF=TUgr]TFsb˜F~HUx\gg>Hw ғzz1gERL*YWOS""Y*Pj@ b1ⳑϦjk~Rry]P):yɍ&br#㚁&/dȽ]/^qqbCzlg@{j.o{Qq>(؎HHHW֩CQ>'G6W{GguSעV JD'(^aI?|jlϲ}7j[uwx49 ,xWMTN̄-RɈI.(m%+dpmHC喳D"59`P!v[{ t8 cGGe6-gnЇJ栝K\軔n9=HMDZH!yrʝ%Lvl;_/dm*QETeiF$K, eX+䵷Yk̕!G1t=#3.涩 }v6ó(Y4<^~nFmYașƚ聂IaSĐ|-\U qz`Hª!NR1AX@a[+B(dH %FeXZ)@*;:U}kOԂ`9"p@9h( !c²6JkJڻWuzӫzAnU)v` GJ]&ʩ F(bQsN++z]m}GtնƏgmȟRc{ടӅ1 ?p~[6V]v#~eW m/:ۿn#Qed9noyuwE{m=;&^4?@eVV9XsˡޗxT%&w#w?laxn;s{m͏ Jlu ?m (AZ;Dߵ&^0XӖΟg+냓l<[ȳ=;ymJFeC[Z M|w{)SsRfݠvQ5Q 6%e9.,cTN!'\f2z`g{7ګyq\;b+UJ΋<)bo&D6[ppt7rp`sJ1Bd

#&10$:@>X U{C-g2SCO) ~}9ˮ.yPXugpo/gc/ f*lL׀rAMVȕ9J4e$TsݪP8 kX2-cA*2 /z/)[lֻj ihZb(l% Vq v@ZfzJG?xN%ֶUȚ{N*j"QbΨb0|4*AǦQKB=VMxA.TmOt7( ݴM^>f:TX@rqtUWA)$fXf/Iq7JpY5w岼٩`,R r+y(DU+  NjLkŋShm"U Pq68t !/8t1z;k*2UTD˛Y&$2o,dk^2ڧOL/[n[)?ڷjp^Pl_fwvx{-|SxITx(B:!bN8/;٨D QdUm 6@lN r>ɿO>|g=Q87<P0G -ٷ߽!|"q!ZAOt."Z:!:v)( VB :l`DC:8/JzɍV:ɆV3Q"$s9eH5"+wЎc6 jϏ]!R ^[J{}{k%iCպyFh k,~g;[m8Hށίz|x7μڣP5z1+ =mG3${BR6OE}9\Wt(Ne3wUSG*g xg0r8Z9[kAYι:c RK@m6XS0wk )Е-A`$sk#jHڕTtFcaQg>9D:0l9{DzTpzDB+ϳry&1 Z/,[?z, K3mx5^֌But+mHe < 16#}0I X[`w>%>SBRE8JM#"Ns@O2_ycSc `dt1`@%G'fh lyy3BడK !ۉV Sͦp0-7B}Ű.؅M<#W\4lD%G޻c)L7pUBYRbW"w褥վ#m͉sVh4A?yK [/Uّ@&>XoqQkY' @䇋ێy^w8CEڝjmj2Hn~Wu޴怶xXsIOBDw3rsG)7{ ]KYNC BVphQAjx^D ZM0"xWO@A.C. }ỹ`>`.y mPܳP62q SK^]dJ7 N&kܳ 0Ȏm:.thuqJ69hǤ営nܦ/ú-wx|3K_nHXɯ=ͮU$_l06ikcӀ{RúpbV&jY v[1ČK6Q<͚js\uXμ-T3i]}ߦ*CN6 E'ގ6N: rqAorpǠ>kб!~cɗg !I|q rI>dTqĔ;kЬJH 4US.-Ƿf.(>% 3Ge=P.pbQR^"MTiFQRS %#)-B1m/F7 DM)@8FҎY | \K0K#hK:xpZFNܸkg\(_%oY=747_EQNhӣͯ]ֆeYLyv w6\#mw @AD[ИR'9Q*b sI9^r׿K 9+TDfEwZI%HD0 AP$@29գ]`T( ҈@?{0a-c6$ӈ( Em`"M=hTb9  ?ߗij#HFDeY4:*5!,FͣJQH(5* yh4B[AjJu/kOT|wqӮFqicd.MR~ڕ> |2qKbb_4):Iݜf~]|j?+o!}̰Ԫ( pf{}_QfKdYqӒ)DR*Dbx.mx<xy:տ@0Ɓ{eM@ZX.*҈Oon=ؓK4|)|fB]&2)Zz_NOgxzrQ}p6V Fk4z_zF۹Ml0r=)#M h=]6uCyfyB`X'`bA&xg7Y79LFZOצ*!ɌtjIs`#^}=0fImc ˓qyxS7(gwu w/Ξ"|?^ݳaޝ}O|4 A>wa]+w9m]Nߗ+ mDOyoҩY$,t69W` 旅N{l&Ũ~XF ܥ/T"oAxGAqM{/I\C/GN;7/㵄aFiu&eX:@ G` }%;GJQ&O7Gf|px=n.xdbpBFE\ZUHdD1̲K9# f3wFjcyQEx^X*VFk" J2SE VIq8`h΄EHZP@5J݅}]7E:㉅MrKdgnѭ;!-Xr vNOc&bV`CՃZ1L0egK;:A91GG'` VB J| (%#- y qT] ayΠF|f: p,2it^!<zɼ]2^ mٓIUBYr֒ik ^>?mIvxqk9 @W 0.v?zVTP?% ~xEm%,ZuTq8m2PI$V瘿 !U$B$ٛgӿuL5Q`$ E~pOf=~㠖9mͳK,Oڮ45)hJg9w{N3?",MyY,]]TUlj#XTmlRl~I9@(-`}S1&>j5-([0HXͩ6*B,cύO{1JoX278Ou9sH[G‰uXTEkUwmFvNۂruބͽZ+IEo";krZr;AfxjYr.`9OFW8xmR[6tTr+Lo&.st-u8Цf'{trr;fwF]Ot/u 7ˉ 0qʌc.xݴƕr2;h#FxUZˌW>˶QEǚYWa)Q޽څvzm^ё)bjI \>0ys\9^fv.Wx_^0bΉ{1w\ɒxzn\Iwzhww4`[/6ͻrhW6wZg| nvM~\RiOxۼ#VWJu&=bm]/vkfn *5{8v+!ԾFD8&tݣWkMdk&fl;0rv餱Wӧ\D.1 <.erzEFIAsLr3NA0ByZWwd(N"{X/nMGˋBìJ_.]M.]]_URjhuNWw:FIb5=XߋdU,x4+7`\GBٯ/+6ҩr4qg7t|UwL^Zi;*3,*ʶm[+#EF3eT]v)e?d?>_:@i(6.D 0UP`)$Q1gQ6D2&AI'VF*#lJ4Rz$Yǩ"WH g,Yr 4x,^L>`¦Dh_3"=sɇ||~tڎ zHs31$kB29Xn^jS"|.X pǼJyivGS5sVwK)ʝyM9yjEM,3%wa!4?*VjV#&E$rBjCh\@AFBBF0gw>cF>8L^%,o#0'+@91cLk̡2Ė+7g[ ɰԅ5nCP=`k U cJ#aHm ^ wE`[V읳X7]ZÄ\ui~ry5__98X}mNNG>l@Bw/5-}6oÏSiXtv{:3x06w˘=ޞ2rnk0ŕ[\qKGE 6>;cVLE`F|USp VQR`*Lzag>L_^1Tk%i#j/qVZju$#qqh0gҫ6Z]y:@z|Y7Wg9]MsQ~ȉR4gKrYn\'[5m:j]b'@4PϺ5))-0'FijbAaNc A=sޕW T z^˹$qܖő-)KUt HK uNk%Tq.4k1jtf3֝lFgZ{"xK[o|)yǮӿhhpq'8ϑ` 7E"I4kSKѰT`&f Zu1xwD(&z';1ѻX}=!k-([;VN`19%r9PwkZd ArLQd]4g& X 9˜T%󐭝`솂y3%nuRY{,k?oX!7ÿm\\#}klt `x fUT*:CxcA;moC{mVlMF`a7nNn^L@Ά9$./64ӒOB#9:-/: ._.{X#9 s ^d=,t]ٰQQl61wӆ!v+թ'ްwwPNv=fFO!Y)rLݶ_(۴8|7e,LܧY[-=Zֻ..g;yt5-K|Uel{xK)O&m~ƻE.HGt.5p6MiWŐU4%SuYbt_϶SyވHd7%z %&0ЪGbyIZ19°H$$\m2?4p0 |N/-B&HFr3hzZIP$O@' vUd Vá#\t@r P_4N }¢4i J##fb9Agt7 0}F&-;2/چDŽ^nm|)<|jrO?O7Ү̟rkߖ<wM~no?/7ʴ?&C}=m[7>nZoza~l:B-'~!L^.W?_ʘ W(KS*!Lj%(+SɺFHϡZXt-BYe͙l]Ǧ3c:39{WMTR21hO*g Uۚ5U&$`6_֧Mdn۠.*LsCC5IKXUd".u hz^b*F+nr!t*b7X~eA߻vy6 ˽n~ W/eme, ֋`{C@_S*cy`xwda@ [ɖ ƔOSLL^||A* &/=X66lF 8s}r˹m;_|+) 5, p6ayXڰai7ma0lC 3z ԧ-6߼Xs=-pd Kpd_?uȪ!!ǓB,)1O8*FVP>Uemj:I9]dht *\%k\lyRr{P==z]p:?gCK7K/?ڵttE f;nO mgs]ӢuGXK7>;i=gӸMk-&|)e~1\|lj@v[/oWML /Gn9fG ۈͷxgu\97ϝ尰s1I YEɐup=ڿ8h_PZ/xL&miMBjFj՗/+pH>Kʞ_bmA.Q)4&u` ?ze 2v!r^ 8DNdg49'q 6 ȸLL9+2U(:C ٥P%HrY7n\dِ1ՂNUly.2,grR))Xo3uW;#UcG$#۠54$EPRv.Vv֭;{ܮ}fR2P.މYƑXZ]q02ε HkFB5fܲ&~񿜨==Bl`dmKC$Z jq6}+ {{o@@tsqJ?8˫Yc5 l9gk(JuHDkQ ?p8&.&Rw*%>TJ|"uC C:"]] Bb1pFa[e|rP ƢŏamVDo߯EN l|+=#\l kS}qEjd!9f[*dk541i2͛gApJH&GJ6DWg\dHLsoj@ ) @IMDWd7N%kȳgs壼Chڙ(ժhvj F.J-x M&F)U؊Ҹ\ OHTElK*kYH<,$6ztknvحot U滣|9;\CM(- *Npzi}RY?JeOTeNS2MX 9j.C2$zNձ1=tE9mu\)7P=KF%cl$A~LAj:nٍeЧ_d8׆^j˨Q/ɰ^·c@#AL/m<"1~Z'VI^eDz\яnE&ׄ, qQ!s.xA #\*D$T4oER3VI+ @͞ڀ SJ:PbTN'y"F/vKoiN{,9`K`H.e$`. W-\*ފieDZsDmEBT6M51Sq(skFwRE<[,}pK^Zr2"lD!V*``))֥ĉ-z4@hkvI ؔPbZMjq|5bS>E.HGt.i & BUR,M ):i@g|`LߗixF@%[);PN66)Jٻ޶dW6wKf Of2@ L2FU/`THʎ~IX:)LNUyJRV`,(S)!Bױ#ұImG~еo;nڪn-Πx wқ'ǻA~XՔ [ .A^ nYn?ܺ jop=˃LajxʃL U*`!`p<*]7gq4

TT*7.@*aN@VŔڵvb]Uw|$K~ŵ'^pyx>kҒcRBi7%mJYHN c{CQ2[b.":}ׁʾ" +:lL%0:;Vx|lkY TL%o>;.4Cױ^^/fZbɁ*U(cl׼| g8NBT b5״^ӞR^=RtH"/tۙ{lq` )ቃ-Z*P#B &:(Q}MD(l*QP(/42BvZ iL* F;=|cw竂 )}}yi[gE 3޿Pu'vm:s{s W]k' KxrK㯥yt;*]Ŵu6 d%6wj.J lo.xK}O_J>2-\Tb.>\[n\1RJV`IR*RQA)'!뎄H<]-Tc)f/X3˶8b1 咉ރ1)@NFf }ipgi$UkNKI#RBBGJX"4k%P>N EAV6pP=NiVA\F6ƃضSX:4W5I|mi4e2,نRҐS,U `EauDjt6 t^/=C?'ԔQv5E$uNNՒ v>:"~B^?#V I}O:,~;$C J)*)%Ћl2TH2d-^5JD`@K~d\L&Jn:&u7>fp.dIAdo ï  N%LYͰL >K׶<"ܷI>գ(_g"P+2S*R@R$XƩ6jIP < $H." :$H6(rcƎ*GgPkWjX#7q-{򳠷@,`@ ,t'z9 P74l a`̫<&ȃgУQ=/CHM(U AYi ƀ,т hHR=@~'K.~v 1݈lK2Q5[ Qf9Í'E,xg>LF8@*%gI=]؀ Րal%=F_Yh%:ER!\|Nc XŐR)Xjc`RV)@&*%}=$!rW9.ctYwvC^[+1GVN¶繡~>ozuqnwX]CAL;}&.p!9ujʪ-R&z䖖kp.KyL ӗK.@z6Mp'k`p `J{HuHg R9/"e}k۰J2we%  L2ޱ ֋nfH/gaL9jg15)W]N%8 6w.aՈcfyscAzi=;R?}J۴!K40p]r!(u(!$Ո¿FA=׹Ys-f;GT$k|iKKK䍯PB7$QȱSNN%WWw㐶d^ytKܔ#:S:PF>>ӽ]yp>M"6]twJN #$ 2ZQw>@X0@Zca OCJQŔerecR deU+{3zYR&(}HR$ $[(.Cκ fK$tc'n}8r>bnp%aV^]uKEJɊ:8A628h)t}v">\{/j>+!jȇEY bU{J&9c%炱uRssJā&"zZlocNYxLu'sJΏ)fSG?nx,j+6_,EK@0*i@> rmmT)W:ii :ǐ!FWQ][P &+{TfʮKk {ɛ~;)kDFVl>8}>U fpx\zGګ^# OP!dExp& M ?2?iGOP0C?RJLq{41.5 ֫,lfL[EzHΞ0L^2̳=춺ը:Ęf<6YrSOtF ʲ7̛֢?28Q'1Zjt$ H|{x?r@W4ikY)$[*N.g:t/hd:$Cl"#THbaqQ8¬6Ѭv_Jn5?14@ϭc2viKn&y3xQ`7Zp:NO^ew&C߯Nȓ0M{75Exu1[Wݿ~lۛ΄Ī3 ?V빽$ƣfmL[;bSK`7.6567fn /)3XgZhe96/g7O[]rSSS:y#a$֪y0.l9Ϗ;~nFoԲLg^h _8??߿|{{gY5h}Su?w۟n U޲iaM^L>.Ҷ2inߏϧqoY&mO~ӵ՜z6y_n_;zUZV]BzWbud3>h ֿ'C=<sV>:^X2>Һ.O@hNM bp5 :E0IxMtAu~ӳ˿qo{C Qqd F+E1AK6 9 -`7tYGHF9+%mE a5*>SPNu$M&b0`IGP@+Ng2JXqr 3 Y݃XZ+ePN}4w *:Kj iB9~MH TqL2rm118(I"2$t@L%UNat;.dR"*ȐӌWhe *,#i UA Ywv$^'#dO#JV8v,O6y-|SOU}BbnwS* ސJYHfcҙ2QFLQlۉ.+!Xu@mǽYd_ -{+=f po.0Xhu񨁳:X @\ 䚐 4hW^v裿M;c KHJ*2%!)eJQ) k;ŏ d-hA1:+RƢU*#Z` !\?v֝=~&M_~{q*d}z99a͘<"onsl^:|N7[8W;/- }L %M돥^; |*?>>YfC@*̊-fЛԛe<} & e)q7nMiZbs7D5#O0SeͼG3xy+4t^;b JJD/7{p;Օnԃ\jk|F ?`m Wuzn|Vv`NZ,}z[4ΛynqYX1o}xL'Pl>V7hNVZoM[pOiyu@Rut6Y._3AVśT<ۤݔwcJ}= 7Z .,;mҊhNXziEn%uX@yf(j:8Ny-4X}zydPј !j'GL6c9X@DV|UXd1YyTi=(]%0Ol<fPW=NCtꤓoE 5njT{ש+^ͷ1)n~zEm;# BC|6oXNؑB֡ ~s[\=xN|ͯi$:J#߆A<r,X'y-UW: &,hgqzE5]pc@nvy1.x\s._NB4Za4tvq8RR{weF?{=h /mdB)ǂmBhȹqr@AhTfXKV%}M.1P)Ԭ:C J:C.6 흅X1$t~SySJռPc$e|fȓZ52ifB S?b=ga)aXkyՋ/+ő1J潦BZԍ?k$ YAc](:]6yY8'f #?E ?tY!b}0 jY,0:J"Q3J/ٗ?]HcaoCAⴋ"84w糛!KnS:š[tn獪;~R)]JPt]KPYÞbFS/ZV@fB3`kb|pQNxjQŔ cPmT(Q8Yvɲu #Ny:I X"4:}`Y||h-dʫKL tv 3;Zx$Wj%PXIa54żrM^@kR,%dZx<$#Ę9~3Kp[ %Ј|H\Hq@] "H+2*t%UN$] ozQ>ZNVFAGKXuXeé,jdJ2Ju 8 et!X4uU=Š!C8˃[͒!Ϻ?W-V5YpQ$u8.]M %Ežo`Fܧŧ |l҃Yo=`]=2؃s=,)(e @oOV`)B,h7$]*00}u7Rv^~}+n86h|^}@}'.{շo$V{{wdMl}7]v.'0{챋 lkۖF+FEv@ZGBZ^/e{@YCml6ܭMA^esKty9_|n@閠<..ov6ZGs}99v(H(}Ή@jJ1y P+R$Cu`m)یdw?냽&r:{MCEu7h"etf%( ڊ@b(6SNP4ng+-IJQ|̦$bd-堢)GL"ț gUz{Yj.frأǖ._3% X6 LM {J*.E{ υU޾]opShѕ QMc'hKœJ!',5='+RTkl8;$mD)G$㱶7~g} k {[ӸIzta<>6>^naL5<&RZU-c )\HL/ |oKEnn-<l/ аb HpaΎ)e;h%R1ID YtlбcCl@K_:9P*1m @\I"б5 6h!5Ykb $7Q[gpyYC:qa`$?你Nx; (vA!y/C]?:kĤmZ8Z%]: %SĆAt)BS çMwSq<(df}.얮^xۡAnڼZV]w*N oy76YAyUpv+̲v6_:n;=_7nQ jJpy~UsrﴅvC=o8|RVz +\>'qsZ} _O~LGnCl~\K}؜9UO|&c0Nd:PwH҉UЩ a-!xv-dZ) *mB D+Bڨ)29rѢs2dC/d1k=e$/Wʩ ^JGuZ5Y쐮6^|/gG PrBD2[_Qub*U\F )fͿf mqb1*0+%!'#SmzNig RREuX,M:JxD!w QxI[(A;|KY"['5+0;煶]AX0gǾzl8;YW|j,AP P q`֗bb9^'TIY *R_SД,hkVIꜜ GKIJ;߷&{oo PkgS> OG R*Vn!@e= mhb8Nr3xa+jF=Ou #]ܴ+\&Ȓ,R,&) tBPqY” T@Qd?nmk"crukdNZ7*m|K^#T ISmԒ6 h8I*W;)NfX^"iu1קqorDM8Y&$ &GJ6@,A3ì`wdӱ7h2[5ه~%yTcnW>O0$BZ xe1}_`D ֻZD#XGEdD ?pCND4)H!Uxg,JpĞqM|o[=ee4K7%vtk『}4_HA d8{Z oJϬW:@#A*pUt0pK8- ;4P"}HM]P:wd32eT˪}W}q6pV61EaS/& zWzզ]{kzt;#|[Y ^n\?UޑJ.fUڦ]Kv QwuHuOQ Aŏ*9"gyZ ?>7;%'k(#% 2Zрw!ATԂsŰ P0U>*,S.kZ KրM/Ays;d>JIJ<%`E6@(vHPpv80/ -mbV'wmI e7X!@X8766k K OE*{ua6$ӈ( Em`"MiTb9 ,?=?iZGʲhtTTk# hTp-:FBA WsGr/H &YZ[Ùc2̳m QY.ST#ڸqSe&"\  L >|kxR.g EG췟?ea%/CVųRt0CV%wQ>gsd} g+pR`  ktx.C3+۽=@^ΫgO%P0 *:mq^YK URoa:Jhs8(F|7> })tG66)ZvR6 .2^}ƹh}4Z١yj2As vG46lfLUlBk|xz<>8f њ 1Wkrn۩ Q~3~ C'&{b|qOWMݐn$ifY> f0b^]hT4zM6U Itv3B-#cW!,#|~G?_钗 ˗uzu/woߟ^o^t:;ӻ`q30 .Zo"AoO¯ ^?v ᶺꮩbtu5+rC{'3q[___:n6fV_A̯ O9UJ#*w 3f|Pl@ExSLmډ{PzKm-.㵄~Fiu&eX:@d-~ք%;GJQ&?4fxs#P1~]Z!^ntɈcSerFfSGFjȎW њȂ GGTQDmNw-,?$˯뉱WbI)ـ4*f 5XHaVBoH!j4EHaDDL `bYeck䬹3~'I8,sSq`WWLHl 6s3]mSx:y !Cq&qA2-8U ,0 XFnMF1VYE 6 ;6rQp']n5rVKB ŦCg/Vtr>YOQxKgLDT=M/:TU5ˋHI̩yr&<06]G#.;nB~5wI!B}T3]g͆Z@:y+)=.aGQ{AM˦kv͋2Yln|})\ysڅ^%~$uUe{NͣxHjJ{N1x!0w^x]Ci$Zbk Ӽi$){\ { g:zf  ,n߱L)tVb:]?)ƃްlAv}VJR" 'ҝĂwHnԔjW2aߜP+R3EyT&zG[,tm9YeS<7̙C:<Ncʠ6WLQdtn ~ W͂D(5;:[XnGVdSilgQӽcza җȦ֘ȸ#ҏmX^1W)Ib_|"\JkpP}rPP.+>#`|Kn#XG'C<ˏ ǃ=~نZJj]uuو{tnN \GKG /;1Y *$N)bxЄlMZK3KBYrnSs?mˀ1T)b+t@Hm <pSMSc}TSt&\̜_.(uq>ϩt},]rsfa =ߝOѴ޻Ew*Cۚ 6ݜ{K° "{ԔoKylA_0b›lb涥.'y2-jpϾ%I4o;^' ︑^^lB֦Up;oB!1c&ն3@cĜE NVF9S4ۤ' }Z}`Ay>|W|jڦ5,i"Zyj͠Ȋl%];cZG&VWdu0| P#j&(JN("UdH)zw ?VU MQLI :p(QHYuVZ`14f,էEk1Dk45[!-*5eed׿Jnx)FΝ 9IIvA2f& ֢f&$xAa#!7ƌ sC#1z aCA1HH/El Vp4ffz%0d%04Ֆ .@m(L'e7܁!K3$HuScT2:aRw bbR BPr V:55lC}eat5.[h&SitKP[В.L;('􉔎IFp*% 62Aq&D9=gFj| V:>'X6a%QDj&HXZ L"AQ)/pvRʃgZ{8_o0dQ؀}6Fi.D\Җ|q==\>CRdK$.gjzS-&f =jX'nLI:Ų '2S ^Hgx Ug!HDh5"fB4`,g pPjsAvɑWmS,XˈO}h唱N+"eOc- i]` k K4К5''tXVH,4LkTRp0U UmeVN h_W[ 3]$ym%e akC:[j:8m{{yk_贚0nӮE\ӘIЭJ`b+Z`8FL-FTa-~)v`)jYu<.֐Dk sQS9)  `w3wm  xS9njl> -$ʮY.1eBAt GRDYZ# RЃb-2' [W*ϕ+"ኙ$r i&xykeNv(h(cx)U,JG52⓪`btt]B2=BmX)+GϊSZ]FʄFjmRcuz@pkOQ˃>0t$ƀ?!\,Y#9|5(Ws΋noƛZ#@KvGQL!#rI"VŪR盉`ebBv"JɦJID a:tbISlI҈565)Ii,?uj+z3!z c #W;͢./DdPAT-uFWc*%j@C)&"S\ݜ:,FH9VK8;?- + |UN6<ӕ3n>U7ۈ%.4fEz~'0k0%ia7/G[xZiVjNߣ3'_]OFujm[-a!}: " ,1D10 Ȣ$ZOzo6!(Q:U6 pl2w .U%Unxi}Asw_Q葭A߽?}i؇uӘ|[|&ٷM_Y4Vc78w'AҚ?{+iIkNZsҚ֜5'9iIkNZsҚ֜5'9iIkNZsҚ֜5'9iIkNZsҚ֜5'9iIkNZsҚ֜5'9iIkNZsҚ֜5'Ԛ{\pǁR~V츃b#k`b4y (!32mGOг}~]O9bTOeK1 ^`dӘmZ1g@ Eg+lFvIyx8r]Ŭ=zG1k7snsҰiXώO.W^yޖuS^~nG {ڷ?r{MQ7n Vo]~r\^n|ˍj}G?W^0WA )n T勩5Fj:k#5+ǢHHmwz}Ok~,Z.xiPeȠ|c:jy }B =ʫٯb_9M:,Pd9,=% .`ۗ>|ܚXz@m WU*kkW#ӮLndz9Vczm^Z8Cqq9w/\gFFp4Ki@jȯ6;e6y~z[k />m5hkvp份Eos)dK>$+lɶ#{LJ+B!̒sn:q;Gj eQha4YSO;|*|d. `^L m_r m L 2ժǣ{3rg:JP#zCiMm6yyz3[-~z<,Κ۱vrɳuT[NDo틋m7Fˍ͏vZ ώV{^-/qg0.m|a(,\=M`!:Xq9Z.Xij>_ٲZ{|\{/3O|x'p? ՟/~W߿½yWoKd?`JBR؛2?ބ<>__}w[;{t>=z3f}/'Hfir $w\?Wi<>9*?~5[\#'J_٘}j6 iA %T/ v޾J}A X)$= JtaHK7~bP[t()`jbJj@#!28cq~t nP⾕(ʾݯ)-U6l_m#:>J(U %V$s[d<^];.F`(f݂q'4}߽N}p>>vfwеL_\vɭj)&o:JxmYNӔA>  iҖS{/,/< 7.m.̦Ogד"ݨw?Nrk(YI[pcSQ ߐ(K*̒-+zSzPz|QU[\M)XcavL[[Zb{7I6KW[jy >÷!|\;zk4(fQ)Jd]4t^;bJX*:l.?uL+ItP/rasmtm P.L뒕[Y/n޶;Mj,$%4MXEjԉZh|bo1к--dޡ`ϧmN = 6L,MSo`f_&G[2%}0圱`52¦۪NgYRMץ˙jt0{woGN,=ScͬcuWëL\OF-6'ߟ.5]\stu.5u&Tl󱌃(1eHN\Pi797'v$pPFSz*UWrgԕlQtIS+x1QdHOc 2hiMxsxʝ&ZT#`JM ?VꖔU31ׁ+ިS %8C,PTP>J@,:cԎ6%CuHFΡk2<ݸ mU-Ae9Lˁ)/^ qύg( v&ID*uD5H[3E w~}cZg%mEgS$HofsMQs]KCC/Jeٴd Sntnr Ղ 6TM wkaPM]r|hGVo 7V7mp7>6l&\wBq{Y͗3parhO\TߦmZhi=~7^?=?S/oM欼̀qPCuX6GŜeLvWkgfi& Y"]|?`21^Kָ8=&9 gap:5Q-Ǔבl v{ >ND3!S C260gLTeQ3x6qvhI urGɯf+kW͓Cefy`2z?m2%-,Z~}ۏœ4GSkybɹϦ4ϯ_ \V{{6rQ6F8 \NEOWFT^ۈ #V㬖1G18{$mE)gfơPg“SO1Ӌ[c77Amc7|i|grĻAcbE/I+:RVYadXG!1mD阤r&$5б 7EӊM;]@)vpM3͟NCL7=rVJҶd%|d%,0'NV"%CVm53YW;k0 dTJ*YI+xv 8R kR 1!Qx04Â2Ja#SҬ+'K98/Ja.m<3w/*ڔ8?IWߛvsfީ[~\Lo;w93bѼq4Q1n ﳞsz^j}>-ͮ&ns~ 'vco§`ĬKO|jtsқie8+Ɵ/'kG ٛȭg)G,lL܌:XΕ5 Z֌VKd1͸V>U 7hD"eB`ii]KԑȜ@&&m$,FI+PR2pDBJAY&D4rfb?%`CDƣE|\"9C4?;_(9qs4ǘy.Ph&ADiႋxHϝRsYI+Msk@5tg︲FybaAf5zõRa--A8!& )-SXFTCEC%Dh +Vl%L4QAJpહbHEnEFzM=ѫ~j6e*&n1 s+ s *gh(:0LTb/*S>/5{s@O`(R0*xcDV* %)(GArZx%]+P&.MUnSX6*1=%F jjEH(s`%pd1*mNtk^8 ! ,Cx7-#2XH18xDȂ[ 3 8DmVD@QSOxѝ9m|"XP}Ĉycu9M My~_E)dPMΔl&VyA͕eNF]$ Zjg9FR8rP!7:*'S9QP)RL971Bب Qـ2e*G#R]g89 V RNDd)%Uxψ7Y <,|P}tOe2α3#qi0$a/ȧ7>'a[Vf/\5fߦ&~Ojƿ f/jvw&xj|ՇhU^mW gV0Ç\3QW/,{0wHK&}鵟gQ)#fј!Kj!l 10ҹJJOr'E@9b=WmܳHCT.%d LEa]ˑ9&r(ɣȹM ZGEdQ0A[J{J!RHD#͑K,XR4{Dn dur:Rr05T}4>~>JD| B8 DI\.%"NJ8Id_aDLC_DcHRH JR*^+4W"?vI D]qFbt\b^bB+.\%\e^JJus\r.+.@UX\JJy)抧b²*Iu1W\ (_LХMv*muaikV>aon kN _?Fɴik /V]3t[%ۇ? zL?LzAQ3ݫx;20rpTa4_x1p6P=s3'=7Qp N1'~|*\t9HkSl_.Yh6f6MV1SRO+%:@jFvJÃ&5슩 #|9 I\z1#I{jwσj)&qI3ۻTǪfRZ)kꀃMX3~Lh1*xH0Ĥ*aL@*IN8bV]zH+z1!MB FOLR]*у"BRX)uJaRX)uJaNѳRX)uJaRX)uJaRX)uJaRX)V :N)S :N!'Dst)uJaRX)uJaRX畣#T(*JRT(*JRT(3$/B̽w tBY|9y8lM.`ɼQ؛Vt.yOh 8 7vzoڴdSnqW毷a4F7݆*m0&.G0>4{oa;]^6 69A+l \l?tG,7=0kyPbP^08& . |B-ڱ\FmuR RQ:9T=%EǸoɀxH\bҵV:F ea;T'墫t 8$*8ʝ7\ LŘ8P޵q$v~T 7v ,K aT8mpUϐIqH=ZmPLꮮ b^KT@A!T nV<g=gur77Pmwg7/0q/QΫAf_?7Uw Cgz6en+;y FcqV%,x#+5W9XE~:k#~z\w_\*fM=u̚z5Z)d/p\CTr&EǗ0B6Qx)9}"Ev%7!mJ3|v>xU#g^vÑ#>{aj)Lm3A Bo"*a=^ajV;))U*D>rN*E@xP5@s y1ҭ'6&w=e$6E|OC>O{۰?\3c5tFq ɨxp7Xowgga\2M$+vISr!#RZ[ɤ8*yRTVe;\ Dy<{nc^ȼZ_Xz8[75)bk q?%z˺b׹qq71Yxu1#h9hсB |l8yεjta-aGTI)G]3TGqdO-I\qztH''JyȆp6'6')JMla`q>7t6JiѼ7V~}2>T1B Ό9#V8FH(|i8=QmYaLփy^b?_;]ޮ-Ȓse8NY;\"`<x9kbj"-'wo@vbH]ðafYޣІ84 g0bj1liddYTzJ뼛2>L}IdqL`nw6NѩDwpP|oϟ;;'铷}{whq30>w`N/#@?C폷ZZjho14pEɧOmU%c{-K?^~n6/_-6;?yj{`1oMWly Uߠ fI? Cވz4P AxhZ>)aZ=CkUtbqD~f 感t!JbgQGODk.2ϝ S[E'@l^ĨD('DJDym(r&nJl=B?`1ȁWőKbSQ+s4U SXOYGG7l@TUb ^d2sڼWWm(UnOx坭_&!^1OW#XG' MFe+=\f?Ar&E ès墡Lt{Q.zԪ5Fpeݛmr^n^e"MZEhBMBMc۶e 0vzS{t_ܜD_ xQ:~gv9׭} C4v <."zH*{9;k۬ç}3yqt 9Nc1 Pz LK5:?D Dr 9]eKzCErY#TcPt4[mMrLbLx- Ub8d2K]1Y{_|)j~Ox;޳Y,Y)Zܳ|W+edzH%u(c+j4z*z?H}~>d?iR8]?UiV8QLgga0v A\4$JGs7n'7H待;3ݛmnE.arϙ Y[~-YӲ*0+^nRn-z.tѶ) 0ƟjavJPYnj>$rr^2Hf/`>`nxֽv3 V/O,;N`hxy0umW M>dRKO 6" Y:..FUb^X56Q}pUdY<v6%h4@YƹXAܜ;6{\oxO`ayyG&ɏ0:&hj?{/) ۄ Rl릋lGj@ s<μS*LD tңMmnsp~+CQux[h ]i\Q&<o2jϣ"h^Ϭ׉3O 2, 3!+avuf l:x]uBz݇?S=H&/^Ow\/TTzVTP峝Jx.W4-VqCဎ8C#\%Ofb(C Y -GTN~),I`H.8bBTbs\t٬=$OWIolVC} ZA=N#`l >Y`tb2(_x45F~ r%hz{.)pU9JA'U4墜6yǍ4 3[f-X\1Y.¥uKġ kU6M"PejS<%C⭑%OE-yCHrA|GK ӐID45z 3oܲ}1_|@ID[cQrև=}XߖCN;K0J wIB*FV|*nT6Vގݶz@Y JrJeyR9u2M=f>X5Aԡa~rq5ظ~;wC1F;eeL{9y>b.Ebh18o3 W"!(OTbvW#^?eD.d&]M.jetƩ x&^Z.uM2b8/.(ÈJd|Is][y#88J'Sl?x9E4JMhfjCb.'bV[y`<^]L]dV_ȓ .Olzj"kwyk,WT+ߌRjnȴKI yZs=u&"x FE5*{htH2T=k 8l :%*I% TiX[D)Ϣ q}u) M =2/wio6qoW>mʎFϣa=56MC|FX P4GWKYEbD$iFi66X ^\8 $q&I{D\45DPٯdVCŸZZ{_{<25+#zHYDHMED $iA=dDЋH#sdD: /XģS-]9a/Ngx(W#׈%BBt[NM?{WƑ@$zƃ#{`coo }Z\KBR U IDª#0I@cmxm~'i;IxIj XURVmv|[\O ${@2ń)XdWQ )L5^ߧ:"~^lJǑ^y>yeo/"LE>zzk,s@>:/a=rQKR; YJ6r 2(Pz)lp 9кg*ii  ʯCLVl֯^RnT<%x6 ㆽ(ZMuX>~j~KVF 5zZTz3w=7?U͋I#*"޽~mELyw z_Zc^~Bm%3ԖYW:k0 d.uy|r{B76;*l\JJhBȕJL -DԛJLT*s%>BW"Q Mj,uȝJQ*QЦD%G;u3E &oJr-*Q+6]%*]=FuŘT>: zggnb50A~%a5]#b2V^d_Rp442x/! Ҩw;_7dE宙8xܠO~򡜛IbT5IP5Xx[R3JuiIwm&_R쭳A-fKf&A[MmsM ein3M HZ%EhjKi#urUKˀY,C2i#VU.ڢxOqXt$>N-b .bAZP=eݭ |\촗;ka)EゕX!s-vul1a4N-^>W9vvՅK>YC}̜7a9{ym8BSۘ)XCݰ-]CX`)_̾Ms`7jj@l{׭<6$+'(y &pcEm(`}ޟɪnG_{q~~' ttTd2NάY( ᒋo~gb63Y񽯭EZee)]+߁;:ύ͋k.Z7FZO L=NV&1;z virf߹1e%B>٨)uQƫFki fk ه3pnCyѫ#iWq?5z2c S;^"MܯAG3&߰p&MO ڧE)am?EmCԛON@U,4^[ (t,HnI&\Vk\ĖIAl6F8Ĕ}G+!9oWs 1}ȟ2Ψvk5+WNaq4uA1_NFԠ?Mްzl%&%!#qɥS^}܂)sE(^ LޗA76T`3a`a>n7D'ǩZ1EFU9_.xi19?yFU'u zc6g LNXv ]a#2JPӔoN 5`8+Dbzg?F0RnA[2z[UmJTN #RW8aЏ}@SX}P|/|2 Gn2L0f#b*S) 59+<;K&{Ȩ >eb1C N"2r& u!gh)솷5[2-?zIYEGbft]%_4/:O^IKҩ3m%U^]4ۥ^zq^67Co/VO\kIKfu_a񟊗B礪+us2-1KZ6lJa^iPd Z{1Qn ̫8wrV9q$Lm.Ee,bCqKp_|> h\)n(~&99ZOsfg"D۫"4<ZG %;3fBr9c)/޽|\*EwVaoSeEWip IUp; ,R3jg>LLj9+j a;&Κqʙ|f%>|UωPf9IiHi(w=ӈMl A|mPS/i9-t=dKJYwn\7Ujzc^kA-.9QZ _DJC άEH(Pʍ"VFudXD -J))$\5r⺮\sp8Ao0'[W5д 6 fo;@-nnL"˭InO_VM*ڕPx13KNۏsn;ęw&g-g1xS+P—a؝yHH]鍯"-fR Dr8X+OPCK"Lӈ(SZ:g7‭iU"bK(F[ :aFhN![\[#gZM2Xj|dMjyNj>3sȎ43?juT:WLu0 bvהmՑ%ϲG)%vUkPJH 4 8b!݊!։]z'rcu3sٚc֭"5"̙f[+e=P0 lRj.k*F:In<&y$LViR쥔OXRҳvk,GW|c@Xt_sZ%J:f Ƶǥ]dp>hy >Y+- |M~]$sZ'Ck11 ZGXFD$jcSzyd~$Χx>(EJ.x ͌O&&#RNcRɜRa0* M%1  Ayڍ *!tK0{gStѰ;=,a-Ma6$ӈ( Em`"MWhTb9 w r&dDTEZb> J 1 sQ J1uf0g` l?~6(\80 H ˥CR*]0ow\g ʒ8&+F^gU '0D WNZt9j00/UU<~>rmŕ1o̮_'Ói`6,03Ӎ'R=)'Qu__ #`䞆3м{: ߍ$sf >t{Y?bliuO&z0s.?-u4yLtN3HFR~Nw1B^Rxn6Z%Өx_5V 9?vˣ^ow|{:>|30#AO¯ F?nܵBQT&]6G]Ox~eCO3mgfC|wv哟t45Ož\WlqV[W75҅=ʝUbc=2Lcs9WfjھXh!q_:^KXX`V'7qD-@>9YjPQC{Njdo')@?c|]Z!^.ʈcSerFfv}{FF̢&zG!Ѫ^ њȂ GGaF$8f[DyIa4g"$Vaxq&ݷni'8EPC [! oz aVaV¬f`]Zqwwʢm!*DhTH8(-%  dwֹ>7s3C,AX(kAT z"[llY+ڽύ%g2DO<"k/%8%b"XZYh6ܭ&N.7"t/g# xP1A.yo+*o|vVr ]p% &S2b1LW178lG´p"5zdwdl~ek +A{%5 |oڴohsF0ڿKrf DLU]ܛpn R3X饦%^:pVQHvWhWBHz,eS8{ q0fGVGmL0bQtY+냉4FDD b -VU n?FNó5#o W‘s;57[k͎HgЕΜ/zg ';='g"8Tgٻ6$W=b\vG0桻1FcqMjIn`FQxLJElؖYb0+/2## A JZ ~b!"$Ii\x|XՂruB5.؄*80Db9Q$jPAKGFs-8-A t2mIkk_/ٽO%h#qEY>SdiU&FjY}Fc qZh  l7UpG)#mG(t{6aP¨M޿/xfՋMdPa(3~0+ C.C?]i^r}^KfZw+ ͋Emո!ar/vz)Euo`6n:i7gxLiJ7zϖ4ZƟf‡Jt*E#vfE",?irz| lwoXfoa>aދ{ bWoٰzxbYDw>{p;ŕQ.6m2qŦ]hf:G.ܽ%e8_Gٗ<nXЌ,S2.dcӰl=97])δ"rE a-_ ܌Ìwyǒ2[(q;^~`{GN]ᬋ5IϾe_YTov;CY cQ~|[i n J-iPb@kN6䴊VAd,l"pQG@+b4{Ph`w(T58( ӋBwԙX{Bɹ:OT{4v=Y͖+4\#d M9УF&3E`6kMMqJGxS)K,(DTZ*L~3cIY*1eb%t֭XHDf^tQ;Ҡ!rF$k⁦3 Z E.Ek٬%~2|= ×6/2!Z!_JF#d<@":>M>RYҒL+*oC_ŷ^ɯ)H,EpKAJibԤBV6d_ܹȬD*DJ\ {rzV1sʐ pފL/1Ol5ٹ"EUyT?GUWxoP$j t0"ґ 'JC8k-zWW^hi P_O_x]{BCrc~bjڭ%obnOwk`q+?nǫX=u{3g }K0Ap& so^H_&7ӵ ϰ_:?_^Heub7#fOB9ܨ=kr:BP]H4ZrFAa`LjBl!x}}pj}'WS5Eg#cVl9N~y{N/7e* ˬS18|,F4ޡtkㅮVFn_ѕa-}l2T2JʉOI'm$<-nT{1_J˪?G$im%*)UIX._P,H |06Zg4!pyvd!JF5"ښcd1K.zR*3z]6gp&E͵t+p@62Vޝ,YfX ³GYWߔ]-wf_fau_ hh4<8bZ<3&"E^o%Z.BC!P7\Z. F") lJM hF)cK/sHTFjGl?Ǎ+]M;Dm~=j vdr6zRDb7RjS0>%P3,ҖIGTInq[Ł"db($#aML9 RAq$YE(N)XMx XD#!L#gɐbY"$ܡ)t0YLV2Uc!S 9fmD.8cٟM:r|߆@#!>"|} 8~a1.uVӒcqQTEb㢷.`{. ¶w}"6Vzt_tZޟg#?S6$D3a y|/ CL 8|hwF;sc YJN]i&2*D,M6q2L)Q!ƠgKO{fqיj]JKy|䴽gaK7K/|[!wSk]!Bb$ p)V=f1bNjÚ0lͯz-ޯ|;Fח՝7Fևd~;d,?(y.oѥ'㍆gn{w(Y믆Ze{C[~m٧Hl͖cB6rcg͕~MX>V*+eiXBӀP'l熹 /|/_~|sHN[sIHZ'yL.Ihg1 2\.( g5*S) :{m }C!HFr?vrK5q8Bڥj{Xa&ɷpJy'PqP/Qq*! Ő$@Lx8^&`-(RT<j29ߜvFYlN,L6OM9Bkh`=[cI#HSʊa A&u W,0mJ|O98YkWlHo6,]i9i̹+o9'@QL"\b}9To5MӷK>[=9铓 U"raXIԊ%%'G! {zތF T+G]RIߺ]s,G<%ژbR8˒f횙='"IV`y)  @=pTNWGů;UJ?&πq&1C^!UUrØ'H&$Z& @8q[[fЛtgjNTScOW6U[CNmŘ9( rSuܣ f B4AyOQR"MRQx_hk2֩T4X박üEЛYqѐ jDI8Y%$&ϔlbl VQC0`hz?5@nY0Ր7p V%Bџ9 F8rT1tR>߬a.=WA ͕r9S؈0[1ZhJ#"ep;,~jwt"J?«'>#fRodY/eL7Ͽj{n[VسgbG3GٛdDy-[ d`^4^B6'%?@/vNL |Q$D˥g"d %ɥ.ڇ5-R*$}RdY)_f:C*hm$aa2=kB7wKڅ`4zjeg?'jһHkS硘\puA}!V^L_".K Ʌ`\v/dR/ )Ѕ +X *J{)pUUpUԼWR&z1pURH pUnH/)$"RWEssH +%+}ApEK.H\  Z%•[LJcpwSk2ໍv[a ցh z[&bk/߽L'H9|j8/o9D ^o}ֲ i.Jy)0]:L)5a%´޲p2k`OXo_Tf_D-[9< qޏ'0g‡>6A.|z=ʏP/6H-՚K ϾC:j08-WS^*7E[ӯ?~ Q CȢY0_sJ<@B)B ɳqLGBj*bJDs4>*DCjKuY.sE)hޱ@^dbM&IXa-vj\c6~YCzwsG/<'|oۋL'Q6zf 5B܀3ƀ,}iϨf ^|3)D2Tf fJ ш0ǔ!*'lfe8 H#vACHE mi:=KERAj?F73d4 _6KG7\蠲F:D9DK(ȝdҼj h":FA w!!c`nv>gFh.Of.x~4:moFw+#̫~WBOP*vQK׹87Miǯeݖ|r=+niz>νG,<&9&-B>]hty{-ȿ{ozwɧb!x[Rvd_J{tXn;~jm6㳶i.ݚ{#C{-v@Z/jw9ҺAcQ.>lq{^odϐխˁTG) ^ 5 9"u2AGZ7H ܺjVc5Ajc,V5wtEHT^$ R,)'0VWҲPV|Lr=H&s+"gYb.~g5Ϻn~_>kO&/M*oBwy?\US=V}{g<b{>$kFk$)XtgN>z]J,%,"k,jF]ɹ @96>-J>yr rK([iu@,Zkf.lՅ. .|R]m~^*/V'jHzen_E.ӋBy~ TрZ.:AJ271D ɥa uLҪMm8ZN oEA $*HZȹ[ctl.Ek7jm%je (-ލZZJxZIz#C5d(CSȘb ;dFxF5xK%u1EM5غ(v3rևݝgq_4b3xF54ҁxs""(WkM%'h\1`ɸa3e(IR zBJlS❴c<0Xӌ5F֋n)\l%Ջ^T^6lZA/>^}،;uc}@3lN'*_ڣ~e"GA7CP8Oz8g KN@D](C}ﶾ}}q&a. _*q@(IW i-10#,ɂ _{0Se "vyigEK3{GK+{P%[IM'vw]^?SocU[wܺm-Zos2uۈ1n_;Cnq˕on-W=_6erzr;?M[[=ݞ6rλNsH'IPޒ!a͚鯚nMwsn}'~|oF5VnM_YHU@BqN'+5wxx 3㼴(!҅} L v  xCp1ElG<虵%5J<T-<e16x$lppIFdEEm OVɹTHtI-7V9Iie˸Rl[}g4qg5]n8БN(Of {1/R?EPvPb@0 6Bj|ӛvٿ7- r¬U^j_$4SpB$b*BlHIHYcZP@D(_٢[s7:d ]Q¿-p/ hbvc^ί1_bKT,_zM==t3QǮ(Rh t{  (^%>qV+o0: ft/9+nl"0WP+$Epn'ׇ%͞8}X&P9&Jz>( ]^Oet{%ҷ_]Mͥg|Qo p>Zzjg9ۉ#6tFչ)pGl=Z4.FoO[ A51'񤌏?-V˵$M}L;fii$GsH'Qqj0WLix*V+>zv9|eTbInu\urIg4V2R6>F2jPtqɿ-Z.:;,X}ݟ~ן>R/?e^9x8#j迾y硽VCk mZz໌ZNyøeK7imr{o'dz4叿'?j.Mկ0,'rdzQZvST"J'W!^1ٌ xI:h-_z›7U*U*O{3xa(яbq23r=2^ |ҭp[^1Pbju _<ٞ)9wug2$1P@%<Dq%`SiU$j0-<o'[/ޠ7zymE6,gg\ [Ԥ+ۼB12U 0G 7b<*XϣR3Xᥦ^:pVQH2kepaBHz,e a,fGVGmL#aR/1Yd!R&b-a$hYe ǏVW|sy8f~nY\5Bt0W~Uf):|ΠoCksrLr A6-8U~GkREJg# Y1VYE 6 ;6sQJBԀQELܙ8-xed<Ŷ͞"o`YYodr|=*6ey) 9U8U6]&#G=ǑjqiR9-Rү ,9;qnif>L_G3}odMӏ@ M}@qvruwuwy vEmݔ[/%䮕h$v9,z(Y@K5%_&^͓ex<…n`*S\NX=>HTO'K7F u.t=m`mZ^."ѠQÅ2@u=?2(3޺Ɣ2DM*847kjNd":cMUgtּr`Q48I)Yvr:$ KB]%q&=wWIZ02Dž] J8[(2)SXE}Io9~,7DuV/ ' /z`!; e'QRŷoA ~_2R{S3vy@A*ޭnB}~`Fl<+23^fP,3ux:#-GE▱ř?g:F-D9a074C+Kf}j ޜ ;*ao{ZtDhQhc.L0&wzri,#,ry LqY,C"NcʠvpL#(1B 'q19`8IKľIJ*s0aҞC&䔙2SfrLN)3^&䔙2M>*)39e>)39e6l2Sf䔙2Sfr29e&䔙2Sfr}erEerLN)39e&䔙2SfrNN)VDuj|09kQ5(1eλQ 5 JH 4TKD'\s"#{Dt#}WÔa/)0gǣtCl0ʣBB&q=FjFQRS %#)-B1m!8Q ؄ A2 ǨU1kQRk RҳC27x;wBpFV6sui>I2pL0ąq~Mbu)^i>=\uZnh`W,6:u` @ 11 N鈓ˈ(Xm uPO1l셐_upë)\w&f64*szGFǤ"9̰4$,"0 (O BcPga7X_XK60UDEeiD"rA J4xrœR Y_DgF皚1$#,ÈQJ wҨc$vyh4BfC֐; .s5,-QYӪFq᫓?ì{g4;3L{'1z |ߚ.]LW4 $bI>fM^J>bXjUܳWԻ$4E#)D0T]3)(4\t( uQ?kV@ʚ\:$UX-Jq LhI¨b||Ͱy]҉%B''dz 8W#U)h^p"?Ty&+'0}"J}#5VҽxZb_˥_'\hM\ y0Y[cHG?^%'>:{7aeDkKn颭ތ$,+G0ZF?QLx4izhsp1jFַtն*!bHic!46>OJjic/c|{jWZ]iRl07pG߿oN|uw3Gm"Ee7k~ܺipWMC{MTm]mڕ]v+]m fn>DޏWSr nr'm8-fM@ăp%w==|3uRܑJR*wKȧ3܀*'քKwl_ڱ]nGK;ݍRVcaNYirͮc]߁r Ep!\ *0wHsIMj]i[1n@<218!FE\Z!Z!^q2"xfY8j3 {eEiȞ KhMdAI0UQDmNlQ,bR͙Iyހdjr:ӉG~.s=xbeΦmiG^:jݚku)jE s^& +|MGV+_O_rBl *#"_Ls"lk%o44u>ͪܞ{8fuS"3%RZQ:4Cp=r -66`,Cb$w*>q2ާ >5wxI;űGϝ>%nϾ3^b,+ bIh ly@y3 5XHa㭣/vgMr:d{S+YŖfM| J 0G1_<7b<*XϣR3Xx ,t 1*ٌp"9}a'RtBHz,eS8{ q{zG#6ɨ_XSC*lbY_*>sQC :P[d]ᬑ4BFe䘑cFMB:B,=qSPjSMAhHAJ,i6j9;e{ZdٲǥGtH7A\Pc"HqCwIGLgW Qݦ97 #$ Fזi$`y;AЗjY4JTϐMH,zfW_W%hDp*&!tCe[,K91URƶ^ɿBC[ RJ&q6{4L(e|b6"IcA:x\ ; V1yeHNŲ:`E)cmC,![aM>;g ßSʍP+}3lIf-j~0[td@YPҐ#O9({J)Uhۃ}}~Zm{2Mi:cz5Xt oC4~47L/>%БLCqFΧW ~s[r,)n'8o~;i͎a٭b;)g)g8'%Ld.\xog}@!jl0xbb]3@ ? ŀt. ]\zs2oC:aW[7,:~ק dqlE-ώ.t2 `ݛGZP~a:qtEx%jM~V*.ߎū'wV# E8[UƑ8p!2( x;S`SpiQzr;iM+Mߢϵ]ߢ6j̣"Kk9 (Yv!gЀO184I1[ zG?wb7~qiZfry+eҧqCJQqN7}zќ+9}Qi{/H߽{ûɚ*cEޯvOK*4^=wtbiSDx&;ۅ~54a5r3+K>F_%y>;eK15Qcpd rJ,K|O#aXJڪf(͋&.Vg˧^DٍoR0ݦl͌&&e~LP$lT$fپH!Fv#chmSRQؐ.%~<ҎM[+M b$t֗Ʒ3~bɓ (Em~q~Pa5Pnw*V |/>}Frv 1L b9 Jh:t S8 Qxr+u|jNe{G[s((UCd΂L99KYF!jFpQ@Μ-c1&(%68ƜF 8MPk#9aGkOwVT&TU@,kV[U/آVx|p/E%(> J.:2"J@ lY O/ CTi](0bj@_7;{l6 /D@ =ٔI!yeS/9J,Vy4B [(w(4 ',ԑBlb{@ɼ8ޮ̬WNJ-P^̨]΅Lλ,9&,–AL |R %R\,EaXIЊ%%@!&zߌ)tFޏԠP+LJIoQ,pWH?l)%)e' IV`y(@!5Qqp~yq( tbUJ WF &1yH<+*r0# Z,^GY8jVoFd"X:b.j*Ʒ,U_J3HM8UGɃS!;z,$H&(cAJG!i'qMW4;4dS박|X%]ɪ/ts$ eIHhd=MVZ鲞:eDR\ҩ r缵SE|P`6N#g6V|~q.W@̂+sypdhA;SHZ Pwag%F,C:FUz4aww ݲ"|\i!W?XtvK  IKycut&W )9ȘK)Wy[x;w"䴳A~D@_gR iǦxK/2IvI$-P:,6J%f ctfbVA4-R*%)T(Qe:C* yڲHц,jx%TR(6CoUVO B7^kùd%E]Vx}/BC.$ef/Pi=ZO{ ҳDiš R R"ApjQζZO 9rizO+`*)Ww!w *5XD0&.U#QQb*&֖iTq%HXf2WeJ:g̩:q_Z>&n{R}|w Ű]Gy 93'~8^m'0؝ srۉ3Nk;vF~YddI܅d$Ynɀ쎾]`#:У":P`H+P)eKocհ|tsR(a>?p>_~8w,_ht2 `ݛH˩/O4*[#6{UqtEJԋG?z+ֻg۞]Kpu~Ln']44!: ke. :# A()(1! ECi-W2R[A9* VØYDQDq bFkfKٝx{fnVm#=j-SSfA  j:?۟ߜ,p:30~:K'GGL%ga"J3bЄbpTq,39f%B.9RR1&Q )51hS`c[H Iw ?{;}"޿{kBC]Lv` Cr!O9sN?\sQ,xtrelùal{ˇjYrH42R4}>R<(q_x~?ݬT-W6o옦_?{w?{υ{oo޿{8a,F]]¯뀻Gn4o~ڻixeҩ}5z1kk>rG?nhRh[evWKn+@R>wpt8\'?r՜a9 5y_n#߯3W4 f‘n~@e"ᩭkOrYpV> Zj8g0\:.] ߲ 3Utdb)2sNh΍ LߨV01(!=qTm0؊Ha貙٨TTPH jYzōT+t3?qw9H˜Wy 6 U!a>$c;HGsT `܊lmI{ HX86J4[E;,fx{>Y+(- mZaRdPbmjM'OyGrOh3Nj/+yjs7 >T|S2߭ ݍv6cJ9ZF )9Y2&4y9HVg^x} ;Gi76O";z0x,+,J o~c~s ֈD\Fv=G:eTA1zJ5z٨lQSfc#]K]%:81{fTA*!b,.&uCB`4 >{Q(d2FI vf]HYQd.Z#"(ugϳIؐSx$GgE'n>v t]4g2k1Iɱ9;~< %OŨ0EPKG,&PI R:Ɛ*s z΍ Pz]E2&jH ʑAe ֝g}r<gauG`ufE5˯Y:}sYK+bORy sv/h~1BЦ7}3]y w癆bH~w>_x<& |Q8zxp._w3g\AH B+g kYدU^l/ЫBO2 Bto.^ 9y5,]Qҽhذ@z5y~e0/m e;x'|_Q7?8Ct Tb ˱tr@qp1>Z]b8IZak>)mry/B iKǪ'4ɎF?)x#$!PxMtIT"d[P]ƙl}Y|[BuHnK8dzpr7>MEZ'8m.>"̇JTThr&"cR!fתG-PGL\y#Ap11\^Lz` hAzc˔ECB!/&t؁*I z.T=ׂk}%!8B֖įK<~uti#ټLs>^ MWOw@[mb\+v*6j{Q:(]aeZ=WVn;*2i?]XimŖc`[ŭN]2FGJ'j_ޕ_ύ8/<8>|`3hb<.r")*Y 'DWl lVݫpB F#\R%.J*:sFstY=O&i;zԫze9Ɲ^5U}w_H 5MxRk:=KF{ٗۊ͢£땒i[ *B#d΍ Q5% JC[2dod?W:`@R0o.XSZZ2s yV\w 3q(AA. eȬ"!^Z1qF@摆30DxMԎnR;2[d/4EyI;9&e&)6Ƀ ÐyRN $J:j Wy8iGiڳg!b6FrF,+&D`<;YL/9eq@'BDzYfU䘵٠VJ:F˄>j!rsoEKY#)%эY81eHB\ꔥԶK glT칭ZzQ0 cXO1`fIIGr,A fee!j **f`nHY)!#@Bb&e-պ[PCRLxA~vq(]̓me໦&ga,_vVևpᾸʝmMt9;Qn*/>a^Kr6%+(YιuɁsCHBL 5ٜRX"R\`ti"&KlL k@62VhոJ5,632 <|v^qfǻyz1^t#C ~DR:2heUp* !8tko|Ls> Qy#fFRug7b(fQ[GvE2E18T)F6 aQf8&Z *:DB]w߬4db\5 Isĥ!HE Tbug7Jҽ bqWD#V'wɭ2NSy Z$#EHM=0-@۬Am"" }6dgB*GhS#iBY LFH:.Ni,+.ʸ(z\qqیOQ[c"$-r܁A&)4K@25M\f]PV;}rb]#w>c'mfnF=a[?~%Gil/A 24Blbkcz3ix=Z:8cmb22sH*lVJƵ*dHk9Rʘ11Եp2*8e'vmk9Bv//'XƘ/y z97bj / Ȳy0Է;#b12 "wTOǣfdDXM/x;).駧c]M_I7׆9Is*b*{v6 mOޠo\A Sn냽^oF iΘ3 1"3@ɱY2a4 3-l8km8}@d2&@7bea)W4G+T%6j&%,7'ޕq2(+"@պ Fi`kYOё5uN3w^(Y;*Fʉ$¦j`v iTș H#Dfhw!mSDD/Gq^8>rdnyt7'7w ,zuNX#Qot Y"a+pV#5j'itD@h,[*+#H3X,r\'Id76N$cH79q.S II@e$gSG9NpnFȡ M|3N넲籡:~<*^}Dzn[9( 1y1[9qb8Ǔ@<σWl>u~`yx&;oK;}wp)xu<O.7`%|oHosIpx}{V](# $G{\:kdp8 oqwsx!Kyx* 7|mC+vݢ,[UE9[kz^j>h;oWjqj3b3h2}zUՆMlQsm/Iw+܍ld||yR_v7w N-%8ZSKpjZZSKpj N-%8ZSKpj N-VSKpj N-%8w gJPSKpj N-%8ZQ)ZSKpj N-%81LQ q?oOOC=Ù*N,8 Qj5"Vz;\w0Ryw9\e`vT7ڽG)>}}&x"Fxknj_:8v\]u/GZ #X2xG e͙nUm>9|,x488c dyi]+WGğK~)fYvMXcP`pa`pɖj sݚݰ]+J z"H ! qv{n Z8AC8zT"/,#T'vX9̩X {_ X]:瀳lsJJM$b*kh1[x`O{ nÜfyu:͗mBDPzjs_nv'5 G\ּK\c5IIO\Sp]qz,5>k|hgٿq QSSʙ_1D-b LxVN%{9̅PSvwz8Oqz=\njMj!XAI4%C)"0&o g'=:Q5"U;'C"Lp )jAm[Ԭ9£I P0|*B2" (`rM djG#u on `Vn SUخS݂2Q"oSsJhM}CٿV61 'ZR6tzLGHlH4#$|0x 6Rp&I.P9k2B"7<0d&6F .pg5^EUyP M-܌76%5l񇽎ϲF{Ry缊XjgT" bi{z敠*X˽f9+R@-83fo6[.ljv%˩Mog('P| F1P鬒9[Wmlb_S흗͍fG /ց1`N( ڵfkFj4>5>/p-wG@HWO ;{quO']ݰnf}C q4n`^,ǣb>'u.&k\Tl|4;}6iƩlcjw:/mã;^*qO}`ޮ㟐⇗??_^~Cዿ8|{8Eֻ`?7}|_uךо]s Z9ߦ_;}uE+mف_܆8_>(01^X$mN~Jjg`1L6I[v0ͨviT1Qа烘钏t*!l?擸/JQZ>! Zj8uff8u\у'}]8jx #њ s~bG?ǧő3o[4QPN D+,sDJDym(r&nJlQ2[Xna8qILt*9 j4y[aiut(iQ.rVUmO|%Uqr 3_tv_ZIʏ UUYaj rݰ[n.P[t2>d\e{|z}( w|oWјH]5.S;l-)a Q.M #TA$[ pi2)dQ"U!6KQKHc@3CMPz3),8,Ypn'N[\GFW+>dY-zG*dέ7>y؞);]sYGn@<*x!yE JcHrxͫJ=R%#=o/;W6A4jMn0fX 6PFj+m1<9MOHzbʈ9EU]l<j8iԃ#2` !U(@$'%÷A [ibE1:$`CT8Xȱ-n&{sDž),CmfGm5tox$OO$:i:xJXoQi)F⃁\X5s2KPvh(Xj#2V[\D=)&^rD<=SpoùY_OYZ{mgOQwfy^Q4˧,/^;\Uo4bG=葊K&kPviT=Hzf=˫_bakМѬ&#Ef 5AT3FI\2|_۠thA7i෹^p?h`/]Keduw ,Ae'x9 ȽTpˠYb`QP0o2jRL9Ypy/Ϸqa 1<c R1`CmHy@*C:Fwyl9*%FK$9 %(s) G%Z55JDkmXlFKs'llA_eiR!8:߷z8 /ilKLs믪EG:3>Ğ}y61Xh x+$PL F+ qb #O)|݅aV{z篌'oIۀ.$IKlk hhMi8魾a#2XS56+A &ڞݕ. 5,BVK5@p%D3-f_9;QA3ˎ'GAc{,˩@g tZ0`WB#drl(atYk*47 tupruMsSt59ܓ.`:"P"UFCk:k]Y!\݌bY8(`n7hA8bݩKWHru|g}Llz~*DH %l3*9#V'zwO>=ZN_߼f^ 7<ǝ 5(B: *Z3R=@ӆ$9җyJ'Ob+4F՟ٌj&R,0 ㎓ӗ -xĨ2spDLru4k, R.SS˸×tWh~} e h0veL#BL`p- g^28-?Uz}p[:hW1ͺspRywIBmK^ǃmϴ^iiP͐N4.Gy7l.ϫΚ ,2iXV`yY,m9eqD3HAIxJ$B0LW%D6&d.Y))Jz1K9@)bl#w;؟6^SR1i5P/t62ʴO%" n(+g\I/s*/w_,ܵf\wt=~'FVt͵+Dh{ޏvzFbп \ `bn+-|~k߮^vS>]E6W^iY&-n/H:{]|˅Kƣ2",Us&^Xu ja.\s}W͑kK߼CJg6'A7AƓX@n_sww?БO7'ayZ8 軋Ssm:P\'=\}t) ׷;y-, й4ܵСs] ~j?`/~ѝ .nxQy>3s |L\yA&Pۇـqp;,E \ :ޯ^"_BpU6dઈkũI+8v*Rr+@"WtUTHٱU+ •Bk8!*]WE\W$-9v*Rj+* JmuqTqM)]^ ]յq^MXh P2 &ټi>.!/$^08րeV>skaWbJ7aUԍ/x3g eG/`9e)D$0>LqHZ!UkW(벪wrRN#@nܰt޷zw:}.dQ̳H"Q~"&x|=Y|y~Gl~ǕJ](>(\(;D[ƒeC[=гЃS?OК_: -(AtFH.S .ynk~=T.5h{uJ 7cR9&>ǤvIJ L&";f`6 1:DG#LRZ %dsP97{vΥʣ'#W u^HTYiT\;7{-qS <䍫7ǟbIu81 /~̤ϲV#˚ޜe>VILWkU#_g)Ԫ:_]Z1J2m MKZ9φK 1وָ,|=i=J3SR:9Ә%׌Vp3_b[kܺvC,st+@c"[eRG!AʶA*+1ܥą`Q-R%ɕGC㻏y&+!UgnjK=?ҽ/O~~*44RWU8[o!r9[m|kwEwC;Ue>qtApM,H4rH Qd&Σs2X!*ö:DI+N0@]_L2 N] dc99"i ,"92*`@o8 [*?|SJMqymcEVy0Lبd9+-xHY2EJ@pYz )UԿ] l=dJD˙3lNeƜ]%qNސ>&{0 2w6im Qy>F /O̿^sX@Sϗ ]L,npQ6u\ ~7&?A*F gP 0v:xH״53n()&8TL]K24}۳Ybӝ[%U9.{]b2"SKq-UK[Ѡ4jr]cS28n׮E/]ExwUR φӳS+Ù~1vh hHgʚ#_wJ#q䱼c#,;XDH 4lΥ_Du7W$AIi@ %ǻ4j Kl}hގǣ.<>o/A4\hV.pHt0;SVԺ #}X7Y?"΍\\,$3f'o٦Ѩͣ^rݨ5eqy,?>4x'zn˚A>x(wG~zǟjo?_i߽w88ܜ׉ CkCۿ[o~څ]sq*H䃲[yϼbOWBzTC!$9Ԙsݙrb_ċF>! WZjIZog촁qED+n wʑUBWTԟHn٣pTBN؊)EtB( $`^ A{  >Y0ox:beW^!PU@le9 )똔uO'B*nr5O|[=&.DDw?,/Lps(]{ՕnnV7pj8_+zKuO\|v҆'w;)߯ _=\žK{ț:|>ߩ(Wuqed ^+=7'=V<{xnh@&!$W*Bz*bԢrU޸)9RTjs\ b"DKV+pxGJ:$Khʽ㹻ݪL<=_qwxfG OjEݬ랃b?.\`IdzUA1fn>%;2!(Ər*hR&ƮQb"Z{D(e5K&VK4㲷$JUM|#zz|zrrVd?{klѲ`WLY@zg#CC.:(=]I-dG>܎lG:MJS=T1lf j1lйh14ٳJ6W6[=-rىO|X_8~c۔3ppMɳ}f,0UY$Ra5XFcR&DxDhky'` A%@!;FmKPe:Z5H>l\4hd6'wyv3<&wK&ݭ)հCZlfF"y:,xg gh[!kLtR;Tc0G"d4kv:pA gSf2m\19f]r-Wb=Xu96M@ aRv[wJzQPSgG^X#J-z|D\YHSa5?T6; 4TZH@`R6USn6 (tʣEVB8HpVrC&Wh\:g+ӪYg'$dUSYLx h&@V"Ǩ2IEi3;1@/qSU]'G۞< #\g,)1FNU/Ԣ[QR{K_ )k|/+ŠA!VRLEŔ3b] = ښxG70ԫ5S-\A[1dR4cU~(r41T UWtCǔzŢ_nD ݞx;PTlk#VJ-@>f]jllG푎@^qz67|4>qV^Y`QyYEQYilZ8 CaiB?]v[Dd#{ߔ>UQ>mȑ,:ύ _GɌzLND6Fa&-F0)'U,d5.PؐӵoZw {Tp/o|t\FO|p|rggҮm_dG(<ޞpppeG߽gN\6oey?ޣw{+V-TA]oisBw.зJÅRĴiiQLix63>MZbfT1VFӱ<~kXt:cj }KPT]o6UULQ#sIm*RBw*eM5ZcՇd|nx )V' 9?{ TEEԼ8bɉcȞP +MQeO+J$ ᮊʠ>i/P1gۧd I$+G>BPWnٮQéX|ni"m_F3 w-d5;;h,æ/~}#eM2.b=^7wsIw_EgcbշȌ9KmY+8[ j} RsԕM jstTLX[&-.^{KZ;* l7 b!vB38Ԙp1vy^VKy :>>r|4?{ 5%gp5AUbp1:X1r_mmfQ1JD04ش^$%,m"VȀfJ3bwg;bY`b jw[wDm '.2)]l7YsY,Ԫʆr>*KBՒPv)cI}TKJ"`Map8 F(bTs޵`G~ bqWDԝQO8!ƜVB%gHBlN 4&9CR!Ʌ ]QR^vj>h[p&vɞ}9%m X GLʽ|Ipq3OgV]qtE3ℋV|Afs3Kh vZ_ 8'Kp!pq_+xxla[.{gq c-(F&g'rֶ}|`ǯ|CN\"@6 RrIR'7x'6Wأ<_a/Y9)45ZBSNn޸X~\:ٸrr޽~E[Tj D /6.YUhEPv+ѱv6U:4z޸GMUZn^''[=^?ewX;9\Re[:.Nf]/d-6"Ym.\ڗNs1A1|B8ĤlȖ:妾Q8妞rSތeZ+ {NYɎU.iUͦBjFт*} @s._l-cP)4&:$ A˽C`>%#.e9".x!=rQívи_Qo5AeNT2[ <WM(..Ns0'ӖEigo\L!ce5YTaJi (OYؤ};dƣ uvblMPV *doOlgQF٪ ᆅIlK8+{js-*ZЫ*6\M*S>p׿'ɂ w՚-kmnƲEO3MWW$;3SMefRxZ˒#L )۲d<:$@܃s ܠSBoIINqJ{~oBD4?\RFu'~;#^5yN(@3IT)G($ 4G|F<4=qBG>UH]⤖BzOԔ/+ui@ R'QID-/C! aLreJ/ 1#q\ތb"gC}4燴E5>ٲ-Ƭ/D D*= V0q9`#%0[4;1vT'Cݲu8 zS01,RTE>w?R%$r" 6y`(iZJBYj#G"X "  lAٟDvɑpG E;ţ\7CP`*' TƂIEQ9p򂠍Z;ϰD eiiGD+钟 ^xÑqE|m;d4^_gtɌV:z%u8hӺ: a#aW+K$k N2^ӭ*{[ee${ EnTДD]r"@1i#lFI:c, *e 9)$ˣI@e$gSGy9gJ*kpֶ༬%>&yC,QerZzN/v|,>z:Qi(P XΤpJ̌dyUi3aL]e;Y~3\BWqvtQJ;+`S!ʀ% ]eu2tB{zt%ȃ:+V :CWP]Vc=ҕdJ)!Bvg*e+tr}tQ=ͩ> B~,/j]>k,-T z D|#'/?a0E?N}((-(~$1H*d-h(w3J;nnk\6śq82_g`&27 xRΆ _slį0KedC2؁(.*R?i -@ Lq.QŚYb oc|k:zha8Wmo v_$D ι+YR16._-_/E}e"DE N8S$&r"THUp,dgpiWpFk^ #J^ C19\w2`e:CWh9#]e9{Sc6%tЍ<غe}_4Ztkk֝ wFe^۝?3J!z|3h-|udj36C+^)no3atzSC! tg*}@оV f(53B!bFJup ]!ZͱUF)zztōZw00p9t2ZNW=]C {ڣ+ :CW.UF+űUFYOW$0p) ]eUFUOW8@c`UJ\ U"6Aӯ+NB r P̦ף)^yGevln;W$Y(ךY +kE"]Zπpug&&5FFIPchJ+ ]ev2Z2J0=]C2@#Xr49ϟUGשvZn+'O2Q'_}]~Aˋ_LׅT-v4z*KƳ}M'QS|崱ѝ@?9l?d*>{TяۇrwCl'ۙşwn]셫Szl٤ɗx=jʌwQ,+ei3JчDKQBsT35g9BYXػ E Ӛ`+u t7On&6=:TzFu`hwr*-5DMOI9F$F&jS/:><$yBǔgL+lŋ{<93Kl'{l!#=W y.G0GE/֥OEH|r66mU z' HCGP)2Nx9£I-x:0 !XnTR094EU)he߬V; Tx㩎F&--Xi#C<,[F9/ jgKW< &2 8a;nm4V>9. ̎ WҫAMm{b5z1FCtE=aH<XrGjxhWaa&燃%#!bD2(p/X>F2 zS>E{3^!+)-+@ʋהRQBPoC p x*]p>D1fO3޳"yxTA;#=V «Etx'1Zu$$鴷&iJpȥTj'hAhQZpj(ax!tvzYI==5!U 8, O]x#`<<))hE g xJ=JѡwEYJF[ Dj0ՖYtoXY)"VߚOFX ]W!z Fmn^so%g =~^(|)yzonvY8HY>s,H3kӫ8-yG&Β[ }:Hf0i=U*4J?~M  ؂3jPK]"*̏r 8M֩]Qq NO. Zq>ik+S 4Y#zKTIxg7Oeݦ((>cЊ)޼ݩZQUn{΅<*YT9 6G7 L5hŋ@^$_~:/N;ɴF+0G7T~B'"NŖ?$l<_j2`C44Wv] evQ!#HѕUC]՟U=8-{[?TJ,]Q]F|9ofXk'UqCv8WM?UBc0zJu:nޕƑ?Iģx 0ǻowbA h+(sث<仿b#FutdUWbp+ywM縉}I521ͺ5čh6^Zے^7s8g[_,e?۞lb_I&G'3_9=`>Snrח,V 9XFsdkE. J)s" .k6$qf<qJ[H9DCʒnGB$ EzB9suHL2AeB5r h.qXRlHLJ#|phL |xI+N8M @3]%Iأ0 bD![*:k87$IPKU$ >W;*5q=XzVºվV8o`1a ]Ai#fY JLґbU2H`^? O._ji*&_g 9ǙSk)rRzFrL#a^ k=Rڬϊ/x5 N}?4ik/"eCD/?59^R!WYtpREǥ_w4ﱙoE2w87av6HH2x0$y`wGܒ!PLLFKe< ~ɔ#v~ojrtHUȜDf,MuL-M4ng`=-in,OO#T8~͹HD7mKU?JS[HӸ|1{F;Z>p|z @g0ok4/;"gb6}7*0[z6̃rl8hl|Icz'PwaU7bu7F*&M8uFi4\,zrه麗Q.^rUj#.tRvIq!#]:_FXM妨c~ěQo7F5Ά&#vo_~ӿ˫߾>žׯ^Њ;)R-w $|?=__uזZ]SmV5z>kj~Kٳ\r;M$EFQ7YdU05i?9 M8o/7r [R?NXlLzd5>2Ե_ OG<}֎rn8+5Z"HbDn Bpaghdb)2VhJGO}?(ߜArPFe,R&3  Y/rfk`}iB?g#{Ъa09 h %-U[ N|$6*0000;:0fBR *P{Z'v d%Q5Y}vm= ]~ %4 x ]j.4ktidǨWoqx%iY+ -x[Sn SC;B˶:eM6ٱDw2Sm eJ "Yn/ɝWs?߃%|[ۆWko/HH?.W q_cGϚ>bv`Z-9#Y3<[AP,?OGOAh_8A!kT$ܕT c"ɀN6Eg2XU'R^,,rpR!8\dR{N BWvU#gKZO8{i)˼/b/] & GL6eQ{Xj5*8Q"& .VzyD$$FHA` V#gb,qIjd_싇*TNk,p:6Wl9N4ē<ݝMnr׵ԥ,X{ER{U6u(NIr֥ʤYbm dCٓ('xUͨ)UTF2) dsIs-= |&%U[3V#gfUj.tu;OgLD:d"PK9hr1d~0}Mq r1@ PdWo")'͑n;Ks&-9!@P=#EmJM dK&mǬ9bf@*kjla4LWL}ոcG-X]=ovUD4M3L+n\c4,4S2 2Bt)M6>8LȐYёh$],GAX*jXJ,Q9ag/RŸ/ye&Ґ܀`"Oq<"1iNDZLVQcewDr2虐ʖ'jH1%-HokW#gFiɢ>:qɮzQT֋zq'xFX$C lwKd@)bA/>^}X;vՇ>ܱ jKU 3a]ApG~4}_Ϳ>Z!њWea ^ֳў{Ƽ=9>0ǝ{0ǝ(z0ǝh<ӑi-ILF\I/H@F8id\ !,D2%!̱v3O_vm:_f74=_of$r-iQ| \[)Nd [_vQnlRzB*#2 ‰X(Itd+g'07>8U!ꗤrg׺3O~5 ZY Z+0Y)xmHV4S.l#cA֞owe&ػAw#Uw丛"7MӿtS:R- ڠj ڌ h:\cY}J1broɗssZ_ 4gLv8ynuߟ?뷏#ϟHW^Y8K?~_E#?s1ޏ\%ᄍN?,Rz$eI;9k,qճ=|х.E?](+$w%GǴ?.2ӽLua.}yw?;v`1QM*>3^Pkpٸ/s10ė1Q/$̠&ЍL0IEkAHH4$䶄jƥnoNrvKu+L2t@&}H^9kEX;VYIIwPU}yW'vy{S`*|-"GIQR ANQTȎ^0K *X eHډ`9+U)jtat6) ky葶{d@H%Lnp{8)e1)k7/%8J )xuV|P`wlq)xfr7P%޵5q#2Mv3".൝8u=8}8R*$Z?MH8Hklmq@nt7_a aSBX0IHТiT'H^b"Qkv h[[˵pAb"hhs0wItOZ1 Y]۵"~BڥVқ!=qA*q!׵ĝpz6_,2Cb>_2$ [lW ܷ9vXGQe R/Zw=s S x\IL{2ZBj2u+ɺB{\I&/3Sw^]e*;E+G*G]erXWH-dUriQW _ -q!?Ӛ:\{TYUq :JG^?PiӀy/gZU}DUdPΙY|˅q5W%c;ne8ˢx5 ebfۓp<ӟFVb'_a+ݭɀZD8Q0U xU //Tu_uhoTd|WwI䜻!cS @læʐ'bS\<Φ]L?ՖZ8A$ɓXVKޱl,6g_"hai)<1Δ\9쫕lLQ{c #}13lL%h VK>Aި+$/*Sv.S);uՕ6}RW`F]er싺j*Sit^2DI~JSK;gćEn)Ocuc8`ggNJK&KPƖZ;nTE?z{8Jcj 2_~yG$6JNw{39*MTk'o#202@1FIH]P,Z(KEQ %8"DG+QiVN& >*ly^U:g8oNKA5e?Gg?JiR-NB!&fD;{DUz?XWܟWjyUE \csVԴ(P{/6iFH.3xu\8$KZ JH5J!5UyWo`Мmx-+=+| 6;ULRc=0b3UpuZ5ma z(I() ;OfJc,$m@xUUƪEA, vi1Q<8uvb/lr͌:@0kuB2c9 *VcgaݫŬ^x-f+vI(aHI)dRb[rhyRuԮ =w*sC R7s֟F8)55+mrf=\'ڛk9,x#R oUfpg/sTˇɳXy˹`9D!PeY^ G28sFG ' Fq"(9Ў_'/)g$rUQsItN(OՆ9׭ۛǙ⠊OR*? 1wDzVJjQWz˕|.ӛFpYFP(=( S>xaFMcc2u1mS*Tiel 1LT\/A:Nז.pUV⃓7~=W@HH G~aX0S蟵,3J? U̻pp~=[bErt22rI64W-[,89ęϏQuja*CZ HvѩDsY0}\)_P{ǯ}8L_\I)6`'[pkz8lhaE|qUSn5.rf~V׫/m B"]=~6"̾d_3 |EEՏ72*ETfAxGS]n+ Ѫ5GZ Frm%ݴײVS*(o4+}u# ~d}74W \xesֿ#vb 6E/ ڨD('Z+, \%<@6Khr&nJl=uddwP,:qILt*9 j4y[::3#+P\T+P;O|#9{t{klSY0 w־OO۵p*&P1@CΦ 5A) Τ&rT@ dZZ6Z#~3 uhK7t|ZoI2y#T";ϕno7xj1;h '~YZyBCJir e6Rȼ,ڸS-U"10*:mExk gSJ-7') _bWطLuS#jXVyb8(A Tʀ(2Ti2TyjgH#EYWt3ye@^e@7Fwdw.h܋_jq/Y?0uؘ:m L0(3A?&ê QOԞ_Zz: /hxQ'3]e( [s1cІ"RȮ)V [&F6,&kz24߫y3KT[ N+w=^Kdv Ҭ^_w͂s2gH AUZ\!(~PǪ; S^ ?@mn7].7]pQb:l2\27chZ5: +ʄAFyR-kr:q橡A!"3N鵢2tR nk ՔQp=Epu"pr L.2 hgQŲyKӕmxJ%8e)UR#wȮbzkU7Sv ZB|*3sA Ʌr!X4j CLh،on7uΓdysU+P(ADP)M6g a9(n!/2PM<0G> EK4!!%G i5iI K"^ O9W!H-lT(~@X|mHH9[Zc,/ɩV($"q>בri|j#P5*Y &Έj_=%]w#KdQK"DgP,\ k"K!%y(Nklv3 "ܰ`kk^k}0 >W8P*X_g;΋y<1䘠b8Ψt:UY$ 4:~ގ?\%\|sx1]oE0 m|[Kz5ۑo?ݎZȋ;ÆۊJP  EE$)ˣ.zTᕜU 9VѲi$:TuAޤGԋ.J7uZ7] 5.|Y~eh೹=]oGWc%X,gӤBRC!)JZm+2g8ӏz|uUCn kwݯfwl,yˉZ^m.t< y'.]N*ޙ^Cv+񋆤HxR+Y qkݞS^0fG8(}v;idӬW~;KhuT{i-ˎ9d,[ %l)BV'7 8zc o.Fjt͵sQ\Jf4)55Wg8nt_AWdvdMs4ȝގT6M}?7?ߎ'W l;"\yҧ8xS^х\)s%_kf?o;ʟ9>uC`-0t~Bs].k}꽼uG-z$J2k.Pj-MB z@Yfɳ!""8gƱ B䘕Y$ᮤb4LW%DNw):Ǫ,״2,w)!E6K{{N BWA֜=w!C:+s Ћ^9'|v2 Kb@k‘SdY^ji'e# QI[UQq9& .VzyD$FHzJvw`E֜݊]vJܭ/>ijrr i\ϧ2vcӷ$>[];bwONZ̖ܧ׿U>eH6K$s1DcJ#gOdbFU-&C(D]r6e6ZeRvA$sg2R-cAUjq_[*B7Gϸʕ1,xqbu`7%fqu}֏Fl[lp@*2 `%UHIsN9ui 4 %I0jM 8AАpduYCp:Ǧ He]9-vZy(Vtj Vj/7Xn&c)pT)؍k& RfQbJ[&=A@PU. ` r(:R$Q#(Ubj4vjm~CW|EܚCrڂ<GdTVpw:Fg&zenm#ij/# ɀ΄T!p#w~_prFё gb?Jc?_>T9Z*9lzb7sϘۤ9ޫ}P/s4LkMb22rH 􀑛lFJƕ2BPR44C*x0v,ufwrs(\P *PS-0{BO^ R$.1(Z2ZrrhBfL3|tHo8yax ܕ:l)%)eJ HX^R|("tb88m5wU8!n6ȃ&@ZgVe_)xp$aBKe gW[]N-i&H5ʷ 2g$pBv,$H&(cAJG!i'qfXꯨ+0vPƪI}V-.^Ίc~&F-+~؛/R$ dIH&D61]L98}2hwGM2/yG0ڬNޏcXȏ5 Wo_|8zv͔J%43?Naa%^JqzѼ~;in|3ƫf2Ox3>M10'GjyΨA޿/~wMN5+}؍㣛O*i̧̤h;7:`UY^ :+[/936k {Mw&>(:3 t}*`N;!U5.py|%&=izwXhsBՂ#9p=V!h~¹w38cK_z?Tp@8)24vQ7ˣ>E@ko~e8 M`R?S|XsS}=whapkf۽Od;=|pI_}Pߐl_@?v.n 6Ӟ?-FZC. kwݯuQE~C4uD-6}NUr:^мz}/\Mq {tEC|\NWMNޓ].r{ tOlNa\|.^B{hӋQ}u16 -HZgZ:^ 1C7gtBMw*柽,8ugo6 Zj#lʯ std_{bMd<[.]Lպ kOB(u#`Se ͜+,xŽin~sۡǦj[]Ioa5'8XO1μ ;LQr%Pkb?Q;t'̩]ڐLSʫE2֪#(מRTMnCn!2A,Uʅ`p؇J0FrU.} 8ˣ(UL^-ӄ(Kl6d d1tӀ6JW6"7y czׇ狲Ҳz{ Ű=< qz("m֏h\ph*dt+-xhYAGzSA1O}<֘ С+sI-2œ j!8E1((n[l.mx;#ɼأƖC-3'>RI=g:vr3Aa}:H|6~Yѷw2&6-!ӆSFPJ)I>+c @BJ^zdA\2p xF ~g/F yH\EE3B]HAvJ0C2.]9{GWX"JvySx>%nm46Ӂܛ暝1^}& KWV?+<j;oyL'[dCmj{DHwrgocޖ=A%/j<g C,s8*EڐNB ƙ1J3)MX܅Ou()2{D.K `]-^)D3C2g j'>՜=KoVxôx$Yc`v;ׄ 6v0RXE4̤4罌V$ʐvAzm]ʊ-\~0I٣0)Kc9:x izRR 0%. <}>vU? FϞw'nV6)/` y}Aڟ5H^g {tLʳNec* YH22{mEHI+_mo0dX~;,QߖU" O'j˙^O4-d~pM,w_ ܒ"P.tR1qpxg3< ٜ\<[U9}$*xdNe"37k"vMnyƣ&AQA8ӄs߯'>ajiq/g"uz^-˛JRVI|4;;!x5իMs:'Nm-a"y?NMybWu.>4YfqN1pGJw9kDd}Co5#F:R.iӶaD0FY&OjD3Xfb9ѳ1(Q/6jۻZmΊ5cԮԈQ&hEhE$gDtLH'"ș#^{ꛑ|ۑ=/a09 h %U[ N|%10000;:0fBR^lO|%e`cK]{fVkTʢn8Ёfy,ӶUW3a7$  EHxFdcJʩiF:>8ȉ̴Pm$7@;3K< XKH+@A+/|f`RdǮ$;#vD2q:ٳ-5I\J87%}6S;ć?TWZY 1 #,rDڼ$[ 6}HއD˼B3IByȪ8yJ HÐbe}Q2i F2`܆>"b@̅ƒ@OZOu;#gG4X+HSp$܇gdž]&M3trL?,ڬç}}vt K9Sa ѓ"ZIKS,&Ry)`bȼq itkN&z׷lfE5g,-in=\TsR>ս](iRO|u\_n9zRiwy㨉wZZ9xs:I0KKVU%ꏭ.H;ѽɡ-j>N>t``0R V1q[qW1rB+``pTפ/0)6TjBR )EZDQ#h@%Qj)ϭOzcTCincnc1N/"vM+gxKC \?Q:"c3k x"&8NK/˘K1U/E"l^0Ӧ 0iLq,T7p9IuH!v;%f @iq 88#$OvKzIݵU,ZlqnVO IKYK0\۠AQVI}e3'^t#x&IEO%&ydB `*ʻ(. +ъ'):`\)T%O7uGe_+f݋%mt>buBԒ)+rvk`ڂK̳d~d BN{01}ͬWE ? Fԝ@Qך0N 5,";K]M]ЃjVJͿlP( B&nidl)sGnT>9pgOKKH IgpA63t\@Ǜs\~-)Ngyƞf" pIa `yq-OG׉fw/0^ ft0>ؚ gulgh"o[ org#pidubttZAL3sO_wXD<!QztexiOL睛WrMRp(Syz ~NSco0pMZG?ju_&Zrޭgzޕ;~7.`MLms֣{[g ҹfdmjS*]5._ἽouVz045?[J/T$gLӲq6<2 *򺅢F[k}{l<ϫmM`kv!i$"B2AM۷75,vI1oc9gmbguEU--Uf;n-BD7M}oB3_-"<=Q'P(09R͹`= R${l(9)󾿘bz_8}8:0g"D" Q.TIV*ERDxI)Ί N^ف1W{UqWC^a߽Wu ^V}Ij)J%7H>%3q>zO41.#Ǜǣ G(˲~F߯'JoLڕ/}EgM|h[?{wi+SOx:^<` yBb/|#׹'_~Aje0$.ZMf(+ {td u`mB:>(@-)'=$aW.=bWץI܀F-ۛ^s]W5P,:4Ώ˅Z>Տ;x8<_7?8X4/ ӱ5kn[RbX*07p?_K/>c/IyfkVǪj!Aњ|%dUɑL%(Cv=We/Lm(U8t*׬=ͽRYERZ*2cKD *qFdH$fHї^nC#.y,2#::7$\c@k⑦30 ZDdswEWYZD[Z靳.D(N TLC,Khrb%VRƾn'p[ >q9<ҤiGH8=t&,ŎA$b,aMN;^!>Vz.U!y S,6G,#&U:(TޛQ.Ʊ?4 xWhV+`3[PYPyEYtylxO=},vmJo ͅc.\IpV= xgxCE;D72h0.Pd;]Msξ>kC _>m,9eƍ}vWDz촼]\L Kk9 }:qIC~M2ET&L[1]|JɅjuo{}qqK`R-FrԮDHX{W WM\KAmkxiUh3}-~L(?˶w{C- )Fӎ}Vnbn)Dv^)C}3!\u_|D?*ձM Oڳ,>bv`*-9#Y3<[A(,>D%EYEY^6ef"ǬDp& w%a˜H*!r2ƠMљIOd 2Zea ;ːq"s>'McTgxUdSp@/r66B|v2YKb@kdY^jq'f# QI)"0s:pM]D/%"I4Jvݐ`F팜l#Ք`)NK.%|}r!q=f^IZy ֗o{ ړڍn[3=աMn TKЩT%))aCκ. KtC46Yo41y!J&Ij$ީdrDF2) dsIIs-= |&!յd쌜<:*v}eX^>zUryA-񧋇um̳Yf퓿hPhi4Ϳr r1@ PWo" '͑;Ky9DVf(Q'i9 F$)ؔ L&ݎY8sHt,;#v&㴦blwSj ֭nv`Jrl a*eHwM5FICL(P1-Z (*DB@`Ґ r -:$k(KEMd}C*HJ5;sH+$䤎GvØb p`;P7rV8:t{^{}9OH,k^;7Q?7%K@!_zgA$*ʤI0/1D0Q1Er{ܮ˓"gGB;%')!]Dnֺ`њjcDZO 1KL|͒dbI]bffJsՄcTh]޳FΊrտ^ue?`T+yeMLX80EiRG5S XR0$`Ѧ9()7/z=r-$p)gqJsV-)) W&DD6ZjK쯋mS8jq6&gV( ij&GHhS|F<}p 4Eq8vto0#]jxREG j]Q55ZD٤Ȃ;V:m? Aih0LR!lJ6 &i "ZF劜 2XL$j :֪}<39C iNDRP.7Q9mZ%Q|=M"nh-VYj :z&twtddTOGP:o֐8V;8HS# JrTvH8ʩ$ܨ)׉Dc>G8荍t> X0M*e 9HedP IQu$qN",%(at\`R 3B̫w HvN:E?2+xӖx]"&rLB?ir|4(I!@SsdjG#u ogRX'MpN5Fgc7.]asl'aV cOc"ا@9 7r\%P.S b2@0P7b 퍺r/*S*S)DQW_O*,ިL2B*)!E]} ` N.)+Lqc޾0.yEıDŽWIׯ}4l?C_E1H!3.?%[Dqϼ|6.A :MřrBqoɧyUMޟ_^_զK1Sxc,l;Y)ZQ09ꨨx0Fqg''G˶N-fSvޟ?U\2 ^>P]2:.n bw V,tHG(l-!ԲJTbNCW^sse*B_~z73]RfA"Gps'BuZ2029v5|LL@YKF|.*B9;Ml1Q 24?FIh \MS'>@rڼ:Ylw8W/o<$s`q*@v130P3RdE 8.Ns[m2x{jx^zE9Kn~38hn5qvr῭)G)ULihm ?F|o/7#J"{\Σ*QBo[ 4CʁJ(^ Q\D*ԙhDQI"$+&6C /sZi k"dpΚ%Q`h 30jePMpbH9X*!bHR*!'~ H.xR0UR*!bHR*!/!SR\WR%I_R_r()KҒ/)K /)K ūJiQR [R%I_R%I_R%v Gr:.?nIQ>,mV6!1|"a.I%-4Y\D)!s(euHЈu1RHp;m<8vqrVyuri}>^U_|To%??GO-d|]8;W:Ào}Yb1@M6D>i Twɼ5ɩjta-a𝒜Sd/ azz%7ZDx?1#7|y;ӚMm&O#b}T]F4O4k^AQK5!̘7tWz lh%΋ҳh'|F)խ1M#,Cs뫳׋bi3%6\؏48w-7]X$߮-~ 'K{qO˺a˻Qli76O8y <8zhpz42p lZ*6IȈMƼIFjjt>獶i4*'~k;+~yN߼7oO?~wmIr/!qn $E37IJԒe!EIeO [ StUwS߾>~op/Y9_Kx~~ܻֆRoֲ͍S}uGVFi[Œ[+?^}"|w6eεO~ͪل?|k ̿b/:ZܑPARPCaAt#A}U\ 5;WfٿX=m>5q$ChO .ϝwhHl..עJvJ‚51yD"p‰Khr&nUޡ{ꝑ9>W~#Oǁ^>+bAhATq0YO3͍#\FGV`P@MӷszH9qkSl%#}ܘ2Am1Pa89j ZdX+JQqnV'I  QhT3%Lo >18qSs'#TAZ[ )bQpj!R0F[3LBL'ΎEy`oBo?u,Ny8ݕ4ڣRJcrxk=TĄUP+omAwg}z#eӥ|`{}}7kN0)y-@q~Vg)@)59[m@wzi]~;)s<E%N?\ؐQZ[BT9)F)(:)IO鞼! #P-с PdmhLDFTX2)\{#^!v3Qi# @7$ģ Ae\XJW DO qG(`5w:jx?EKޟnGņuTy؁-wuHW/}kGBRL\C .*㤣 55HNVK34@p z)<4 ==|nCd![1Lzm̎𤕤3 T:+GSG2ȓuLN~3.FYَ.;9tPm~V?iޠDqIVFd$ݲ0c8 ;ӎ7W]SΠlvz^Jَ%.{>Y2,A]SqQ9yƋ:ӫꓜII ےèF)[F;mz7r&pc}6Q+_#ƨG#vQuݪsEn&UR-u*^HtԚ5mG9^6Qɰy?go]#2{lg?,_rH6I~ Wp0x ~gy%eUԪ,tSFwm:d}OOdSqMmc< ~_-u!N[CioCB(yheu'g8so>)"riJxb+'yf>Lgӡxe;yf]~m \._q?4=!_I?bp|uu_=,j+{p"l4|w5|^qtG:.߾:c_r%/鈴|<?m.'̭䟙}g-vv%W%=+LIҊ"xV'[AdyOHA[p7,znX]oN FB+7h Ei(,TC9t8ŅiRn 7~Zv՝tn坲G\^CsHm ]eʏf`#YvWs9cD.r!捽c>oߓ'ڌ,Pp"!.TGib;myz/-6tmv i3M̴"ڎw]GrXQ*SʬoJy.YM˱? 5I+sB(D~﬛e'S٘ %G\*.mCaHR?a5˥CLh@̔"$ـfDs͜mNX%(@Dp:b &ȉJT3l0z_n N9Xkl<0 /p.Z]\' )8Var&!i5_xsѡ$:Z q`KsU rf]vz0$(E|3L #P 3:$b$,P)YQM?^~5:DhDD乡Y$D'D1E!%yMQ$1[~ϡk_0'vWX[7g4s N.ptmQ^} N mЊt*tSO p zSy{i 6`v֕i]ZU`-\f-1P@ N_ԂoY1/|2*X6CsYopto{rájjwq05 O`y+Bqk]^}n~b|}qe_~=,Uo~ Oˆe׳ 6)˟Fg n1̷mުz-gu~5G͏ Bè]n՗VR<7Lީ2 tD(-IDGiLp^_%c)haUZGՉraI#)sa?%38Cm ]BD { ?SZ>k1ٗ JC(pF{u:1 >!^Wj42 4q^rSTQ 2b 5 Ye24"0O}D37*'Ӛ=~gƚXY'9 =ˏ\l Zb6Xul7y˯o}j+YBf/rcfcS sI"(ObT=6 , hADOur9R&]Iއ9s<2bpgiM׺l솮Ux7B.!B)zsy;muմ.X``;ѼGmql]7zi\Ƽ祖d<^borK=7t /\ed",M%^fqݶBy1}r?+z|syS[=΃yALsI])2A7 _~U|O_a?]o7Wrl(!]rK.[i+5Ȳw=yg(k$'iYMVE2TBߓN<` aDB%r#W߿èy.FI\ti&2JS Ygv gy9zbf|=\.a}QLO|rg$tМl"= @]^4bݯ_v ͬ2!bB-63Rr|>)4@ꇿ4a^dR{ J۴`LHK}UrØ'H&$X& @8jz'"D8:-Ѥ;" ,|1}c̜A9h©:HQY*MPޓǼH!jޠ+Ubjta!lmj&mdïpH|7#-J=E`*zkH2N%ܢ`XH!I+^3 t?oV?vKϕP%Bs\d6'F4R@8HFH%k3Yq{&g7)J- 8/ Z(YFig2oEdt8^բ#v3Zj1『 nXqįy 9Ij4k8*go ٹWP D CdG$iGHM\}*B)P<\ ج} QC\J)傤"J+a3d `{֖K F&K9!i>&xx@ I[_427C`jT؍z8?ļ)2=T"خ}uk^Yèi4%`Z`쳕ˍ2LgFzռ{g%kj0MDocgw3L"~Lӟ/4+so\?C2KĩccyX}j,[i7g?)|BPurgʗЫqO_V/gmnk^HvY9y{2@;a{,6MjJfW$|iR++v蓭sf}.퐭Ð%u{ܳ}\)ۿ64F|K;(?Ew„eW^?|oꋯuxqL˱hqB9,.ZKvR|RK6+wv}Ϗj%_#+E񍨅CocwbTF|[m}FH#q Rb/kL-q(>uOb3#qn( >$r"T|*Y5= (ϧ'jПuyoO򄶇*uO/T"g4+pIq/XǡĝG?kiv0MztwS8˫ft6~ {c&zb.(25N6* Nhڋ*6_7@̣\HY{*!,D_:V9![&1ײv9:SVs<:;Kv W*+f ʵB܂bF$}F=Vj(+ˊ L %5!JK?XrR)A"2CRgn*2Ff^H;4yZx̂V") ^9@5rvgk/Fg㊂='U`Hh1HR2 A&!с m9%VTV.}I'1ܥą`QU-ᬗR%PH Yyyk[V6@c%<Ĩ#EfLB]Ef} T!%(q)I!Ip(}.T&ȯX9eNM[aL_1Ol53]VQro-n$j t0"1ƒ֓}!s 2vD2V)ާC֞p֨7IjaАQ~2J|R,{0  LH#Ds~\me-#-J#~7w~e/~I:&΍6\um綷7i2+NԔz,ME-v7Ϛi;-HoZ||tuN-׳Xy]\)0r$Orӻ{t]55[ |ێ+,u/H:Қ_-Ӻ8YOij".`Z-9#X냅1<[^(ifn( 8Yf&A)+N 8 0aLJKX3jcMnHKI\ x1K9@"s.{߮~T5rzmUGB^ !'Ӈ.c)ʵO٫MTrxeeup>nR%EZ͗E܉'ZpQdu{Oa4Y<2u֎L+LZ[JJE FXvYFRYb:k ̳# Q22GZ#Q[c7l̒ R=)`.E38Z:LJfFNjX.B qЅw w^d|=I]w̓0/J~g Zwv6*iygLDPً$J3,%RN'z 5Fj`B]?Hieb( (jSjN5D#0L`9H.SVۏ㊉y,Zw쩵 Z{@8^&s"Z dSs)fA$ŔL:jLR>`>JUB&dȁ@4,GAX*h cV2q>Fn}ŋn|SQǢǾWֈ|ЈFArڂ 58w4A\sr=dF9=pwߺwf:us 9grꑃ0BNr1*@u>Y/z-EknuHǖE[W淫| ^8[Zpҷv134}nP#Beg;Zp LaQ?"1M6)I3XI:)>G|]ԲnOzny.=%Oܸ~>Vʜ+eiYBӂOs\N۔*ϴ*yž7`4w+dۢ$|ACY)>)sf4 ;|LL SAxA` Abw7R<ف,c yJH.Xa&6igA[>iI\1Τ&]TC*RpQU D'B"',g_YEӇooKfnV"LQ'C:0ք$ X)pX;hn s?']HcP)Z$(^9MF{;m"mw$GM>4Bj%!$ 4r#P?8OhEp`x-R>-/_&h?Q*RAN" $ ᗡ Ra HHQ {w&Ok$,ɖo\d!$J 0$T9uFdp ΡG')A9e"{LVkB2w?tϊ01|ć}x5~* N `g 6Yxzϸ JRAj"XK"JD @ҟΉRgY*xfVGYydQ4TN4&6%4-HryA ,&hϰ9ܝǡ]M'"Z)@8w3(5cC(&Í{Uj-j;-Өw1ٙNB$.bGG#3Zxð^◟_~4 ;ZPzH?8gFho2na2hN28[1@hS^Ȕo"Uԟ_ث~Qwx&sCnWZ(;~c?պǷ߭[qzԒU$سJG%PɗJug~@a)^@(^ 0DT.{Iѣ8ړYOdZ^i?TLIwT?Jj2n`PE {~tۢ>4 g:Jg ;Mzޠ>ꍓ9mmXЭ?fv].+ OqvBX#qJH֖@+@sHSv4e42|=s"7*hu.9 ρ6zc#$O1v A 9He$2A3⩣F뼗¿LIExbm'o;(9 MV\ފV|*w<m@|xwɺ |>vgH@md|&ț/.1A +NBikͬ;vIʌ\jޭ?*v*7"jZP%x|laefZc /rqeUS\t ?b@q*b#r ǏgS ? /q+9MK_/Nc!%=r䋇|1Ȳi0xc H6!V/#*3jtncAUTH5_V)r} iqّ6vÁVʹ5Ri8>qlߒ{3+y&c:" w[{[ rpI1Hl,P"Qs2(gnwva0m "&|LB jr|4(I 4E-WNlY{-v4RSL1.5.'ZZFPvuYZ.1ϝi3?\_56M_/;$@$<A>^G#S}oqJ*W8r^ִ82O Q1B2tT)8ouZ;,.S{@}coiDyҐwqI ؉|fzBNUſƗ**ZZ"y:.]=H9w[yu>G\ }3 b:+WTIP^%w8n AUu3j~yvmm1ۈd1-hMR(~mMgZl^R JJ_~nz<^EfHNj$-F]V59YYW;L^=9⨞D8鶘00 r3ZO`1`\{T9 ĩv!ă2Ll.y:鉶6%xZٶȣ^8?@` 4TQtRN>Omo^FTA҇*&hq򠙡*$|$/iGb% z7N|f[:?=h$Pvnwn-";/;*mo7Gf>C߹%-(p~E|Dm^3b!Mn@ЛѲGHRqcH;̼X=_ rrh8HKU60)yH9&,wΫzF%BSf Sy%.%+f{xT$Bl|d@6@fLȐy4SU(MB7ZQ#EFj nEʏAݣ`o-]n>{yEȟP[ a: #8dsY7(XowfGB9=ߎ7a$+vIQdDRJk+|p< _E"#b 7'~o^?L{|7qT^.ջj+t೦qMkBj75,Ҵhsԣ _]}{OSs?QqaNs|q:.TkNρ0~l~5>Gz;2CUU^y#?#Wo WRo_YAh:tx-cRAI^Y@T[`OIK;#\:$xbD Zsa_7:{}{-_) MK@‚1'0V( B P 6tsYo@ Ze9.N%gD&QUq3L1c=ͩ":JZ#@:B bT?+ozS5s]";?;5 OB?KG- Y&ߍ!pwC Zi rI̓:!xCS^fZ֓%iYy;Q&tܾi|yK_[\W܎Gr͋v q1_)4w kDWl7cFD8BJ ce%Pege`Qt(|%2<`U`BIO`y:B#le½`#YQө`J29]4IyluɏL0'džǮvġW/g;}S19r 9CH5q&qA:-8U ,0 XF;`1VYE 6 ;6qQ p'53rn~7xj~}`A6c*W+Epw4"N?tGJ*`NcU2MlA#(TGC#&F#E_xT<>n!$͹"Ah&ND5@ %LוJ sqs(ӭs1{5)QR?'_.gܰNH QD^2kv'8H(`CT\(gc!̿D6UH0X{&֒u&`eJz 1Zf3~w3W;Z;b|7M4*X7 k@ F4&͝WhɖYΎ§:;˺zH"{Pb0bx=O&u?R?NN =őp."Nglp2}P. &MIَ&N0W8[ 4cw~˦k(P\VW[%FS!1 Ɲ!D!lx:eˆ'K.0Kqʆw;-JMZPLj@['moNw]=G;Ωijt}W}u>wF=S&*Ɠp13{2A[!]Znu '1bePl68_J4 XFՏuq˛P8wQ7 ?0 [!/>->;p(}IJ+rNk(Hc4 ŋi3=e+GZ.Q?_EzU//y~oY}a<ޚ $N=ߗoWdƐAJE”?/4$cVۓ22O?˃U!DK Ƴd4Jjs:,f͒ x0_?:Q HW&U-JCBf(Ug0=in WuأAqSc2eKX^|}3I=90  cfC3;}`-5YΕonn= zb5$](!^maPsw`pMy:Dͪ!Rvh.L7VCYv]$~G]NzʌTja<H𚘈id 6PpoPD>*IL{oʾsظs*sTbN+EP#W9Z1NEJ y Z^'F/I'AWt&{W\?|x{/Mǻ9Z/+8=n9T47EeXq>o%~FPW I  a *f PbaP՘gzzϞXr_t}تhv=٧w|w>1_ôy%?/?EMK!sŇOeXO^2KBQ00Me|S;Y \{NwaJ0j$p ;Iక )#|1`A,(e 2XP `A,(eǟML҇F4BK69c& C&'*11̅^m{glmrGm2Y6f,cD]#KEf9f,celmfm2Y6 ,ce 2Y6f,[m2Y6f,celm2#6˸-%dܖےq[2nuܖvakaͅ]fAq2(Nɠ8'dP Aq2(Nɠ8'dP Aq2(N>vzxzRHOo/4"iO];6K+ET^ÊİÝm}.l=f\MOXG9G̏VKBioL(: Wz*<a$%Z\(ugr&LI`ViǬQFU58zqkYzgxr)X׽ڕ͋Dg m-ˋM֧HJC/-:J9,!w:k|` amcBc\jHXFD$jcSzyd+(|. ofّh!"RTD2f4!ʓH&nc%P 'a.X *Tm5IQ2.D* s&RK2\g*`߲iE2"*b:J0"ŨyR)Ý4 ^$x;YU}kIP0Ag)E RV#lb\_]䛆cgi/GhU) s=*4ͫũyxB꜑1,\Y F r_ܞ@ %I`MPD J1u=33;`za(R8 H ˥CR*\ |*c;:9NFFT_ tGe11^oxg?thNw H|V0=rjJLE}{J1¯UʯՅgW hMВ3i0i[x$*FQ]9:yd02fbu&7t6 iC8ͪa4p0q1r՜|év6}CJHrҪ?,t$MA|~O+tY(ƱHfp1%ލ\- ?RUu\o`_>?g/_c`p`, 6I~{6/ ᮦSSvw wWv-}üWJ[\r+4z6/3*$.{OpZ5!"$+/#@6Y3TiĐ^R|eBy6bfK;TPp_QA<ҏ8~T[` ?n~] *0wHsIMH G&'4K+8x+k4TF$c, .gaL`&t_:3e9ȁW KhMdAITQDmȷQ,bR͙Iyހd*:*@wxagk߭V7:ZmsՖZ+pV;Z?0'#`B(?yM(ڒ-Jm|,S)S0&3lƆҲQk\¥;m%|j.s=4.DZ1a#A2?9HysQ i_Xe[]\oB&/cYv9 WQ9t>H)0QqqFK >TH(\YAͅчb$7@r 9g{wyu[H8v\HtߖŹڰVVb=i{!\>׫Vk4pY9颼ݸ5PibA`qTYC'#p@{.QD20CF8Ès) ^22(ﴌ2:MɐD!e!x0k5f,`cԀ'%$]g|jd;q`8om ņjlP҉`hP`rP V&Dq:U$U>a <* ?s~]`D +JXQ;#vE ' f94LS 3M`Cɰw&'޷x4^V-[u׷oԴ8ms᳅htXX@:ʻjN)'(HyT˨uRye9l#4S X ;B}޶B))&`zSNq飶][oG+3V_o}>$XajIkERH]~D"ESM'gf8SHٗZ埲VG) 42*Ͱ Cc, n>me˫ʌk=c4:I?7'7L "Wħ 6d8%$$ < ًY AiUaS[&NS!¾(DE04Ff|78z~u>wt1>>ŅfT/Ͷ|i]B v" j%F~17oQIdiQ$Ue0`6hPJvZ iL,ŒZQ٭(z""me=՞'n=ߺÝ+Eׂ.b۞nOޱ'꒮ ;o7ZW7l6^ӥM'v񶫆;]ܢ W}9lޯ1׊c?q>[% zE۩9K׭׺vۯ ד V{mXJR50۾=]nv%5 sj)]JPBU}+%ءaXKl(X7~sTZ|3x M*M$W5qFmAzܵltQu (!G" BB6:vB)KɃ&RVL5fA/OO=mc$%e0:+ACHL!"OXdy+rɂf}ܷ̍d4/lyFh_"9 kϫǔ@emR„:ؠ Z=tQGHDX¥ZFzU{\x_@Vd+1(btiP%XMot<;~_S$j1{X(IRB=AH4m110ܷfGm-:Tr`>/jiR/m%`:C Ta:(TCΏcpױ$4iJ LFupKĂ&CT|JB$v-bж1Rimb,1 ($%D(duBxU8O3KFBS3c%\+6']}fw19v[6c;A[SLV\5>7K:AulsJ GDL*\"a/ {p ֖RHi&VcMpQ\tRTmer¡K; )e[zҚ DP) C:-֕|F^}Z< MپZfG˒OFU]{ 1pHڡ F3[pX8g9YSwE5/LǦ? <!(`-YR}Ι wd%ubv#19<]lbG5j1u/_GJ639Sߖ0~oA(ԙqUʁ)@Eג.hPL{U*46γ RE^#.97'"%^*+7!|rȶp׺|7ھ6ͣ/K+SY.mSਈ,( IʚxNt( PNZ2&b/W>Y lpD_ /zI3u.=<V{x)GoGBmϞ5T]7q&1iNbFhD2{mb[UicJ?Տ-7aGl/Gڟ\QҕOAZKPLiXlƐ-HʧϠKk853NWUvpav8.̬u0887@U_{A`Z^#%NyUx(N>O`k.sVϟj _Ĺ>'q=eBȖb!{LCg˯,> OsnP 3]Go-hz3NNEsy5>L=?`_GӕɫqhT[2ڇ.WˇF<G_ϯƷ2f\f/[-Z^L>Ggl:.cJYʾ% Aw98m<32!vE@gKޫh9<'St.Y8?ac.pe)/dW+EjX壇oitgI[9ͽ@1.Jo3TލF̓aܲW?}7{lIe*k~}?߾?9Q洭?yEv;6Ħ/3z[iJpG3`kJ19?9𧇺)A(G 8)DّAC)Q@TK'eWMkY,8<%mRY>s*$i 2E(:{BZCNO_iʫZɤsdԊJ x5TA\=8 kC>f,HUQڧ [#WgL ͯ[F/A7:/GK{6k"P(eO-qd)konX_e$s֞CU[`8$UPy:ȒP|۳%[+ l;靝-]M 4L "җ≁F+L[wd3^?P%gBk-QgDLZA3qvAJ^^ߥRkci>*nJET l@8l gaEM^<.XcwF,{ǟ@?ugC\:SZdtXm(0.;Rdѯleo?oUa.^tN_70γ{[x2Bðޫ0jEU ~zB]!)Hw_^H1v(odWkHlN;>>;0E $.gR\T9'5$ڝVd<(|H'Ssښ[o<*~|qKUsJ?,ov(tawAzfZ$Lyθu$B$YaӞAo'Ge7)Zx򯪿B ZG쥜0< Э 0;0/euѪSUL0.{uV6O+U{yYqxt^]8U5ASb.̧v7vgR}53N!ݳX~1Cn$&jGb|HuÐa$fv+`ŴDl_GNO#AGu5+|q/i c[n@dOI{Rt$8NWաM{r'_/ÿ0sUUFF )U—]:NH~PEP}C@`_'-wKA89u܏Y3AF=iR;v.xdbpBBV!8Iǘ3X/fX0t`v;ѱRt2ZYP(lUԨ`5DS(1)LX|NE6& ÛwDw=8wm鵇Nq t}ęb-ZAȹrja߶ث9\!rAn)1H@W ڤ4S]Z;fuO|f%Ksͣu`TAE$؀ !W}Q#ID xGɼIT8_eDge^5!ygb,+@cI)ـ4Ù6`#aZ8yaJ{xw䎪)P5ysM/IVCbg.o7S| a8bz*Rn& K¿sNTz/`EP1f* 6xuOsY&)QcX,PE( p ` %`K" KFVDhL$xXO*W F1Je  Zڐ$8X U†p¢>FhhT)f5tbMc3hz@Ri SvW%J8Xiga4 V0#A93Z`|<%ot GiI9uf}.Rߧ9pM5\-Z)l15*C Ot@`0d>9HysQ ũփŴV-ro{>WHi񅌷!KBF.maG;̲o$R{}{y}}֠%._v%= ܘZ.NgpaߩB7]_3[Vls77۽@s8/wy) ˃iXiq!j&sA1-ˡ:r@&㨘be}ډ2١ev"Cj 9QR<;'GNHiAyWd30MvԈ vD!'xk5WȘt,j%V[I!' ؐj݉$:NUjDXDg' qfI$0bXXAƌ95|ji\OɝP絒Qw)'m1Meǥ.Pl)A!(fa4S:2JQG©GǍJFd>Z"`P>RL@Gm7]Ms7+cDs\|ܕE(3}DkHYMBRd; =08koG Zy e\e.ˌscXE7~߼xOܨxϛz|uzo_p/^\7l~Bc%t*%tfuEkxe˞ֵ )` { ADn逶sɦrg {Y8#v~IKAeq&joGnd<ɖ@72j@o-%pRpYv<<Y<Z~:^!䗯~DYm"\7=pwC*^]!Sn(|B8r6?bc!c/աٕ4޵A5 с̠׮HtZƤ{m1K8JL[驘s9*y0K_WO. ؖ F|Yݿq&?|엧h6O23XV;T o.IEIE D׽rUU^gK*9CĐ)#(ҲXJy\dc/xnd@!>КUQhՒإzWeu,'ٱ 1Q|(QK<8z1r)%`ز,rg6̵4]XY|+O鯝`OJĤ)ʠF4c~>9N'\kp7ׯ+d1=yi;za_ͯrE_ONRJ>u)htoߧbweܼn Rݝ:EDW=x3-/hE u }ưle&ἬۭT씏01CfcԠw?XρH5ICPR[(>Ht_(eĔ}>y8qWCr}Exu'Ԁ~w95]cW̏8QqbY(4M4F؋Ǿ%:S+j>ur͊z ϕۂ\J Q6C]?5 ۄNyk!!<(dQxG!W/[5gƀhI hWCȌS^ۗT[4Hm%P!{Pl=0CM"[B.=Bbm!DlRal0F+GI\]?]Ԧq3Bx$rOI6.7EB/+=6dcN9#l B[qa燠1R;n4J(1jC3k5G?LHB~F93AW/rd4C• PJJaj ʚ{t Ϫ^]/ch>T5 dɍr|=uԾoi%VrjZJP(fYH\Z`&ԎKAA. U;f7 J v&EP n;X[E0ȵl3 _@kb7-UT g((s4e`j KGv`悙T2Xs2X$$r2XRTGYˈ.X93f ͉b;*[inKrB`? }(0 ί1ˉmTT=*_F=. mF2*i } =K8nkYɜȸ((]ͱ VTXgO yRD (/  /bdȃJI 7. 8#af`Yq/) NM_BeM>%[sx+(@g6 񣇂ӥLv:Aסk ݐ;qje2!w2};` Ə9s5xҹv)AVFZ|4yF!a^4{} XG~ir[MseB5;7cCZ$蠈=H PRun"؜IڒR(,Dp%@z?H3AL2sqhHq7IE H 22{Wݜ(-ZeC5qvPD%4hނ X(Ұ.;ǂt# B ƎrZ#e* _`!]:>9Mv <ʬIH@)|\PZ-92 |4|%KpC`n>֛ֆ%Dz>POy~7աvyeq*,hF@0uQ u!r@3Gώj7EH3{hC{gW㪥cXs YkNè<{mggh31/ BEȿپbofOs E z-P4ts ^!7 [|("ZiiQN gZR0$/ȡZA^LyDP@շM1,cB?}%Dϭlֆg@zrNt7iE7V2 -8")*4t>&.:j<40diց huHyh6+6}4h pNRܘ4UTnqHZ Ok&_}>33I9)HZ)#JV=<94œbzZ6- 7y]! o@(ԖXk.9 Xfjbd=`e;>?-nD*c~ 5jQP sJi} ;p@o@.`UR4gs˥ >N&!SHS,MJ(EP$By huSQ- ]Xr^Ty>]!d07@FUK⻥^s!_^kq{[̨ zvHǡA A(}o t`vT~Ǫj%~{qmv>_t#|o}'嶳ʻY?A sRA7c7-,C" Pzʇ2WuOt7;UhxDӫlzϘ,s>賜i?֟0$M!z92R09Jwq4WjkGV<]ypp1 ]y^Vl>x:}yt<$ttw.A3^_ oi7C?}_IOlI#Z'(O'(^̵7G)|Vy1}7D$Ϲ>C6F0w%KylЅŝt^.:/3'PBiKei<}!ܫiRE^{r+O_b} B_٩V7ron~ir_O!o񿓏]Nc+hw+݊g٭xv+݊g٭xv+݊g٭xv+݊g٭xv+݊g٭xv+݊g٭xv+݊g٭xv+݊g٭xv+݊g٭xv+݊g٭xv+݊g٭xv+݊k#-=(Xg9•l6ereJf?f8׿<ܻՋA0gzHW?b;ԳX,0l,<-iRMҖ݃Y<Κєx͡V?Lke^/p`qQ.m.3_YL.gi˳hԞR`Lk-Rql̜kˡE`? f0 a3@ f0 a3@ f0 a3@ f0 a3@ f0 a3@ f0 af`iͫo%!<~|(9~"їqDS t|7( j]2-@ &⌒$Za\5e ]:ӥ>4v^*JًRa]+R9b\*m$>l@\'.ePuO[#=uZw; "^̞*HYT|L!*Y*OJ>?ԕi.e|v[ydy7]G};l)G+rǐ%4[ƽ:͓ l&&N6w}:H]N'H=)9i|&EI1$HLʒBͯOlOw*Yڻ;}Q0W01)yyL%MVZZ!d!mSWdCW8]]>'PN?-쯳Aj8?p$Xtd e.N Ru02:)p2Hg|.}%:|S-92V,q!D ;MOT/)r) 2Pi(<<zn>:birvbe_">fy? v[;Κg/Xh(xY='Mf4zYbs|*F}ňqP4#EPj3V\)+AD$#sژlSz>IGV8հb8+F:b>bꞁ_9^N_8e.Uε:L0HF텦>0ᆆDҮP8j8*-c d ֨hS`&zM^I pģMSw+ʛMI5^nZ[Om؞+dڛ/`Zj| PpJѕPmda(XSآRCL6ӌcmkl WKMTzHuU$Nv~7ص]('JY% V@\e9W9˩z]whɗ"(C |DW)48N wuKӥŸ&VgaZ 1j7ӎ#6mzhуݢ"M6pVK@ф!B-%ϊ*mv%] =,JULB ! &e9RQ6 ~C*<6f췇]P?~ER bDZ5 -"Z]3R[i"KApH3:֚d&SeW$}O4yJd.]<{L;=W0a˚-Į#P7G] msf ݌YEx@f?( ُ?5Hs&U ԹLx"홧[Ls|`l`c_1j͙ITD\ I_<12{#3p')Ud(f%9Nsڣ,]ivϥn?SI%OtWGhU?c>&6p6gٻ/s8s9A?~Lg.'d+[ƣю-?!:Pb_|υϟGC2# r٣{TsqSw_!|Y8aMr$d"7[؜_QJcn|U/}< /xsKcwa\yE)DGO\q "D"[}9G|vV|gWu Q, yRZ(W35Ltq9+$7"!RznU`;⒗Fj]{O](ExZ벀9q8n& /Zt7{6>?^'jUI4KqxÇ替b\-/}Ȇ2;g] P"q&(yBv6rhVRM:D,W03㵈LxkR zPY̕Ud+zC8.L=uAiw>R)0)iacMv&"OԆ%OBbNwߗ&|„nPxk1S @ȘɁˍs?93>8wGN1H퍀YFJD<[$V0|'YK) !R5""/2jK\6&1瘴WZBd RǹY:LP2H`~֜fl9e& 0K dq rLQv IYa41Ȥ&bp{ 7IEkRɜˆˈ#% o8Vu53"$p3Cmއ䕳RLTHĥүu6eip 93++P6je%l`nYPDqLY7٢E璃,|kK֮JzWpsCn8RINA?{WƑ w#Bƚ Fh썱= +ud dd5R]]}yFL>NT"KC~j'e@/A R,EaXIЊؒђF!&zތ)tFޮAoZq><&OzZv"w.o2@JI oj"w< HX^:*".Z:ъrr,[GU^gTYa6m2MTyϳB2/ c<80%2uTiW]N]`G~5y閕o*v)e H4T%NT B ǂ)CNLoQ=`&!۲v z9+k*4K{)5(q'.䉂M3-9,lk!2EeүB\TpU4*!]7%6CI YX`~޻RywZr\׍z[<.r(Q2Z[Lqל9|,8ƵY#sjėMH omA!.]`h=tzt8.Ի>QoNvxM{B'zM\>zӣ_Q$w6ͧ W_w2&*!SSFRcMPJ LY3EP0C )yELȓVBs1%n<{6ZmC*-:yR%B*ȬSvmMg:t ڸ[S?ҽo!klm Gfe X> ^r[ǫF?Su\t5V\2x?\M-c9hk/9*h+i)у< Jv@VcZ=Gͥ} [rsd!*OhMLi]T <353 13kק)@BIAP0V}8 pZf%. -pZMgVc5?~:"@+!yapJ2q۲moVwsO4¹;$G}Nf@ S JB] SJ dcJ`0ē`pЇB"I#`Rd4_L6x-qW㕹h̃F=Q]54]UAϔYwMQ@Q9KMQw-w}"ʝ3Y46_7YmxT1W0UVdeEp.U3_BLcgUd`TqƷTwd7Yc]sBf\8cw&{.,iŶ!$hMFL@rqeN"''t@0)Qࢷ 9ےr ?ځ,FR[y" #y=ʄz6=sIUf*]O'BYĞ4g#hwZ_9ղ*8 B5Nmv(C@QdrFA^Uy) YQ>ov 7<{khܥk}K^@v|r&蒥7-wV1'b C9j(Y$]YmeL J(555 ^٘r*u >!9%|@nb|ʩZ N"8YsG֬m3phq67KM7/Nf&>YH @k"w"hytI" Ƚܨ,ȸꔷ]xhCE~2[(v܁o;~DF)lAz>ԈP>_돣6ƁJtJ9ʡ 4YȺ: kI8: @H3IBye<%PD0(Xpyߥ2 Kjd:ZNۮoaUkb|1f$A\OYrD$G'"ZIK)KW[KiEH\C6Ol=9=ATֻ>f$&jPQKOJ#k_EkyXY/@<7qX{}(1J)jfy1K#V*]R?pI3b|elW1rB&xx=riG~uҭ!Ne׳ZycZ^N)+A ‰Ig4,ݎVgS[uyޔ t?J?pEո{g]E>;(yȘDb@jB1 qp2ws)D,^,B)a tcUP d2eƱHG'$8@H:dB4Z jvn @AGgt5#$VҫNPqi 3ϭtՁ qg5RF匶\^w5iiU{-wF̄ɧ3<t~:hz:,L(RA1B$Jݏʻhœ0 *uz$"g2:8}bG nBԒ)+rvk`ڂ d MZ!;oMo7(6h]G&Umf7OI0E<fm;AդQpAL<9.K(Siכ: yN*pwhx363x1p7?,j2-b}^%r8zo-cv懿>y~m{bΣӽyyYx{T"~Oۿ~a.5 My_)[82WO_3Wى%ȉw{IQO 2Aj*]}B\TY("`%Ęhl64'4؃]ͬ,m^b˯ ~dL`{S쿽.z>E ^pK:6|όxȃga;] NeBˁ{* FULY娇{]Kј.Xl!&vVtυVS|p>ImI&}G}:F ` '` V y,V=DS:~;qE1 ZzB8w`x9 ,DiPRt&6JZ%^(|YO:)Yl r֯5riq_L]Lr_@ZąGoq>= =oI]eo[geuVQۿVf(ߚy&1E,b'ښvQ^Ne)]HEP &8+ 3> xԘ]EzlY,8rv*cW1]BJ6"H8=( 1d2wwϊ\FG[#KKt7 } OiOf.ҋ~%\/s?J$?`o"ΙuKY`^ ],Y '()&J9X&N&h)'~6Ë0t28V߭fH c.p3#s:(xtcj=m"A89kfLzVҝ~OH=~lV-W[~t~19J2p39;Kŭq8Ѵ.Ʈom}NNro̻=-GË~]uZ/^o;&`\Qk+n8zh9:{M^!^cMWxw-m, '+:i6['%ndIJnܢ+imKdɶRR>8ə3bIĐb$Y*,'`C+Q|]M^twUN-oF:2ꪄ$&~j#A0~0IAhN/ BćUn';<uׯϓgyū3Lwg'_ em@m\YBx[AT4Ul||rU^"slΆh@//e!CjW~‚|Ԅ@}+xHlUDD T1g셋iX)f U[vZIwJttũ5`ќ`0 Zt("CT05-67S.N*UO'9S5YMa$%,Ai4d4(ϐLf:gN1C;uj! AT tLtch:逖]RM#M Nܫr;@0} ̖-|W؂*ט#*7}%1k>Ÿkk/N{Iehm2a+Y z~Z *1[OtWb00.%aCZxYR|[uTJ6?yfIRhj4Kerg cr#<>hZ@H}m=8e|b^DC/o˫vVN7HFf R01QJhSvR^F)q4CN]SJ1AHcڇn*쁮4S7\BW-y*d9]h6)ybZ.OKW}hmXFWk$hǂA]]ݷ17`)CWM+@+u ()=qV6XU]ZЮUDy(07` Gij% hjWz(8Xc Es\BWNW偮 V!6`CW!F ڇN؜$t7t%d4]SSpaEYLctFSc̲t<5Lխz.YdYr,KsIa[-9KRp 3h啈1>F@Kv~} P>R=m=Oe !vttu߮ DWX>փPS*;OW C"ι÷-=B| *iV{-=[SMoA`.֋۾z>8 r`\4U h%uW-TFčUYB:]Rj)P&G؈- po ]OJtt58WM+ B\*|*dGTm# .;0a0+[zk[Ec\Վ;k7sQ_Jk;[vjkl=;l-5e'M^j9.~j v>NM랎:r]o2t)*ѩv3G뿦 ;4c{>-џ?4Lw'^t]%82bl[9Rɉa8<׳,y^A/'^ ^u{-:\ߔ~M^ fw\7SS(:<97}(xQegP$wɤҕNp폋Y%Kʝ*5J C{(Zf:fp^ϮI,ש`vyˏn Xbsmc4+9*a$%8ˌRs dޱMU$4:>u=FpΉsC6,p)e#Z2-ˌ2*{,75ݸu.3.%W\f|u2PbU~\&\] o|VM/# VhbVk-1 frk(Sn}K3jKZfӪлlf>kc F)˄[I%Vr,s%16ZE}%5] MB& o&-3)<cJE$2 3e GEx _=FUa+n|jc]Ehɘd帡 3qA9;X9a4 $B?д!9 ˤVFXgA!`WCe)2q^gc:v )!o>4"WˡÈR*euuJ nHx47Bΰ/] Io`M|P;1چo_쭢tJX02 Q IrnzO´`_ B0D ?8&ݞֻ;Mn ,,Y@XBHHkZS2E/U^|Oս@ 0;C>3iqiT~1\i]AuYw4˝XgW͟hWDOn{>*/P_ K)a+NqhٕoU i9AĪo'anTDirNIxrizk{Wyoϫ ul |^yٶT۶6I'Af eX*Œ.!HTL]YNh"?A V ߻6t1ٽ煮Zތ-uZeRU IMZx;(O $Fpj veFߒBćnW/_ ?O^N0Qg~#7coaPh-7n3jDpʢOoC4ț"oU"WnKY wn1ِ ɗ}[\:\ {P.Ebzd>jW%0&'n 3%ت4beZDe/\b#cB^}wG׽1R\[$ d=C&VhS#is8[vi߸^mcoaBFHa,uRӔNsze B-t[A6i2 X#((38y/B!)9KGε1[ w #uB0,#Ly@1-arne [wgťQfl C͗] {7Ȇ@=u/++cet,jk]X|>f;ǛBNP}+v3ɍ5V?{ƍ?6e`5"Yu9IcgÚ^<č%6Wd=Xq7Z:̌iW)`cW=u8:B &Hp: T92ٯyWǓxd=X:͉-`[lv#l]mFVݨY337T#4¨蚘k89FS(:;Dv;s;ҝy#Kӯ<1o7KqLBZ;AAx 2Q( Qg-$gM| ]ΉtHR?(vJblۂx3h-J[{/މqA;AypA%*g"2&"YɪG-PGL\y2Әc"\^L:``` t F1HeX"QsADnB AΣGوFC@FW˾ݔ||}Xz3nso-vP,ӶHDW ܃I3 KK2hܸ7AQNK])B2'HTnEj[v_@jXBVK5@(]Ѕf:opod+ %d[R2l٨M^)Αڡod^e^'xn Ez3|{1,&?}7|,לpµi 7+J}(E?YoE ?_ " _b_8m.CJ%Z s};>c}{=8yco|:eA& RC?2[%&0N~Dd{̮E62ڗ$6w*韗8 q˟^l轼ӧ+."68b@$`p9i|-;[DPC`WrsC^=}cav t~a<闼a4I^rĮm8l@>~۰, hG7J؀բ- (?OƗ18Ĭy <%o62Ɇd 58{߿\\ (W>&jyPp"`x0Z*e\F?n7vE;2ióYtsa,Yϛq%iiYnB~{_8UV L_4_|Gmאinn8t1WvMEZi.Gy\T|}bZStOۜ.'P"ТE"m`/eOU`ΨM:Ly=?O7m(vhaŭNݭ͇"Dݖn7ݤ6Ȕ'CB#(&F \f'0y,ieI$3ܯ?=//&>/JGy9Q(-:sɠ4W>sY=O&i;:Ыze9Qϯ==TgU8T¸?Oꢷ)z[r'\͢'[oUAEh̹Q! jQr+QxE; i9E;3ʚ*eͺg=(*z3m9-o[=n.þ}O-]7p8|ֻ"tFmO ZێiެՏVrH4< [y*"1{e#9#YS 0dt,۬$九(3:Q,0Y2"ǬELJ:F˄^j!rs,.Eo3ʳGRKE(p=b ː"R9K{Nr/؟AGR9p[o'dǙP/tmC1r, ҧh0`d#9 H F)²tU^0o7GIzyyDU %B&e7ﬨٯdjh&DKqR٠s:jX`[=:pb<%8Ƿ; !\D/nrodcS<]N-(_(y!ߌ^Kr6%;(Y2̺uɁsCHCL 5ٜRx"R\`ti"&K":L k@62Vj\Vb“0X. 3_2؆?c^M7h`8w3&9A+dWY p^{5hc]?HdV4 ؃$ () 'JVd3v,GMHeĮ&~QZ31͎#Q[CnK[6cp)pN ۍ&57m@$ 53â@ʹqLQ@Tu>Y 18Y "+:5 rĥ!HEo@HY'ɨF K)Wg?F2 bq,"ʈ;DqgVl)x2<-S"$s&Fytmʠ ctDr2Y%rʑ'4%HPV%>RVFjG6e"2q9uV%⢨w59+O+5&B1R3H,diLs?Zux \<<6;CYqxxۓra]#>cmfF:=y?ޓ匟l th(bX׹9~nGw9سps P#MLF}ID" Aa6[ᥕq h!));7nTl\no,YYbb>z;EV+5Po1A?12|i[W,0pXYjcOkXr|;[^_`19\SE{:Noǣ'2w$DXM_6=4,VXoVwWe[gߓm2mNڜڧJ^=] Co%7eȔds].h<-sF!1Cdb(Փ8KYFU4 rFru 6gRc229Rg a*Ϲ ZJ'vT&AHWE [Zf;$:a{vQYo/9, E4Y4@/ t$u hV܆Xͤ4ӳ'pIY*36i6I@{ D&nr0 )[DTٱfCk=ZN1=dee=&΁zտ5[1Q䕕C<09P yILʃO9y287為S*k >{{(it"qɲ`$jE֒5B!6zތiV<V iN_v)^%\';Y~9p"Zh}Z8YU{6'r]U"{|eu L۬ŠCD<Ri # 0Q,CNގv9}6CMcTVeC"Rʜ$TO5QuȞ`Z%l!ǂ)dnzYe6CWY;췸 z9- kv71jҴMVa/~YHCBlrO&%\9,\^|9?.'VG8]vt}OKJ;鯗b@oߎ'Ww1wrLRM7&ԧvYnR%^&Ir^H(YR_g}C?fysV7 K & ]4P|ͤ3հiŶ}*B1`>}~O[~ћw_\[`W{l]ʺb2~G<-X=xwLgOViYZ7킊6[_rI4*5J1yF" kKݓ|ѝnqiɜp!Ѓdd+m) f#I`fRuEOVos R.yGd2 'iJQ"[BD"藌os舽Vnű3ǘs\zQɿ;8r#p"2_Cp/.7[ٱ^;AU͌,ӣ4f,gat]$"C&s>IU{1 +Pj8(P}FҹJ4љTf'?:e:mS%Z֐uBo7&|OucmL46bl,RzZJM=Qu^ֵGN{-}|;:dCWw Rx%Jdd@СXLLTSn֫ࣄ#&6(_?Cy TW}Å!w7 j7` b B_^嫶ԝ˅cnNn܍^;+ۏ mJKH5RH-bup]S}WX+5@m CL=1WШPXP{6 mXFW{-RqV56#8Aw--KKiBwoK8">~O[_?뱤:=)=Ho}6-%)#N/-8]V0?w{Z@VU]ձ6kMW?%5sYsYsY;WLR jm/_e >t:tMSJ} E-W ^-e^lPW|8nijz4o%){賔"*Mկ@W]KXǮuG>kj4ԝAkjZ YkA.:Vȡ`+?yhhiNθ i (}ga[6dx֝[Mh6"`[7SHb~[6 g +/בFy-a篏?jѭ'Cmeq;@6:)8jPQhb<{L\ LZR/ N Aw?ml\F r/2Ls{o8[nzYǦͧ 6w7^]Ͽ555RkWkm#0 M5X\rulĔPz:8jg#rнҝMc{dDne9] 4Ԧ){ve\[8:7SuZij|Ŧsm Stz'A;XWui.ٽNΑ8@92ܑg]ﵟI*IǽM ]<^G?G^>q|][]NhqmSl4c]`jD?l̛s-;S[tjG;5`d҅g౒=i7[`-8(]H+[d `ص51e+v^e+N+ J۠]MPW&#]1p6] .P.Zu%]MPWAN`ѕl:L@+ܐjttW 9Jt'] -j:BxP ܗ%?v%&+xh2\|ݔ6'6.JxX?4'~fWC[>skҺjfcmorˍH *1;{0Z|'X(MOQd}:/>[ij>t$?|VY,KYdfgw9ھ\(WP|ܺYk42QhaZg_uͺVͩ~~jݽd).oVvr 1kҪr#YS#f+lt%> u%F2m>\.ZJ>JWF2^h˪cE AdȺ 8H(]b ]Z HW T>\sѕJ]WBij2]]y+i7%+4j@یt%ål:BMʍEW8CNAѕYҺ  YcѕzEWLJ(} NQWd5Te%+=S4<:e@=sn6?N7xSL€* _yw]n׿VnPm !k\Ϊf'U'-S'8fkzUYk b*ĶdطxZ:R{Sw1WШPX7hL:WFzg`l;cJBMwqط{FW\.bZJSMQW#tSuM6YJh{J(DWSU^)HW|pp_\2Jh}ѕP2+mYXMfa/!>׍z0ڱ6hFIEWn\ծU*2kGWkOjm)A]MQW&88`6bܠVu%]MPW?S6b\U.Z|gP(Q]MPW P1FWDWB;P0`&+ROj0A6\rѕІ;LIaEWҕ TI'N;Q7dp#0YܖSQj(ϲ}u*PKM!1u>C(C 1bx|$dѕ:SʀEWԕJf+|r \t%> oYXi`! #GWhHa|?@WjתFo来+6]ic6JhA+D_t5A]cPѕb6ZGJ(S*:/Iug_̾㏭f?E?ǣ/UTwXIo/9kl^9lY:ۯ?u͇<9\^eWzst$7,;pװW۬Z{om>Btk-35a<69ϧ_fof_Zv2;5W/ rWⷧ X>Hp+&b.㟮lEGXekSS ]6/]yy?f}BRPEHWwJ8p> -)n3gfMVQ0Ͼk}/?^o2;swfZ{ݻu/껺ʶ[![=Y4b2UkZԩB %q:gQ*B>LS9ז;gJ,``t*] nйiI'?$NQW 9%0:] n&bZA )z9 [V}8uh cwȉhǚ4'61 U(ڵ5O*#]1pАf+}U+j2ޙtXb_܃=IWB .]MOWC؃FWcSוP+*xEWB;(acȢ )JG3w K;BYI@[LxRV'Vp6>_O͊Pnm,Za,n) Q3Sk@uzh19M61ZK!PbM1pz1>}Jp!}-+tTt5A]yGrz>oR3 Wg] u%&<ڜ+&ѕ+%]]ڮA+!= ¸GiG@Ju:j銁Jpm6Z4J(I]MPW`銁+ !u]1e@,8")`7d+uZ%/ 3Ta02*pquŔdj>PFb%k1] -u%Xt5E]hXQ7(XiSqЯ  J;f =6eHxs٠e`l|0(s2& Z5Y%Aˠvj)r٠eX-ޠej2 0@63Eȥw&:at&;s!(S&[PqqђJ>oS(}uQb>y#/4\yL-z1[V>VЌ@`;jw@0ңAZ5]颫]^Jf+ErѕЎD@+.#] 0lt JhM]WBLue5`2K}}UJ(j ag+EMt%&+Ut5E]+cq5d+5>y]1EWS{Uo7V&EJfRb6*\eEjޭ9Ǜ7o֍Bu_ͽwQ钋AQҲ?ʗ"_)ǎ+[,+zǿ |~īXϻB4n^Wky|w ẃl E9PtxӫϏ|⓽5zWo{B{/.雷xb//^n^;R7➭J3)^ka劁Ƭ2:ּk94Yٚ,0Jzxx/dp" g@x҈ "'&7u^-c7Yxc*lC Q%C(d[5B>q.9 AY55ksH9QQJr-^4kZ)*!w3 `FYu&^fC_e?ij;Vд-3򨻐 ܦG2^-RX8tLK#K_^J.G[ -*h`1yL΢&iA|Ѭ`!Z<%EPi!2Q Y4VLta,9ѼyT ݗTe2jnu›Aq*qt/dA*X?<{UbΈJ! l+dəj&@D4X@Ozc ?wu2W)Kb,חXk#xhW(cѣP.)6H@/f9І(T`%\Ha("Gt3 7@ᡔJr@Q uw(K^/B>J۝K`f&Xj5YҵD/x=EȒ,B\5v(άIdD"PlArQ+HB5@բ=`&qH416ʉl\j% M{ʾB:kϢ;k4Q8FJ"5@؉lJ&MVDEK*6 2߬r$S5y+5U ~cȔZL/mmM`G<5@PE^0ަ\W.ڮLG nfV"ip(t!lL"ݷQdPf1h֘U}q%JV ]"Ș43@XFvz4L@ҌТ EI z XmڔRA5r#`bsU.j0uFAvP ,)aIP B>^roުŰfP@>-7`E]Q("VR2P$@᪝r JeDy ahaO1;JuY1@jJ tT#!SsLW=g@Ul?Y tStrgAVA%ob P[4zC Y(m@4XB¬aFZ&] -L9EzȕQBb)Q2P g՞ QP }J*]\0n5 D…a3;kSFr"&bQpHl!(tst\Pk\7еhDo2B [!= T GVU[z G[QAWt(0xWېr55[A jz @>рz_ZmvUvK~|\bjG)voA.M9.}G_jz^^pYj{}n 0 ⇍2׸XXoҫjsq~.{& ܷoϥ+m˽ח.._naжv6./Ʊb-;vb~PR %Gg*I98Ρ[+گgQqi:N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'u!ySz5us:q-ɯ Qv=E'PmN v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'vYwbN N ʓq- ^):(; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@N ~N 8'qm_RJv=E'P3 N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@O z(㫋V}iooo ;Z6}Ym?bd0攌Kp0'c\\+:ZvRGFO¸Gt'mPIW8t`o_{JXNWO1]t0N= ?-ه #f/ߑZ0 $/fq> aY;f\HyhxF"GOo6B^r9SѾUI_ǩwBdǟ7 c/"/ }~vЂ:\_VPys~Z~Uaau0}D;Cj>ۣ] @?fnޭrO67˫+W-yw:y~%g8#sJCSUA2gh+ >~F&8ҍemv֑tLBI"t 74}E"UKRV|:WtNYNQȞ 1pM8bѺuA01zĨ?tJʓIW:tњcUM禫8βzN%]utQ^%ʑ?/}U?OͫZWzo]xj{st}v537RwW?;m\ȅ8xPeҔJ;l>s΋?Uz}NjLܯ>}@~C?l<3$4r7V 3׸=˦"uQVH];m>R4"6I>%i[#Qcūh{+߮M/xkM\w1Gg)ԁ:.μh||jhoיn_^m@Z̫Gen#~3\řH9Ay" )&5O6Dp X1o\uf[qyZ>#o~ޏf?OXIj=#>~w.f}z.{PON穗,*T45^F$Mm5R|8i<_Μ{s/ }w}:>l7[.twmdjg7_zZ^_o;o˛7?o؛ O6ކDh|үֻ-^3l߯\-?| ;r?NЛ盚7߻sz>| rmi? ڏ@#0OA@AH꣞ ҳS1)^q#?dof;kRC07dPHVG΋ΙfoPKg?LJK8ݗ2?1PZ5f0-S)QGK)W8ЫD".U )RpeOJiNtwUS]]uHPN·3$>K/tm%+Oi]%||}㶿{}9j 4{okr:Da<,8TYhӡ;+. Zo1϶Z%j2)i:JW!Ź\ imnw}hV='}4ϻe2:$9nR$P 7#vIY^aHbFĈ75DPywc̻ m=ƥ;WĿ+)ޙv{ t֐Q׍(Vڴ~-/V+XSN.eSBaRBjϫ76=~ԢC[擭ƢY v5=M$Wƹ0g 7|ύ0UNip < a+adTKPyTVdިjmqvѝe~,*qdVRvd|KKyӍdlu(\43ϗ ҫmm·PaF-k- ʫ?^Q9*gRbr[ܷ.W~raX$QJ=<9e|$z"[llY+Hوc'QrH*V1ܣNH"NIǃTa#K@4LcĨqXXo-F#.}V?>c@l';T5-yUc wϊ8V|V9\b88*"7*f`49U=4bU H1lF;Ў(SBfswZX`4OokU75ǣ  E\2!."Iܽ#xo+Iaa]52-h8$0Q;]nn LC#KM($Uo7q8ոh<1@cD43s`eJ@5[eEȐQ(VB=p` -֒u&eqɠ1,Ča\qZ'*~x1, aMM^\f+}}ʹ7dهt^+< QGcprqXl6$F>fϫ|ϰSKby58:;y[覓7@γ?s3`O ø0|@U> .Xǯ`X8eanw}MlJǻ *_Sw{CY·#s#볧O/N3L',h\ 9'5Mxߟt{5='@z/5, yzՃ<M}I?.o Fn$M|gь:׻u;k09sƭˍ$')>q $L'`U:7Zpz._Rgu,}.Oβ[h֫"ݹW{˨8'Mg >TKϯm&>к׵IњAdC2Ưo_Ya,Oݽǿ_1=7.kZTQH \.7MLjR{n(vtPUjg0( Yi#5ͨ{UaNfbф%Ro! A8GDL#S7=C!*k$\FF G%)Ppތ||XBB m5(mV[[|It{Ry5\yao ~՝L{_@ ?I^Q[s}ܭ\êwkXH& EвTKk7BΊ%{9COatO~…,\H@E-h)4![%USe_nie  UJ]W'{]^*YNuM}Vlfw2;|g2ǥ@/3W;YAyܡ:e0tའGY=gJ6a>|@G)9vޕp4i@t80l@\|{ y>zf٪ ur+{BTfC\LZT3ʭDŕ O(AIh)5y1K$\̐؄ A2 ǨU1kQp k|\kR*$kq{6;3}#)d#H1 "!iK+EcvH&S8V@!::RLݑù]'ϏBOսO%P.:m1 !* N IA_X|ȳ Ep)\eCS(D1lOgU+Ps<9/^ ):6:r[*g0;/R_:qG\ènɏ;Ux? SōgYu"0[ 7\v==+]0eA>7^?eDmM鼮R_$,N >~LVAw5ud˩n$h}]K3>a&#b߃KC6YRgr>50'ޗK,ug0翼5}}vw/Nu'o_<_R2&7ᏻ5@ݡT߷Z!TP߬j6U&ߺmMrM/W{-[l峾;vɡjQlTUq7>@A}8s#VF4bH[}B6bs=R ol 1҇tN$L Nh}K+8 )H:<5QF8H5ÂLK0yy!>D`e&$QPԨ`5D[r(1)LXidS砜CrM:Οvw(׺[xnWn)q٠9Cwxvq)H~5i™Z":Rjsfȵ1O&`LYiVkmDod;6ʉ4ms6$ݦv 5*Cz0 lp6HF7g)Oc.*58l-Tud9P;Kq2fαn0Bې= (8 ly+9ʲoC'Z{srl?޵q$Ч]GuWsr^ek )Y.C%J4ٲIk89͙_WUlFSi˯p?}_AoK^ [hݻ [i^msi裏ƹ4Km_+l'&{4/i9 3f|쯎lӾ6,y?ZSvVKZV]:ɲ8Z!S?=䷢}/<e.Fb)4]ĈB!Fh* KoKch\[S.8sN)R^dHzIJ ?ANfl|mOT/~i\DQLڧhs,l#9-i+(j3BѮfU&%Mq?Y?L*ʘ!|p6#aAͯkm7a4Oj0|`þx؛ɶqmk.&5G:ow@ɋ>t^[>X9S_b`5~n_?}MQyio6(J1%..9rNc2Y%$Q#ɦhJΥbEg0T\sj، qƶXcK.{^\VxMžL1Fxq:]?NO_9bwgB%0%4UH(rfp2eo{$! #Dj 4TcLҪ¦8APm;|Q3x[b.1b7#a㳴bb j7-Q[=߀ڃbƒKAjWݘClaETcĪZ`U&} m)cbpF[ȬJ`+:25)X\^؈L$٨H0F|h}Af-"ƈ(DZ"Xe 3OSj1H%y#O# ,l6m>"ezd=R'U*΄T. 6mP7C>=g#7*\{F=n7D?~GƹG~|z a Lҡ#b%IBfV=jV=RVs $l*I(/!J 4 Gר̦[9Q<}yUUwoz o~v:56ǷOΩ*B@{ϙ KﵺhYGw nStljܮu-/2~~oqᆯs߀7 !?pDYyGYC7-K֦lۛ.DU&{o ͕Ψas]kŚuA!uJk/:^֡x;tiq-=Jn.Z; U2V_@j$fۂEck Y+=rm EQ|"" Ԩ%d32&)02tVlll7#gyI"/ iPΗF'e"x'!{CH%l5V2 ~❏n $)K /x5d#R+1{ 1 %B61^,ڽJ6fAZ4T-\D`Ava!*^D$iPKsR\k/'t*;k@*oʢ6.$gCV`,(S)>\AH&yV}=aw߆`ж`v /TCuE U2Oktn!uX;Auӱ:J;Tu:6 r I%VFvELsqb/zl HIIc H[X2HE֖CPhQ{Ef@} A =ӮO|zf8mnqlW?g%qo+TLULT%>5:K9"v"{xB)\((*ҩ g1ΒOقJgm)%+ ]3Z#I@JڿXO+0`@, ه+*dQK?uiflP1t+|{|0/Q?,TdR̚?d ̶` l)b=(L&:$9QA$vLj#N9{a73 'ilM($29!Ǡ(<ۥLk\D!J+!gYy2FުyoY 򺱜5#gC9ۿ^~0*_ޗckMEl})8 %@= SIފ*SB -Es4D:dҠтh E@Ybk QhtCdAAՊnME@dD 9QzJH8YJ%"G8^n_I*6s,eb3//Q?0Y|tjśѫ MTyocc7.et4O_R,nm{ 6?|\.\G7d=\Uq~C7p@䇮߫'Εd堙_ZGw'#-J+.G5]=xj\ؽNS]0Ѝ>R]}0>^}4/OSo(qZ<#U+Ք9c>jن!AۖŜs#2"[Gj9$npnE8HÖ2*l0̶Em]~xMs=*ڶ_7,wO?]_ld8ȉ ({&#{=%XĂjcc3כ~'\Է34EV5 -ltv\CäcOX}3lg뢤&ۈ#BmD"6MuLxTF e$PoCa 1a,ЂzUרE=GXFj$nMUZW3l9ҩ?.1K7bǾvE؟1CsVPa=bZ14&S$cH-@ZD#D]#ޖngD6yɵ' yLG:`8l9-Ց&ӍrTQR]tl7I<`kkPtO1VP9ĉ) T0_8v%aұ= f[Ӹť#wj}މ 4eIpse??W?9Taɔ" թ<9i{EZ 9i&V&F$6OBj E]I?9Ns<lbBl[مwzf[ƥ]կ4>cB~2OmG'>zyz}[&89Os9t*?f}f7r<,6̷=";%(&e3ѽ|)5,:Q]Jt9 ui&(ւ1_l.@r6V}d2(`g{rvH?vыN\Izߚgv?Wr\NY_ٷE](WwO/NˁnHKɥU ^A*")YYa^U屩&R!Ez$}k) 3y$Ptcĩa10OrH~sM{JZ %IIT+8VK,0n[ꃂM{L+^-VU jmr]*rn!Urj8almB*a*֫`Lpލj7B-l$'!s$qU˭eYHR{2B~p߭6ܷm߆`|p`(!p ^|^'&!00 TstO׉biرQ_gs(04/-WtNZPT[ D[ YͽX~( Zՙjɖsm-͍.R-BR3c\/5O|<9n*ה3|#}[҂Iorז۶7j9k rB1?էxGxPK !4)iR<2e`uͩ/O/ϩ/skC,m&CQ;MIh 7_] ZcMTcN J!Blx/$S18@6Q[?u4ϰҫ6:ʐ_u]S9bѯ$4&[+.KJXuLnYD.,~ [Ύzԫ?oT|(Y6 6e(9s*O. n\i[gN+aV*N+zBO|chޱP^zK)zO&"})?3V<9$ϳ{ -϶Ca6XndHd{g*@ICGp 1rP8^wu8#] %!KQi@-HMWRWX0͕,V0k۸B` }/VU.`l!*N[*Zl>ԋ3}чԀcqJCT+avڑ~ 跋V?` : H ` 8)|wse97;fsGJ狞;_~`=To2mؚLbB,;iFrgPls 1V'IJp4=W-GVaC4lv !U acb{…Rw54gM?׬X՟mXZ[ lgt6rxytt|~b}.4 DB-qr9 DHn2MR7 kAڐ($")z^[M>d|k=QE`1%Q j @T6S\9",koiO*;0?oX^` @AFogFǻg4YE5|_/.?):*NAsi1lIO"O  ьwfrP9xxAlA20GWY,'$ ]o5dPKYJlS59"`,CA9j9;DǮ|ϣ{7ro鲸nV &n4-f~Wx`>V fƫoWnv8^=~zIg ʔpM!}I6NG)Y8e氁+dzb/dZx z[5)eBdL|0y5Fqff48 .Ah we9N+a/^oYA)_ U_?yu^D_7'{O~9~¹;U%=o⛍$$HH5;^% 4Öd߮XE x:l]CХ>J|nxܵl_>{8t<<>uzZukVָ}2o廆X;"'o Ae_.ʾ27$TBjkJV~MON!d5O L'O{F:#9ꁵydmѫEnlƝz$>]_}v4F7O*}z:+{(߻|.E:84,wW(+TY˯{V# =|a{ }'! c.CDZo쀋s+\֙/)H8TC fY{s".9`^˩>b}ݷbu*LυShjNTR0ȲW<^{e{-KrgjAH|be:Fd/֧^no]0L^tODՑ~q}?RϲdӧZ}|_9zzvuZiXvwDv+PĮ_Ԗ{$q+!AR 4}q N-y?cS*2}L@h'&L( *8J2 j`_t5BBR}!ŘMTLŊ6yPoj?og7q8q:É)@gM8MEL|j1U3SbXi:xJ,EHZj!T=e6 'U`83 |p&9E|qfn'd_"=BWǃx~z"X5@ XQAPgu~L0 gGlJ͎mP 2aREQٻ榑&UTU=!g_TE]A`X",͌vwnIvJ!,`IuJE)E$W 1;`a8dO] 7y P< pgq$?1j~Wo6ą2EQ1nV;?>%u.{,_o?Y {?xm~ y5;u.,GGq}$.&fd'">H2( F:@i']29~a1Ѩ3C=Bϗ%f}ɀW (|O}ӝ;&a>~_/Hgg1X xc8)"7tٰ=O1bkچ[>z?lf?|ɧ8/y~o '~TNru-&Ǐ$ωr)/S!s:-Y0?1φ-vo`<&BҔx6uŔ{abP~|<}?lw|Xhҭ\"*.``#i,Lw-Ma?/0l!?uяז7k6IĖO1]k`k8_{\5Vc\uxשc~zI|xͬя+6yq:~ռ8yZ W_~Z~ZuRޝwgs^~Qhl&k9-!03d紉ɵ_ r T).N6HE\Mяfxkr l8Tw>F;rAɲʕk҄2gSVqUC% '69} w *;ǐbՅ3u= ?ag3jϝ^ȎIfo]38 g(XO'Nv=q\X%ɉ ^טrR$X_-A,`-tO =C@Ct >4Sbpy GE'vÅtsN5lV LKAs4OMJLAR ߜޱWmrw򃺲F$Ozˠ7:18 $Ϥgu53P;Lg\CQ@-估E8w1AO©|wmЃ;ȫ,-I\=՜!)!Lz0ѥr1ZW@ov%_%'M~I~MlKP Thuq">% ItՏЉ%ZԀ!~t°aik9Fxw[1i+vmZߛ6DEbU.'KprwF)Usa 3T8ϓLOQS^BYv0>^JY Z~csC`^>|xG/(S27zwDc{WyYxz6>/*f *UsuzEs|R-s;$Jz^p;>Nn7Bz'!t kFnf6^'AE\dG}vN;g{ צŔ'UDFc kQLEq@[4Wۨڨ4'>T1AG_GOzQ2>|v>zV ǒM$Ixf tKwmTߤkn2]Me՛rC{!1% hɷ3u> (7qb u/$a _7QR%,N򧡄p#)H+  DK#^](E%q*p{#g{&q QsEluBD-<2ͬZ˘RȦVn'n #7:mom-(7[^4[.Ыn5aP%9ٙpK$W] DjpKR} Zjw`JP8vB*+!j;WHPWHz 9 WJ0!. vD S2&fOʒ4|2 c`&Dfg'EOdدR%fXGCj2J3"ZPcf۷bB^gN (=^ߦ iTr~/^93S;GM{9iB/p-VWƥAmg [T !3sT2Zo*={m҇Wϟ\!v!•5]+[GBr@-t dlW/"2cRޗ pl 7Ka'AuI~ښ+7ԤB[4 9f$9ZO逵|.hKw%oGJ*WVܓj˰Ybͷ}r̘BrRkRi \1+%%BBrjQWH qCp{5;]!wXJN \IBd;WHQ;WH}+D f drWqCp+;خRR)"\ g oٹRIqRZʬUnVbJ;>wsz֟8aǏ{鿻˷ŕT68a Ic<L9RR8 F9+ΉI`rGn t[]J Qx!mz>6sNAW߂e~|5݂ ((t4R>ĦmZ0r/XY-KJon/IJh[sYZk"x&:(9ybzn<s!уV.wDk-YlRa.WlbԖ52mJX 6tCZͥ6 E:iM\ʤD3"` W3~ەꭊCjn#Q(\ki+ 7GDpbHЧR S=o?AN/[;))eJD>tN2E@x,ܛ#]aR{зq=劭tOEKPPZsGk@9&`:0 !XnWH$4)rЭBA,3xeH]Ou42QǸָDphiJ#A=SP5rCZ/CN>.ϭCK7ǟ_Ylrc?5n7صF]h*A> ^Eg,WWq.QvK&RFZ{B%6'9Z:9.e`uu#mrr@CZ={>UTD),xa1E=YŁ:v^R$fe:Dڍi2`P\{` D8.DxPCɌdXGD"DNL@5rv:v֗Zext̓MfjsOP¹! .'E2)O**.B:MK3>V*H\@D 9 43чL q&e(#*G9K_LFv!^ saұ)S(K8w'en6?FbEvQࢮ,r]w-~Ly>k1DmG;6=:6_/A {Οn9@H\kJ=!&ji.*`;c$:'-uk?p->6i &,f%36#z"HŅ7HP:kZR1:D̖S4"Cx]wMG5 9FbI)'0"0(\)KD$C(a]; ĆvLNj%9 u8'^q' |Ԝt3򱃜vxB!ja!yE Jc(rxD`S-U"10*ۊKݩwSfh#\xY_2^aŮx:kѿy"41ׂmD-7%n_< P׽%JpX'kuNĊhct(q%jz>&A_\$;iP%ʴ\C@.,皹@S%jZ`oHN;kd4Jc,FxI.rўU /EJؖ%5rKZx8tsV%jKz,7GwD|gyZYwԙ ٲs+]Ȋ}%XGBN{kM.R;D B҂SkF 4,O;g8iT=Gi" ŀ2D. k0i4pƢB3~y,0}ep,4~[-^Nsf/[˨[b}9})P^65uOن#~?RYNWI§o_FIwɺta/G?k׼m}71| xuBy]Êůl?ey߮+X#x, yu1e-GpB*]c uV.\\HB".Lѧxk~Y5rLf|ˮ,ҡ9u?Ts 4'W).Q&9d5g|Ougqs G6`vx+:BONP g t_VX>Q wF̋r|wx6 , T=$f2[xfSY(iN@D)H zgn8>E I 9BT $ 5C4 1#<{8xJl\"fr=(M?k*AM̂6_ _ܓ=V@Ҩ6w=fߞTx\f+sjs/4#$g_1r!"43I.pg԰:gRFH HЈ|1RHp;m<8eLehBme9;n_'^vEK4BME_bjᇆDzژ bʝ2K qT" biP>e^ \1댵k"6! ^|@Ykf,o# ) ;O9)9*IBiR- B\d P A]b(d: ! AaUO-;)Xͭ ,'XQ=x ܠb Mq V#? Osr͑4"Hf[r%\#BdLT i!sD +H ZkOpcJHJim'ԍȈ8=ϓ 2wnrYW'#78-~xSMowp;hg[NcHb&Ȭ{a:2WŹ|YsEj#]aY/jtr`\099SsL_L/ObߔO-IDL>y(Fbzr!bAnu*b;8;TgHR*.>=H\Us-Ғ[=_rY>T2 3&W:;K.F{c+T.fsVfg:>b!F\Agk.p0`@[,n?BJ _225^fY^І8V W0bAzГk.{M7WdUUKn2/,}1<$Tdѡ'K1 awm$I3.9`,zln 4z0}#Xe2*6vXw|HG^<{EOկ/;g˟$\y?_p_ZZx\ZPW\׽rz0.v sןysķZhzMt= 0 F7Tj5Q ,PO TB lbn*_b 6E >") C"ED2.q#6 {ayU>6qơQWqELtM"j4yϭ0`zopB9B-YiTU1) }GZuǶN++nt_CWuK_Kqߏ盇eJ3q,b8֨ؐjj !+VtGƵeVǍ g)ms!-K|ν!I<CW)!"ur)D B-ME Θ\fʤ9.i-D,~>||ˆcWFv1PZ25ZXJHR(˖C=BB<4pj2ϸ<{I3\@0#s?#C^37iʋ0/;^\l(OIA);66,h&‡ŬcW>Ƈ@aϘȍz5~qauAp]㡢)YwaCH`/я.B~uE,zZpRd75~Ҏ\&|?ŷk_? ./~p^hg|3~(؍\N@8"W葫k?58xЧ#NΦ'?΂zS~eˈt袭)塳S>w1L8ÀǮ`OUsÌzQ*fG~v~My3{={gU§^Ͻ]-?95.Esag&5߻z*~@T"JX*K,XEegTN4*{'DQRB i 2aD"z 9R&b#wQ٥~ݜw]ڳngMbx?:Dunh4t}y |v5=`j|Jl]xo|ur"\oњG/: l~8k9kap>Wq^59 []%k gu ]0#n*@S Sr*5p5kɊ #r!B[1R%PW&[V%"\ =ȶAi])fbNi4,Hťבfy q<9CL91elBO VP p`Nyc|ㄡ>PDR-T6F `,#hZFy[ M<0F`p.]9- )8VNݹqnK&y+Pse*\\'8td4$ka1$,1QWM-=u%GK<)m$׼jr D4d劔+,c._/R{3N/n9αSGRGB' A&wU@iQQ vV8V9:(StG̬`3H8@ДD]r2c2lFr>I+$*Q6ۗq.x@.J& HJQ9@Vܯk $yZk{[zˇ.D,Z?9f=G'1$1|<:{ Eڙ* TAjh*ob=<.nMM.ҮN `GJ!RJq5*3`) {WĦLD!+ AJ%TrJGΣqQ;Ca XRpR8ap*ǗsX 5 Cj,bOq\ZcEAK˻ʝV)ۘKl/JטER#b4b[f?̳.(-E&n V #zbp\skӅkYX,ƙ2xI#"]33LGCUK-N:FEN ?7_cx)A}&&B2)LC;y)XR0$tM97i\ |^FՓ.Z.8(Lc@sV8Z)v}"~3F'Z~2hntX ?U$D3ITxjhHD/ȹ̈xY*V5tjBUFHGWGHW-+,h ]e̴2ZINW]!]qV2`Э Wwtut%PM 4 Z3wh5i(Qҕd@I3`pEkUFOW22R]!])f8- Xg*5QWVPt(yj-h)Xl9tZ3SfSzϖdz82V1#L3{#g?.F_whE;挑3N5TW5jr7Jjs0E̠˸슗p2  \oGIߏ?VOCMxQM.i W~qWj˗3 s!6xΏ9oinjc|S޾z;W%Y*RT^kfeٲlȀip56Mr9ǣ14Ӧ!EeBWC[t(RjGWCW(.ѕ#~ $aW(OO)|TJ)K#w֣jދ)u,98y{>mם톯rumfc CMZ +Ҟe{iV6^e;vZMH%i]e#ŷB+)i:]eMۇգЕT"ʀ%o ]e5UF#+R;g¨4w΄?{OƱ_Y搕C@pm%6>c>mJ7?ήZے։)xvn]\Јsf5yΠFm i yJDwffvVa[TBYr %5(XIZY')& ܌qX6w M8E]j:R 7Zi} DR|Wc@(O^WT6herwZṶEZ*ׄkaUP+/Gp1 fʨ%%y?GBM qsTU/"MH^E uNkW Rgr)lr6AdzaH,6M|ilПβQ %6:n}hr{.B Ogx LqY̙C:<NcʠM#]Q+)nqqTSM0Q%·6}+Hjj)ڱNif͊Y&,w&+OxVݻfk'=i=Ұ :N\xf1#h [0V|>PEgk~}eqO " c3M5 'LGd-X/^ڠ%cJrXƟE:b-;.$lt/Z/~CQɤ^r/6$s+#k%M$iqiYB^B|gјtz.ڇgţB=.HWM?OB,.Փw#_R0-{=f;{-1oAUJ`̟ m+ϫUDЏGlzP.7 ‹oQ~5Uv+P6dg9wGn,5.&Q0!S&pZ]ufZb"GЖ]F6oNb.%}>Zn+C G~E.>..g7MN'#A7QU I$w6 Is`G">6 XKR]Ye?ׁ.ut?}~??>:|aP`VY߷C@my[m -599+ۜq_w)3wՖ[og?S7拗¾;`-wZ%$ؓ3x1?DW҈ iyMt3z`Y.&\ U/n.[+|zDI=p')xdbpB1Ai xƁʈcS= f\aL`f$}hB#C~<դ`":Wx6xn˃[KhMdAIQB〿%`h΄EHZw:Ll%.7vxR5 3?֖֟恿ﶱmsSJ کROG~J(:V#Dit<ߨ.XiWש ,l@!׫2L^`GRgq/7.: 2Q&AYe@OfxXpOKFSgM`ZEe/đ՛ +T /H.K=fXt8+K٪@j%R9ac-$o=儰x<) ]K2`z) Kw)?b3V=5?m!2Dh6WH/[JL.0"4/ Ro'@|['H6X1}a.[ 2d׌;ee)u*o aI{ytMD\Dp X:2>ziիÒE*L<2;SDc%h.DpCLDK9 Q ƑՁNR^ń W;> OsnNJ9j;6,4^-Tt1->ȴss7oK~?N?"5R XLt8ـ4*f%j0-|'d̾lxyw5׻lA;`me[DanowoфkԺsj1s]W%=O<!fKMsu, 1*ٝsW@ˇ|;bt/KΟU`B IO`y/B#leܽ`#YQӪH ruY+냉4FDD b  Z<CV0tck4]<$;rlxعI|i͡[Uܜ\T78"8WggLSyPXlĺU vh5Z`@32 \#*D%! j@Ce n 8UɍrgV?xb_Ƨ(<gmq{qEbbZT*ƫߤ/+x%!FN[XF }!M52hpH&nefc-l^jC'wӑ; \oKRܸ+#OA?CU-ǒڰy lVg*ќ f<ގg\n7/%gNdz7/^ yiyL]cg3۲x_/AJY}gs\G~6ޟ5lR竿ʖH'_@b!XvXKh{Mf=jxEۓ}MKca!5[ce [̸S`&,{qi*ڛE|VISɢ`Xg C7KTeVRĔ[7Oz/=|a&Y?Nþ>29'jځ}3uOC/[ roä/E&pN8h1:ERfcӟ)(óx6 c;pGCGTB!k8e՗!Csa$8@èHzw)ZOE$ŽpK!j3쩙~,bq.ś.%=Y"U~.g4]>0xvyKЍk_O>&+=?.[_,ebo WBĊL{O3z$CL~:˓o'íMpU} M6RX/Φo&qPoBb!=vl5-nE4wmK7!VXww8;r|ȫiuNw`b~bx0sFO,*CUd{i=c>twVUJUՉ'l׭xW#$RMFSB땁!;mWmnT!Z, A.,*y{9o܍ H#))AaGRNVځExvQjeq\o.?V//.\Lo Wy2&Q$)x- pxV.*g/Z%dR<:&Nj`*A,oG,]W}ťwZzXxLZ^YDV}lR+bt)]N%ا+TpPyÒ0BeP+h" T89*7V@ 1*k˜+ *߶r02Y>w|_d _o9+:ZpQd8.]-’olO;ܑ^E1Q#: Swl0A2"jL}HHşR ![ R6ˤVۥEݷEYr>gUi}|"|y3B~oC ݩ۫_r~q}e;~_n{X~D{<^ou:5~݀oND{LvDZτ0|rieu(KbuVK(+.& ɲxHю(ۤPRzh;j@; A($IŨX,B3&%'s?Jv$;-G>*6ݒrTɜ)  D91'1HJ+r^1hdg]TT3q\鲇| KFJ0`@ۢCNKb1ɢ#[41SNdoj%ҲS&TI&F{6w7mo7~V"姡7ӓcӃ]c)Og\ɪ]:KWՋO]U{K!ʗǿM!۬hm(K/;a}{RȂl!Fg5lw%c|I碥bC'o\.!}ml([i5Ji ;֞8d5*|a38ƾ0- 6狫ԗ,eiz_tzq==/^dž%FgBe0%*р5U bd%` OPBrm?aVEPZU-'S!±(DEi챛cNj|#nGzm%zx춎 E+nRr#"H0 "C21ۖcJ(:k2qK%}1E⠚fAn&~DZQ6rGY䮥%c, "(U-2&pɉ@+,m|aYYwG :Uny%[KL}JI+pSJf}F]CZl%E//~q@dd*ڄ٤*QgT;(x,fºEFcXC_<_?lC \؞nw5,(g~tZd?&\|4y.f{dkݿ o_-{S޾3zUQ&C.f'p' _~YW3zR'|?OVJ.x>gҢR)Tɬ,Y$Rb_Yn}>՛e~?(00.gSo'n?w91Z>cú}= +oL~X~%r7#?v}Z,0jv>_:7|?el{䉘v{>nfZ~݄qjrvl|_ʩ'ku[lRThb'BPEL],Jmfe%i|rʎ(l*IPb6?8 JvZ iLX8%ZY٭D9ȇ=mݏYC4g,|> S%c(^X J IwS1|(Qtl}_0q}7x-8#8s>..v>?eAyls53r}~ĤH7zGjo}}b7>*0ü>r[.%csYm9WύzS:AuS:\coꇇѶ DBJbO?3zB 5hgA[SRSv":Zq䬖e,_(ҍm&qDGτna{tRc)?./l\nvP;Qu0 6Bj~RY;T֒l)UtpK틌% ESf8!ؘJ`m/ic`|,1J0I$ MQ:jK8!{<|x;Ἣ)zҫEݽj_b>Ąx귩'[.-BvE-r.yPPf/|KtN 󯗜l#^6;z6ҏs_tT,iNu0p5ZfկVX;o"@I`B.b5TpoPv$tD袉!͞Ut4.LU^td)i2g1I: ]%88kƽĵ}utwNװꜯwK1 cGLP.vam$yۮ?LZeä~XmڮzWl۳ENٍӬy `;E"FՉ?M".00=U{[d)s1RjPB*5cTDRcE$^4zK;of^(yG'NmU~/jH"5>Sdl}Zf&~D#hsArOhG3ӎCͧ ca CΨQL2;XYHCcق*}}$c&$=[dTd 11 o!]Z3̶j_~Yk$t~qxry (k`鹛)O{|]>^Bj*)lv1.xNHcm43{((-}̄Eg4(*Xx@z"49Cq1u> ЪGt5߯L]YsƲ+@}alYTTrTY%\S A9V\)J"$Z$B )t7joveؕA~zQapځEM#n!BxdTX鸰 Y\ 1,Rg2D'Hk`}HJ8%2b%0 냉ȹ]VJÍUs2r-@R -ئ.g QQ/cn[zD/ʯf75Q;":Ap*MOs0|n 2FHё= =TkXQʮxؼt#;Ź7Mm Hjݷ.'Av^+k)L@y %a@xTQ͍QmQ'($;[A65I <"C=@0Cdg{DJ`OVmIT>9;"`r팺"݉CLR*QXJPK fxwD.ỢlJT^]=GuŔb퐺JQW\EwE]%jvuTr{u X]`IΨD.*Q+嶫+R!ԫgկ Q.M +M]AvrP+GXT p:@4_d;>gG=^opV6 d@M{"T]=Y1jv*Sz;<\ VkJ ?~Qkdl pr ޙD-cnc$*mghcH)ƠCXގǓI=|QLV7{5]TtF*6~w~/Vk^$^|c&EMo;R|t:5K@rzƗ9\Ha,O^9^:bTF!<|<dzxB`,FSf)3F.t=6XƣWzG#6cØ1Y6)y2->4D\v=۫/u3tmܜNlLٽDLN/y#8";$78ȴTGkREJg#֝JӘG"h E͝QFG@(GT DԀPEL9KZ;+x<-QvVw.,yOzLQ;߀+zїc+EVJ}8RR8ީyr&<068prG~q,T 0x❚q3%8cZ@4h&ND,PQD2^Iau7׋6wLQaoN4/dxfeI}s?Ǐ7, )BKf J)wL$XSxD#$^jW bx :!Trf8 $RLHiBƈQ--ⰰZj$$,a6{ ^oZݴS~w'h|k3tY1*b"R~)N#F)8jJ,20(S)';/A(  !"1jeCB1-PyMdFz%!F^;^[i"]/2Fv@x0ّԅO:UFnߟ~lٛ7)8jf/c<cz6-;|_Lwt2Jn>xh%+~6 ௣jҥOW>L.N e,bqt6Phogwߖ;坙~Y׬CzETܛ-/fR7S`w>~J\Pk?OV.-nJڼ}U?dϟgyɁ1Ec2:; (#֑Њe}x!(;e|foŧ_'Ʊ]Rc㇯Ȕ.LfW8a0{U:7/J?8Nz0b&"h~Bt9Kkۢ&;mL *u\fSL'E9@q G8h=E/`2'v<pOƥq: dO؍G05s\*IX|pC1D蒋Jr48|?Iv߲<ھE3F\|ٟ'a>2W7 x}RVR\K(ʑsar@ gӚ?t0>Mt-w߂!S:܆ ^me =R_]ӬqqAU\Xuk뎓6'o7ejj (5-ihVί>0 UcN}R=zL~C"Xn Ɠ&A7P rT-m^~F`A̓L(e U #kb")g(D0+$\FF G%)ߗf_.$>N:$xC'4.XpǼJyxR1NEJ y Zs7J/qyf}_+tx9x|5ZJiMVs4thnPj>_'Ox]I0UB"xq;Y1װb(i"1Lӈ.,Y ;a@GHUO#;ў+fKd%5 BH\LL(: Wz*<a$%Z\(uaQHXGWb&y$LViǬQFn5zqmYzgl/ٟX;+尴e3jn/voGv/5nb\jќlr#.k@>mXAD[И2'9Q*bmL|B_f ׀H#RNcri"+aQiI%HD0 AP$@2q;D Tkm|<@ok/|jah#SEQX&FDyn("hk,J+'̙H-epݳsc-Mi)"#" #<LKヶ.! RtMP#,`5!,FͣJQH(5tc@"шaPoހUrJWqAce /~h(0 ct_oRI),ݖS {0;17>d~5'N?8g7%{;ʮɴUl+ic  ;RLݞpjG{ ^鼹T`M@ZX.*'Wiės BA)h9ɩ4yօvγ)>k{iBtz)'8tuJE6z \uҾ44̋TˡM`8#LfԼZ<.y:?z/f'ͅUl 1ZkĜEG|nەMso9P`DkO鴭ͼnVևr\'4iDW}i$^lmJHr1U:v Is`ӺxYq̒{8%SyQNohTO1o(3WLNhOo_O×oO/a,8XH'Aĥ, *#1Oxf\aL`I۰yy->lÖ{] KhMdAIQj"6[r(1)LX?Pg:;y[.}v-_Z(/]t_ZcWKO.;S6KMԮⵉʕ*}SV!5d/Cls`Rbr-h7kL7.SbS.ݫG}f%Ksͣu`k*,=rlKgBG#},M'4ەXER9_'^<|؁-y_b,&J:x>:4GK@f̫H6`#aZ8y/o6M_>>kepԺ֊wano혧َQXHE;7g#?ȍfTgd*wr"̸58[XnGŜ=yr zm= =-c>QD20CF8È2=w`C)n &!Nˈ)ӨSQ`B`*:o;cƌ%{hFs+%c3r6"uN50$5I.XDw7 qfAH$пbXXH3rnPΛtֶ(4w'wMz";Nzr r6=ZW_ ;'gGqmk[T nJN&hS!(RYh "PޕƑ$Г lxM&YDQUQYq|Eb )iXYxzFC͹2H@j# "L%K[UqR=c7qJ7_M3 wN|E{!5Ӟi gfx~ t||zy|4;0jbgdpM G+yUJuc4pI>Eh+0dQ"M8{`Ur;@_SJHgM=6)xnڱ6^o#]GZ8f֪Yr7QjK \dUyLq, Tc2q}ILGA2dA$N"\$s4XXI"suY%IK ?'#vӏm=G |tf I`M%Y-Q!Ъ]J$_\愂G"n#VcA`^g9%d xp8TTgv$4MK_4_⃌LXөV@Ԉzr-wԶU2 8'9~q~Pa7nw鱵6q2 _Q̫7(Th2fՏ_/v(s!!ǃB,!1O892ǭiVq+Y92)FlR+g,&@!(I `U*H*i2eC`J~(4}*sz8IĐ6pUvc:>:7o~w{_ƕ>fyF!syGn|^mwo<yWzեMiiu0x9*egncnlFwׄ. 4[H6l[uk6^^]7Qb7a[MDw{Yuȥ$LKETLUR>!eOA|d@V"@Q9d8YEmZQFwNAA(};ІRG` 1bMY~8`K`H.e$`. WݘUYoŴطG7n)+XkC^4Fb n()֥ĉ-z4@hkތz<:SQEp4[KU)C&JTєLeu \N)Cm)lVoJT[PF0ENS:r&kd+GƦe,eD~{7<ݷ!K=h.ɠq:pOӁNR3"|DNd슆yH4C9f4=f-{X ( $uhlTs2*0|3o?yoM.?N#R~Ͻt/S]wEF6QەtLdc:) Ҥ%+1{ R>cz|ӵG;*6W}5o{2fO_s޶Eˏdr.׵y)!yOO Ͻ|n_O>^)n׏a~tvK::,=%KO;d_+V'vDZ{BZa' NKXع  *Jl-?%qI.V"b,x '%GoH\m2WT5UiUCQ<蒲'WX[KdT j`5ԛ!8?=942&.W |={u/Q ?4^ALL9+eP| ,'Ru4jBv)Fp.#t h,wT3mYhg]dِ1ՂNUly.M39Q)JJWog03 (J6kVŎP\l kS}qEyjd!9f[2HG8a| P%<7 gˏ#fM𴛨p~>F/#^X q`@pt: &roUbWwʑ?/_ őtI;S1Oɧk?Ӥu[iW&sd^Y}˄K?y̼% \K%@*ɹ,tymŢ _ V{^k ys߰vš/M7yyy^dъH>sez2[8ߟ &ϓ)'Դ$;&NX9lnUr&K&s08V*еXA昢Ⱥh%L:2!g p<:ڻe/q6X4YW >[7N%WAf荱uGK6;$L2:^1OzFOrzqr'{MNs .z)Ś$ɏz` e!@X# kaYd٭ddCM*ɑ5 br͍kW!_RT d Y< LWTۨ UZ0 zCmpEvk+ d!]) gӯК'A}:z/f}M_սyp4G)UUlc }N[(zyʿ@L7{PO9C(Ue-DZGƉ!Yց5@DHMǍ u-i_ܬGޯ=\ixr14dmo:ER5)enrWsw ˞.jJv8"EDtFTHdRխoN[^gH_쏯~gu<vuh-ǽKxr_J{]㌑ق'jkj 6W4l\Ғh*&C$/\\vLm,&Mc )F-\da8e.ӉyCѐ+^cb:H-t#q/cT[D-muGƞPwjM(!՞ 7^R똝5, } Ct P8>OCLjHEC )1 reܶr~X/gGz 80d]QyTX U;6G@1ؚGէ 1|ϟZ1CAza6# 1ilc2}rp `b⁉M+`>)^ٍn[\1<7#cx6FFsciYl# I_eˆ%R3$K*~;w!2ܒz#=L4͂k(g2_ 4H# }'oNcj!q1i-ΏIiaʼn#-ζ^+랁jbU7⮚]^]Aweb|A Еux)I DJcޢX1۫7|;I~zz6M2B-}7'he['(nTpg{tw0_[4Q8@!;54xL <a ȰQ:U6rѦfǚ\s`Fz. Fmg =Wa=ş{N~f>=!F0䚂BpZW6bE۔m@>ڒ4YU >)Fb˅Wm425L*7/y>7 _bo߃oǥΖ hLĈE\+9 /tp|0, l>$52ſU02v]R7͂&:rv6os2}359| 9NcRh#(P" +qny >fKe;TY먋(&Z 8|Dp1Lj$(Pװ7Fn hg}r<g~+_Sj1}ʳWx uxCF?RsŔӁ@MGh:ai*O_D~9p?ϷEV|Qsм1c,2+ae%Z:QI~s޳R>}mcin.*O`aN?.YAiaDBʠw$V%hԖ ^&6ŔH2/.C 1@4Fz)di<7$Є`UQ{FR4F8-lFMMqџ%Zp[T]Mb{7Lng1K;3$(< Iy&C \JSe$΁VK34hOggZiKC" ي08O>Q: 6fG3suܔ]ԣ|0 7*] u}Gۖ_rĪN\f37N~LzT.]Yl'Q=>r%BCɼ_${20O &xD-~|x+.&Qy3U_enXI%!cY. |S|4ݭ jɻ8:CLARJ FW|{9cH㋜OWs-Bu;orgM͊-̖o43N`,#eFRUs*nd.C~]25uY)9 (GOw@Ays7hӹDc=* Ov=w E vޱ٨6OKغK)h$!v` ZޫS;:6~»1's61eEԇj@bNQA&@I7^ahv5.' yㄌ `K'a\dAЬx!8q$fkL6.L\<$FV$o=7h5eDkLPN`$W(8jiPAȬaF@/s NgU1W?8ro܈{Gwt1L|He^5 ^V[Rށ}7n~׆)gx1\rN2q3\9\D :&I`5%׉x* '!WQؓ' TۋM3OTˏNH RFۤX2 3BRQJeD^$TM]18q{cAxgULZ҄(D~tcs[ M&•|5)@U |Ye[^..xwVtoz%y-7 _Yuy LAdU0 e s2ВD8yxT&+v_uYܜ%ۄLPyʙL ȋT8єjP2Kjw$ {dݨujfSZpЖk3K>w##~i;|opJYEv\+~{\]nF=7D?)Q[=G?~Z!a֦1V;ƜO3q+|{ sJeu'c*0בA`fe#"04Rfda"W0agNak{Ϣ]+ߺf"e͕+B[~1\4޳o ]ݶ\̮~8Wnz+5O!]Еo~y5taryワZC8"W>'~!\4qJVWm|!jqz۟7岷77 Ğj*?$uj7JIZEͅt$El]C]Yg$~׳>CmaS)3a $iK8cbڈw+R̩)#9%nZ9 L^r|9#&yAUdI+;?}quyqMJ ZcsCB*"$1!Z'kϼTwa1GiEMPOh钥cT4"Xdtڗ.2C@A@'{g:BF2sRNB)7H4&K̓!atv.[2rvՄ'?9-AV"%f"RIT9됍u<(5S>K4=4,i9Жˁq5v,%cǑ"Ky\oa6IgXz mWk~|7>.||W7ߛL^{rn-;.ɰ\rLNp%|?r_\v5EL1 d[F>"9{[El}7_B^(_qO?z9 _XvvL!?齇L40gv~~c{m4c7/XtT}f2`O?fn7WZ4*}RJ~Dqxy2~^Z[ 5'6J/6y)M KV756s\bB_gЗ!e 5;gIsT9 @TΩjYDƠ+^eF)DB2 YY!He=:# 3TF^.u*{4u]ڧaC(hoT&ō>sci\bH@e2T֞Y Gni:` J-(y"%˭f I[,xlDQnydSRD Ik cj/rِ\*Ga*gkv,;>%niԌfaIZ[*T- 2MJv9 53k姗*ݕ(pYHHJEqZ8de:se2ϗDI&Pn<}OɁOBP2Rk$[ 8MMS3& vZK ~q[k ynK>\63b3,)[j`{2FeVK^u_u֡OqfZ&.VQKJF+$J" a;Rl\zɢS'><}c(šuɄilh6p F&b1x&fЌ%wdmhx[Vϥ0$PBmm6>1R. rl$Y2:ikCe61NB|(୍1`FE\$7vZp^vO Kdcx!"݁ bJ]KW!7z]sI/sItj1Рc΋eU`~6 h[I%%>$h,g=LKȹ8*d1kl"!#%{BfA '2|Bo &fJ#j/h)=XTPWGRAO91S]H|z=s#T׿W;9E%KMR>^&"GIH~;Cr!9GWs8 Ar.Kc&I g0FNxB'PZ<'5(#0:e`Ckנgձ/׿_-R{-6`,%]&bji,!uc={ cc!bz$W/oKEb>s6Wn6jExh^( !lhj+(C+G5eI붐kBv''!B́im 5A. R8+y~ři˄S)2)PxP` 2cPr-'!j (H@ + $ۥ!le![ Κ'ޅjr Zé߫/h/]ZǓճ_Z:wG}e-dHңJ@Ayy_hRd #3#YQ[XNCdH tls쐠ޘu#W]<]+1XҔӇvt+vEmSOȻtOEl+\>K뢮 |WÙl7^ q[S՜qLCL ɹ8\ 1qy4S=ڦv<3v0Wu%A?͸Uss84zd`D[^>\N{U܇gSs4X\4O׳`ŋ5RbkhqK=Uђ 1[ <<0`!SAEo38~Iz>;>Lu$hH. 1EsiLޟ!4CGd\]8NH&$%6dj./ec;Ki.ωηQq%^!}{;JJ({\{ios=?"Ө/}N5]W; Gǟ}b媮//˻\ xm& wuÚZZau+K#%T:HH*ɄC}}yr=;{UKi9ͧE}dNhM 6^zKK PxoDSXUWUB-IdNj S S ѠetE(9Е]{&t`ylu\KW5x ݢ+]فzL>*pMod杧k] ]`yo=Fh5UA]:F{DW0U+Y_誠UWW] ]!2ݧ+q Z]S+pSU7tUZ"|a-UA3xtPqR{lMo誜:q ZNWI.=iHC|Hqv&ՙ49ר +mn%G?ķ*vYY2Zwqe$F(u4]6kX6Cٵ&80+{DWp\"yo D% +8HWV!]`}9X \0}Vv~Tz3t%z]/{&0G 0#mwyC8y 6CK t%t5vmK˔]`~ ZtU *hu J] ] %:FeCR \%BW6]聮NȖJWxg/tU:]J:EBexUܣ++ ]]@W'HWRKL+Uq ZNW5] ])n6J%Z۽Iȥ1]iY "~kl0$:X;mcb\e0K {v)г)вܢ},(wV9Qfh-| TA C )zg֧:]sUzIhbRA]"] GtU*p-rnR [uHWhq ,(pMo-·(@Wo]ߓ'S$lXTfpOlvGVf(ttڶuZ{CW. ZiNW] ] Ul7tEp?tU*u*(?]]oIv+v}[ ؇E$;/\ #Dy )[ޱzu-橪[[~t56|;<pX] {Ah/J!]2vQ-pƧNW%{gHWN& `ch1t5@O(ztb$JZzt5lZR > 'Co] GEvD8f?ӳb0|^my{8py㯓BCP%PgwԈߞ>xחut @^+mJڦ|nQ.'hz7.ޢql}\"Xzgs9n״!֐"vOv.R9~xw,*_~~_Ϧ_Ƿ|zy w<)WG^[JGjw~6O?[:Qů 'Gy"Aywm~DLç8 E N|D|scf{#G z#ϡ{>q^Aj޿ޔ{zր=׫ɰ;ɱ%C5z.ؚXWˑUl. I%wu>l ^#C} eovuF30Pym|KətQ*jzM^ؚu]f["]9vp=C>2Ь1{.ԩXRUmf')՚nq`Խ.Ǿ Nl,Z}SJuU*bMԱz`vL*k5SmN1h)L 7'ZZj j 9Z(\ #j QS)9;D 6FK yj*zg.5|S5 kJMSjJ1'ZjXBuɌa8h\\D\L5CёBE{))wopODk3E27:d]=lҹj;wo+d,TD2; 0@.\mK>wh0FGxt! A 4QEz$iK?Y*ohSآ8A1#3X2.Lf>O͹9 AYUW-wrZª!+F[u΁z9K u}6:9DkQkNҩYJE.)$qu$~`ۄE)%yNVCJ 6ZЗ2l-7WqzUIOi;k^,HU.C2+הOtOV)Y0f;DrNU.Y{V+ RAvT4JkCvYw6*yˌWh|v lVS |FnQx5Wn< u<աmӊ#5 (%ʭ2Tb]vut%,d.VN5!1ѕ<q'YgeеИa'+3\5˶ E@P3 VGR7fa=g5+sBE(#{*%*Ze@Pɔ2)O-` \:WlCD&0VjIQ221LhHpu 8t, |FEtf(MJt*T[!z@g3 d2ymB AvEsze] R uWh%W&q2FN[@zNB€("29{i]UtFtQl<t,,x:M;*q b *LF8*)S|B2. `Ʈ pqXi Ttgf(@Hq`Q <ڳ<;"JPAw/uzU qJ7wՕHޫ0I]l-UuA"F_R48,f^n!LƛF_ʔAIP 6A Z"b C$uWH&мGwU+czh2&M ԙyA;t^̈KUU+f"cEͫ2( /rC|cgu9ڶX\=x TɌ4QG[S16$I#e56<NGзUfĞL:5n(!/[trM1WdsC<܈QD{;D+1-TP=`C(uFAJ@2%fdiPweT}fQ `cWP!>+IiMkׄ:H& 9?"oQ^1 bp'ƌƈi)7bQw<A lU?3i^'24T VqFb]Ρ~]ۜYg5kPnB4c5޸2 ֫lѫ8fTڠ@4XAVL@9 Z&[QZpi䋞P+ čtzGzӠlEi8 8؆kCW$ЭwE< \T*NYc6ՔWNDL0r( ;fAjr% T$b4uIJs't\- 56u4>]wT`]pJ;&^c#f/nl} zum-k*>o}K-wF7[>8}㏷ϏAlk uݚ+ْ~{ɯ/z3?%~jS%4In6?Mzë`@o6z|6W[\v{qu6>ͯ5el+\oj\i{vurip6X'_XœOnnqжý>>}:;Ewm=߮v{r:zh+ֆ'\,9, .. (>g"" $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@zN ([oaal8+u=1,'jﮭ-o'ʮY?r .v8hb7⬿xeA ܴ <VWyc^~:AÈv4q%p/bs^w AuLű9[-,bll@bc{66f&66Mllbc&66Mllbc&66Mllbc&66Mllbc&66Mllbc&66Mllbc{66「%9XV\p/ Jwedq=G'PO@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $Ng|k.0SW/~<[MGw v~lϷ7x2 2. v9ƥh)ƥ=[ ^na,K ؓyAtDRj@3+sȁS=BNW+޼RQߝz\/]= mNPFu(zte5$`|\t5h9[K&-QjlaDOU{ZUFَ]L+0TV|X,f|YKM\8f{*E`!'xWh6O68gDAu=]%Е/dcYPޙapƣdC dņ쌆蠘o֋p DBGYOIfpx ;O7 9_+ND!cÔ:`q3`Wi<ƿ+(X#g2l?7y2kdalHE]ئAfJ)t4ba,/)pɃ0q8)Ҏy䂝Y5YMY`Q2 ?!Ntwi)J $Yˀ-zHs*%ZCA$V)B'R gjY`NE}Q\ӤLqI@5&:pEcmfJztQv%FgI$ʀysƾ3\PMVCt$teݤU13<卙Y>6>2JZz6tK@Nٷ .Snh%쇮vCL]芷tuߦU,1t*h =tB^ ]1AU,ic*U)tjstQ +.M "1tUM(y D!52\u ytQVvkЕCƮ2`CW1M+D CKW/$h]*TtM>q뢘-QM{9, h5l7"Y^b|,D>n<xv0韾(pQmkJ޹" $F (̊hXqR6 kU^*ܜc.4f'?ޜQ6"z}J By,|Ѫ+"(* o]!`%ec*5ڍh5;Rt| % ܿ NpY]톖iKQ,]AKWmz*FUUvCUF)HKW/2&@c bBW-ӇNW%Ж^ ]qD 0QUFءUF`%hR4aQWWv_>RȖ^ ] -uƮ1tƨV|0("ĴtJPyp+Q_{@vHa&lR-ItRF찙w^vzcdw>N*iZlЁ;2Wm>gQWRVVkqȂfpIG[`U&G̺Av(N  *e;3^y?򏠄\̴:1'=q`rϯyOxïln'LꅯG7}Xt/yA%)3vs&C\o {ޗ%X/ÈU6_T;_u* nVI-%TڜySJ]UN~/F#Z)Ն;!j*xJ1b5 h5D-b Lx?C\ecM0Xxq4gzb8scc I2& OZĜSSRnvprvϧmnɲ{34ťܬ'*efRv~kp]f9,+jbgk;;YsH_-x1E*!Ǜw)ɵu8dڿUo?Q: ;9Z%DaS(&A}~JJe e'Yw~C;#I ΢kv0R2_/._ǝ (f+XE9usYfFČNf) eA?3 YgR2jnqO#$Ε W9AyT~CXg eIN)6 egl]0xKU?@h)Eҧ$E$t6xTş gg5"Q*@1LpZA>\\ `I]eR=mn}&M/,KAi D9M4`e>DSQH$F`9ZI|ZXh.P፧:c\k\ "8O0+`y`mpn ^>$(ggZaޮ(G/`>u۟Ԧ^-UjXJݥLv)]jWY=FSa9/.KlL)\VKH&Rd4f"12Z >"dH'R%=] JqR)Ꟍx`P`nN!8HNj-FHݏwsrq73w߫)n@%O΂|"d`Vq`X")QTe2zl"~aP\{` D8.DxPx,h5T"1wd뉳 /9S~(6ǵYm|)SOU]a[XSi`HO'*eRt 4q.9cՉĥTL$АgfP t$Τ }Q娀`\Ϧ?=Yⱛ$V֠b?%2Ivf.H!O \+9|`嶦rXmhwyNag 4ڔAm=2nYTF)vړ'$}?XSeVS=Mհj-Rt0:z.|BIDM` ܙ!OD$4BhShuuV{k}ZʃC|g7O3/?ʻCJ=I=ghbןy. Z^8l!)a FV=Dቧ G<96#TAde.m@|" GTyؐډLDN ,ZsZv.V伞}f:SϞӿfQ1 (MF}WGO-THOh]໨)-sIn&wb joзm|FmOpGwq}1?N.5D8R)3 HA& ("*II%KK[bXN+ѡRb`i@`]8g4f'/]dBoS}(dN }9>y1RByb RpkeNyՃ"9P*F$V[\Drў1U /EJؚ=68{F x2Җ{e昢,W"KF@jۧ䴆8Rq$x]@uB#mFOEp #)GC#w7ucA4rl3cd-XdF01ü%!rRˤ}mVҙ6n~Its+ 6g,}=m{oX, ,d'R>!F*gI. Q LFM)9hQ!-1tR1`Hy@*Cþ?~߽]&yƃ@M }ry}\se.ihr|tޛyt7J;\uG/D?wr߶]-KYX9ΉO)G2wmI ;RWɮMrA B?-)R+RW=CR/ISMkHLUDV;ӫ3$"$\jɾΧa:#/p4q;wm+.&٤:}Zb# ׄ /p'U:ƣմZGzD%h$|4>jcHӋ-B]3# Jt3r'yyӔ*Ffb a]"ZKD7ɳ.킇c'dοN/OJ= (]޼~pf7c.wGM}ġ/wL d"~7PlX;6̶ytO)e L*ڏ&u]mE*83lwE?k̚.>v7ZF6]|2܉d"JJ@о/ yhK'a($4+4^N'mI$&Z:G9Sn˗@b@* WI(=h)x#ה bD&ppJF#ѽB)'`gZU5l(W\Fe ]=yzt?9r^7EΑ'wl-Je^9 څUƓtp^JD&?g8xMiB_К^&Z\1v9ayw,YHc{wc'^t/~4ѩpS] ܼ˿5Zx~qy|R̈́篙9Qةy|Ѩ8R8/ڬ y9.puvkFgcB=Ӣu6ϡ5 Qyxu=?_\8k!66s>&iaOȞN@Fxb/ʎ, 2ZWu}wvzYm(wv p8]w(7L*)m$"q\ 0 #xzv PV.m$ĽZ'M<5Iyk[.7-7\io؟%@%@RbdH2tuc-p> 2FJ"14X!HPG6᠄(M(ҙM4&ȠOFb| FM0<LssՂ"}D(N-1Ԧ$JF-a9\A 8dAhz5!жKƀBD&a"c.$d21XKy`cJGFM.郲IϹOQAO;=C3>#\ԙ)[8r&8ţ0[Z_M~cDl{%MZ8jbsvb`B5sA2K\w-#O5ޭQ}fu*%*9 Dh,@C/z^)xk"&*Zi>vէJuV*tYT86:C: aRͱ]^*?1q-}wQ;:&#Wu=U|2џ=޽h(0і$A I2HA G+4aZ(KQ KL V'ʅ-Q@ 1P ༷T,S.ygCOiJc|SU=1pټзKV\:nc`tP׉Y` qEBL$7EіCT$@&4fX4XG!ьFPeZ j1r *4E'.cQuyӗx1>7.t9Pױ6@7E]ndԤ*-Y#)sKVeK4b))xt4&M4C>PB$ġֈfTFHS F꘬gA4)R)\B%UZ3#gf,gӅ8c_]h B;ƒ,Z_g˸#t{)۱1x=mڍl;,Fv}8r6Y1E#}5"-15Z(#t[tάC9ZKU!8iA IP ˞B{м;(MyԂg|[@A@O ͝F{P[4gB}u}"+٠1g4$,ZSFk!ow5.g.!Gʀ6A/B/E}!/~*lK鶢vFTr-=Џτ~ّ?0LJ`FϜ5jḯi#ęe0GڣT6TQ!l:GF pI35JOh|99ѧ;˟tm̶bb޽r6-IM{hԗ.Bν7#$)xkoc[7ϟ[mtNf[:]Rp7f!2{QϷwVCwZn懫tC˻Sk>.<0x.:qtpLw=x YDV|e]tWms&ٯteo<7&\a&uw\/U[)f0?TXE-PւxV{IdI T\K{QNC`>(:M>yELptI\[ĥsǑ"IRzЊK$BFcBbeV GEݲD ]}<~׆Y#~f}9p>6mVOd6%cc01DʗEaYwqF+>pn-Y,ECmy`قeKыw7AUFc[gjR,b3=Ev<гR+_ sLمjUl /ϒ,7G]ӺhvӅ:(Xlkũژ}ec%jȮ:Aoij&C$?<wrPM&J6 ]jb|-'_ jGßϏ%r?ī|)Nޞ_p0_N(rl.(ۏץz؏]Twb s0|ˢΆ;826N,GV˝_(`=~:Ä]3?./<;Z]I~o,.!nXN?,_>|/_~za/-w%`37ﮖ=/X|J7Qh! >[ݺM κF<^\<>Eo<7r;s훏_,HjqSY?@v}"Qlkcl*nI1zS._Wgvq(N[.D =՝8e}H ]:"vi([ *tJGݛSS:Eilzr-wo;5ՇkP8_~,fd5Ϭw#o1L\}-9`e*Q1gH\6C~'n+cMә +c3 ϧ3'MgiJ3'tNgnYX\Wu5Wiu5'8Jkҕ+[tkу6@Fb`6] F{JhmJEW3ԕQV9XFWTbZH]WBjBIٌt%ѕzEWBlbʠjZu銁vJp +u2被i2{Ϧ3U&y]1fjF€iƺҬr9Y|1$m,jBkNY->[\_/mjq>c+ͅ N>pt<;\ '꾸uN_SKSysYf `~YLvmOwWק:ߵy5o+є_|r^l~[5gm|1$ F(b>km{[V)7D[M~߾QhM]W=tPQmŪ!ha}5s1p0 8 .f3#6c/1c e\N ѕlFp|bJuؐ;'\krѕк+]=]-t5 !]Z҇t5RCb]]^!'] sJpEWBט_WLI]PW`U`TrjL>bJ9m2ҕO(\t.ZЩJ(.cJ RJh!ΠP]QW 8@6b\0] -P$9]1n|Ʈ׸\t%6J(]RWm1qNEqNGe~u~gV]T-JEFzվqhNh򉽽96qPp^h_mfɼ8-S25hfdF]u ֡B ԾJ6YR*>h5CzUc cpUm_ !m~5d+mwҮF0oUT#nEc S"S>yT6356ޙPZ_zg3응J2kg0Ip)MJ(Mu3:'] '@pEWBK2GuOGW~ˢ&]~̂qâQNO9rim#t勮v-zH# JW;~RhJ]We(ddǓw<L.OO]WB} u+c2ҕ;gZJ^WB+++Vʪlt%rѕu%DEW3ԕHW le+E%+$St5G]a: Z^h:Aub1Ǡd1fŌ8oI|ep^Z٠e5gc00|1ָc 23t𘓮FWٌO>H(SbQtt%.#zqS-!>(UFphˢc;K`q~(pi(hqq:dI+*ڵ`uFɗkTEWB !u] ƾ EWm2ҕ|t%sѕВI]WBj2|79uѐL(گ]bJa]PWaJm>cWOt%RוPRue 2ҕ^`-u% QW8]1.l:Bk ej #(*TcdM0`2eʆMzq'>MYƝlˡ- 'BUJ֩uG[w*NKqfƄzg LOYyD="( ;ay9e0!]76ҩ5G])\)]1.L8ZH~VPbUزñp zb]ĝ38382̂0BWjע֚H}6\\tŴLJmf+MNѕO8\6uŔV颫HWLDW|.bZJ(f+t.']1Uw4qS=GEWsԕuܚ0#]10@>ѕ"+ڋ+2>G]9ҝdptlX Y1#koL  7hSy=(`粉1\b DBJ1û 78+5*cB 6u] EW3yJd`FFWkT.Zu% 8QWAQFb`2Jp+w 8?]щRwsa5y{4Yr42h4$uLIWBn-ڥ+<%rѕZL]WBL u֛tNC6\\t%RוPnG( a:Yr4}GR6ѕ!TѕPj[t5C]!:j4p'b\P:] J(]QW6(mFi벉S%(z,]$5bB'A6Bʙ;o Z>2/>9 eIVm잖5#OKe?7lW!y{MVذ-8mM㢦ђ㢱֮QN7ѹߧn:1x׺G.׻^ȗh}ksm;xw߬qడi5UlXՑIh}yI|WUmb6ijh62̋"j\(vJ($<]-^+;~]S'åZFђhB8J =BWjע;銁G"+ut u*銁iq.] -u%]GWFB|6r꯽J(P .#]17:] .\t%6ΠPzUt5C]YIWe3.>y] EWѕb`DWkѕbcWBimueƔ2 O?-9`s{씷2y伈ʤ2Uy=Q!.c1ChvĹ31؆lt%^+䟏 eEW3U`97Iܩ7G;TqEWs9EW Ot%MgPh]AEWG jɮ)o.8$,<[rzv&g&xxݼ^xyϗ/_*t2߸ν8r%7 xV]]_reY ΡGNm}Thoxv@0D19k> &5}ڼܪ^kn)Wkq/g*ޖek^|%{OlWcPeB׈Ur`Q~|x<Ӭ8,j pks&4ON¼ݡ?Ct;V?~wϹ. ^uU۽漹㱪[^;C&N;2mZ:3tmnE [D^ywu]CîjGݫ+_vx-˟VU|\P黶>F{쨯=sWu@8]muT֠6k臨XΨ6^%Ա1FV ,X뺅5>{պ'\vڂᓴZwco\Ӷ^\!T2C=W|{Uk_k]9G b w]:@jҡZ.%S FЈ>6^ձc01܃|iɄ`plEn߾y!Cumu$׿2Kv;TT0?d$<)38Zr IEW\OGUNspJk: kg jk]reҧ 4Xa׈fhO[(:TӨ%gC ?{ODF01v Yti:`Fo+.P&]Ra2!C>;ZjB#mFM#d |WU׾>8`xÀN&>Xֳuؼcߞ{ ުa}zd~pvM&!jhX DŽ 1=*ҧ-ƻ9D%@ۈقj2AVa1 -O(v5!/,Z肸V24ET4iPyME;(C#`ҥ`)`h^b u6B[!q*p ]]TBNu~T}:MOUSL l;V+/$ a)t{R꣼B,UǑ+'X1[е2"1wP:1h /&s7DR& +28Œbk ) x0 M=`I+5v5+ڱ!,2/C&D0ZA jhjwX~c|M]FpX2zv huz|}{h&ޔl#۰,ALJ3-VUFi5eHI)A5qDƣ#"LCכEa2,Ƣ cA8OC*B A3i'jjѪ/hV3i  j@gV3 o=ZP)FC;T."ر,e"ԿIIؔM@[t_h5W jƐk9.m/;Ϸ 鲿 6r$;~d`0u0 gllfѳq5eoPpjq(XuuLk昴59%knĘv@99˘p 'id I'? 9k9j~Bg ڛ/`'MJTCۊJv0X4$SSAv =>]ߴUal+4OXQW Ժ) }r7plJ1TDyò*a‘r;p[Tc$T:9Q܌J#x$]AN lU?<66Tǀuml"sJ@:XV=Եj@רM{&=Lb2Վ*WZ@׶wϬu~=801>@|5@RMwUdi1ûC4ZSDàe ̀z12=pe6|_HmhJ7a#=>dNғfZOQ!,JiSܰd@b~;pilp1:?iK2h]CD!c-ձPۊBGR!6.Xvꢀk5#[ U\*y׹]O|w"BQ>J+eI0H5zAYܲZGh;*( [TCZDza{5?r37/_~BѰK5-)Z޷eニnkI%p7ggq[I=`Xzu rr|v'ˏvH7Yؼm5\d5z m±¿$A4k3+\]꒯>{3PNJ'w?<_خs_/~gf ?:3}ʉKCWZh';ϣn4qm'% tN VQ@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $NuKzfl@ih I^BN" $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@:\'P6@\fZw:D'PJDI@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $NߘŪoG9zqqt |Ͻ_!;H9^&p%n>Dmh- /z6?ϫ͞iǵś5=Sٝゲ?(xooVؒ! [5Kl-wɴoo6ۯd*׮SLI.5#3k 8g("U.6(k9ѕcO ]1\nQ tut}8Na6t\FFyѕ#9*õmS,~M_ͧٓŷwNmNmGxzyG`ʝ8{rwgz yŹ[j,^(9['>?޸;j}zڄ2z=+~MT:=rE}:ml|uFOVxI|גGOLZA_@7-J.-_ob`w?󚜴U9×?WNO߾]IӞi'!戽T! ֖%)Ѫ}RCn]cE)b386NOZvMcIUeq%-a?z|*'oeIϦ@19m&10 7:M~߻qͤC"ہ]10S6fRd#:DJ.gDWLf>ͭ:F(}v>p0?7d(V硫ǡ{ ++tKψ ]1\Ms+FkBWCW&X0bBW֫} $tutegDW쾂ns+F\w+/tutOv3+森nb{b\T6M+fh]1}3(tUʳx]1a'[ua+&ٳ{.W%ֵ7+ix{6}4tq_rHv-Ԃӵ0,ɪLѲhPv>4{g1MKhIk Fie5FԜDťBW}+Fv!Ul83\?b_nQ&-tutB]1fCWi.tIコb ]ftYqs; O=s}ҩ=#芄>wurIj6tp ]1eJBWHW&E%=b;]1JHŒ w%m,!~N1 dm}Z|HCJV[=K2)QHAstWwUuuяV<uWWJZu+9)v ٨L7 f*[gTW -3RW`ßBr y6L-=ytdUW_lĮNn) e/>i|J{4[2ir NN=WU-\G߮ zFxkƾ>(9BYZ=YaP4t`Pc> . g- BQc՗aB{Uhvfd4vT!&%KZaڿI<Atu}4Y޺W܀ڲ/zǛW#,4EO:Z(Ta<Bľ "U1~:(^Oz1zL)EAJ'Gg1h,%!qxIl %JVDQmnVStl/zc?).K |eBH9V F+:rNS,T9T:1=29)Z{odz|d `q5YrM6=ן-5UG#4 ^nLSq524] ]m;1K:_–-2lsalaѬS%l؄S']0xCU>_lbd xf0V->"g3R=)*#֬=;؞swjƆ;rP!=(-Q4TN4z+lJ6 &i "ZFy.t򂠌Z4g@lm-ךROm".*`wItOZTȬӮah9wC d"+f| ڎ&r\-["?x ?k6rpټ? )Ql~gUjN2죪Gr>*QɠbYpyg;glAНƱEGAGqtFe 0 /HV) 'itr~D #ն:cO\ڱ:cPs|#(3X,r\'Id76N$cHhr!(\Xϱ@.[%Iψ$s 3%5#ntO a òkzIyyzI ml]G 11)g41Z@p-|NRmbqty G›-ت':Yg>,BoJƋLǟ[\{=}ro"3`)kD8.DxP2#֚ͩGhk1;!=ꍮU׳K`uP'7l ^|} 1>F;AZR0IC21))&ep[䙳sK{- mP6@ۺ0Grar8lq=Nncy/5/QA!x yrvA]2')Wr'W|s7VB? BwYuf\C*+= Gzl98_vif؂nZY*j#\fՔH3CfӐ] ްT_X]5ڤ瀳lsJJM$b*kh1' $* *7F+BvrvIyxo>Og7=|vTޢ6QzGz@Q>82tx.gqdjA_Y @k Y6ea㴀H5%,x# 5/W8XYrIO\ƪuCuCSTOJP=*AZ)rx^Ḇ)+$F&jS/8gOij65#ЏHsZ OTˤ SQqi\Ÿ14*TA҇*&h*f"P $Τ eQ娀-)rt,W_rD;GsscAυJ8A +80ʆh RtYtV& in 1HɈVx2)1-AӃƫVJzo6LK!̘8e#wպZ*2RFreS'θBP<ϺQw?~?_xWo~|)_|H0'ᗇ@gUO{W mjoY5װOբV/:|zUMQ-rf~er+]ݠyvjy$m5'ZV[y mr',_AQeETTp*CĿʟJh.O@+=T/ Axߵy^MV]vHQg9}MK@1ãQ p;D(.qT`[驃 KWT) aCW8.N%gD&Qs3L1c=;T9EtFt(D;ȜNgɩyOzhz}γ?:Fs*Xզ'mGsZͧ픋io׍x)$ J/ Q]҉0.M> <$jx޵q$ۿBSr `,`;PH.Iq [=RԃIMid [LtUO 'QLpfHN8t4ڣ ch#* VyBegE޾:j@iMi0~P:o&1yQynS'8G4q^tPS m2US\őqisyez٨k\b˾7 D]֏AZXdVJ01˼[D FϕQI.n>D\OJmro,mul=Y݉zz&Xw "WZ#!\ĈQeUc10bɨM1Y54F:)@2;n(I(mGz˸"*K(1zlyga?ُ;nL ٪cs15J#!)g|f=\JqQe$PI 8t~:=|mCd!{1LR|13ÓVXP]yCfҩ#ǣ&Ϩ8d (AV̉LoY} wkcZl}Cvv u#td7pĆyo d!gn]OT_9{LHV2 PPQx)qE$HPԃAm= TtJ&m~qWe1磍N?ټ!^ `'8Η_C|ofvyo?d}84\|O5gsz&~et2i>˯e_8 ?׫l4v6kC/&L9:w痫.)9bT#oȚ\Ϫ ߣ,ae]28 #o@rug[CIjL=-|:9kAxvh)ty`J_\) ؑ02N&~8hj"gkzM\ V|uo^g /a ?ɛa1s}a8￙ ,HiOAWZM/og5QqQf"E3BGE\~;1 =}6í jY]⾩ 2HJq(`t6+8>tξ@GP7U~(tMoLx/9ܝ.`.,P"!Vn0ZmnƆyŽΰMxsl&fVw#!|r@sNUu,Dn! AM%OO<$g@Ȩg1X$ D ZC@ hf3ݺ|uY +I Oz>:#<הI/ F Ԓ\$<4 edְЁ^ӻUչƭHz/h;I謹[q4 [Ux+@&/m~\RcSK|j:|xbڊP1R%UfQR#wɪXTݨ,ZDx%b8\mUђcT.6%0b$p)YΙ]0BL@ ̔9I<9-rmJPh p:b \ɉf)%nWuYy2|[|Ť`OT@0S>QoA=:׻m#M~l~?6ddٯf]naͦ{O,[8.;N ϕ~E綣D{ɶ (0Ԝ[O?] ,fy[-(꬞? xY|LVJqJrވ b2MFx:/۬x`X`hE7mI": Oc%0=~ViMV'ʅ%Eߠk)!jc NP,"W'!WOBr`9cbb+diwK\/lm`d!9 E:E+-Иf8/)jQmQ2%͟B橏¡hFPdpZy7w6b6XT OÐ~E8Ԗ 䶻t 89O?MƃMPl^|wNef]܈]Tﴘ-U{þ ~QTއhRn%f\J*3k õ%-D@Q#ТȨN!$r.N&$NBOC̺R@B*#aR imBaWV\o8YN}hPp88_9b k(8 B&"wRh ǸJ&1"8)Vـ-Xk_6mC 0"02lrpJl c;blbFaU1ˆ]L݈ƣbi-Qסvn*uGI H nTE]&gQ ီCW&Jmpe T3 3h2"ք#C_+td ꨡtb:_qd`l "ӏ""-Crrp2B{EǐW &6:I,:! 6)MyЂ3.WQEw1fBs0Ko+&nDe!.ΖL;nf.糟 UE ,]5fϕP S.YVȭEP+ѱq4T9n+=7.QhZ0<~J{^!̳|GxVEz J5l/Of} ”M' ȭ&6B (kP`)裡*X(k5YJk:ʚFFpu# {hn&`W7Rv\n%5kp=koƲȗҾ!qҴ@z[. @اcYRHwdmYm%e[I.=3鵓"QYn?KA @ u*u广MТ>r]E#+ZDlo w{yo2\Ds~zbx˳l}=@Ukx'><ԋy %߹[4CmS?;*ʱzfn>@(^S,>Ʋgauӯ|wP\=XKVkwl9\!r2[JL.0">}ڼO<(w^a{3%RZ,~4g SʣO |'uHFt@bwvh%ފ9_ol{c;goxIXLAWҰ8Z40u@^ŌP$KN6|Ǽ%Fro ["TsF{foKyITr~n ;`Hn>9ӸEt\BQR3X9^:pQHlkɛm&o*dxQ@IO`$Qo`s/XaVpdSbӨH rvY4u @ȈhA #BO*e@ 돍e[H=$ww2|byUtM;t9,'3kc*}19y< y !`wInCM N?#5 rC"cbf{k̋~{Z`3h0刊,QIw P.v ިg}t7c]սR7e.a]ZOdYjhܲ% |0;ӓHI̩yr&<06~vّjHvO-碓6zuycZ@4犀%8R@ %LJ [!B(ҽcQ%}  q2_>w=99HV}_K p KDd*uXGS}f Ҏ]1`eco9ƒpe OD*n93 $RLHiBƈQ-.ⰰZjFR4F-7S~&N8d>>NmG!J' 'e<`"U0P[,\ED[Z O'e=<>f1;l7 툢Q =%$h0yRyUt3{-/yٛݫ7*ie1ٖ&hQ8 S''~0g 5P~[ ì5hҥWS.NՐuR,}:{ ݿWC,=n4>X:44 ^a@XSqo\[K?I1`L!w>~NXPko'k@b{U*sRBPY $̤܊ϿLzUbQbwcЌ\{gfgT\goǰz* r3wO#;,w'[QwAXM:8wTFɒl[7_BUX(i6;1d0πqC8pϢ( Ox➌G@G Ίd3` d:xf> ޟs73.ԏ0, > &]a1UzO$;Hb^LQ,(au!#5GYH\1D_,>{",[Ms-yNqUc Ba/{yYʧ7).Hإu5홋/ }Cf0d\iָ|cF,fJ)ˌ֥~q"zd.m!צLM$UyyZ@|+[ߛjx|zd߰13Oʒ],+]"@Ճ RU'|Yܺf&Eu`ߎxSD{ԇ|ˊ!;zm &%#hWGz}PIf62 ^:0e jLGRJE%7S"Lo/?ۯ]EsRCQ2 N&4.XpǼJy 9b8@"ez0.0/_ۺ)ر&zVTApyW|N`_'5F %EnyY713c]($&4K=.X=dNԟoꝱpHNiTY$.&;rVGAJYO' [k żn8Q@؄ A2 ǨU1kQE:p ^jKxְb8[\qӽW|e'eni D UnoRq7/Lnziv׼˕ tVCndSG|?V9&4fJwIeDJb6:EGv1&_uW+R|^K;"E @4T"K8 J3XE${]@(@*a0˔V VĺS024"sCAF[c ؅H%pTީi8&fu.^ۋj z٤~)'DTz \ON4:EҦȉaڿPi}viMYxrE[ypmFkVOQOʽboל#h<*"G~ PLL!HR;:0 v0b|ѳ9gMSHͳ.Y7kZbKg4&>?. l6qrzàpr_o^&e~|sD^| _@(t@ ᦦSSŶ7j÷W6y0V];W" JS#z0s)3|,")r!5"$+'6Yمu0\5DW҈!ʝI/XzDP}1~C`_D\&d#Mn#L NhPVpP*hH:<5AEfZ鱝 K>: B'.6۫"՗hqTq$7܌vw+IQnmpcN!f%iV ԶըYmk Vz"Mq״C-ƗR x0w% Q3 hY)<GŜ@~'tbOw'X.5Hp(gQл HBa~Ű[)r jcN蜒WS,Ejr/dzzFyqB&U^}7׆54x1RQo]YG+DKG%@EIv˚FVC$"lTDUdTVFq,J_.=ޫJ]Zj+Rs>A]0V8f=qyWښcFdVyR6flPDͥTk9W3 ڒ9KzX,3,t,|4ۼ݀ƏFOlKlp{$|DQ$ dWoXJ$4Ot)];!B@)@P=#ElJM PC4Eɶce.9HTy\bd"=`ub^jvSv29Yj ۍ65)L,4 ")eTC2IapJ*d!2@Vt ID ,4Tѐ,Qa5rQQHjc_+KDK^"n-+mёq<"R,^-9+ d2YNn4HIІh@"g0#ϤmA KZjV#g@#ٲ9:}墨,E/{KI38FPjcFCw$g})g>ޏ,zhh{x x.WP' iڵmuNk(ڠ'~TL*~\,]RW-t\y ,_tvыQmkыA,%ewMb%c:wBz$h!iaf-M/gӰLߧ+ˢp`\^_zVNƩծ]llj$:giwXefqrs΃Xhs7 ZbY'm=wFnd|˻k>.{~{치;t#'>x-eV[sˡs[?-i9幃0y>wԙF+/p8o^6 /7\T80\(sdF5.;Ы,'}jяZဟQ}R, DTZ*W=3VT= K-O5^غu.z,0#@'CpyMΡw̑gQ~\"W@̂+s&ypbbhA;S~sPx$YZ1'-1El1@VG$_l61أ'BG$ e?uIFc$FyN7ZـgؠGѰLNh_HdN̩!9$B B\ɜK܃&s%9B ] ѧ8UIqz( H[μڡ5~bUg|"^ 牠MjAV({7ͧ?@?V;9EЧ& ! D9 %xv5;HRJ`2y1Sc ׳^x4$\2p xN*w^8ey .pT;AE j^;%!kwEޱN.9WK$tyݹ0 ӮK`xJhl} %URVTǫ[\-\85\4,|psU TC>2ɃAgG7܀^+v$ -lÍcCPК0R978|~͙DŽ)3 PFPTe)^|8 pZfe"]zYUٱ }h??d^}^\n>Q-뾲Z|Og-C>1Q!pLxuI!xu /hhb@`bn Kb2< V })1ZyA< Wb"{ kLT#q&^EsH?3)F>3Cu|)ed]̞5~&jm0Q,ۮOd4~m,k#qSR5qlC^6b=% jozK'4U2Nf>lz";cW{+ g.gm]Wso3IS[!7ʟr^6:]k'`=+ƎkYD sKBё+cv ,O+R ސ\Nj9ZbOwcV>͢*5+ @3^jd?hMI4 YlQkWmaIc9[Rǫɐ>u n5 Y+LR~h}4O$cj+ s݇g+^}nwa<4PZ׺ٓ۷'g"~յ'oV΁XIF~[Uio>h@ {4HEc6%jtx6V!@\.~*`*Xy8s *tMg~#jz#=MA2Z[ SxǓD!!3sYF qs{u ڸĵ9rN:Aeq>_O2sAzev={'H枑{1}B/T:;Cg?(yFٳW\*:}⊨wC{q݈+~V|؄Z⪐{0XI\j ;wqR]oƲW~9ͽ|H4-~1}:%QJ,E=llӶ-*vg33ѭȒ+_`׃>Uy9;!8F}rE4_d&Q ~~sor@+ ap3R@b NJ\4PzѾNT]iZ9iPeT0CoGAZĒD%?eah^hA]֘" |[ƆKRUċ)|\ lZ뛗[﮷Rg g&rɳA'j|k\"U"%rų 8'juuԸUWߡBXW\lԕ!*Qyu%|jxk'%BXfc3)[͒媒=GY瞒΂jsuș67ȺY#! X$ZBh{K"z%\C~4g?JnH>~4DmFx8uhzp獙$qF#˴.x"DrEDJ}A"?upf J*vzy|7R62up"'y#803ɍ3N"2-8U!"đFIU'Ă-qչEy=t!'qCN,1AO\=qTv'ގZD[QI1ۭETTQ;z%?+!9=9m % 9;UΤ6+388̽a {{]{:#ǥ:0y4Xh-CH sE ,DK4q" ` B(}e6arQڥkB-cзׂ*snW1\ޟu8† ɔp!BЮd֠T1pY1*b"R0NhGRp)!A3;0ZXl`AKi_[tEq=xRrڸ`<AW$F`H(dȫDfW2bm1_[֋Enk4a8$c&ne5ƒioAdCRJ< p&F^i~jIx4EXcDif@Q{0Õn8ʂ#C^`gD[5~{5^WDQJlT,~Ͽ?V GMF38|a |)}<-/7l<=mQUULmgSX۲^.9 gAB%.~}gs/i^# ݿ֨I\~PY&1,;X\)c4vZY6ah94c/Ԕ5Hctš{Suo-.fe1 6 %^xviA x%+}>+WSo* X&R(0l͡ a;$lŗ uk30C&@>E˻~x?݃d7J4 ǫcYZMp^5q0r$xɹf'?g?b;,z] CĎ&=hhKR+ d'ȍ0=uc\*IܕK.*QQ&~2埿+/sDu) 3F\N$t>2W6f^̾( VR\KCg9}#â倄i&%litt5w߃!T)3ټu`e?sMr*24ki>1#Ȫw_,/]B*'{u CLMLqNq"-BT͍!'hNq:WIg`LLS5Sx$T:*v[dž m^|F`kU&$Psߟ\3AS7y~CcNxt*&>0t TMܚwڬ MJ@]A-V?޻~0rHVƃpFL p2("$Jn D6.L\-]߁y)~sb3y"(G+H9r8)qXc/ݺ]Z.hf, dWG}VrLĥK9ˈ(Xm uPO1&gb|B^m`_Zŗ yG!̰4H%HD0 AP$@2qkVKƭ w *ZX c νQX&FDyn("hk,J+'̙H-epݲ*Vrc-MiL))%-0E94AFxNfm]$! RtQYaX5!,FͣJQH(5tc@"ш,8B3 hīظ𢛽 l\Y#g)=xoB e鋫%\7Ktىzx!8WHwUgv3bGټd #|5uG dI QX J1uS;Y' t^[|*`tiAP& -,I64Kr9[JԽAcqLH˲i{<\oU{vgK9a<&ZS:80T+S-6 :qGVfԼ<.z:?Bk Q9KylzR]_765AKb0'z5kl8a}`%C'&j{b|sOguݐn敧6v, > G~LVppbg?np~ zH1y4)kΊbl4{FU1kۙOv/o^ޫ=LTo~W f`, :I~{pwmݵBETmMmM_7ҙ[  п.!{GhYg~ l_djKTiĐ^R|Bd,H}PnujhMG" ;G&'4}K+8k4TF$cQ p" f3Kg;6 њȂ GGASDmQ,bR͙IyȞN;@vx e{^㰝r-lҠlgYva`;Eh>"c]"V9(cB y&*d pai!!ȱHR'EZ1N4 XakԦ-gw"0$Z+LP.pwa%@f]ϘHS9Q­Riif 90˝AT@ aU (Wq'J4, s[i!;.SN`a KqDX28a:װ`#K C$- fc} lT,Ĕ'h$qQ9 Da>Z!=A׈erR 5 Fq+:4[24*faC `2-¼F޾MY!}ֳhŖ5m<6:Pls ʔ?,r'&f2r ^㨘ulYL k vYq'ZFE #ʬ܁ Isd`0 6wZFLn6-C0!0JRVa S띱Vcƒ /{hFs+%Mo<5FΖOת]m֬.icfNo8BƤCw6'VFA+m#IN<@cbm4vv_a425)5I,:̫DeC2*"2!mE\;7XvM5pjm,c6k~md}rPQ B -D  j3r juyEt9 swj'vtZ=Oua[v>'}eېRqsb-NO3m̵!Gh6 bRl(:еZG'.FIEbtFM5T7OEKŪ'Gq rN_JBa%Z36#gflUfq.ƺ]x;)󛊌=tٗYgqBO؀^b|&TSHeMWDά| ZL|2 {$"xLҪMm8YHNar};d@[b.Hc݌&鎋y.Zwh`JdW2DRn(gP"l^PdCRQu!7Cfd(4I=GX*Z6b'2Dd:fׇSڈ\4b38V#QqЈ;[5kzpQ&Bv00,1bM e@c0.MJ#BXGݑZzcA)-l}NdOZ1Q[7#l:a8j9:qɱzQ5֋jЋ^zs62ȖYkGŽt}M $R`0)fq>ԍ>N@ ȃ\6 HH)Hz>7Hs. :; (vA!4ǣ;ac/gx$-8E`d ]bVQ'ZVCYE)g8ؗdA)mAZ%]: ~L9@DҘTgvf+D̮;rloBbJVo7=(+,-l/6=dtq3^̦qe*c0iHWoR4C,NG[?W(*6_,-%j2u~* ٌ\اgww_s2s9{b^/gv|*Zew-WW=yūWsn~^]T|uO<sᅡ{Xzщk͚_zhǯ-u?)d>˩p}f=SJ/i6F~FIeBt)a: ڡy'-ǕgT+aht : = U2VY8j]6FYu b2|2QLiF !ʘ`Iƒ#SKơfP^u}5ڑ{ynKLlA41a>SuBH%l5V26|C $);~.%=J 2: ESHS61R$V>獰1`"X"a %+JUN$] 4 >%GչQޒ.Veo G&| -eB2Q;#;YvEmlmn g 2dmex;%uSŜn洷y}NZD5YpXD(:h;vKW{b7=Ԍ#3k f6t97mGldlO:_]z gߑ1 u:.ɮI0I!h樓}wAӗ(~ 1YNb6!RVV{3'i)Ml}pgsX)~u3N<_UFISM')5?F?͋kACӿ/NO>M$T7%f B+\/gYf,A[|`utk@;mxchuxH빐kLo. =]SpٽJRP?.j=W?; ^K~t2HfEӯQi]Q ]s2+4SVTd}Yo#Ɲ?^/ի?, ~d6 ,>IBd]>E^/VVZ_nX%[ hq#?d/X=ѢԌ?3Nf$?uzGlz `qQÈq"Na<]J=!Ne`&c|t:P m3@!S2@6 !Y%d,4Xk3E]9[2 R*JRdY\9cXZg @ʀR:2@'nOՌZO=?6 },L@p#"g7_/T,1_F )f/X2˶8$\2{0&!ddf$m&iտ9 x >T ca22sR0)aCsȵi%Pq JNierjl*CmdSm<_|Ύm;e-r֌嬗~b3]ʇ%DLKXJmk)s":L%JmiIq6 /zSД,hkVIꜜ 5C+Ւ v>:"~BN+;@fP+~`/X4ϰ ^d+D(S!@:(9>Dq\96㉣95r<.ޏ1Ӯs K HMMa A,aje*h(Y70m7fjT39Vlc[fw="$Iej m Ar'HkE*,*rPCX/ٌCX:mjx{;!>OYzkz QV Cn)ޞ aUuxv^ZT"`2@0xӱ5djDxL%wcQ(-p&igTVJa`#Zj~F$3*TPhGKU'alt[w!Oٖn>}(͊ V nV8gBGU(Mf>ȘK")_x;(!t HudE_21@t@=cwFi,$,4iJVe.ER!\|N!G |!Zh;6!j$1T@&*%& #[isQd8jSC/w=nܢ#vc^ί1 bc=5.?⷗mv,9YC!wEF]GdS@V4g8\qf@Zk`B1,HSMRAŔerecR @d ؖG+,)I1J2Uڤ IE!]ZmEQȅ^ ݹ]wnMut1c莊ޞ6U0ubLxzk}sznz&sYvceul:6mzKv3>⼁sG؉Q|DᜧFk jѨ4H}NZP`s)Ytb }E9jP Zjיv l ds`)]O>9#@Q]}j_zu_ȉ;z#ALc$ "k,4,c8?l߸~~]K }cZqBOJ6 VR ^ P(':DO-x&y|t1Y{UTqy  CL zܢ32uEeՁCgޫ`r?yplyw ^ʓt}ukp#DaމvlaT4~a#춘R={:HŪcv5+8kX-b7P"ަmLeJo첄Aؑ!deEPPm^6pVA612(&LrM/@ iFN46oqmOn ɲ'][Ȭǻ{-.;PD8AmW/f]iy%gD^wuH$RdE?SudtOݑ=YGFQSg4:',xJ%A0"bC-tNEG6zK;o1g^m^U?oi8~v񟅞 f#8@g>*i^Pv%]ɵjWj{wnq+B#oii6y!C^2Oic|/= ߳vIVW]TvN7`ISEr<䡪->å |\A#3 vre\Z^=Tv#ps|y{]{SB$7 f^|M|.˷76&~i%:5˵+=ӱw?6{ٮ 7E+/]݋v EFE%W s/J7rK!dP'97 I\ś=} ?PyCng[n7m?ۿw\2蒙˥.%qCɣU\|T&eӲV|לHW=&Nń-NA,t|YvL0Mvr4IP(=;SvL:,J4TiprCW6JUpuJ"h"\`}ڰoUa츊lxEdWxb\-NNe*&0VWqCސs" 2~\׎+U7l:\Y"3+KN>wLW](\A9Be*7wubCa6 Na\ܯz@ģpzW;i'3iprς+Uq*Ö # fR_GW*W,ZkxRo:C\3pi\,R~S2цsĕ*7@6V7hZM| 7:zA(_d7sLoEq$uƚ< v`DR1M++LCmaR.m:C\g״+_m l\=!v஘&Z-rUܻrͷUв$xs3*7M1oTn=^$;SJM++L8Ecl:G\U6WWX,},\\/JJUmq%1z?#,,@Ӹ+UkV;_Um7+CkzCWH=F߅qPn:ιej헆\-UGk7\=Mbi"\A0prO|BqRi9 9p'rW*7YpjWrņ sp DJZWeR{#ֆ/O+'N+k,RkǕt~KrSJKW덝WW?w*pu ^ZKaf$"PEвg9dYb;S%xa d;bƓ/r ۞U|V>e){sb[l\aFfgy[u*wD,Twf';KȂL6!8W*צYpj'T W+]Sy"\ yyܕ5,Rv+ T%o6\{6"JNéB-Ox4+ÕY+Mow1 2΂+UiRk[fWo` r]+U{eTz q&k&[3T.YpjY;Te Wg+CUW(3mO;OjAF&Gn"c\? zPۄYz(N+i\`fOTeRsĕL%!8yܕʍa\@e\i+{ϦW% Ln:C.t"\-SiW.p֒ W<&z\Aڶl2z򻦏l>>OʻW z5_gZb(.ҫx{?p@=y֟#5vWid~n(}U9t z.5^g^6gBWހJ%z~D^go޾|`09N)ُp kݟXgw.ڋS܁4'ڽ7߽Uޱew_X~r3pE 8N\5H|/st/n]Eיz[57𲾺]%d l}.ҳuiOeo-&qtS8)PMf_oy?_~y{+wT;_~[~K ^l*F"ɞm+xɳ3^,95׃>?Ŗ\IIUhrecrs)v0.lƍH-ϿFۙtBzC쇅np5 ^ZPg񮷊2g :IlrjcwDMJ2{C+qi!ƹn,$ã4Dɝ]N]|)cXo1gԵ R>VZz=4b) aH P[j"CHuSa=nDcCi& 5wP, :.Vs2u, b)&zn bXb@QXuAf a1X R[ҊXeohb>$6yābB0x {4 %tx @Z K.p?8NQ Jo.C MQMƱe6dG:.TCoޱ-:;곴R롃S`dŒak#1Ї\j -r O#Z%c'p Z7&WH5GB@6pH:v!$\S,`dAN&NZ2 o$X^uv1kX48!.aowB1alU{Cޕz- wq3Fo}ce dA 2} gp#8童j yk)6xŢς# ՄCI:?%\b.*:!B`y\ 1#KAb=T<602 _`ްC4tޮ4Y<%82Ƹb ;j: J1< eu q hﲫP0H <3>)"ZB`cSO;XHԯ+.C-G b{L, 0 [Lj1>Tdȏ){L!M>ldg|& kܩU{WqIDğ N ϧcCw2 3)aU&^޽y__e~T+ZEGϏG] ی$z3x@\8tȦ9ajmZmU l`VӐ.`LU,R肸!=dXiDH215V2]*ёNumt\|OG2HPbb3,k29rՊ[aqFo;^XduNvcԷԑd VqV dlNTi쪇[ebΐfQ]=Mѱ5@2x57[o:WYDqP,8GMw;,a# |4|'6q$yzuH ;dp~3EDR~UG"EQ#AKTOw~8K 1|1s60Z7k1;5@Zb>%7Lo!Jy&3ҁFA$%\ghI%@n8/?@Ap!`~V;M}6U˥Σ0/̎YY $"H"h B,.LY&u(?@5 ߮H!ީE8Ø2UGy&B4U.n&TYmFTo-9 Rn!8|~FQ2FsSI,d ~|(pRzI) ċSoVŞ0Ce8OnOm E罷JLgg\SsNL+M'o8cIxrݯ[6hզn1n }nˤw}c[0> lS&q{.xUaH b㎇s8:@`G$з! D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$% ;y<$?BqƘ`Nz$D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D zI @hr$ѐ@[+%5@0e2N$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@(Ic"`gs>H8@`%7DFHz"H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""@SS`4|w~X6'JY}/3 Ž;# x%0ȣ07.)!#p5K7lG\EZX'zr%5.`\`w%Wh``;|,ʕҊ9և/Wh<],Tv,rֺJ^\im0G$Wh9 ͵GvֺJOF+/Z mJm>V߆`8T{ج^L e~w=.00T  ΅9r_setd5ʴS,~:ڑf0ї^-:8SxPح~Q7'DKvx4(g󟡫/~̞YtT&}12OU1/Qr?|Ɠ~-+{Qҧ %{Ue*s9@]Y&P&X͗Ʈ9Ǒ+cSfiz iJ}ij}iV{& Wj\F`]C%%|| ]99 g˳Q£aocS˲ŧOYiQ5)s(C\ M%qc͡-&Lϝ_~V;Ϸ/oHjnEϾb .Նn6 ?Ϛ>kv4QɸQQ9X g FR/,C1YfJx.sYQ#n˟Nzqr2ghߜ@7{+RT[ֳh0,^t-i3ٸ$DgwTBJjlxLU0o7pΓે-ޔpã=Rjpb+-(th܇ex!Zf@52T8Ce̜Ǖ;_3-g쫅c-/Kn)zg4Wkۋߎd0g4xA&`CDCtȨUIlܞEa|}d3%R:Y< B LRlJMVU(_7dX;3qŎa^ 1E;{`ݪ\윎v|t5(^_]x}:ʏ#o/z+W<78xOH狞ā(A_&_:Hst6EtlQA&5Ƃ4י渗}/Ι94=m>W+p"l x1db1k#2LVJK+%ZW+B~w`e!-#sz\I?~NYтp]aAFeR5:-&(t+/zWu7sV^kDJ#Oqc(cn糬xlLˇpꭂG\Fejz{.pn 2MOK `Vomw6 G.w.)/h@x*Q&Up x%-v(yz4Re,jNi^wG oV8Cp_x^q౮텯fk|E2xX?\C+mi :J=xlY,:4B(_}h)񵎌o_nʜksT`:4˻qm^kLށL"D190r%Rkaފ|xYL:;]:KgYiE4ɇl8Gqq8 3Z1Xm[me|?~0(D$Ra-M"j~qhΠ11+ČqZ[beޛ6,Td]㓅o"% q:#\2x aGR sʨvy,^5l+!Z.l[ͼN9C?9d&(Dff0Ouq=87t/8E/__q?EGMq؛v"6zoǧIf>MJ-Lt\Ek;Gk,\gl6,/%w:7dE[{=wPfl|7O}Wc:o-T_twg{^ _|o5L/ƃɜ1n0NC1t1jwcH/:66xsQG%_|YkX¼Jh^%e|%4DPVDTu@DVq*k+R4{V:zRܓ<. 1Da aSJ: &i  e 6XL$j PvZtc%;HCڛ2>1rRnN./un֊@~_JT͚nn>bV ,nd`s; U~!yʤ oY@X3R*HV@@d)\-F$ȽP{T%Az\`Aȍ rKNH<&sB:Ip#lKn!(\XJ$!\L*$ <#:jk))OtKR܏CW<>'p|dw2TyBbK&ZFUTp~ >ZgcLW|ڛLǟm'ybX%lz*dP̵8ř-Wi D9 ܴ6 c"!gqkr|4(I 4E& ^Y{񫣑@7h$"_E DK V!J7-%f! |4Ks#ui!!w71VmP@[z\9i*]v_) Qb%1Շe:(*h(èٮor@}k= ٵd)d3CD),xa1Ex@]y(*2Y6V2`P\{` Dxv6 l<"έV5pb b<.y쬯FO{h F{8~ץW\`uot{zQ},s@3qEDe`Q8L񓊊KN _ʋ/aTA҇*&hL9"P 4$Τ mQ娀;EE;8ɾq&} r%}F%8iJ}+`0!#<)&ep[Jy5hw6QAIQ3nA#YQD&Ƃz"P%Z3+bk lGF =t1-QgmOm‡Z1jG rc ]d؏s8U@7($T<-DU*TN8K09:Քk:~enp0:z.|BIDL` ܙ!bD$4BhShqVV_k3@\VI[ZʓߣrB̼_--CͳFӬmq8+ӚW`u,'8'!jqݺNmLmLḿVʹLah5<'jj1%M"^ pD/XBTe&9[ƫ jby9 δ >ﳃ[$#VH-7@zn|+tcl>|AR)I?*TpB  o~B{O?rshcthq%]w8Rw䴳FFC4Rѡhښ"mNRD@Lx-UtCbog`pV5Z;X_=)uPyG\F;,RPǣ<`:eq;GYMk;7i-lR^K&+PVQfiTp8l64|Kף3n) }&yqhYL2PQq;#tc& s^'mWVR^$HDmq@ KK+Dui~J@z]r6^[9+hm.xӜ?v<^ ,v_6N3!{wF;?^SΣx):K&2oS:oFo~!o? !?yI/qݒ4y`J_促lNG9Q<'pIs6?ӷf:jr99͹ug]opہcC܎;I߇xzrp|z77,嗀{e8X凈v|U&٬c,'ۗOfoe|f;g7't'QJsz#IAnWtwz4pPp>3AP9/=C?+S?%|$byvnw]t\g;.?&ꏿ\LWm2kK {T߳8?>?_srq |í즅QEg}]c-"k0/Ƶ~~Ob+B2v~9+HnHr\ixQEݥlf1r(y&~1ԛ&YM{Dqӽg{y\,XܰH*T7͎_Y 3sZQ=OtDžQ);egDb,DIveIoIඓ{"JY ~hA8 7Җ$VLs%IyWr/ WUol:ύ\ ๢LxhdԞG+EF+'Ygd"2YԸi5zj)o%gKƵYiYjJG5%c&'f[ .{_u wt;*rm9vTŋLA_Z5eO-j4!ZQ|K#V61 'ys%+>)FK~AZ?/`=ZQ0Mu|awDpe gdI ]20%J IF?{6_v~0Y'3 nvn0`?F?c$WMRd,Ѷ(Hͪ_롌6ՆHЈ1RHp;m<8BdLLehB-c`9KMSqgJz-WyVЫ9=:$!;P^320b24_ݟ=HC!\y%JsŬ3r_3q+gx[E Zb3%#Ca)0>{(XHڀ@DQP"EA, MÍ X12n7}EQXs R`Ѹ[ a: #8dsH7Xowf ٚ@^r\HɆBdLT eC4gO;AVVqYM1 %#RZ[ɤ8uc$2"NdeKH*4ˮgY3/??b7nY?=~k|J@/'LCas鷃5졒ȹ̘\b/u9;75 EF+'Nq@}(\]ԗg~\TgT'^^]6gb;_TĜ/A>k8GpP>#MG@HH [GxaX03,PhC ++W1Az1{&ۗGLmԶj͒=Nd$,}~|8Tdq@i1 usS<{}ޠ>])P>~uɿ޾s?O޽y;p.ϭ-H0z53O\zhMhWCxe]κYeU]Neܷ ƽʘ[ 9Axp=yX/ZnvM{peu=E|F1?ƅonQQ- bTPqTKBwdn MɄ{bk lW,b GOU}ŷ~]?$OM37?u1;K<]y'fm)Bh;2_y5XyrKO0*vۉ6[]TV{K}Kw >Z6 Qle܃ow37ʨ:oz'y1OjhߢB~UII ayWGK~#oU¥ȕχ2D~?5׷ ׷֣Vǥvҗ~=*&|EDk?|ٗD)@9Uk̛qs.{XD~:;<bu R{o* _ُ~ui"5Cy9;MqP`[hmUw ԼtZh[o_홠ΕEV65J@fVD4Dm2`lk0F&WmZ 120w1TUjԕ2@lJ-*S6$TÝ~ԕ6RiE l{U&otu;t=+DՌ-RW`ŷF]er5u5JNwǾuK>z'Zrê"2*Žƙ-dE^N}8=SUnkOur5ՉIΖFOEpV|[gBm,[wf+ԛ<O?z'ye4O^1|QhYL2$^7ʸ@Q[)tLD6+"$E $u;{E]Xdz/Pk޶[>^n˭ۋœfrڿ{e/-^{%`ǥj:~So1WʋdI & VEj VhS2% ARoL`orը^{Jk)co>թk=f+>.Ofyg!^ϗÔ l'ݽm%> b@}yᏽUr&h)\r7A"Xxf0r;ag} 6 ?ʏ21U8g;e}>T֫Tj^!yrEeX't™21TԻur۫UBz7SyjPh2և[NRSZEEsJJpk; D锨ăCx 9m5W`rʴL4~xզb3@8(.zU6.r면=>8 ?Zʹd5Shɫ7eqdkj=sm:+W*5=F}˹ﬕyo]aA |9w~ݬ[y,=Sf$66V2/ekGzmq]⨄@v%Ӛ`+u$ :u:TsSQOV;ǁV!Cfk rXM-H M"^ pDv2΅H©I@e$gSG'LIExFg|m-Iu8V,:{'^3/R#M\|Q6F䞮dMp} pU^J)uAl-yyE+_>X&DJd(Pԑ Dry!0htP$ V; Tx㩎F&--Xi#x vFס`7 dǙy7s~SX{=5n?Pn]Y<~c_Y t0o3UA#n iuʋR J%g.td:(JGrtRs\h0v; =d,3dWZG+=UTD),xa1E};YŁ:|O%Q0+&ҭOs<S ĩv!ă2Hf$<"&!$pb:dS \omxgܿplcjw?tuQSiSp`o**.B:MK?g} 8;T l9:27VwmNDYMK-;lWIL~WUOO~M6`pPg:T%Tz]%¹PZKl1 hҪQi=j-Q*ntfVc?uݩ{O2ٴ(l6\㭳b?PZ]q09RA0h sLxrynuŐ[:tNj5ɖsSK5j`v"@3aɄQZ͔o:R~*k6˔?6FNdKq.=Tۛ DrGZS 1PKwQ߼3&H|2 Y.q[| d=vݳ2|o\6:Jw !Dt8u磃e]3e|SHP_B4ǡvxSy(&GHߣNaաpnS-z$PjR-xmHq%WfyKgE<֫Nct ̈́RH#…ଶLEtA!:Y茜%Sm1 SOv&6S%n.5rhNN#}=~~y$3nnLP(7,9'm> JS-Qf3Ŵ1Ys*>@1T@C#eKǽ%Ɛ⠄Ht(t Cc錜 H\qE}‚mtk04&&ޓmlW<"Kݴ A0f"K$'q w,9Ў-j>gQL z 2yD0Bp/yO,A,`98j=<7svn8#or7aPns1G@ZYKx57oہ{.Dh6WHCsKE ?3NTSOVbё(kE <Ҝi*I=r -6`KgBqp:I8Kf6U45:Lir+*G=jSGcA\IN"8XĖ/= hpP%L 0⭵/v[2V<]W n/Ů lo{s71$Ǔh1Rޏ'AIPߏ+C0Uw=k ls-j_^'8d6egy`Qt/`0}ӭ/`j2<,FSf)3F.t=6XƣWzG#6}Ÿp:j>MgŻ||:I5NgQG'bzt+50;$78дT@w^(7*b=&V:bs<{B9m\I"HZPLK!B^%&2##xoOQ"/u]hٻ](I {)E][v@@v/u׃>nO{q% VM ;b^`\u@{kD32=?He'xc5*+ WЁs,X12vFwse|(rH 0q)|\+?4&̤,H,x!hF;o|X|-*^N篿2tuEL8Oؿ-@eo/!t:Ja>OO0hNP~?hҥNW@/^YNр`9H}Dsm'{3- oS4׾}1Uhe@f.!aKi9`L)w^$,(QvOӕK uߗty9 HvaM(ĥ-aGU⥀83b+.~'}oF&+O"χ?Mo|rt0-~_Nz$1Jt|wT^fI57G{oꍃ-r2i1><*C,pE/pȳq(Odv==>Sْds` $ռt6YLdV_ 7s.t0 5@XLUa&^R_իL_Œ?O, zd";:;+|wم`n%6͵yȁ)cZ@c(F쇳YWڛxAjy">^Kl=MԤ_5Ĵ]򙗤wErW$^^^2qy۰ ~:,s{F|℈RXC`QUpWsM#ly*l6=Wш\h8e;c2!$f04#I$ꫯMXi0$Me .e;)v^LL;~I&UTbf:fwzo0Ok =k~\Opyke_I k6:W;IP4rf[^ ʝY~;f2*p`Fԭae/ޜ`wW+]QܮWJiTY=UZ{{ 66?*د޼c;/FUk Mڄ?;e0ڳW].vpm*zTW/-caKy.~\ NP.`j֚ LJh"ʼnZe9&&b2xBTHRf TQɍvs|:߮8Q8ю%: vp`sTbN+EP#W<9b8@"gz0.(O)Cgǟa9߃Jtn=z3JeK%1>S} F$ƏӡvyGα.g\y CIaF}&}iU}_ږY.fʕf[+e=P0 lRj.۬ͯ/풛}ikfwkYrB^;] Q|?V9&4fIeDJb6:EGv1_uqFפ|Ay wD@iLD2qfP0$PD0@AP(@2qiݗ0YCЀ->7qo5IQ2.D* s&RK2\Yпm]=N"0*"X[KP2@Da"2-ںIB'n !?#,Tk#BXG)2yKPjƀD=sG- 2c5c{qaSXdMwgIfY .l;|Օt1OXfUݴ0;x_O|ڏQ9$DV3{V|f.a짓2 C 9dILX J1u3S;:Oa^W N[ t5iatH\_a wTG% 1˽O1s) v}}γ)))k:.N둮T&h*% ܝF9$k@H|Y 0;Rҳ&gGgZgުb4%0C1[v9'*z}|Qj[֕߼i2}IZYEƓ?]LR|27=Z8-7=jFn^yȶU۞U It0q2si9YbxIϏ'IdŸbz/`GG/~e?_|&h_"8 6>i;/jiXY*vyO]o]֕]>r˺6{E1[- ҔP] 8otr%5".•gI\u*晫]ACz ;% I9`I/XvOC4/x [\y3S뭋275"Z":Rjsfȵ1OQ&N{lRsKzed#"-f'F=dKZ1a#A2 jZ_bZ))rK3?jJ ~D0zvGƕpt?"t~~Xa5P6tuhȭX~YagimIp] e?yBSa֦c>4;Ԏ;ƜfN0ќ3},i;R?f1BH8QsΜ02!PpLYP,I1R.i XLk12& 4V+gB~oWH9cgP{jAY&kyt٧I'3l2 4;zvuy1v1aT`i*ƃI*pW_FwKcl{I؛җѰw$mbqS XOxQ=tv{#6_p'7-kٟ|UJ --Z1 )Z!c: #e@qٽFKV{B6:|I_pO+y]Mv:+5g߭KZ~RLeK_Uɇ/}벶tVGEt[ҕfscφmاKxEcrv3;b[:M}5nW/ʷ'hmʙx4zzۀ~ *y/³nk{<#VK\VJ +e6,YlhB”6op2.1]bp4$HsIHZ'͚9]:-:bAd\Q@μn 6<Ҡ&IO[Ƭ(X+4JjVg$C)VZ,űGl,lˇ/T&'%ce2L: G8/ *JE )*P@,6gEY )w\iHs>:ea[i4.{AʝiNB[O(-hjVhԁjelRB}YYΪvW`f.ST:|,9e&9 ʣ =de+:se2tKӋ:iׯEz ?OIVYe<IЊ%ԒD!ƛ1,J;U jtq-;r%1F)aI31{N@S dtG1Iӫc;{/V+{^ 7#mì<$5R* ~Yy2poIÄ˄AA$Nꬭފv5Nua;fTdWӕoM;3gDp{|Y*1()c^JK/)CVx-W4%;4dW0"ha'gS%/amD;Y$$t&/l*N%Ó)N-ru*ũEZ^ZSdqH*u2E&O]i>vwUtWB3'H`dUWSqWEZyE]Mw`e%]-?Jݭr3> ) Թ7OCoS$wџLff5~wrZyxcLBI;qu\fe,$؁\VdL@N:b&les`BAtg(4'#Wnՙ{!UtGphN[%!+Cjlt;˯lrV^AǓ5j^Gz 9⸣C{:mb:2x@[z ?ajlDMxhn>__~ї cR Mm+{xm$ Aw} d%;n +ٝ؁N)h5a饚mP¡ [ fhp:]fu4/#y(i )g0,#Z ZBad"4AHVٲwػ\G_{p:]~}@X_w's*Z8w (B>]VfHʃZ@„R{ J{ü`Jy[ #Y#`,Xv } I =Wb"; koLTg/Q׼G5ݘu`ܿ8[k+'mT_oc.&'mo:D]*ۚn҆-]B+GaN;n_Y.~кzb빱zW æԻ}\Ӱ"p{8$nkGP50%Bه5Ƒ.uC#XSP~6.w@N%6uz^vsr r^)$FZ<2.ȠXC$'GA(lf|v1׋\Rʓc.*s*uEZ^\4c~u4٭kr0@f:"*oHbcQAN%+nYcR8œי.a ֫= r w9i>lb,}B୍(cVy1 r/'r5hSZ)M%?SOҍz7Ұ7 !S׻mzɭ[Y7my8!-(힞=-tkJ:iʟޥm<:x!I͉O2#DBPI5I8Z<0u ^ŌP48lG´py[azx,j6Ѱedm֬صoɿQI@GTy~޽)53)7g\2 PWk-%X;_˦s k$=e2mHG#leܽ`#YQU+ O85ULo+ >v1mw& op>7V7gbrt 9CS8XpT8r^(7*b=&V:nv܏TG;oFa~;?+7Ss R\AD'Z QD2^Iau7?\(]2X޾Q%R݃_]Bӧ7,1)B^r'8H/`CLg9;(!`$Rq˙| xp*L}d)с &i2" 뭥Kh%Ec$f 4y%f4uא$d*ӣ"xW|nUJXO`"U0P[,\ED[:ӎOg=<_>f1b hCr#F)8jJi90yHp?>qwCy"HZPL[fWɿHd 0q}k\ wLEnka8$c.e BÄ{ْԆO \5 x /\+IX(W|3>]_?yIˋ/SYU!{$Wi9|09Q(\L)ͧ~ӏb0Z0uMu}:u|͢Kza){{ٛtWh袾ͿG7ˑ|Yk"<k*>Vh])udt|HGN=(*+ uv<98NOd$wTr^~tqT(pu@y׻'sEg}]+ V\K]9}#:c Ja?rZ)k02v^φ> ^ N&kgکs_ۚ 6߼4>,b}݂B8a:uφYK_ms{2FꢿpBD),12ix8;(y@Bo#ly*lhD.E1zAcD W$Z"I$߿&<y{ojBjo!KMS\^L^*]B{/&ŋ$i!*JQk3Sa g%oxI/7i0,o?2 S +Wa|k{` ţҺq5=8B pXJ&SykB.*/r5lSw5We5z@HЇ*]J.qLJxz` El>L4,GVВCv) >ձ ;M;a- 8W7T/Ah<Hi f"A3n @ygl467ӚnL>2H`Vv0`z Vw`;:6~;zLlV]\2P,7\_ & ڈgw뿢C:O*xΑ51Ȕ 3F od 0#xJn Dnn剶,BW x{qUQ w;A9^qp9r8)qXc/7T(z ]}gA~'lb\Y' aZ >-_5o %_6Rd? ώ^%CV?ta`"^~PĎ3r0"eG ŏIuMPT J1u=3 ; /zaYW N[ |5iatH\_owR%ß$ 6`,O));0lbtgZ|xCNznV0ꖷaS/AOӓF?.u1Kbcp6.3Np~Ch⏲| _ N^':L'?/+vr=d0\ō%%T1P _e5 9l<p_ /T›[ȣfRH#>ĥ<-{ƁʈcS; V\aL`:Pv1㹑Rt2ZYP(HjTH"ʃEL 9!i=tZY٩hA9;͜EVlY,NY̪.y )qvfUtAK |KYN{Ze][ɠ逸*T{ OSa^y`6 bRl,:GܮO>z]J,-,"{,zF]ɹ o"O>-J>yr rK.JJba'36~44P_0'/Y.63x b_ČٗYζE.N?߸džeDgBe0i4E'9YI|9`C!(!N.R{]_C bVEPZU-'@ ٩]NP`K"؁4Ĺc,_ 1ⵛiǁ^[^{G-i\!X䐈#"H0 !C2Bƶ=.&FGȌ %pXF58rTeH\Ljr16ĹAMuCP({D9z#$ҁ88 aN{M%'h\1`ɴ=xDȺ;RX'uvJ3'Y?E8VttG։#=P#٪P:iɡ~Q5j_IL@Y*7fvܱCMuba"GcXC_| 8L;?ԇ'paT^TȽ .HH)ꉳuk4Gh!PQc.at6(Dcx|O_ ;4eiSF\%J8J)"BH.c&J9X#R1i X&a. fSS'tJϱE38ћHg VG:51TZ?ﲩ~' OTT.W>U+3bWu6TH8f8YI Qa\ȔVA5#F?R>gY4r/-з%-?(Ọj𥯯nU]_=n_ SٔMg3c kvw]{NO*5b}Os;n?m-Z_Nmwf!j~nw;=_5e=zr>\L;[]Ms^i^7s=/&nQSo=Njt_rfH3>vΊ|S p:̸L+mA*x"+nQ.N*;K.;ʺ;R.7}r@q%˛dTZ4&xvT6VPI8mkiZڑe)ւ Q$I#HNM!f]. ,y/L:QxuT[x鬙8{YV9rc~%elM!A0)TBHbj^r 4' 1yNĤeK(ǽ Gp'R-`%LgMV $+b֨ٸ~Cw@R6 5$5TV) C:-D7كĖ}CC_N młf8e{3orsGuR|^B:2ả4`lq8\Sq^Q.=P64Zn3\Բ.ߜvvljfVu-Dp?JcXt 31q)UC|c2(J |K$?'xjVAyr.mOƃ9ۑA7fig]0^Qv)*_L Xq`![ n;+ :z.V@!(l(J'h-MӷKG~ng5RԠmM*KM t^3hY-y'-ല#jtH8y۰K ]s*zAV$%I/kDб}p0:&cFP5Ӊ(։΍(OGLӮs@-R,) tB 80e52m2PT$y7[a./&Lf*rƷڊ݇9"dYej X %Q@.qyؠ*5ӐC7=.VBZt;1 7&kF7HԼUB˜l\&JAX',Id֗v]3N74uSvIpQE~Im58eWq_'}^|yA>?}+el(q_?NZiVy8)Օ1]YYz Rb/Y?Sg}lſՙ()^meӛ %.L~-?L.x꨾aE~׋_d啟z V:Y:ogUt?NJ>тޚx^LX+Q//_ wÇފq9durΘw Nht!z1xG;!j o3DeR=J#iH1[1}6ZE*QN5OH\K[;jןorBi̳skäyḬ9`ΌC@]:+V j{jՖ${5v -9L}xVYHKYrBؑ$Mƒ#~wEp_IsTWbJz%W$uH"jk) k\oshbf~ LN̘BY'a2pr~&fȵ9qd~E_b_i)@=\u>ߠtQѝE߽e?^h)lG xp)lG"v)lG ۑ^s)lG ۑvv.y]B)OCO3~'iNvihjт Et"lw!}cc/>ih.1ŒY?-䗍՛'V6̞ kGgx\Aq=:m:^=(MxI=7|~Og}y:e~!S0dr{/ȿ=oٖ6U&۬v`ձ;RiYtl?U˃MJ{zzі Э|[=9Y&iYw'i=R|Ǫ|*Ww{T>Kkz>Vځ1ǒ j.+/YǼPxuS;kʘpRpPDCCHY<$cP};Qzh 0j@ 0 QQ#HQX,BsMJN(4J `#SpEjm/Q$(/ ؒv kl~ӳ|-nzm%ze" a "(oc7Vg{6W!bq<μ {{mQ4],pg,l{q!EIvLY)[IN,82zbD5I< B(\zՇIHeӂzȈ )F5!Kyu~`]H"NuT򳽑>;hcWH{ֈtЈFXnSR48Ι!j&PkVKWDmJdWhrDzh>$( Xzƅ|e9#^{4ŭB<œ;7rֈg_CX :}qɮz٠ドQPJJoAፕ1wDm2M2  zЋЋ{]!YT2*M5絍ܪl~ډ`GvFQho7k?S?]!Q+$,w%ΗY%-X=9~a;ѷ0G}B䡄9HKa&)@dJ<<4 NxsVcJ p/1c\!wiDJFU |.4 iP!=GtK!cE !ů ̾e J N5d܁@eX9;caq6Nˋw3!4"'`KJ:UiQm :o+7b,X9I~1+ 䩪B02A|Pgxyݡ~ů5{@*Tf| gv~,UMʟQtMe If-^,G J3NңUpQy {,!ߚvd66~.?.{]W'ytVšl}:,~auȍUSݺLya[5d=8e3zq9'7]yլo#KoMѝW"r;tګKy*x.nCW6{1-Y}vUmXf5)7t6}av:vףϭd~=챙^$UMUzfyHCzCJndA#J)8H)-Vp;,%!qxIl %:t!5E2JHy\+ C7F* c/G@$*ʤ^I0/1b4*~$\t$xcՁYʔ."7qk]hM51"'}ҘQ{N-$Yr/vS|ffJ¥GS-4t mX{Rj fm+ = +(MJ9J8Ő)5!#IBqmʩ .gE1@"pBQo=Y.E,P"Ni %%9EC9|߄hgrZA-?Oz^v59.i $JB<54$GHh F<}p GAq888xbewoFۘi@ R' kiQPECw9IY$fth m m]*|gk{@mOB@A"NSgK0q9"%0{%;qȮҡ{5n e&Cz2>9 HM)D&{v6da{ۀ[||b[Pg>LPuW&Y (ךY +}d Y5TP*GrRۿ;6'P}/{&ކb0O ۿpAo}jUq _)5ơV. '*⤮m53ɣm\ZwW߼LOדCΦG=炷\:"Xϓf֠fü hYp=6-j 5/53zkn) 3A**{T:8N-JJK-@ L.zT"/,-#T'tX"BfNac T^sY6JP %&154rgwx`OaYY nt;+X5>4?번d뇂˶{&Ǥ/+̋-*npϞ805)L~#8@g>K>by1Ob觳@./|TC^?ym7 bϧG8BG.'\ߟqJ_jۍQJ9EM$)˳.zԀv9Kg3),CqT9* Y}9 u/|@u_0gXÔiS=N:N"6EEvQYju;e]wVWwEOOUQB%jz^" 8ˀ-<'TؓE⎓'p?uyܑy׽&ȡ%Mdn߷{ .xWz,T^l|vTh#a <$57:Eb xC||&9hNl|\gra{n>:1:IQ>V6[~Cc+}8Q;U4FO #0xBM™$YB#t20%J IF)#FL"].̈́RH#…ଶ|p"FysiB-n)}ϰԦO~ ;^˭;K4:;ULRc=?3Vi(D+ПpuZ5Sth:BS:~Ukʁfqv~|X썀aBfjn(0 ,'XQUx ܠb MbbYؿ۟4m#Hf[)I8HvP sfJ8N3p` =RK/:+vin G cI)dRb[1yRߋU6[3ٛϼZ]XxR5΋ߚ?}Zޓ:/Uwn%\T'䤺U#2QT]wJrN?U樚r>EywǛJQomdNNV JBh19o5 lѣQ8.J/mD;q#NW0*?LE3=}ĞR$HAYP^IPX&a(Q]EIWp+. H^lXP^^J_M\p6=n="6y]xWdEܙ-̺@1Ȯ'\&OVJδ[ ?uAw^}(PjBJ()ek}t˜`h fE1KTr0H& %R&Xx##Wr02?xt#> M[Jl28=3ܞ,T=Yk֫?| s'sCEr<\Q4|JV.@6)Rf]̒;K% VheV [),W.L^D ԁ}8PčkinG&ۑGd&ߪ6n ]c,1+a`eA%­V"qcʨB}n~ t?29B>}x(wYrzou<gp!@!J B"ZxG0e!EFTX2E*LEc'a7jtfH/ l d c!$ Ae\XJW"*sѨEc1 r |8[/7Tw{m%w@<=@wmKc%'PzJ L*㥧 $$HN Z-MAEDG<h2;I=|mcb s̢t6؄ B+I#g,YAVRiӸLX*TJ&ihP46(#f"*P(:ߜj~jj({8uo%"rZ+SQ"ħs=5ƴnb?<鲾sCyitKGkOdRB,SϱsS8+"\F>BT!(_Q a>Óǯ)^ρOlmu ~Z ۱tl:>f0޼8=ƛ=!7yfO$3p)| :D?LǟkwƿNgh>񸗭oPOj/O{0O4o>Soys)6wޫ0c.a*e ]*+XWۛh{?%7x+mqz0GE9TF{y_?XUTb_5[iK&[Zb~uSw%V= |}/Ѕor>>w_a@[_O8~^,Ky'Ke 3y&v^R*'jP:TCe]YD_=LrTެ>ddO?ɢsAszS1@>IQcp:ǘaP_ F eW+PSvg/u6JϷJsbX(2A0!FY).J/A~_z/~78%~{Mg4] Z=4HweP%fsMGX{;L~<}bL4!T=*rõC:y7] { !C-"& TzxxCAl@eP&'U8SLL*\}P*#ԝ~g`ؘO Jfmvs I/VeʿͳtaGBE>\W:M>n'jbW/&i+CrI.6BU rQ"S2|JrVN,?Ty=WcBy0.SɕLnЪwOp獼l=[jSrbc^R祢+UwNd8ox2T\Li Y\{vh>.&KxyU?>?LԴN >h؞] ]I1-+L9m ]!\BWV5^Bwtut\|B=e+ANWѦUӮJ_Yp[ ǥz]_ӭٹXY+o؟Vkq,g"#'y~y8fZsSFgT'WZ勒rgS7~a1EbO`E:?0ݙ+ݻh:2W@o)JD>%q8SN(`颙aR^d(sMcVET\ jTz6闗?•-thh:]!ʦj't%̶5tp-m+iޞ%[YpJ)iK晌Fw&=J67AD(*4L뿷&Pkhw\R5-AKcD0DP&3ZCWڶt(+Kh@ƃKB5V5~Qꎮ2kv9!,{m;fpj#f_6C4cl@Wv=LY"-thk:]!J-::@bT &ZDWXw}B %] ]q<& kFZCW%\BJвJ!ҕ`K"GBԴM+Dٴ|R]턮${\BԦ+ QEt)o1pekA@x]"] `cp0:? ˃ A%%?p1MN~3\@CV`z92JjbEͅw&'ZQKWqZP߮p zw\gw!y3&\Cyaft|o퇫̯߯}8u:|\g ;qݖm9k]  EJ*]!]KE|O=+v5XgGk&w?x]6σ&ۮϨKZvwڲ/hh{y[mHS8S.ԧ| ʰ=)jA+kYP'hWJ]wtk:+Kl%ho%(YꛡgF?&f`>qjZǡy(݉3*tuhk2@tZaj7BWӕ`ҕ!3aJ8 ] ڨO)ҕ%NTYЕè+A{st+]#]+#ѕv{ʎBWVؕ4n3+fKn+f ܯzJ~i-/+tV:CC. CWXCV3xty2x%p{n \)Wn]eWL $ce/^wd;ddQqr! ) ƙ{( C~hxߜ(+y:&p0V7DRS%ϑ"pG;_\ѕJPu+kz^+]8#?̟ 3qL4jvC^{enn`0t%pC;u֯tute5+i\+AkO^] JV+]!]Y@tNaJ7 ] Z֧NWr5gIWɏ= ] \;|fPP:?{`4CWwu%hJP:Z VS9e^ (G`,X|S¸1u ;39Z0yȮtZ+LЕzǮ2蕮Αh@t%<ލBW6tթ=\Е3괏gy\?~Z.;'2+xy}JKQB ϨKm.ɼNR 1gGa,'|k_5j`pЕX"Z97U+]!]Ec IaJ-CWfϦ7Qq,6 ΂Kț\CKGpv33ʬtuhkJ= ] hG+_zӕyЕH ٺaJ2BW֟<] +]!]YK_|%t0t%py3(hOe+]!] چ c91NWtu芢?:+0tE1ؕ+]]I;ȁS?(k^_gEmf~Ow嗇 _OA!]n'܏W-?nPu_7WIdmy#+W5]i͛?/?uIw)_? Q~@[ms}{ w.no, 9b^vS%֐"v+Ow7{B*>~~r.:vw>?D_Sn4үŏ-}o~0ɪOī _H@t9{P-4pXs633ixwPO~e&,t.jWz.[^_awcKB%Yk8]5&k#9r:]wvE'Kx+/wwڿ~=!>A{M;~Ï  u}s]m}KətQ*jzM^ؚ (:)#(ҕc`gv}49eҡYUcUSV )jJ[\X5uK_tlƢ}XhhOݺR+UE@67Q=q2}L9*LsL $ZZj j V4FDMFߵUS)rv>7Ulr~;HXꝹlS5 kJMSjJ'Zj#P31N#1;Q76{PtPQ-^rJꝼDF0S$3յ:Qd]=lҹj;wo TE4P&sGSaz>E~w!UlS=^bC7*.!~DD_ZW*ҁ6w{JsU!m [=:fDyKƜ؅),~`'9IZzU^əJjݳA!+Z[u΁z9K:>Ò|XMѵ(ĤSB\RHHHI?AϷ ҋJSh#%yIb(DmT/  eZn6kr5J`џvѼXl)6>]PdTW)鞺5sS`w"R ]P֬ViXkCvYwiJ2](_ \$0`gX{5PQAm>)h-a]K68E @R*C!VeW*qAY,<[]:A8քFFWXtčd] Dqj 1YC[WOWgjm  E@P9'*HJƬC6P m "׎]SPP|(MA'SX|/H1$؎zokҹZe R5RcM ̆v6!I6p.JqI +Qr HM餻P TBłd2 V#j첊޶X e=ಮ Z dܠaS`d 1K@ كHc':uGŨ3| ߚ wjTLɃ!64Xv=M?\t}s~ KWǴoߵ-:q&-:P`0uƝMN,4zR4ظ`O ΄NdFEPy֚BԦs4RVO ]Ę\^%tkp}EeFĪSC^":$bC;"$6b]|r3pM9B oQ^Ѭ b^8PE#܈ČgZ{8_imR"3K#5E*$%[n>Z2)"#܏SUN[}OWY+rp lKa<׆x~$!k\+\OvOct&JbCY*LN0p&mHp"=j\OE 睦[Vq aנLʒ?xQJ mʑnϡU:0 k 6P v m:b$ZH 0_dpciޟz)a:tkQ2=zr88% M'e̷ @iHuPu|r" 0e| 4[D"H4|teI٧Gt Z~UAylB_=J<![NWnp_;{7ޏ톲1wt%v+tܮDY'* ](/ڹ?]"] ]1b,] \̱UAi4 Tp#*_=Yr77UNW%HW'HW `c{CW*h%3HW'HW0R{DW]7kW:Rc0xtь `՟`5/tUڣ_S+d'ڭ2hڱ;lo6O5SZ4/dԳ 1n'SwI2HX;:t J]D|;[ky9*-{bËywme1:_yXf3lwq3'} <i#ę.m% ?L\W>zw⬯n!~c1u oP K\?׏096+TBwӉO%ZhGoF'܋)?Ns;k]AdD=nM,qehRV}URfͼ4;3!$I E)q^ؠM@QLKU:rL1 %Ӫ߯m5L,1ݜOU}#H Xmv ,\JJffZ༷TDY&1Z)ϑJgtjp]~Ikꛧp CQCu$< )Y AD TCwZ9@Rf3É;WXҗ!B͡<g]5\\#48yRp&WC;3[pr0fָ7׳lTqoid29_@<=8o\r1 +Mme &[kp:~{\-/7i'Z4Kn/ݺXlj6T>OلcJh:qJ?T&粽 nz1e#hT}v|d1\h@mk/zRROnuǻ+x62AYaǯoV?~j :qe: Dc7jݙj)LpS9STlW^s>ߞXxͪ=K)";h(b>AA>#8UR}Kɬ4ͺ9̷ݕAE;r֮l\mM|aSMoN=.CkoLz@ПXa"؃5rSSӴ^xxݾ,w%N[)f(g摑&{\5Qj6rD@ijb()HyGFL !$y!(J@c6w8ƎhMSܤ)z~1Nxon]mp9u58硖wM{%)ԘaKb+3kVN̬<_TfF[àe!-aeJGPw0wCi OKJ杣%-Qy "dꘘtϷk&$17mw6fE&ףpOhRX}H&7LaFvgUunhSoxKBO)qaV6J_cXx[=X^Ky7hz7kne_-ֆqOiEn2"Yލ̜V6eQRB׊S^V FԎ'lD T>0~loQf(ڒLt,*IA + :*Iawù[(hVV{T, ` qE`0 4ArsЁіSpK~skTD`vr'%OHJ .kN󾏜2J%81<:>h34;!G5"=(3*sSԓ F꘭gѥhrNR)\:43 2 fB{`.ȅ/ʅwdtoԞ|gv; l{n4| g8c k?{ƒ\ ?& ݏꗀ"#vQ-1I^5S=|EJ)CpSL$P9A`%UH9nvvyĺq I0jD hHFd2v̺,!8#fFR5vaZs1YǞ-X]^Cnx&c)pY 仑QsDP3, TLiˤ'ʠ ѥaV@21CEGj#aMBq8FRVErx_5vED^y"(mD#yE2b*Z9+癉9^@2Uc]ֆd@3!WQS#yH՚s;"~Dzԁpq8jV/.ʸ(:\pq'xF蓊ЉMX^w>KgĒ1KCa5ee< IV2'~,өfJcs*wJk;;U(+Q_%n'{Jl4 Cgit 0)wW7vW2u]V3 Zk 7Y2 Q3h4 ș f&1c^i9hR hc4 #&(൑ak֜;e"x@K?S˷<@nU n˛OT&{bc`.c)42AM1+S$L#1؍j#wYHnT6x"K1ӘR: Y>$qg`G"z,=:s=^!vtɀ-\hjNX41#M"J杬<Ϊ5Wxì5.C"%}k˹D`9!d]ii8@KG(!Ps\=_pؓi4HKLc%Q+򖌖& il4!fL3RjUZq68&󷝟^w%1[ $ 5c;rD$iX^`h~:S}KC5+a%\)au |;4m2MTyϳB2/ cxpdaBKe gR FYlB8omTCo>( E5wd}.xY}<.WÂ+s&ypdhA;S$03:9StDwJIuo97(J&43nE<7,L3a>nL^ݺ`ɱcGOĎ6Ύp} P'5iOUޖW d`A4Arٌ J|DG~Md.Qر9 dL8Q$(IKyA%%ctfbVA4իR )XV*Ye:C* y$C Y֫՜!!tDPпpڔǦ^f<|y:ǜw/Y 䣣OD1aVafM_ 5bc% 6cK5mNpD;!m%u%zp܁dG^ݫD@jt092cUS:c]{s(9nIU??w]Q[^ח1e3pT+;XSְl&G& 9qeN"'O\>I(y{jHr|txgNdwEh;\0O}1.ǽgk"D% Od<O1&  Hg4Z:^uCP˽ ܖpڒ }jzxp#@olK-ܔǮ?w!K W2Bϗ82~&Oئ6MbG?650}n1,m~SM;AϟME+2Ypk%FԹt Z3pYn!ȲC֗_ZeXE;w;_5ZY ŇHZh koc9yHH=rjElrO2>IG<S +0D,hhbl~tHI,R!IR!1!UqD X1 k xVdPELq&/͉QupD(|l;`i#UAW|  -)6ej8sBuH! URovp3oh'^M;7]mJ+AG>B(mKhoɢ=2OX:]єKPD#*ɽ;lfSOMf!sCPb XᅤFMʻ8. +ъ'):`\)Uꐧ*ŵl~O'?Q|@ލTZ2eEz L[pmK7I֠^+O7꫍ ݠ2X7Ŗ1{WQ82t[Ej{vp{hd%T^)hG3-/g@NU%14. Ete1+,B4L'Y[a|[n<֛Wos{7b?ǴQumo1L^ʹ1ԄkXtuɿғcZz gW4Fu^7$ؗȘF=Ę`ӕ˿ގ17|f4MV~[ҏm`ٛh6WosyO.p~z6v;m<^tyK?لz7?<^{朳ꅲgホ|/_aߔkOK;}4ij|al϶m+X~x k{f7A?~=?w嫢xo>Y2ු vx|iZSi;j4@[K_g 뷷l޴W˘͛gVS XjΥvCQ3 79m~S,{KD`Gt"|iF1Pwk]nI:9VdSq K֌rW*q˰|/?7k9zE:]ykO/6L-hxBRjG#'h0J1-nO|"D>d5.-Zu`n=;6rcwKOW 4ه^x&gkBؤdG-ݼW\Iw3b:7;mOQBK|qK'k{iDՊn5f՘_FcGAQ`rg.3X"LІM63r3V=Bnq~Q@^Ux "ȫ=$R( *BڤF^dW4Y>,E'Bjb1_4S}?]fכie?}w_@YoWyc1'>XO겟( .GD=8y?4LxcC691Yx(ȷg?.7X+}̹{`GflOT`?ڒ1rY\n$r@*kC: )g(Cζm#_7{[P=R]yڵ\*Nj_\\``,(%v_7 BDVvE?03?t7z]#L5: )IDҍ]: .&TI'ɔ&JTctgr6\ig%WḓY8H;?+&zu~w -z۽[1CgyM"`ORHCOKx )ZeV҇Sk3gX8{7S|pMI&#X,DIP;HҌH&B2yX0vB\8˱JoEUb^)MVɺQXjrӡ ^(WaBPQ2Zs:* ZD@ GN{G<6HgEz|-}Y*-2Ehpx哭8oC|9$\?v; v%Լ"U{OB)kUk_J IDvzQ_(áN4bo2l8 qd}d1W㵡{ F v͌?uF=> q&6bGAҔh?rPތϪYC x32ґ#$=^[k:t˳5||1nΈDuy&RDu~ \ʭ>}y)tqdJfy9!sDߧ@9鶓_9ã~4bȋ0nt58g&ahX ='#]ޚQ|m k=Ρ5 Q8eųwf^a|Q߾[oJpdp1溩枥b]K`ֶxKQ뛱jm3uɋI'|)>86/][r\kuƪ[:p{Zni}rԜr:c9,ſ??)/ИyvA___ٷOrg ݁T: ~{ ?n'm!೦vBijoѴvIӸ˫Mڵ5I8|6H7n><^3>.6oh^dWwN[? GRVHNx2 .qiD[ **9 Ɋi $fŀCǓLujH΂&RU>DECK1 BزNt*hgcAIF4wpIF=Ct;]& R{U ńR8uМ)K *}tBH*!49 duHfy BTՕu>PBSZdUE*kgvj} %;rM\kJaW9ypsn^}Mwo3'wN˔xm,u#GNTe-Ngpҕ4%12\\TNZ0hRQ|^fx]_7g,WY:\o~}wN126ww}-n|ⅾ[p\faâFg5>iRpvv|h`Qo뛫 #j`Wof[ jހ4kkY7x888ZJWTcdnq}aP(k%W}㪝\vj=WT [ {\=ne`qEW,W\pj:X%`ĕ X7z ju\Jt=WZ#płW,\pj}qE*QW+W$ W,W\pj:X%:D\!H XpF U"g[WRW+ZiH^:XdoY W+uog7M*'v)l !ݨYYc^er_Pۋz fwA4` Bq}uyV~v]z%ENZHfo-Ƴf0}?|)7 _(]~~WŲ,jԬJEtNL^A-e\IF>ecO~Ke'GL#GM)TUe%-r.蓨ZTk6C`k]5ψ]E0}o!\h7z]6#FtX P$;gCi?ovm6CϠJrHn(E\F+MnA 2 P?zvhE8 IFH?SCkï*o>3og4dy` xq4(\>?Y/Sīu̓u[@cr"`D6f9˵"պGXh;ƚpE'rU6b+V'$T^g+Y*d-浕B JlU B!ʙ`+MYZ\8GŬS %@}zon5NWWQVa- X_Bk])| 慺 nճFvҳ^: $. _XVm&+j5vr$XP]溝ǜ2 BLb❑Z;J;V:/2 OkWb՚Uqu8rRW,f+V;f,޺:D\ykQpłm>y,׉\pjUӨJ#tWvîCa=*w\lqNKSEB+k֕m+]/`6jbJ:@\)`sH0* Xu\J+{\ s6+g; X$n_EBکԮ e+ Bf++E.bJwWR! =h lbWۮT=W\j ע4ׅ_ cM6Z#E0#E?3C!Vt]⺽5Zedc`O}ymVu>*6rHW,dpfuW.A .'\qB>mpj+ViW+a;v,d+vr͞+S R黖,ZzhK)pE%lpr]Z;+VJI;F0X.d+Vv*IpE?ū]b6 }׮TZqJn'lpErU":@\BsrI0|pru6b+VP!(=f+l'A."RڮU>~2Z,7ԮnQUJ{g9 sde+ S;uWDĕgfɲ|b,dU*WW~îC k[ -*vjUJ1gзqЮ bFbd+vjUq*W=W @ X0lp*k+VJq"'C> E Xmq*qu]Hdh'b6b^+RkvWRW+D6'gy W>5XWv_ZZ4]ۍՓ peW,77Dqը=WF\I ע4׵kQ'f:\=)z 90R0$Q\ehl e>Xֹ;awp]6l'+e6.u\J]Cĕ7:/{W$Wl+V+]q*U~uCaA;~r~%h߬UK[Zlٱ=XlF"^a6b{ޓZ]4JYiAd+W,"^!J[`2 3rs5bW`W, \XWbz\ ЉE/׽C\=oRRm+Vdĕ1ʺgjzH_!>  sX`[{OXJIjvec_4);T5L_fFxnDv9xnˡ<ԂՂ7t^(3<㠏o8-Sr ^ E?\1쩥O PuDL>vOw(J?׾_buS& %mg[fn9ҤHW)I2[ZV>mn6sE+(Lzt%BB[+n܌RbDL=rĈOmX3./=`Oϵj?N<`ڃhSP^Е}gV (̤HWlpL+VJъ:]%դBWpzÅ=9])\[+E^])JIW/ ֦ ?e?mm0NW2NztL`ĸ+VJъ=tJb7%UۍP~e{ l怖}7 AZQv VDkY8Zhw$rvKph 6s hs%/Pc$bmg YpVr+8@ ƚtJ(X+n܌Rr%Mt+ȡ'Ą}`VW|l?ݼ%{Еtԡ' uCtfJڴR:])C{&狼BWr!FЕ!9tRNzt66DWL7CW 7S *ZLztt;\9ReEeJ^ ]yf!])\BW8t(Ct*p4n4IǭЕ_jJkf1"*Xx}dטE9d0ḾZM{ς>tl1- 0 7_gg`~Ae8{wPO?!{_o]Aǭꢾ?ɝc_՛l}8-ҳuiOSHܲf ) 2JROwu_>^Bcć_}v~rۛ|orf`..޵Op]vd4JnMٶ8l<;b%` ד>?Ŗ\ΦI.\Mnzc.ƅ܈\;JKM2|P[sF0VHZEg) y[ECEN2fM;"&J"a'thF8Gg Ȩ&/Fk5TRJH5nvp*֭̾H6!b>Q ^5/V3-D}&t6hI%31K,1 LmtAڳdhXhJ[Gvi>{;&G˂B i%6y<bAn1x{4 %: `0O4_ h}Yq$oAuٕ,dF:cm̭aXB-K`1wuU6PŅu ;&cݢ?K+5:x{ 2UȷndF)Fb5h[5c-r .lk])Jk'&_ol#\-"+wOθ dJ!CEБ"p#Y, |r%n(i]-P o$X^qvL'*zҠ =a@B!7CAJȸCLAAO@&X,D&T/D*Qn|R3Sn-F$"XY=l n1IzUús5vXhC%:.bFhzU\7 d&AB-Vz 訽]Ke*xc(@Hq`Q [g oA"=dH_(Hk8uiwU(H1H]Qv J*HOQ3TF$$C=DPl! eA ni؊M=FPQ>6ic hAB0~Cw 4#b6 914VUhJ;L'"B6 v3z_q~ZN bcGϣ.0mFf3x:p2@БMW%sbq%BHAhiU)UcXuLCɓ]@N`}DV .Bp(z+R|$Q*Lym,C٥b]t EѺ<=K`sD!YeuU;Bdm;~XTuqU% 9աk]wtg#xaQMZ^ v"}[u//IW.!OnTJ#W`>]b="eDE6GȋU LPw}E)i`\GQxAPA>a-%|sFE ljdž@HDG7HT}@.u*ڒU5H#VF/(Σ/p&J"(J@]c "tV% jt*C@ ;ʛoj,=Mʰ ǂpN0H!(vDƮzXF>+h?tt՞Ewv4DT-Pլ›Gi*uA^o5{EPED`!-!ZvHfX@3/i4K`Fo ֛> 6b8GR\U-&Q`TKEv,&fjR*Pq*L](xk5c4&VkFs-F(R0`j?6nҍ|7.>-7[ovdPc;8>K󖳏(NO_/An*.w|ߎ{5vEnϯp8z{G|w#)n6]vv纣MN>=qB8>kǓ1o=NWd?:W />\#+}oS<?Fo_nmد~gh㮭ﮏv{R_zA&[O-:opUڧv| W6HdM@@IӸ"@Bs 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@/ $ߖ@7pZN }K+8ē8N >1c 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@ 4@=@HWnɳ;噝@My@{M<h?d {8x:~tM'ZN #I+p%.f !vK_#L U ɑDeXm83b=ru sLGLLLLLLLLLLLLLLLLLLLLg .1@`eq B)mf=G&ǤLLLLLLLLLLLLLLLLLLLLLLLZ)5;eUz}j'7 osλb4.5;%@ZnR Kρte$ W(]+}+\K!^!u2BiwB)%p KX+k;]}+g9•*!:W((+㷟#\I(R0ϰض+BqW(W({=GPF5`I5콙E?~WdoGq1GQJGP 8pl>pbR!(Q̧a\ZwlQk ]mz6{? huP7{;LSbvͫFAP-ݰaM76jIE+'Z"ÜLwy8͵ ݙkQ\ՙW֊}kAJMkg9bz[N7jMo-5\ wt{J'Y0V@Bd_>὿tD)m^ {p5@TOyʚ|܇i `KE^2iV><f^ /)LFh[s *e2&$?>ꢈRin~n>SVz*$"QQ"[xmQ24]0Q-v]p ;wYy8`WR c9z }HxnCYk7m?-=қj?xoa5ٲ=AA֧,q*e /y]0P?MAkW|sHBU ]U'XY+Kխ\T膃RVpu 7^æ t=XUs%}*wu½; {=]B\풝<٦_f}U%`jw^뀉0p3B5+2J+K% ),W.&Z@^Ť^V4Fs0U<&IЩ$Faڨ5gR əxnT{*Eԅ+´qCmxJ@DҴurX39XsiV3H.[`6X_hV%XK[ ޡU/٪4X=ťg>mpcx|1\slzR_}3hH.e_ory|o)f_4^ưm Ҁd#M˜ @-e@ l6Q"OF8ex qG,*9ux?xpy]fo2 ެz4nmֹۇM?QP^%yB=˹;o]89qW78hYW2&)$T:# t*-{1]op=De7R`ȅQ_Jt8&'/${JoK[3qbܙ80.G}<.!?vŅlx7ХC,b~ViyIq%yZ `JrNKǹa>Ru%;`,[[K(&Yq ffLU۱L-xt<͚Yg3I^Pˎ2G#JWMե5 "-ob=tq}ۭtN>ָrKLQE79|R2N; f;׈ΤajӰQJr^3kԔ*Z_,W)eޞ)L"5"0/{=Ҳe/^ `UOzQD5{. Nai޳:ȹ3K|?r8KJiI3˂J[DH \Ujo\4Zyci>l`iÁU q2_yovP- [kE?i3.k8WoPH'ȃ¨";bdIMRaw2F2QuQ LpA LsCy,Җp!h KP$P猲d73lFΖ4ԣxvSO͵U%nIL"->O\gbkpRy- %0R* (#9u.j4%aFxv<,f =y41^9A&I:l@fx3 Tz+bW)#nuw6#G>) CteJiXYZ S 7p:sJ&  eb؅mB)ud7 !]8pSzp4ah&Rr,*TE%c$m=~&".M*M$}DEJe\x7n%~qJ0/TFMpnolZ^]&-e]ZR5fr]JWBI Wt>?v˭0O8/Nf_~ǩ \}<\9v2Y}d:_:O?>ɤ߻ӸA=MG~?Iխ/O9v%j}(0v;7tPvbnt U=z0d<Nn8?{;]dᇠ|ocLF*ڋJUN{yOw>o+a3oםҵKɅxsp8-śU .S/?ߚˀ0~)P׹ +>}hѿSKR_bV[QGlQ~$TNUkRoZy0}.{{~-SJ9T.&?̍buI'7Y|4eia`TPX.f7W γx2nX$F~V}a[L-U;SfƇAEm; ~-XV@aՉŃ 25-a2 ~9_~s]*z2KR[nvx<$,h&yysH(x|f2y213G 'i7HԄҷF[\t72Q#>["7tPSv^Rw3ަF)fYUCy`UŽugK'jE^yjL{O:F:y #8, oIR̕¸Ģ BÅt#s3ng藗Wa;x ָ, eHވ5e21Q%xrJF#!+'`J4(ebְAo'b';ӍKחie8G}SʪCpﱪ7\ p׆Iy٬e*w~^ZZO_XOe?SZ/hAC $&Z1v%sng??7Xgn{Tn0%ރdNT՘^ӲzQNEBaw .-+4G"PT{gRn99DOSeD.4t 3T9u*&&-8LGmi (ŎӝsG M&._ʃ֢=J2L) UD}l[^[ tVUe,`I*<[2MRɂTs͜`0`bhp+*ya"V2mWp4VHO4,"90D Bp:fNӨ9{JXS)ft0w^*FC*D)r * $'{em\S] )R++8J`(e6&d}$b!֐Ti(ʭ0&Rkc Җ%DT?=K?̻/٫4Fͣ&.yY-7,7ĕC_շf[.f 1^L|z:Nb^}bv?alε#M>3d<X%#j pcŽ&T+)Gn6K~ٻ6#WɇX]ϻ], dbV"=俧z,(e1`K kzꪧ)x 3 Is'9c<20AȬ:`zMnueq , Anެ׾$$\gS7Vw ?{v8;F,.@4uMG֭۠m ;qBgW0=V~M/}I^SvA1![7x] HWߗ_8}u\dq@+HF6$r?=d꿼q㵦VSKZ|3[aüWB:6IfŇxKfj(-WV+tV%iVV&5{`̆=*{:J2%IFlִ]v6!mqv#0i;ehۅJ <Ĝ2 g<6*sאsR # <_HOzZ*;$lIȴsGjR'D_5*`5"8EYb| }ث֦CO"2 W\!.#wd&%-D ӚĨ`=nX$X=cSNցI3&IXw)\Ѧ, ȹ )= <JO<3npLDSNpR!9Ճ<̰Yg 4=)v94ĚĄ* \B3cvi0["M +zrF3 bpܕОsZDW/% -8Wlʋa!( 5R(rS[PE7D=%>t羒iCzL;dz\;B}TH 㞼yo1$\)*&ꮙ J #7__O,!vqZpԕM'8-b*KeɄ-@x:]6-r7MW߽M~LO B|l@n{tauA[u%*KX5}ۺfm0-nlHHːQ .u?:+O]_{Y{~QG̠lc$g"kBtt2N`@ m ٷ=^{-[`YTc"dZq& w%eHj!r 8kѥ651!Ep㠝='s =?cFΞ!*O ?zh.kgЎy_^BX4^3*QNqC!Nﺔw < $$VH)nNZbeAF킊mfFgo7/'׳6Y.*[Bݵ'm eCZ7_J^RW4NR.l|+Yuy礍̳' 21OZ#8S!0=TF)CcrtBds|&%U[3V#vXTӅ8P]u! {7,. 2R=1c~9nhx5} |DR:2heUp*-C$d8e` sM%yO)̊8{> F$)ڔ F%+ LsT`rHTȹ]c$}bEkW`urAkv[6ڌ&pFFL FC̰(P3mZ (:DHn18yȄ "/:k(KECK9r"S։Erz_YV#v}:E1E#VC5"hVl)9g ɈDHp5xfek2"<˜!zx)2Y%r3!-O4hx{ '-!NF'H:^:)YKՋ^^"yYY~kL,1mX;/l%O25?$7ŇЋ}чոP}(+Cy>|vK;׮Emk#*m/]SF&=7d?>R=t*9::4 eh< [g̻!Ls<LstϘKぴOslEPgom9BD8p|Pm,L#fA$GTBHm4ǻ9`g&2c] X^a^'H+%Zg2w$_nF|OKֆ֟3FvPoF/ZPIZвsRy\7o(F^ [H6IvtE^/WFt/V7NlSQQ7啟 :;dJ-LG/aE^̺~mJIaF9L;˳UϟW`y|,xBB[օ@+-1\Fh&ܞ]ŗ-r[Ml]p%$z댗hO(39iɇrfLA&ctQ tVeq 0 0+I$tXG$jЩ_;BԦ,-u))2O1dBV"e EXYȔ7=@BY1`$|qޗCxo-^  0=Ĉ'F8ӚWwd~Ծ=Ej"T@{2Fełhf.Mӻ7ۣ7ۋ$9Q$(&ǥ<䠓^1G 1 D%'>J)udMV(`NX,BWXc̢vIZ܎IC74tN{iqaޔ8>`O 5N$ؓغk) #ALa!& {**.چ!؆ %Lи-Qxم 'p$H!Θ3 1"#70XY2a>i*g^#FUcLkLơ]@.C۠UPX!W#gFGC* fd_bW֣Ե[Z{ybo )K|O%B,l] wMxyYؤsJ H` [MrB5AΎtf&Qb!Rqr=& q>$K``Gbiar+ϲVHa ܩR22CkITkђm(YYΪrW f-NbTܗ g!@ΥVsRN`X&q(4PAjz=Oˀ^*iFg), IV-Y#9 i]!fL#X)A8Oz\e(+R;ȁ8IKa-XTd%;aPUG2ė#}̴Z y0HOxzcER+|u{v?IP~E]ox-}XA:ຓ\l2Q <nPGPPa͵zuanm"!t5( ,ےM.uBΥtd}zlz^ &P *`4\v؅UѢ V9ҘSm3Pl]kUC*^(ІhoucThk(оiMYjn4ovOR9}G\St] z<}soKJ>vo]st4Qt;x<^xoFoK+LP`*+S\%DDX!?vHUt'nV `B2*T @kv*uj[FiqzjB=gnW7#rvc]~sk(߮( E0>>k>k)^]\ynލtmlUEk,(EUhP#v`nj{TЖegDua3W ^*늲kѧSmK8gY-9w(wp߽?Ko.MپX?7FTVա CA\S*Ĥ4J3գg%&OUJƎޮdz7udq{:,Si$ƘwY~/O:5}<;McIR?86荻O>}N 7`Ib?sb8 Еwm*M.ZS-+ʮjPu1/DIg")`QڜL~J|]L}P]-nx-7#eX\]NHuUiM"zE)m 6'6se/ͱԮˋ3>xx@$>†Hoq[  v1]ϛЫvEo>Ï]?ZtxӴמ}:Ym7cj˳_Ӟg4k|<7@{s?z3̛F}2(4TTg H]8nݼTlVA/m,^7ob*۵lVcW`bQ*ReMU|WNbjMgCochkuSZ\Pb*2`sA0*,`+^ wZ)bZRSu5]V?Ab`1b\)bZSS9 54wщ]1K^WLb ue ]1btŸrڙևuŔ5f+)VtFΝA1SLׂuŔ.jr_iz媩.>5W)MٚcQŪqv3﷋~篻=5(M㺾Ug}۳Km.=@ӣhSM;u-oT|{&oL.nc[.^$Kþ) .]ZX*Xz׿aquqhCxNBKX@nlzAPQNJDZ1m1S1lb` !OW q+ )ͺ"Ej`銀1z14:ѕq)wJ%۩kẉu56LT=2V7@W.jߡtE]1.z2Zg銀+5JlbJĬ y]1p1"ܠ]1-uŔ6d]PWJ(HW q܇tŴSQj9ʢqVR2H^91bܣ+GWLl uutE^]1.h)bZT;Szu5G]9cu&>[hSf7 tPnm}=$eg=1゘oA(]ۜcv0Ab(pQfZ:|nVH_WQNW :1uL}bԞuuCOc,-] Sjn8DDQeHBW~|վC銀9W;)bZuO˺tA7]-h` u6z@AbbtEA))bZmSSdpBy芀]1piO]WL 6j#i-IMiu2謫ꊫZU~@K۹[-C^]tupG2OFMZ`ZotדRn%66[;Qvl1JX|3)>"Z0gDL yy8T2<KW]e% 9*zZ8D1b(VhV)MЇSsԏ%;aDmF =A< UȺwuP ]1btŸI>wzR9JRt('bRtEVcbJcf+V'HW+A!u]1u^VtNN2ȸAva9l%%l!b+u]1uQ11FY1̈́j6j{1k[ygбi cS;FagUk :,?c[blwDPPvFL&1UR3:(!0(ٙwC+E1sL\"J0jJ+ tEz'Z$+|UqR t  &0\=a8#Qژ]Ŭ}^+銀-]1.]1SjMu5G]rtQNtE]1-uŔVg]PWW HW Fi-+M`d]GWѡ+]W^va9r銁]+bڙ6ue9`ͺQMfꊀ3ARtŴ6)]Ⱥ^ NIN{DrRC^2x5WO]|h_ݍð1Y mi5jg+ FNJ m( mgcTUk\(94?Aٕ4﷓oo'+F[n,^]ouݒ>]_[n}X~{]9w/eXop,ћÑO!u(F},hu[k]Sg_|8v=JJ}]P3nA U jܳ*,X+'g\/&eq~!3 oK ء[k1O)c$"( '_wsRtŴ!+YWujGu=yt5 8N]6MNtc Ki|6j| tE]1]1)˺ NF)bZ)ɺ (%HW DW;bhјuŔu5G]! 09㢖+)}ueEoʢWbtŸr殘6+sWԕ;1bZԩ)mf+g[Jk[g[ЯN .3쀴R^[oJI|ma'뗉Y[oy]1eȺ(FW땘B(H]WL[|7JCON_5ԅPp=ia[u(11]YWv&h-HW QwꮘaSu u5C]ZA2212ΪaƦ+u5C]NKJ81"\]1-uŔ6jBBtEDW;uW0ک@5dp ]11b\iCsWL!j⮘^S|k}m~l5n1k 9x=K{m(e,Zj{C)1sqctL]1fcx8Abp#13YW3U`btŸV6$_,ɔ1f]PW18-*$`r1sic"JPYWߏ̎CoN]sqpqjv0J]}^G+sn.h)bZgRSu5C]! ]p4^DWLk!u]1ejŒYWGD銁m+uQ6uŔ1'ssNcqJӺ+92h%]1uWp3wŴRS˺©/!1hÂV1;v@M#h!)(^f!kC^C̺-F*2Ȧ xt'@.t\haé6^YJPvFhqdgL1){B擝lЂtEAע]1K^WL6gut^6əf(fhNVS\ vz8Ul0gzJ_!Ŵ 6fyHObџkadQq{(k}Ml$OwW:uooZ>CWC6ǡ GЕzғ 6B= pp,t5Кpt5Pڗ ]sO-~EvD8f=ӳ1b0ROEZwo_۷owA1~:<ܟ{]bh=6G<lyP33u 2C6 _^٪ugvzvLg˫?;co?hVl-gTN&>]?oΚ>s IJsɺOa%].!SĽ؎f~F(ެl޿|GW܉φ|XX*Dr 28 E ^<|}V{ qˑPONƥۍ@ڲ_g~=wI[wcK(=d7rdL.GvFF8*vBpW?>yڿX i #oW(.[}K  \ \NNDEH-쁛1YqvRR0٨S1X,ד>?)3fTR1FSO9W]Jܽ.~ ֎d>?O ܍+o]UXs)V,1&3Yn`̉1!z-@Jvs:*VZi FJM6;D &FÏ zj*zԬ*"n 5J66Ԕ1"'Zj{@,:d2v!1-.bnL.蘸pQ-^rJx+ ф`ݘĐzY+EhvoE2 IWDISKQ ޅKB#I{'cv(,V!(>рW@ghF>]g >BXȃcFgdɺ3>h}jŇ$YUȫ;;]SI{k <JIITڪTw$#W@o֑VS sad;f7E"b'QjShіre_'̵| 46G;'XrxQkU {KFC66WѓvzUI'rdfŔbS9drM zXH!sS2`w"R ]0,)%D:*D{jP]j0IQ/3~ E U4S`eL9'XEA)KX^5TTlPtԠ-;< us%ps lZ(vGQ7fa=g5ȡ  "ջV8V1*SO)x_b ~N`K(J J+.@:0 NsG29G!:(57EP? TBM'Idd0*zbrze ++eq2.rN[@h iNBB("2m ؙ\mѪPFtѷ=t,rMLc[oT^5>o9DqV29{H\!Ze1-Op(v5O -tF\8hZ R|" LԴʨ >x!kn 3*]k4/= KH`e Y%cԭ)cQx $nGep6/EWgUS U_"bygUcۆnZaBTD4X! wO.N! =`Ygd8 ]{ #FKzuI>h@C^ @oB|}z З9t,'pt@R@"q{|tE`ZӬUf}JhB;ڱ",P kR([|PS`y&˽eg:|l- qR:fW`Qm4BgVI5NfWbd@BUq@ơ#bguEת/Eì2,L]%1#ޙUW(!h6DpPk>Zl?4ΐ:64MR]#Mtf% MݢrɎKoU5^"`!-#6HªL$` f}AO3Ш\L CjGih\ƹy+.W+X<-7iNWmϹnt&|5`0u1vMr63IҢG:4'GB{'UEŪcԶaZSZk t/ y"<4f&fcdgrr0հq|ͬ2#c0wCRDluHck* FDM7&%\I"MA;hF L )cYQ-HOO(hRC}E7oͰl+TOXQW4#ROnFz7(XV1^8p ֣-*1"@΍-ь;$zށ U ::ʺT:FՂ19MC)ܬxZ+=֬U 6Rg(HqͤyJ!&ci@Zq!; tmr4=B])4.j-klpgYqE OU>Κ`*a2¦f@ 2q]HmfhJ7a#d.iScLEi8 8؆kW$kЭNA<` \t*N6fTS.\;oN8 "ˡꘕq祂БdpIԅ]0q$d Jk|A;oBb!h92_H0H5w[Fv/n^ۭ} zu)K.>.o}K7-wo}@s͛ fɦ%KlLq~{bX{Ϸbn*dۤ]KV!x<|}jXܼWWkӻk_:ͯ{al+\] u5i}vu~nݻޜ,~*}jussGcݲC[/~Ngff@פF??Ѵ6*='Zh@@k ^9:XG $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@zN F<0p8Ɔ$uN $ grd8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@ l4ꘜ@87ģq9?x'Pj%Ncۚ8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@  tLN T8 \ 4(8(RT,N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'q}^?K6ŋ7cf{}~_`ܥzw(W@"pL%d6c\\cƸVq (q9>Zz-Հͱ@g zt^-]Հ@jFҕUd>"= ] h6C+t*]=Cr5`Chj壹@@|t(t^u7j싫t^M ^s^W E_s*74|u>-Wo+Ziy񕶴cpr}yq֟~9ۅ5E]}8/#pOxq3k9F|6͏ËX릗lT]FoyYBɶ8l_[j;@eqv˱}xKxۉŲoF/5/=Dſ*Y_..Wxؽ.I//H!NȽ O4[;K[wScQoOfs*O59~=6?'%bYKpݬ 4MLyOެљ Vl:7],SnjIj yi`1v5ۙUj5`oҡiUתԼjآCB؎:;n_hݓ-ۼܠ,vCmu]-O/u˪Zq\f㙶]ѣU[SӋvsӊF:cN3]Nqooxs(@4>y>1c.ӼvA'&n :^15U`LkGJ?pY\]_ xmgO|{trG^N9]O9|pΫ+y9M4w q=Wq==je(oM8KƝvtT)oZ- Ur{c7wF8gݮ4?+y9~[8v:*FjW'k#"gEM9x_U3T R핼sH*jo:e.i)o7}_uv\*p8kO|<]A˫sLBە|*{@<ܮnsj 'ަ]ED&81Q7Z'f&8'uGDMy6Ĩ;Ӵn..g/CK hy2hr#/Hn}1ZcHv:;C{MtW_=v>ԉ+_| ݴ@7'+y995nPQs ZP7I_|[ sgƳcͳ4ux߸%n=:~6D!?>6Ipr_5-7)-5MsĦ *w0ؠnkk!x:sN)tCآA9Cf4#`xʏ5c\+fǴs1%tCaNL_U=]2^8ڀ!}# 9XSn |λJsNS;ϟ=GQXhny&mj"f" ی_Ro 5Bܢ83ԛl\SQp7\Jk'4ܶ /C^Ǡ銀c+p9FlbJ+u]0k)bZp);d+B+ֽyIubLS uw]1e†U]}?KWWIN:?f~,$*]=u΢ ]10(1bܾirQzEWԕN=imu.z+EWLOptec6G9b\D)"ZS,nuEڣ]1. :w]1EWԕG#+b GW Qm"J+οALbagkk K,6\%3W, X%z単11e,![%HW ^7Y xA("VOUtElq,`ZrSZWtltV=)fApYp!4Zi_m9$ ZZ9e ]1pTbtŸhu_(* {AbqQLtE!]1.ue'ݑ[]1pbtŸ芓ɪLjjIA}q+uŔQ] PW^{c ]1ppC+6w]1%*n>  FWk@֩uR uL<l-iul,z~?Fz}F EW1AWU銁#"H]^WLituctE`]1RtŴ.+Pt5@]kom5W*7,t^V 9npN`֛D'M-jMe1{&Ȉ^O#5+BT ul[=< S%d+L]=9AA2/GWw>$ZBb6,z]Y ]FU1i1(MnDWN++!*&6.] GW^[A.$`q=Jӂ]WL}utPLdH6*z1A:w]1%Bu t ןNbDΞ 7*yOR3u(H?{),c[ycb Шc0c0Lt1cDcl[U) INuŔbJ(ڇcy"p1b(FWLꢫA*^J6o ENƛ;[3LҔdJ){5_־im(::]Jt^X紈Wb8zpO[k-K(7x4:c`3opRFgL}348:z銁AμMƕ0i1WuD] PWCDI3 +A\bUGU^sDẐ4ʐ`P'J]=u0!`0D1b܈RtE w]1efFW&x ]1p4hr);勮+С$]Y~btŸr+ +jr`PEAb$z/EWL )cyv5H]!~NNj\Stztq8 n->Y_|_Ї/Ӌ_O{ ?>B31T qVNZMj]+U\O{t6傘cÌQ ,>^ Uۙ`Ol瘋4/)L,M-bvFxy|L `Lۄ-z~e}Y\]ƿ_ x-;GPw攏NHԞb 9T/7WWsT5hUԨP pV̀q2cڐdˀova.! _R LqxqG) ;x صR KzF^ЄH5 %`ڞR%Rc 1T ]4Èq+>w]ePjh]]ZY)bZ)!] PWD=p&h,7a\/3ӆL e0|te6zm{$`0 Si4ܢ++StЪd5HLVWiA+- KՓn& ]pr+ EWLmbJEWԕSQIkmq+ҫ+fK(V$=#` 3A)fZr0S5QA"u^7X)bZ~KF颫!*y>K* vRPf!+&ɤh6\d%OIeɓLIlUY&&`Rͺ/} LچS}"g}ta$h̘ \b^A岟1A7et6YԈ^ qWLkMbJW&x QWhL HW r}.(1bZsS(z6Vݣ`O0]ƞ_եbO(mg,teteZD ]1pp+k-uenkDWjm ]1pD1b\D)2<++4j.ite]+U_W] RWLBFw իh>;j\OgbJs|yA-HJH* ?̛P~9EtvLLJʻrdz~E[g9L.ӶQl[_|q2ZL~/VK3~v>N3z9_n og 9OgxYMŦ>|^6nt:kw{n8)5i`/5))!Oo''ӫgt~1!՝zxc@I6rUc骑zЪ.ϦԺ/NjzyXVxu\KT*;\U'f=p*ϫkOӋ*Bs UW,U쨪ʭyUF;3{ v_Z]vh%2I7tӌwjflϳ%i6-SA="t֫f Yc:?b acU鬫~**NcН5r߼o4@'Rgu5E-<ѧz^k޵6e`W)5gw3g_ kķN;*q,\vj'"IQq_Vg}cr_Dİ@˥H\"q]u!UX< F("7lP F߲_X]`+(S`-h(Rj-/hgkSH Bg8i'eʓ\ȁj%ӞZd ? kCxz|W.=8{I~aWϘNfA2 Eg.O<uft5ӫr{ S٤jsdE8zmVH\/K簨DCz{ LR(AGH8Bb&Oj{y@]'Xoq,v-C6ˏ6l_ unhw>GxNhŴFIQ("Z]iJR"\*f,m(ʶ %k-ܾ3?NזlqEOEafuy1ty:?#ߐ PF& q,!)4؄:<(8!8NEl,4ϓ4J:ɄƂҺ:.Kiu~n᎖nRɏ$**>ˍņWmx/~"]|0|ǯ|ҪR"|_>/ Xgc~1F`Yn(?UK#+7&\ c7.w *cOwre;Ţ8,-ږU"/PAP xu9Bx&PDB-C35ZKD&J֖~yƻv-!kj|yMKP[֚~2;QrkpJIy-kX^>+d4(-)*gKLJktIRcRuI:;p*k! 'eX݉(MA\~4GpX>aiE-?|bQULj /o_RKL%ϴ; ) 䌝,-dL()RW2ꈵ*%xN&K1VfOfۯŁSRJεѝD)K[sVucV{3R<9'BnJyjKLf} *%Og<[ Or_\y3KHuƾl1xw+no`WWdi1b)d-sks Ie/62uԴ+;9߀G-m-:[SKej+9M2]_X΁\<Yd4\'̼йr#F42!]Tk*Wd@,ŗ}D sz?_+HbȘ̘c?"'(#0TKl5Y˗mUU6yphs;2g\ƊDEq<s )XcH!KshFri@he9_N2Kc\rH;E?Zv;nckniu>xRN 89Izuė3 Yd ٨joWO×`RݹuYb oc/xzu3;#nP*17H~h@`TE,ؠXa(Õ{: e?M&*9*~am<Wv'?DE@c_#_zƢE폐2Α Bu8oI<8rI Q8-C}&W/BDZ2=M[+ā;I9@ L0pKӢKB,if'0!iAUϕ.RD+GjWzYu[H!%)ϩp"WrՐ0qB5DŽ:HS<hD(6;j0,AO]qZjR hB6>Y N\TKR|.|ŋ|;X$0~pdںgݬA] T0L$nn}*+g^j/ ᜱ N ;$.O7܍!ZR& TR;6D*whW{J=sH ixObb4y(|wCT8"4~6]qHJ4ЍwDNn54R;@7upi('f@B) Zxv;uC.}]nm^ky[<.|Q P{5I9o_/ڄ[sZ9wX]\&W aVk0|ϧ7NƉK ́1!8MG}=Yܗ  $QMm%VČ*q^ScĔ2#K82eeIYr͉%`u[6Ɨ{惦mϬ  bg&N 7%z$&}EIvӛ&?^g?XD'*FAyϒC Rӆثu:^$ᬗ VY=#s Bk4$N~ʟ {fQ]t4쏢t@ AзM4g ^4.}/m2P!7rG,XN=~ܲ.F[JsPNI_/>&Bɐ"۸M-Ys8څPY˵zEA Vu (M64 ڔU“K&NL0Mj"€6v?;(T:_ `!vh7UŁ)7%㊷77T0tF zsY%IM`0.uN,y,8g6?~yIU<ٲj߷Gm_o{GUռ=DB0qjDp@fa "?9.7/* 6/)p$95Q[WzRd1H1eH"Qt,|6Flް&Z01LE/@dH"ou;fvʴsѥ-A3r+biHĈ' #-3iTD2nogXa'0 7A,wuo{$9d4E8cE'z sbIh%RV X6/ V tEZz+\6N 3ٙ֋B{>سrlU1N ( Pe(^Qd)=} dTYjמwhV)jz6aB?ܽՐYnClWRA Af9jbj]^u V6ډ1;=ʦ5*k/l>cz{q.VP~soX;Q7S:My5%mOwN 9hWEa0+N&Fh{w.*AlJ\(2GwP}T1VBʤJg⭌ 3Fo-Y,1(Ląj#i2%u+c [3RnUem!Tj'(yNң:=4ˣmUeT#Rf4F)M3&fZaoXClk BSFqdҿ>rgn]!IarAE/nƩ3k`B4J+/Z:w~ԇҼLc[sK1MjG#!7J6J>O<΅%&Dpr80yM ͊A-5!M=4ps (ZEڋ;?A }$1~hFe: lYHS- c8-eqDM&\УrPǘ.~܅Aq<:f$(EYB*0Κ!6Sz!+O*: rrrz;ߩ1kL>!S!V,}EaEpvڬC.c Z@1x8pӷe==#2qlB!uB+zl3G`@1hy:BÃcI+F3P }MPz~/ϒnY52-kF@\ So\Ԅ Խrk_ R6u"YFI5 Ér`B̛ٷ0.7Pj{elIF`X]vN{zt3{Q>(oO!29T3M1DZTW`}0(n߭ʈlA G2ͺ@%Tp"][HAD-Ɓ@AԻY6*ֵ~W0<T7^]SoQʟxꭾ-Y3oPhQfY׈< kw]64M6B=d{mԬP2۸ `8(T!uBk4$UMxnM+[ܠӡU| 6=ȩz˳Zx0,' > 0Jfk*gu2t6o;ޝD%ewq D70Td k7YeiOYļ5%7ϻTvHfX%y<ƴ2H ,.߽ $4ձPno8 m7!T`N0ķҵWB 3D,Wʼn2X8C&pl0.Ѕ>A[^.c@ ښS'b":[/׌x7#kmF_E0p{lh0`fEAwm˖dԶgoQ%A͖14YbկŪ(b87mYJIF]5_|9!afL KsWm9*LԆV8-s!G˼э|z^OM$?BϟT'+*.j^ \[[ Q\XitEe8m압#_CT0Z|׿21wT43\+amaVyf7ȃҦB+PHU#,4sus~qX1 U4"e.Mc}>é9&-\ºwh BGbҚ(]a|12dzyT Ur+Rm1jM!)"lWw jn J ,kĤܓDT[kd\7T+$ Zmj\;m+ii.770w6V9? h k".FiFUY& G ԅE)JLLNļpL3X!lY?T1*) ֶCn-}M@$W(^GZ3GvAbJIGKtˊ|V Q]bP(<@hV#wW>g%~e*UQ1(psRe^ID-ݙY$'9OzI:>J^``U>Όi^oajuV%[|\2/$*@SI6o\VM ԣX`)2 QF%vEP ɍz#0Wp;\|J&inPH)j>7Iͼ͜ƣzJ:%R5 rD)M,/+=zO뷁 {q\k6~$P`7] +% 7Wor.4OY IlJHVr<1{U Z6Ʃ;d$yR60E@۾ U 3njBM_%.E/X6ߘƴ! spL i^w}Mjb~r*ozN=\P!akHb8 Y:kt,4IÁNFX#(8 .bkxqK\pI$i{lcjЖ>-1Yηy!Xi@ȷ錒(|h]?4Ie _;mN̮5A[4mJ -i*4gפF10W i(.X" QqQR`nعB>>2"޽Mٌ1$yK`mUsZjx;}ʯ0_Z%J ZBwO%41N'tQ@ajѼ4 u8kybfU=qSvC8CCSOΓ=8<t?1.4vzh0mR!> _"HqPARtΐ@{L&Dpwg/DS%(8 mmbz@-ٲW"RLt}X᯹G i1f]ݬ]^) 3+3t s<. ]?g9K9b͸c'fxܵft_U=PLvUg8ޏ՗,K̺Vo#\vMU ǗI+VK9\<0zꮗ]׳?]]٬s+?nu%0MU? W8bUֺUDS78<>!41!x{ &ENoF[b#)-P-C?݇q`U]M("E^]^1_GʍIq/YRyܕѩOs=&7v SMz; Kx|NzUamKF%a0OG$c !6\T|\+ɉE㲏clbԽG`es-8LR$,ŎhGҠ)Ԋ\v~8a%a4*2XQ'vΏUV=;@ Hh&0g5SݸڠOӵу.ٔsQͻG Blr6"8;O"¨r) )vوZФ-h^BiJ΂)iY>Y$R nH JW }s9C[j~=ePٷgɏ@\cYZܺɜXfȳ\aGsR2v27J3N:;-fI/jI 5%=,඼!AcqRc\I%U Ŵ_j]y27.[Vi.+:c1U:fʕ,4 #mJC_^3PUK l9L~_\/OE:;v0Ԯq2݉1ri6lt~/w^VD0_̺͗?TÇ;<0닝sq}z?.;:U1|Ow63V.Sצ{Dԧu3]d:/6pe1qH.KsolKI"  |XQooջeĜL0umҦ`\gsOmv8pOO${}I D8%5ՠZE$%U{K9NY8DpSԢ۞oۤCOQb4MH<wv`.vA{͝LBqD@ii;>UC$1شj#`q5 8kvȺA^:?ZNXd{P"j4Z]3r~ơl{_bpR5Ya]M~~0;+^]7_޽*1;ʦd[ܱMw?^hSZ "2 g8) ,BmHXn-qcS +ۭ-n+, [`QE K*_RQ0+kzX\!͕9*L6_0ŎEB~_`'}1D(ܷA+]p>2X{ǯJKrC1 ) $ei_*"97W\X-q,d*teFqVXDrYwidl.VBٵɓB os$YI3PT"N\ezVw'J;:_ოt v*ʅ:u1DXƁI#c ,2$Z7@nPiZէrԫ+dM/6  śF1PK]"K8sd6|G$`ynU!Cn׾@1{!KGm0--|Ramr|a !2nq,Fk#f4oKԝ_26GbPqR-܁/O+>5-J$t 8S% v$aOfQ+4)!*^vt Gu@/:IS r<\\02GXQqiFs5NOk)K^w٫q>lVkY?#0?}yaxr/5aٯSvMx0Djo1뷳ťm?w~סJ S,3&:O:NϥBYaJ8 ߜO]vÓ;Uӊt ;d|Wi.A!'xe՜/;t-o0|ػq ݤ?eN^Muv>i0Gtb}=s_vԹ gGB%8돟{4qT1f,W92blAQh/!?:kƅ09/t#u<2zHr:يÝ@FYkؙ@ce8V=_/xZ~C/^=8 m2{-n8 bdž1~_lјVJGkjKqYR1o @& KL0jxKh4V7B_P#;ԓ_}('FkBwEF :d0M+vrl@\B~[Zҫ}rh7qp!iU)hȭ -)syGD񜃥.A0ƑVL" ~ƨCE4VR9'CXrh|7~#XkN{mۑ>ٵ>)pd9 S0mA8dy,cӛƁҠr3r'X'(*Ŷ$ֹC?ЫO=UW.!jW9z QoRh'  3YW(ESݙiO ?6' x^WQ^F _B0EkinGEѧ@GGa{鉘>nM=ʒewMߤ$۔DJDR2:] $Hd&2xxЅ6̊bPrY˜ *V,ɍS~٥5^qJkIŏ6#}Z.G0~,#}/RkN r0HQnEkj~&/bRF\J(_5:nRN rjɜ"h XqMbh-}>~[^'4˚2UJ+w> ) 5:)mqll4${ 2v;G:)<|MQqń9/b4D1=SM"MIYqN)hR[~Rj)}Uz' %5 }FѴpc:UDEz[<'%_hZPe̔QqL/GM]LN:e}rgz%fFK50`1}W:RCPpSʎPv8T&VѣK8ԧY6(UI1j2[I%FxQfo/1fmχh׫W{7_Yg̕ ;ߍ+#LGp܈1W9B 2>g}Z<#xpDp9oTid4k S3ǭ 5sB>JCU^l+(ס+=?;ѿ#rwxލW)Lv_nt0{[.h+)s.O3SL28e¹Mpj4˷Yolɥ[U6s'!KHWS0a=c&٣wPMR8WqF2*a2 ͑&#nD:OXr-r 2m:h3#2^H!k,-$5> @/.۞Y^$yY,HzUSJQ<~-<5#G]NÔ_R VBE܃Iiu,4<ǎ2&uUIx*X?6R֡ I(sg;KcR <(T$ZG4x]w% KIgm%>O!YW> 2A0 :؀n\Wx!ra3@AFPqң-9P:N~a'#}}RG|߼{"N-mL$;I0m1"B<(ӧ6}d^i(ɮ/c2.?ۂ/c"/Xn8gι2Za$2EdZcɸJpjwfgL329֣òfuƺ/]LzyP2SE2,LpgPjWok*˞Ι^{Ac =a z xaj1}&`Q;ۖɨCIEH fGn/ҭc@W.lO[gS3 Haiq#Bc5pq'ug10|lhY52WZvM=׀p)[<$08ci$)ϩzV\9~Q5$}d5dzn@,dIr3K6*E[F!ɒo[}lJ U|D(g.o6ڵװcYDzt2wQڱ=x NG,oq4+9\4m|.ȭ5)x$ߪ-{Rwd݅4Ua9͡#8zѴHٜ1DY`% +$\7G,%+%٣( Dd8^ \3dG ;J|Z$Y%x7/E4p++Wa Mzd\a^+«}-4SEbsJ_aK3Ӵҍ9%rފ+mVW)=wfbr?\pkҬG~LQH^8cB J)Q(IDV1I3$hkwg %v4\Ҡm6T"hQYw"4+tT(cBkqmX(DAa|t뽶 :YZ,R䓂 #CC3f=4>Bz@{C_w!d2 3" J#_-+(te\eǧ O`k/&Sxf7}jbX (TUa#Jr#>|OOρx4e߿2n=Fnݳ׻lܳ燞}vף/`g? *x[Kϳ/OC?/? cpޘê?`8̿Ai0o@se/HiͶ{p ]o$*g,}X|!b˸gp|ʵcŃǣt93Zg%ODAƕD1V,lҧd׼uܘAfvVqlW`X eD ',~` 3p,t[\K0k-R"SXv.Z޵&*˄̑yY(VT w͂NKZlu;||ˊj~fb isŰ z ]+) |_QQeq (,묧CjHzW~qe*dyf.8=F_n*Wf'}!L,I) %9 MQ])ztYTPX^٧M=͔AZ.Q'm4{LM *JwX{y ݫUڪhÈT1[C) ֠2(,q m]|EZMpSng֜y t o KHX(wIw'1+a:j6Xj#Nd-b1L*w.Vp 4GaaN>OK3>ߦ< #nK_-mGusܨ9蔔z{NN)zd:)d(@3fZgI(L1|$Sh(^`5fq8_ X])*Jmgckm3Wz֏}~6-cwJ5 G%[||$S:PdX$iO;"{uNkrE= Tw,4C8SŲytR aH -3yG״֒ qQD}L9 Tx|-^}Lv:ı$v+ p۾| ܼ汗 "ߚ+yWI}ֲE #ʫsngF.1[~<QcCc]5vYְ~9ɍMF t87 `mo ި[E8WQRmT"Xʰ[{ n4V2lg3P5@UCAO0m[V<kB,Q;ﬥD0Nuݭ&$\PqfHn &{8+/4gEEB*? жC+7}_?ҶD?5"-Q?G^ab2άy08[*s<Nn$|'@))&MyNWO>ZBX|_TД9q-v B[ "b60s iHzs? dzyR>@3ms1JM0cR{C:S婹mp?ikQ1^3s*{ì{J'\3JլǨ?pmgo3f}r:DbI* σW2'1X=χE Ào t\ Ĥnի7@k0`^']koے6(WN r8ӊa("˴c˸$v< -f/jY (#I*FS)X0̏Gh; IA|L@w&#g;SlJ͆KŲ>/u_:RC{SqL&,z^du^ 2۠ ؽ0[*sg_-W}_xK>1OjFa';f;m4t|p^nJb@1T] "j#??~y0ۏ/Z,bZ;X Oρx4`c~%O6\Os7w%:1ܗm rm>lڗ|=nN8+0|ןn4Oc؈?x~`48xo)(XtF Fw5QkL{?-p]~pÀ'8;sv7sx2EA #t'BE+8#֙@54sfWEbzu*r/j'қ\;mBx ɐ~Hg+E.O0DL{@s ~\C$8s%b;q&Bk9$ѕb<ʿ]&ǨNRAVDZ!F9]ZZ,jo.|5…`\B+"P_6!̢kw\p%w|q'9)d_ ެ>C+ ͦ7z^Q<%-wwA.e!Dށ] X,p0s /ph=X}O J(6q ޴P}}趧h4{c`Fg$(d,A "Ꮫ{뾚1^Y}IW^K*mP4dOtRzG *Fd 2P&m|3ub cGWKvͅ ѥDұ|eQNV4WI^^M&$<8 d^֋]2t75S:$=yйv\nܞ8cr7!Ig 㽏p+[L\4VIF ʂd]& R5jheچ໺m dB ÿH2WrԢd8q!\(mvsNRnmc.%gTb^d8VK¡7o높!&!RR-6#f+ ƀ1 Vp]\oHO8I%Pjkr$n)!跃Ny'4nXRSYDfCM6Ju]-UE 5T#ЋUj=oHdd$yMaD[#)q GOY.o쪘^9x2kqcj6jc΋ ̢' knt^ RF,I\c9>9M<+qVK҇ͮ.I.E[qؐ/}q,U`LE#q֎]vXZлhHj$5^YH{=AJwʿΦXs ʑb8d+lVzM9iR[+xXgsC-nƒjhctmAPM+V|ri J"\ܟ& k tZa"ucuxa`w>}7(}gimT3~}n Vm }}L3SsJ5c$6[2vQI)n[1mdM}}VWuoTJguܳZG%{~`:(pL5(zI {xg\f%?~bO ؚܓ˞x< Uad0-lOw5S4рb*&ڔrYOw ~ Fo?^4/} 0at:~T9sP\ULnx²9۝=fgO 8X{揃b>i>^ <~ L~^ds~ŧcՁ9p:`~#x$.L_~\tB|BoS6(L5(5jkV ,Sc"A҉PJ:QP0xY Bhoų[G,FJ na[(v~G N|E&2|O)4Ls#q:H΃}.5B6L[ز͊4Wc A#k(p9 "RDp$VE/ lӵxn_O j9J4Rm>-*0@d0+N;_26<:}-EYF5æܴ%]ed-``KUf/gA#`(8npN 1 2OB548З'zdY)o8Xtm{j;_HzTIj y5 Y<1CQgMF^h]-z~^U)9n(C2'㫭= BtY}Rj[$ m%bbɂUi_D\ RU*Si"djڨ2r8o~܎bE)R9 q^> EEri`MUfEt5KWDԙ ,(B9H`ψnj (r-ĸ{In-ӓ?Zd3D3Kz7 rYp$18ҊI-TpП`ШzkZG΅;dnb9VfRk$ݧ5|>7Ǥ7N1y&0_q˒@EٮEpg, p6[ P=k/26szWY֭˳/ê0k-#GvF*cGl"EXAP}}$v#* xdQ'OFFBJt+HI' b"91h0'J fR }F/ a6QNSO [`0D")CUAZxmd+֩G F[A?ĄWqA(tdцIu ښYAң'%ƄA7}M)+'Iņ˜UA.6uZ̯S6Fj]nķg*Bۜl,8]`V~/,'!06{5Y628C)*lԢ"3`Zh$X /+KIpTA1ШF9lP¼Śj%Xjt4bjIox8sy'Xʘjugr^9x-$48{NNx#%kީt͛4vF?t ]tfA.  hƯeӞ.xx!Qk Zڀ-33wm SkhTmq0M" cjhmʔpԊzm)G>/`ԕGRG &\R"y'="{3ͨ]РA]D::0 VhjƳt$:9NKZ-S.5U)sqܝ$l)r9U[ތki%5:J|DKviz@W<{L 9l1zS1ax/^(zOͼ}y;|# |,#. iOg.3qp9!Z%Qw%0d%ADqFņmxh'uyv鮫dE] A?+(Xٚ@Æg, Ҳr?"1xrL5j蜵p.maB6 'KLbӚ8ϸSz%$g_&_~ZE\x{7bv:C=d%0"z[v%ۨ#m "c歓lȲ~9hQ}ޟlmRJoowfA2e&QFF|NT#vjAox(S&ˤXEth(!~С/޼hw1D -!fjuKZ=7q55jdy0rvp,mg1DATrT#kr_R0_?k0d =WKX83/ξGs|_%kl<@F@Yp*E)gu$-X n LL>*Z\}|;_rwE=)($ai ? $LeіTUD&0p>~t| Љ;=5>LV͞NDa86*]&',9Onbd#[c$LU<`gٶc3,\yxl-1 0Qc@#Jݮͯ#Ago_6d +4 AP6 T >|ws^6.ZN!ZɅbI8 P(⬩fcKdXi,ѐݟ,w'Z\!9jjvSCB56t Ƅ J%B@l<&>B-_#Op|HEnәrV m?eSdѩ,*_']5Db&@W2zcT|鱄Z*G w}S$"ƒ?o_6}=e. h) 4UTS^KTA 0D5)4Knfsc!j-jSc!9IpA}/)K\,&X\ {Vp}%+ǖ!v[a A:} r\|!s_':=;ENRo#_M1A,emFchKNlXCF E/PkFɗ>,ʔ3!硩gP#W̐!bJtjdcmMKN8d4%PuCέR3w-)"ӟVdJ<[O%*5`;͛‘c-%S)5JUoaHޙ*sˡ[zZAv\J*I(NI~9Оs`"ԗsٿ 7թ_Rig+EԥM5$`Ɇb0EFD]lZT>dD%Jzvj{JhG='@ioA|xEXn&BP l7MH[qk*Ѵ*R )LAYT5:o铤H zA@ql;Fo /-bZh5+0]X.`F­L#UPhfUN5Eq1;=IIe}Zf%ۦYE#Hgqzn'UPp Ѳb:E*nQՖPUǸ˱19Ea'qmEX\,I3>b0S"Ż3ʺ**.S.?_GQ5?`՛( VR{%h3m妌 dJ<$ѩ2Xf5FiJAI'{BltHXșz:;*ϰC?| ^WWޞ֏BOj5$AeYI4j=*wx+FťV>B*'ԨjOW7S|bF 8_`\K )ٲFy-1КG+-^lJ AlwcR^II?JelBS=fSƀ[Dva N44f:/mu:luNU׉z)?E7S B[S5L[[ZplЋ م--XA&dhHݚrxLjႛ5>/d}Ļ =Jd](lsBvBAc{h MnD~G,zD" /VNi2NZ}cڭtϥ fzg7:k#6%L|SF Q2j.!qFN^?Xfa2jœ W+rpc>ot:T\YδOnukZ?'t=>y^gz^ob{hȐpH0+fs I6;{{p"~;ȟMYoNkN``> r=uQ]w&H]S% F!HX^tl?La}cK(uk4AqpO%M܍c#k2 *ss ܓMP:"Ƥ g}5BX8tjTOnkͻ'GNk9> o\+.f9[}U7W.. O^9o{˕5{Dy_99\/IQG}xk#Bŋ|~:hVNO?two{]ل ې^NFr|5^~ r^c;w}ַ 8J30BN[O>88X. }&:VeCSp?ޮϹd@B*\5s%L}k[eo8A't9A;#Vd9d{ Mn$gI=d/uL֖^Vt\EuPћ-UԔzdLd]2RYrfbG%Iц@UV^"cv}MXkR$04i~;"ӾWot 0 6^pZ `J`DVja[jc(%!1ؾRXjeőôR; 7hCTߚ85ުkmBɜ#fi͗`"j ͤYBy!EfK#NNt2_ir@q[e`-Iծl%S W/QmF{`׌]C[vLNYIǖ {N#>z97}b"IēЀg0qSˍKb읟1֨꾣laU#-O56]kJV9ڦUͩ"SOW)^L ܂ !q Zj:;Nel4ONQbtg@t /] 4 #[%Fv5zKe`\j7*# -]t 30;y Ww_ԃDPH`J-ـ#SC-*QG,WZbҵ%k PUf68.]Ңp?1R,{eQ * -}4IzP,H XcҬj)m&VWR1ބhtPѬ&H)%=?=6kH,h@4H:!4= I$E {CnYPU:_#,XjO@vT*[t@vQ]] \mAdsXe[iGQ'&y_CR.g_T[?{g\`pLHdlώ<9lڭܭmM}bbJҧ&9N{Z!e޽>[E~CCl 5<@ O}\'F>;}Menќr|[~ڐ'?xNn9g]k&0|Go)dwi@S<,5z?OX3X3N_j ~fx Eh8D}դNp)$s)MroV8$D'drB ;#m?`rO*5pL%R!*Yg<eFE2L쳿̷yiz}xX̻FOu\@ڳ[Ɜ_={l\+jO!Sbb]y2 ̟BPHXWyLźvR gvw>Җn?{5Աxŀ2ЩJ7u )ֲA&o6/)+5?ӯoJn8iVBИ86qTqnt ACY9U0bJtߵ\.*ǠAkkPs惜tzГhY"WlẠ]1q`kcϵlmWV0^$Pw,Ѣқ1zv[qn{Lý!Pb=mі!O&B0_!-;ȱbڗ_q5#xK4,[uaRz[HU m3|1ŋ{ӹC? ^"ѣ[6O,%7]==R{WgDsUX ν\MĖLĦyWx+0cvC3#}T5uOq+x7tGw_:Ȗ |Ԩ+w,&m Ȇm~"MhNRz"ҩQ`1\DC%#%Ecv_=k˜]۝."Xѵ\;>]t5{-4*2_mnduk<>y,F֎˰2)~[ۺv,"Xz{c#vB:L7jӖ 2dcv'N- ~+ӱOH %gx#X~?[XH ٷ@"vy6ě6 LJ#;B;Lo{u1Bԫpur*f1+s;}3ůVī~v.*sQ,ܿ"d~> }~ѴxEz4;=]ٷB">A*"1  oёmod*ڑ>1i葧Ęt6ptGGztSֺ#Fm95En9ژ,f ..y_i \[)jW2$lrLy;]hZۭ:JCʐBs3Cvm6^?^g~㗳sejytA];--gi0'}v5Tg)wA绰@׏ -5Nr~\;O'k5?7!a%'MRteD} xɳI-?AF"1[ v3n9/x&>h2UofUxY}˛$w;ךzSkvB߉Uh\!ŒY\FuQ#<1fJJv2+-3q6Ho,>raE,,jfaQ3 ¢RKt&9^wZqAlӣ&D$ka"#/!vr׾g"fgt(Ń47TyWg uՌf[\ڜJ;Ay"Hfw!5U:H%2( /,v)t?JxN.Ih>*&~_73?(ԛ*:Krv4@ԴSñ)e]A+R}|i7WݨW[xL\i( \ǗP d0?f܅]m=_ F^fHZse*S̘Ga+h\$q:n3-f{_؍klt+kk=tcV 9h)|rt]-bcً>7o$n\`S`.4lFǻ0~<0,Mn25~nSi'RmGA+PyZAL=t`D2|yTR!S2nA:O7v #Kփjkdlw3NQy rhАs*F5[}f,J g5b]H_ߟ~D2D5 E{#5IC;%`{]砢=S{*or`N[$C̶{`PD<C?N׻`BPY܃lL5Ȑ)a{[?;A]MؘLeN^ATf[rR\=\3S<٫![5` Ibr͙.p“ٕyXA^ M؀yfa@`HQ)Qf bg5R-͈:!@ޕ-_iWT.V]ndCOK]ѥz3:*2?gY]7CYn_'wy/R?޺eދs+O\;V\2g̦ުNs MH:3oWq'(ءgj 14xJ&Juu(1'j >u'n}R{CgBZQ9@BNnj~ӦTrm7vn.H~ZUybπ 1VU>.]+!ԫYK}dֲtS5X/n% r@rI7QVrLzf$YQV[ TrC9 O气Sr,G_ߗMճqa?j3jvcx\+BNzk#3+?^sGgZLr=LQiMi9=!'xLd岅[f]4}Ͱ #`'_4SkCƉ)a |c^b3יc Tg54wKAv }Ȭ!@f!WX\&tprwNس.AfV(ܿ@I<1;s!sL^sBN̉I{19Ck$Ee$eY p1䍉\#I7Ԛ"Sh}.4ZFg@7b 41)5|.6+EWq:`^Z` 5ߛ]4G4JV,mVl΢mw>tc}= ֚d /u7[4V {EAdfSQhS 8,҅By,PQ1ThkEc^}X޽7VƁR֕2.~"8 %B#Hq"Z(bhjjp?MM +`R&H\msA(ZEI|>ziD%]Q9J KP)%b o\cxxPʮPt*236\wh_ &(0>-C E5 Њ eМjAn(rtPl,QrrI5-hҨ]2(r?1HlYbM=miub*  /| ^ do@ 1`R$= [$co|hP̐翷;l'P,q]O"`]5HkGuخ/|HrhOڐ ]a<ŷI@ 7]A|x39{|Ӏ-r;ď{Ua$NCU34*Z(׈]Z8REyKZ]ɽ';h5&xYRz)Eqۥe@EEylY(C1XRxl<<m` Y`υs!C:b'2_jOEYKDq0 'e\(tM9=oMſUt/ՠý䷲sgzTTg :{(y^}0X/=3S9nr|Iہhu }QRVw\q.-Y(C\-T(n-'\xy ;߄|Na&Ki">~A)y @͗RsaKJa Ǎ@rJ.inM:7~0wUU'Z呖}H=J('SysoYCnLq kQ䦌KQ3T'ʭx5Z===͟- (aUCփO&T.씐©``[ ՠ*%o9`#{G#ۘ7,psdfRS tWQJ2z&#x)xހY)ksps M[+|=׿ ƶ7{x{zmg[ B&%Rv>QT>գ*^%+fU`ScH"iRfuMN1d)K-pLE[}uMVd7ֶ6Qr.eJ(-Q^rp-3<eM!2\ (3P>J#,Z +MI*aI`Z@LZbkƬc"3STXf!g&yۉ= k*"(t=00#O,гIͅЭX%*Ph(hd=O gug;LC XxH^gU']½)EN"80!d-U6'ʚ<G[ME2qV*h&͢"?)&( 4<"1Iʫӎӧ{fT q#; E8%vWlq;QcQ@~;މpl13o<6'1#{Aՠ'm8OvKuL3p9ꐠg`Vys?2wݙOc csaI4bO>Gef.ܩGyRO+8y5TDMnJ{F϶Ks%T7Y{;?tcc|I%=?- n#⧷HeR _yjLe쾂J=)<34 exDvdTL~, uUU 4uڑ۩G<osfʝ =9{"~S=}d=NW3IHR6JȾ*ẋHumy>ޟF_uv'(Ӵc>zm}=@͐+|Q c Lʇ6ָF[2،^ΑGXtVC(xwZ8V!'LH9Lfy:V`إCpJŤwE&0Ru`JYmF1)5FܭV+h\IKݾEBLuxн2trbNÉa4wў>lx hpV#ռԻL)>YmMƉ)mpJR>c,rW9r5'W9랐 zM*UuIѿO_U[c+͐Xb8Nc\kBN̩Ӗ cy1վr̓U3kqdLj[)B/bDt^ZqDI7yt`tjs ~?gp6>}`e)׏=`'ʩuW>}uD_Ǜb^C>{w0TCy ^|Hoߘ;TV*n#Msm bDɉo!|;A Wn>!>AB:|,fgX6!}>};2C?x3<3~}&y0qWXl1ˇ!yV\wIҩZp`SLW}ϖź%h Us upոwb>2ON.)r1rw1 j| ~x='ϨsP ȖP:791h윸nVS\|t=}t+Pn\MDxv"P^yr垐ݹdzPfȝ@{͐ 廙S0̓t: 9d f0X?V5XǨL3e s;!lLiIJ˱rz@A}0ʹpX`r5ޕ#EЗy0Hl`lGkEHv-n[-[-ٲLYU,XQu]CLNzMgJċ~wE} M=M-bzO%/XuT Xa79V7{RQE4iB)]9E^_rBԉ4,"dhi#M,E,U9 @cPa HMZ]Gĝǘ⟩ 5?D%, DX7 !E>Y)G;wq`;4[ue)ݜTGv> \!a".=bLhFF+ځ* ψuP kb %dd\!;D%,!*\7A|% %dW >{dl\CbJ<%4iv>ƀHH4.؊̔Ls4[qCpJ7R̒(;!Cz$=4ttVP%]:'bƒCQ~>m';`4WTQA9S@t:jNRE"TH_`A Z& ^! g:v=&\iƴZo/*O2m%h>4wNiWڹ%!*Zbx浦4,xBv 7X /<.~0C|x^Vd%˾* eE$K >jD)R_\ TuTQp3ǽB3"1^3ҚJ#S^`>@ICKTԜIƑݠT@vځ(z<<`֬)u^H ݫ8gzHc'?$@qn'9@\\d?xGQi#mSu+D:"m%Tt+i,a$0Ej~.PEOGWq]`p<" fRbr X !WAgiEiȅ( y^BۮCښC !t_ e L;f֕>n'< qQ-*6c}1b-cBl;p0+yD/ s5yѽC{>%?to%x,.͙2j%P(߁eDfaB\Of. BDv9 ÚsKmxy7ޙ}6Zvyp*|ӆ e$GmJ50RȤ1#_rd&Gfsx|އ_'WcDG TTDQg%-w,id?jJswg["7~)UZ"݅Wi*LNjxQF89E3@YDXl("}b:^ܨ4!nfPBP١0瓯D;|M<ڈ#Ci Izg{8CZ*|^OC b+v{/HPjB])גbM{M1%c@xoD}'ɰ!ۓ]qtvoM5%BԩꎖHjԜ"$A##`h*ܑ#hUF ZyR+4uv mV}JZwt^T:sKIt (b2?KTKoxLFvg:x_&TSo~~}X3Wg|R-SC[2NR Q FLֶFpݚ:۰'OoSor P a-GgtwH 4p4ʼn$Jx-8lBsWv#*ɳk=-قMXK\\;f_̃wȢߟg1j;^ Tj4 iĀ%L@8-MLH~٬l*66]yʮSeVrPwJ]N3>|èP狟λu׹q[AG_go|b6SIΧuC}Yɯјɉ?4yRcUhԠ9̐*05 ""ͤZe캼k^VioFe]/0oLJ@(M\W/SFx^S+Tqhc^}|6~U=гX,=@Ϫ@p$sqWʽwBr&u\:`Caw`mhG=~y_tyan\%/*tW "Wl|/osv_;K|2_4Kf߅lmɤu(|Fqf"屪͕>7\gؤ5a޷,B$tAS7[$QװQ x]vIB1Ep Fu 3gG) lpBe;B1@i[xzǒ?+u_LD ?iVϫ\V|ͪ1#/^?Kt|5Մu7%JJb6N}+p'v4QB 9.WJy:e*@%zQڣ?.T0lTX M9fm/TH]2^twHӨ$тJsCQBRCm@vp7my#89~E~htT͗sEra3ʌdEf02weq?*jwFdPkĢω-N41:#sl@Z?ebC{8Mß>( &æZ9ӏ1d1:Z z 5"_~+J&~4=uD~Ir{9A4֊zq(ܠɏӖ7RvnaRgzW?ʖ5B. y~g*>Y̼b}VͼoسM Iʋ`=CL)oH(+[4,IFr(VLۮu0{[?cQn~rMX3ɟ7ݡ'%8y&hx4 ;&jBO"f=Pm>D.iAA3GvV1(Rp`;SFOK\݂K Ai1- ̫`%R~C "FON?R,Ѻ!TP["Z[*2dA jssЖuQUbwKvM4!#o܄PS.s0@At!FJ ]X.+SA=ZRDv9;5Q{PN|fPdNkf!&,4!*hr:WaIbi}+[%3х) n*ji/ YF\G!.GF-s6/?)DnBZ&>dw8(8J#E$MKHN&Q?IĞ^jh_~6_)9 ,6Y68K^r˫{:G٦xO.A)#Zi/x޽]]Я^1/&VﮋUn`SW g%j4_U7{/p=%^X34C͕(*Eu?ϔet!#Leoc Oɣ:.j'/N|UrhhXLˉ`)bY :Tgj1e h[N %A1 KG L\*è\k/]b]_ԩrחo/h<~jwi'ĵVmĵ5m(-oFF>M?=4n,851ʨvO!' faf*;]]Y 810PMtRulAF)Ur `A^ij§y.gKXփEtMں@$<:!TL<᩸I]iCQh Ћ.:hՏgOJ$!Ce3$Ѹ@IJ!Xj2ArB?4c(ό,j4k&1#0KL|e{|_õyǽY簾% nZ$Yt#f^sJKJ݃jڏ#ZΛꀖi v?k0%+TT΃&biLa|GNɨy'å`0 U0B{0᝕x-XB4Ut쩲FArC+w2A˥Hi.kwQ Z$\8-A*V+[8S`>$Yi i 0/t鴤r'9TE+Ev}uħ7lT1 UG8Mef=`*H)`!pٞm$ .L p]ov:rnQJETptb2r-I ʻ>gT`^6沮\VwT`v87mQ&AǼE})uu2oW*J`pe>j黫h*1l=\RM ?-Rw^С7\Й Ć>%gMm=sRht#%;XC(S֡՚%1EBhuO}JִK)mT۳Y[5( 3]'[#Ef'{JrZ& +Ð,SQUM: f7>|\,t+#FeD E1T6!rϜV7`glxVY8 ŔdWmDQh~+EbZ 2Sסߙw߻ KtP^29lm6I(3zlpZS0\;OO43pP-b6چؘj^$FN( Вo5@i0SQ2'!ߠ?$䛵L NB,RpmΊq$]Y<[a.9[B킢1U $gO331F*RIpH#EM1)6#z}(:/H1Im2L%dI>PEĠ)?3x hqgiѬQ#X{6 [MS"a8RZZdM$Ǵ(40ICବBQ EIOTK%8\Il*qє1x"3)V"em~P,Nv.YP8^gr&I= Q8Hw [!$f3r?UpCQ-z"u4MMf`t0D"$#SF,~bhd̂C08i=D5j kP @Rf^"`a1_C@ʼn8i|% r0z]fJ̢bƾZ0Nn(FϣU>[>MnRf2~KSBPu[2%uj?~Wxlp|%hsJk!Lʿ{ك;[~fF?<u>B>U0oӡLH1k$CiDZ a1[9tM·{ 1Iifl}+nd ͗_[z;- {5;ݑf/fkY)&A2Nqm+fy`+$r=Y5 MuwfꮿԷ3Jb0t!81a cc RUA$;1mpT Ws-qPi`qg{#~+Fwg'W5WA<P@h3m^ӓT\s4ќР%l8&@y_ۍ5}SfZ÷Zfmk=w5+~j !=\ҋ6n.&MyQk8` {<˷o72H3WFhCkXBYRגl^HiWo cރ>W$5'IJt؄0/kՄlqo_G/ vyL֣ ;῜c4Zno>;[3wy6{Og|Frb%uVlO\lͿ&ٓ}LZf}Dbk;xMO*Q)uhޢ'lmh)  smވfYaq_15aOǟ~kώncwɧAñ0AַfP~?J a-a񸩃4Ds(Ä쩜gCes|srB DrF E'KMWz*Y~/hbOxwZu%Ek=*aY3m}.16_ݢe>1@mp)^˳X>jfder{zit(8ΕVFGdS{E4r_ͼp a4 Jæ,NI~4aRpdkN_9PY,Z v'%+jݬNa`#{%!7zvN>TZL{x[7R?PTCrE_r|@4Avۧ">0F6#F!eMYED/I,9$ Xӓ[TjT+ 0L}ᾐ6*iorYNc- Iֲ|GYh^'n2U( ~D$%|'W¡ȩ}ra&)M(yOJ&sN' yY}r_im9@Tݑ afdf\+wU]}Io `O=gGV}?S}bO:4W=^"L4[q!y;_ ťY^Pll.]pӍ{ۄ'>}hDj{y-w-2۠4o{W x=Ag6UH{_^Z c> %I)a}MnK~,pQ,`\R,R#RC$]]yQDE4 6BFBAO8e!R*5hlik0Jj8+V #pѲE7SaA9: г Nzpg090rxB>]~)?\O3ޕq,ЇMH}o I O|3_$6(xP6Ve{ʿRjtfQY&PIL+ঋizz9!P58\T G|70ёqjmgz#(vDY oލte{"&=ǝGG&W_''8g7{{;p4`?|q>.|v#\3#2n?er0 )5lp: +~yz38M/^/.K_tCf6N9➴?K[\|;:w N><L^ 6>s- +Gjqb/Y.?>Gu1\Gbl- JxNtE:5l(>Ym:b]7aZ,$C_ ܗO >gpQ@wKhN]'0={ x1|Q.rO .??qY]= !,T^ jLJ~Je a]<٫9)i?/퀟+HS|quL0ip;/z+ǣQڷW@ ?=0MRHhP$O$U|}Y!'\\4bhcƃRb_N+f61a&U>~r'5aK+ɔk"t'ZKD7Nc7~F5PT "/N6Q KQJЕRև`8OEgߔg:o1\/-b&m7W<@Ԡ/ZaL8GkRrrLO2Uw<+mȪhWR),hhWH,Usue(e %}1揽4䌲Wڈ=VH#Vg=Υ}IbF)}R=w~+k ;GDf.{j؝]ufF[ 6m܇1&?1vp= ϼyf Sji!CU!T_-w)w4ۄ_C ?>i~m¯MmLo!ۗ>d߅Vmv S1шHڏ8zڏa?Ƴ~l~̷|ǜ&}~XvXS1"a0# F=$ԱDZ+쩌@T-nw-hUgŃ8%-_rܩ:PqSGpiQKGpZ8"??69Ua\LI-Z5L{TR˜b:Uulb-/˴<˂UakV FDUaS6aL3n*ғTO N4z@I W9uy 3m֑C_#:#{]j*}| ~rhTWj~sڵXee>utB~ -vٞ4BN?8%R*IlS"%Z[X%Z%Y+%`J2BtQt˸]ˉ醠E 2%P֔̌J3%sm~Y$*yM1Kåydjav~YJ|JIIOtCu W4y}? ǩBq}*DW+DO~F>zM*H`Xk$HvN9L愡㰵PK)Gډ !uyUd: 1p/.AgrA$E i|k)"g2˘5VQ^loM_׃_E79\tsѭ:ɝg4U։oBLih|kˆĬ"9=XsPlj X"~9s#H%aLKTĂ.Q`i+NY%q)%Q(IĬfؐʯqkJ5HحlX]-^%UCn߁2-Kc^yA8Ra =) `)F0D1U\ VTn-9Did. Y 3 g(rSID@Fp0Yʀ+dH[ b*$sbe6Maٟ 388߇ǴdOp^ǰp'(!h@JH'KlI0B6j=>@5@{@<T1MH{@!HsJO';)S*G4Np"p8},:Nίa0_LkP3i"rs>ܟ< QӂsZ~DُLN@kMX(E׭u ҸS!w.fww4! В(KfkIℬԧE *ԴSe1 ckzRcto0stu9˻N#eE/!^~y]zQ!0f;x~yF[_ +S%LǠ6{_Ζ߾^5>8y3'pgi*iL\I}B ҏ9NDrP6i%ѕJ ꛃiuO,:E תL nTU ٸ?Dc-R=Y('8 yQ#HhR5\c3y!0L$v@j#[L M\eE޴+ bs8]:/,,[\Z$u*J4QT97",Hё@c uK4(SQ;DqqET&퇍6Y,P6!JrnLဎsay :Fh ajAye%wYo=kPN\hyC|8騲aV[-fQEV*Z {BV4hc8"&jx7kx>mJ|LI*> `JDF5ǵ߈0$9_RbNz`IAFl=A Zʥ'=!`qxUn\$#$Fke'Q`\V7TDMTiJ\TfUZ "lQZ~e=&[opF,j?MxL+l) nǯD·~yV9d>;$Z~33JwB?|3Pm:'O=85ܥc4I ӓ3]LR Ё ge>nXR0n4[D_|DP5t\ 4Ga@vu1!E "o$ ج1jLygG|ŗ8no*9޼Wrl.o1+O0+vSBs~+ ÐS{a[1>cSw͛W_wj Z} ۳|7NNׁ4FZ.wT\ͯH͙vƍW6]A ܸG@8Blrn` NfNVD嚶=QDr54DuOz?^=Q#'0^@ &C$1ۤ:Q!cʣd`ӘR<8,Qk !`wInq1+TGkRE,B+X| φ;FbjrZdAv> NNRi&4ba;"%ֆ3BZ 5-0ü7r!\02`LzPzez<Hcf+zVTtگI>fuɨ5w7\c&BIo9MWomXsWvqb6AY D$QI0s>/NL3SL+OhAWq4y2seVbY:ڰ^M~\QlՃ@!Kn^YH^1[hKfnPa֏1^u&z \kq1D6.jX\4>(&!G)BbUlqR)2Jn]Zp[ Z&@H;ebcy=vsJ-,[=ý%KQQ%:@0  4 5ں(Y i Ra(e%d%f=[\VQAiU6 A,z sc!H-wak1FݬDT/uMFʂ\)X#+Kp$IgLOT{ЁV;586XCm:<Ӡ&EF]ʪ4 Sh5AppA(!O&8*"8,j)8 }\Tsnm6Fs{6"n(⒭cDo`.:"'`sf&1`xi9f1b'lHn.6)BPZc)MH ,a\LI Y5L0G,"Zs;C"w9,' )EԎb8m(29l@q/p! a-` dR! ȩR!U.>gQZg \0΁w! EC\g+9@k9RjLy6RL8~\jO)RQR9 !Y` %„#v/M& K4X?'48D˅aG"!AUCD;8D>TWφQ]qy&@33|g1h(@' 58Yk o A&X'eu?jI(,8O DO^ L&Ӏzh!|8Ɂi>UYŅoۦw28=hQFB;:p<)XC!D0.Xϩy4"H. ~l!d&rt(mDP*hip c[ JSDP Ęw\ZQL=pK(@I~%MO1m{H_I//KE{7![ ߎMT!JQwcT+A_ewzwhu{˓e#1M7r[2 ٫7Gq\\gsY&I;bT&WAi}i~c`ʤEI Ic%0N(-GCN JնlC ʮM=&OL=V`ԘHm t}j:4 whΤ7sk87v c\~'7 G`b̫wԬw+x`n)v 6cy0`AaگdKnw^.R=jߌ'VxԤ dw+bw3>5DlBFgm\XTw34C-~fо҂Z4m8>#rFob~1&$D;;%-cܭ?nT_s\!eZd,\ZsΫw,yW`۫n ւi^0jSCnևSMuPn>-p_+xa ^)y=g;(Nܷ<۰߆2dz\T??Bfv~5Wwd Z\1ZxNhE=& X5lWs&~ئk~*y r5(1$$%2ztݤՄRy#:h; j=vK&4V吐o\DdJՓY%+٣lEGT٩oQ# "Tv2*;|"$SwSE fH:M?7xg#U(Th3!iDNIc꟏vEL)Gy=Vt8%bp)Y5B/-ETAϏ|勉)K9rR $Fb5=s) Oe3 1EiVw~Un#ߘd3<(67)V죭ChA#T,f]w>$"ߛGoҘpÄ7 zeNQd<9ʣ=s-#9ss/!rK>Fq ǨXÝ(>\`Et`Ÿ;[1| tG#Z==l^(ܣS\шVOhT;y8j# 1}iu(M%!eD--U%'S.CO̹2X1(Pxi1"a 4b`e&j5%'v+x{' S Zٵmi1wi9sL߮eOT24 )V!S2?cQcy,7gXS>0Wx+G1ego|i>D=}?hLo&8j_+.!9!qs^x\]{FKkd`x"imwO,dΑǷN2;GT1yQw,ZO<4m QB؏A3&K-%^; BF08JV2 (vCQ8eՊ_ߓ+ZҨj̊OyR nV3H")S<aUʃ hdZ "[䐐o\DSdb:}Sn<oneLp8 n9$$lсvLIuKATϬoj6^x}}mR nABqM)Yul绥 j5bC,z;b "z}2#̜kzeA^@(WCrYfN(篵4#۹8e%b5:2Y./כC̑Hh 8$%2+ϨC"Rr+59e_r=_Kͯ84$m=S;̰$4@R4=Osp7* ZH35@1G7K뼢/ F; P:9{U&7X\]B P:i uz hHj` YwU\z @WD4Rk?1vJNÑ I>!<Ih!ɫxr^`p"`ax20N}PryՔۻoJxLCG4Uk@[4}$4@Gce"cfXKNL:FܒyS4qIh!iODre` ɂe'jBH}֪sSh_WWШS37Ǘ$7Wj߼~DzX!2ĺ(*aDYM%<+y]XKEUE ?b}U=*Vl-Nt,3`Je(jb^2 jC})",ךS๴9jl#ag&'1/k#l4saZQod帡 JsA%՞pNBt uZ7͔`7vo:$fB$ X5\|M"|$]1 >PFj.DOhe*)i-I'j,7P&Q!8暤dj;K_*MRQ1]&iŋr8/0]ú#\Bؒx̙fRۃ`H.pQ– Ŝ.2rJ910=1~^lίo%ObPm9]tjvwvrNIP6Pe*)vh=fxgvmNj)sRZDQƞ[YwާIVs͗;㿽^߮2h@gShF0F__/-wsAfh.lmogTaϸ8Ômv} ʻ@) {auL8l>7boL圱v!{b@'损,:QOq"òLYyAx֐0p pxa0h 9<7:5ƥ0eª׵& 쮹VO.{~^┣!Ō[N ݋yx #tpE QO/g_O&Ĭv"~iYyFpͨtke KY YZcZ1O5NKOJ3 C@4xc_^_?zr~\7ץ"ob7ǯ݅*k@76Aʈ !um_Waba,2"tƗ|h|O/!9J=٤WTRS N"!SL2s0l1PZM Y^^\t=8 0D1r? r3D?{Fl{6 u2Nf Kd?_Lf >\994*拗#œ͆MBeAT ng 0)dcNˏ0uZqKzs"0\XN!rV<̦EZo7~j)u6(j%Vx6J$[X?z7b6{½])% S+b_>Gծ~sI~߶e9 LQ,L#8&%:GחHwj]OkM^,,biK9MاMZ ?Tm7n|ΐfffoJ7(vi4S̀JYU?Ј\L>r?܃o}ݗ7XUhHtpJsJFSן/= W(זL,Ɨބ?GQ᫃ U' vC =;- dͯ[>i4JU,l0"ŨyR4 x!zF&KֲXܧEo͢C^bqZσ͕n.i ]e|ίehn yyѰ^Q2s(:zd^:,ˑrrAAK%v,sh0s(}#8OXZgjiwŴdאmaW93I*[.Ky_hlnPPCɐΜ @dHgN0""5?3$, A6d\D#B"k;nQÜ{q}XbU7ޮe|8QS`eqPE%)TSr1žnJ.++T@f;1aeԻqg7g&*{+Vgo; V(`i=rTсRV@kJP^!N`"k0ՙt*ۤ0.𔃚rP1[<4"˸{ ިϬZBͪm,&VLgۣCgۣ3϶!m+W+5f ]7v08e:dQ\NwI>$`ӓS%Cւxh*3 JpRB|U4IR]˥i:_608y9."λHvc"ؔnHxsZnP^@0_yD`hu' gG >Tnk: 8A*L2 , cQkTNrn"A)7ۨzբBARDv;k-r^AHkU2C#w!gHv/{{IvKe!Ԯx ɩ]W0m+Ν; T6k4KEX[stMHEkd7 ,[LIw,&*%μ~x}8|n\Ƒd%SwVOϙ2X S @H%| -r Ls x3Mey.`"ۛEvdFu(WuR "PkՂ30V3Sd^S`rONS=O D;EP/Ӡ.rj9.-߽0:p 2] Jg# ![ s${ ]nLYÔd @AI9:đ \sHIUV]W]:8(aUeC<\eeʱ3dʱPfCD%d8(әrE;f `Ft9Le46Y[M!3PL'DP^A)A{鵦s( q4T*Ip8N*g֠ל ERs%ta(Eȑv:@Aa렑0pD<l{)Ab읔h/6(Mcì"ae]^+w]w&1V0<1: SH)cdEKbAS>kHh^#xTT ԑŀ=e0BFT[% :6&&[Cktp0#5N`ͥ)-'S46jgp@`hV:D!,h0K!U䄞d |*bʡ4:,rAae\Pb^.)4܂HCbSM"i情ZbEvT/f<WAڀyOt"4 b-bi  [P!J+h8dÌVTU VvI#LZD!z J^M zjb8$C1ZՁ{g#cE(zq3n̠gyvPl3ziuo;Ӱ0:1V*hPENpIƒ ~_P˓'?VXdr1_py.5ᳩz%Δ 񰎇ٓC1/pyO ^"O0p> z7S Kd?_Lf >\cf Ө/^@v s6DDB,Hg(@`!r0N!Vs2=~~:>uZqKzs"0\XN!rV<̦EZoIĝ_ 'Vyκ*5 =fRhQ`cbL #c`9DFK$/j l#NX4;@kn% \MkopIkWLlT*IXk2%9~6P۔S jV0)sMo2VPgk˽fSky49@ϻ+~iQ(=;KAF=i0Njr(Ab#Bc6Y"J0KnY Dyv0qˣwG p\?р/8] p;BK¶&8M e341Q֬+4DX1M͎Rxi O'eOg ݅F9,}:>? K! {˗^]@(`O{ʃ<M̊i:X:/׵/O k~5ޫa16ĕ,^J{ax޷,ۛvYrNˌפJux2ä7?"mGn X6bkVrɋ,Z{O@ 5]il|f4Y{<9|T w *Z9`Y*&Y]U?w~2h?кz":%ïƽ=dS%{.F.9fL";g*J ]qǖ"`5E1"T>̀i; ~)F!°9pVP0! ll "9pV0 ZLɽ}@P8]_Lj(4$t+9tϹ[RsIA{׶G\8 }wa=A^=J$#/>٤ȦXkPbWdDsVd Go$}3yw'93]ԓt<;7W'_DꎔzY׽Wu {/\ O n:OE_,L<%k,_,H7|wi>y}:_R>f~b2d'eH'{]Ѵ{^c$aG"-v !B"͒K.1GZ8v]c%c`: p#G$/ȏNތ?۳>~t١;(3b\>XauCZO;Y>j6JsJx'vgz_au<۬Fd~Y|]\#~Et3)-u!+j@(˺b-Ddg[_.x% t؇ T[I#,v9;]" Y&EGTRmKqxWMhbIԭi+eR򜵣E\}j|@8؜(HUӏ!͗.bf]R =Aq=bwD]cXYX5WP---Y--aO9[KG3#zu#CrunүӺû: r.bӍZYqFV~U KQ F#YqΊ d9 9mpU 3Cg[x*HY:O:/֯ӷ˙u^u}BNߎDy:VM K\6/H!>U}J6P<9js\6\=4gq4BA-r"laR֔|Wo!(-4E2 )Z_9yLRYBGxV6FٓIX-VVVGG?}0ڲ:d p>j:A rX8-L![7.t[N\3nBśq𤴫=W`y[Ƌڎ?ڞQYrz^qfgk30P fptz9zζN1;LylOEaILTm o/0)o^\}rpC~PP^y2+Oуr-kE72t;Ρx=1ݙyjt;A9k_PEs43ۡ>'uCt&b*}m S覼ezICt{SrnsdWTŏqA7qP:!|ֈY!vpCE92+Oуr-..]w;Ρx=_ n6G|) znXĮ?vl@8t?.LbnX ;nX#?uGRlDznXUsA c,_4.n}=Nth(gn{{ruǏ,n,^slV]g&rzfoO%8ޚ9WZWO)SyJwiOޯ8Gg'=utq4Фo1yӛ%WM:̛L`ڽ󦎧1;{΃G~aXm}~y'|zs 嗛9O\^nY;U{ޜ~S!Mz2|yn:mߎ^naF'_ܽꇓ?w~}xbz1y?9-sg\86?x?/o_>9>?vk]o/һ)oTmqm~wZJߚc5>_2kKb%y8fC5.ې7-{Ɇb'*Fw?;2wXrGn6ᄑH_xdTo5m 5k{HނѺ]T)mId8ۏH5$Տx ͎{R2\.^r}m)>}IFX Dlr p̬Φ\OptUjgA}2dؓeuleܤ) مIk鮵X1h]r;{R2Ŭ>qItLεFť1RD*lHʘWv:I8  $IRjQr<{: {ddtSZ٫@Ü9trKڢ"*bsjk &ԱjؓlÇ ){!se\,Uá&ufwo79 ;U$ZCw`MHYB?0]=4սr*r@\U3rGsq#'b$KAQPgpP/s#sicc(E$.w&>b@@WCrЌ$u>؅%"Nz,7 Nk@Y=kX 55T/RP  GJaT?eld Jƒim5l^]u嶁A)ާq4SߦY`d*e4ה( V⑫rÕX2L'[@~JJBĔAНCHs ȝ v&>8..棥SJtjAGB Dh]QZI>ޠEc :p4@A[:SJyPR=&@+`V*h`9N%AŁ1 ^bqa/m>2[@q;*Cq>g,TdAX?3 z(rx jBe w*2}X j;.NfKƞ`e1 ^{@#CKzvi%e6X/hC*#EtK4ǎPF}]J0S!v?c,9;k r[TR%di:L BpvAڢ򬜑 #b|GZD7|1=Ǡ xd)nƼ)C  355h:fUqv PDUhVP.Jb̶d 36o-Z8tv.Y N@5weVh;CJVm5 *B|4R oV@ ja@?wQ%03&븶ʣhN!O+aQkS]No@7'g'o5憡L{t` ,pquΧ`f=_ ҆``h ;eQG1]ZK8O%'Gk6<)//MGmgړjÁvCDl:(4WTs^ 7=c,C[ơ0P|_w؊PDlRtrsp-L 9B{F)ahH) ,*yEJMCG 1; :P9Z+H?n:ɮTBմ6vps^=7fj pYX6kj`%ZƜIȁLjiP pFC!@?}Nb=y *]T7aXx3͸2 jU\X36 6րPzo{gmIpܚr倽$;iNξ,IL9 <+b!CL]z"0b{8&\lً$,&<_O̪*  `(·}mͺ,_;2{ #o{"r/ M;/ߍ  (_v 'fdN~L7v)hٕ5mZX5cYXS~'C }lE~ wj8.RknW>y$aBւ)+@TY;ExURSpϙ~O0ӕipeV=¤Wm?wL$=0%76WCmЋ31(Lj،KaƅZWg1dib% ܍^^ DL3d3J)`l1Bx]_c𭲯SsrLt':\KHF .e֜;^ OopviFa+eu# ^ L烠:B"!?9{ƓZksٻ؂76ʡ>.,s1YL,Ž{,}MG Kg~tyQv택'r|Y7|ʩJXuΕz%$$ɏY$_Ƽ l^|hU"f gE7ղ2LCGlcwNƫ*%Bq8ꎎ!v) FvDq˄M7=e‰ve30Kχqigv13(/x/BW<ӫf+"෤XIB^ZlcuDrvW!Ⱥ8ցh[S|#ac$8^)/_'}/:xqsaE<}2>~I޽77firm*~C;gE,{H>\44@uӈDVDQv1KíQ"Lu|鬆`b鬂BD8̼8p)vJ VCUءQsP}4+&d}AY0"0u[ו"e^ZGV`LpEjA%vP"$yǂUL؛EH>b`u_z581s㵧Y:z $*)Vwĉ{>GCFB握2"~p1jeByiÂf\SVjJ. ZKp`^ Q*>DMQ`EpDG"uc1ZiB')potZXш6mһ:06~'cWTcumX,vPwoue;uvPg/ ~;;AJ -ƷDI! hWV_"̃wMKV <, d#L$au%Ԯ녍I( v'XQDy.֖Bg]4D 1呔 bl@(0F=I[l(C,1H)'-0!LQ!cOSh +}>o"fa\w*$*-HqhD*ašϜ|Z(0  gTsd 1$_R .$mbDDpy=a7x@ (w R@s+ PY8pL@$(uT\ j(,NXȒH!"XWvhz^1_ @F0JWZI` . at*|(+.`1;3=bads<8mR\ƻ\G0(k'qr;Og5* $~|E-yFX>CXAȭB QQRo{f}Ћ rġ cq@FXS0&ҡ) `g~9k6L_(Lsr H_ { e>p4rJ;S<&!nb`D"ҢZŢ k@21.bW# ֎H*J\ڄV< G6MP݇ZBݣ.$}kí1+ (:XutE{&KomW $ n]VwTt)">k[-brW"&Sjh3EDseLP@"LL$E xqPrWoE&+ߛ%˼7=κtox;€L`MC+ () Jn/2j6܂yl:Lv57ɗ^! h9ooAfMrjeۿa>g6]کN]`|0 |eƊ[Cl# ,R μߍ_fE l^X&cqYf qHۯ6vỷAl?9ƿ_ۜQړO+>*ŻOkNWk)O€.q4+;#XcBB8\mM2Wu?OԦQ}|،eMS\G?O`۴aUNYӭ8+)8B٩V2w$U>]5PqN@P4^|{عJՈ\y\g5@<Ng ՘? 氜uӭ0 Ð\&/^CН\ 8)$TuZ #d Cl:9u1_Ӝvxt\8tbNkFq}‹8D7}DVEuq'LYhZL([O\*hZO|~UzV:uUIN~U=vش9VI;d'ڤG5j=I}Eb,P`LYWÝ#_xNtŹB hؿ۴i9}\tr[_/`4$ bO}#WtB|q^_hu^¨rW1 4ArvIItKX`ȖK3VC]W۪דe^S 82#`/Q1LE!J `a2|/*^V|$Wm?%\0ԉtAvj ztϨ,)Y!Rb4v>Bkgm}̯7۝q3 ۋKw .98'3\[m{N?JB0?11cR~@K<:TXǀ:30MWg ,Օ̫/;\hxjcC,\ʹ9k "I µP0t[9I#)iER}e.H=_1޹by<2 3H+Ǒ3.p+Zmd'"IPd}†HkI)^Vl#Aā> Vl0( wgQXx]6 +Df:؝yм( B.è|y LѠKE!ŸF-:aUGI&h~I>kkg'xxlja V,-[+)!#uf$mZ k yHTkt+ƿQ VSi,.DOR?l\qԒaVԇPyXքWɖ :n"T+!H˚9EΑ 9{Gܻ m&$H k!e@1qˏll'5R>CN'bpx̀,[~a+Ĝ!tT4CĤig'der>SylKoV(RPlm#Fcs9KR0^FEPMjA+v%L_&ɤM?Gr?ZWit[̽lEP͔Rm]pW}KƝMA̿r9]`DsJtb/c+4c%\ڿT@}-s$Ti9' kaT2Ӈ! hV1Q (P>|03)jkړINVHH?? \S*{a`R]jMblu>B{ܶ OI%~qv+y9R*ѢH -[O@ UZJb tLO_TخV\ru8>)e" 哢ć W mlW&wS"%X۶ Z\N) NȲfw;!z2V [!=yhjfF͆Vx䇃'Z u'.TRm)i{rlVs-Եh W/SJ zľ%\# БnE  9+JZ\nE޶wP*m9aV-j XJ{/-:cÕYc)Ǿ/Р8Kbq2VA,ILL DE!ZyG*~NoZ( R`kKkBOje65 5o {yA`z*+yZP&$eʞ2L¤%ҖK&p:| \ۄkgN6֧]56'>eclo@Ԟ/wf4Q[)?Ddǝv肑Ò>N|QmeS0ptV^Q\,տ>u5p< /Si-u(VkJ5v b>6+_o,^P/TT*fm+pwr'M=Mt kSk8PL EaNF<ՃB8E\GCg.DZ =\_ },pȔpP%JDWaJmZgЋ'㍠YFd5IB)9|';BΪng4[c4dwI>|Tk d-()_}ge QSIJnn|pGCnE 2FJZf_<8rfÿWU`OV _U ~5gU K}ĄkHO1Y]Q3$jQtu!7"7gTacIie1m$]04W5Q./l୒~LeN@D* ڂڤQE/u+á<6-Tw' (N|Lmwc`ZäiCbCҡbymHᎉg0 @9("" X':!^uC |Ϗ=J*>V<|OaY:Oxde|7w1./|)ik*O|'tAbcTǢƶQI"-7w,>ԚWy|[ɟ~LU&P$7ǵnXl}$Bo (;C%g)9eG-SbsP\2V˫PͩTYG%'{);Cime$J'ok0T-䈳{u?$< ,9.I٣%5+"}XުZ4 C+B/@J9"zSRE F-ϞuջU,/q`ŭ{bdI{*=>|Z9\T}7B k%ƕqz#Ҕ?Ն,JwH"F:^c L/j !\E ,XGoQ'X5~yL$bxC"2B> CX |)/xA#iuW[ufMf$wQd<}nsNz7N᷁% #`l|dX/~d RZ>|nޚc>;߇8vJA,5ū^YZEr+B5C2ypurJY<e' 6{}C\-S {Uevv7m(ԂɾKKknQzQ*̤K۔9i۠bp:ӧvs)P/n*м&hn2 _gBvj"|:u1R`y۱; `J(qHԇ d<Ph~3.gE1(s4\gEȁ 'aÌ2zV1OGS:DG(qQ= 6HAEJأDZ08#i1 xUTG>g1OI93z/OFvnVV^[{"R/pw;gLbbZ|7pR6 $lW n_(b<}na*{6B[ʓK&XqLWF:cG 6@i^ݗ:^7k# yBڂ$I(}C֦Uªl]e>|㹤Ok6 |t#sgh|8 oH$r=Fɮj-\Wt1⹼;B#qN>B͌ qVB/ ]tdy't~`:ťif;`\Y;wxn^IM9>kԥ ? VPv )\17f^+$%!}NAEmO+ GY 0 26 G"KQ[e[AUD|Ѱڄq?QMG#DFq`+E QWffbP3*>ϫ#Z6E>L'Z(Q ֡2%5)06]&\8UnLWOQ07%̐; >Ηey<|GgWEKF:g~JƱ˹9 p.cd"p%n~k_U2Ҡb#k Aw:jUqMpoctͫw))g-okQl,0lUoku`m#A?-{eZeV/c]/L+n*7غo kc]o7sT.^f*Q'h ƻo,~1_ zӌPe%|XEUxUoLQaڈ*1mͲt-cCn8vM?ob͞&Õf"J|<;Y-T%eɔng.}qd$ki"9}n{'wd{ǩw2l[gC~؝dm Oi;ќ6GU#|Sw {B/,y?AM>ES'A'ugBy7,;~-0ok( 2dg|lH“g~b^(cmJQ(g;Y;*4NŮeZqf8t;Z@m-IR^[#*o,T$ ,u!t1=v, A-wkPOD.%]Ʒ>5sNY]+ڼ6@Od{їwzxGPp[Mm*5+;[ܫG͈f|qXyi JK\Z>p.Mn x)(66=TIZMlAFV=#E #Ի{m}$Ua_[8-A?|?so,ա֊N ~~<X%7☓@&oLBJk"p}?dL h $TSͱ@b!*B"PIM_ cpLa\b=#Z B!wyhIalU:CKiH`JjtmI $pnK5M .凃xV68˦w?EZǧT?Nq {%[:|ι>n& lU<꣆ZR}3OG?h1H"/M <tmoiB#2̥Ѣ`+`\/Z]׳3kEg"fN8-~|LfHapE.7 wCКG~z1O@8v~B'&l}F ُл>ߙ,x R31(o;?EA c6pYp3Ua7Τ Ta7Iה'}la5о5>`°J ha#"emx8nƓ&mͭ RIcҌQn^rgpR`b^VC İZj]qYff +3އn)3iat8ErSVce~ΰh{:ghpw8Pq?tKύ 0KBy9&bKÙzIj‡ ț^ οy9\<7 Ľ^_:GgVTBGT<ԟWF 'S:eA sԜ(_x1Axְ]Ġ%#%EA'B~D |IB&81,&2)9A0T' {ف ˗W=%+od *W1- 5憢 6O)Ùt绿{4bL'3S- cNLOff |'pvz Y mߑ#73'K0٫__^]ƨ+&L}9auН ~C o.=(Y}Z݊I# eYUs]Z #PWm(ɪ$oi5JT)sm'mLZ0A A_v@t]gtE}ERRhlI4 d}{Nݻ? C*-e:OG.'׋k@qZF7ezY[jHɩ3zxA<:^2&oG7wGF_>8|Kڙ`ub,]< ]tb;-DP߉kE2(-v%Y!H 9W"@@Hz5(kxS3n^9Cjڍ *eQ*. m[YmozCY7^zz*{|ZSO Rzwkpƕ.X/G EZ 7#ĝF74(tUpUyo5o\g.3iCOb5ئmo+oG?>`iTӺ;?N3;7o0:%XS.Qͅ//g?4?py⦸ qϲ\LD06`sNF?;YE$IqvmN$hsqFXw'u6bpB7 Th((0'qW59"SeFVw*J?:C%60h<ﴉ&$p gLpVjcGf68 Ə|ףln/\ / >Q=XHe *m'n"9H1Hw{ f͏ %uVP[{X&୯iҧZE0w2rtˏ`m?uRK*ThR^J_n=t  !,K.R<{eFys3_h>c{FV1FwX3Q3DWi3 :Ggzh R{*nD/ %Xm;}~)C0D#)DZLD0bU y;w˛*jxpySQqyqy#9LyHzȘ%u{llH4̘ 7`tkV|rX٦N{V:䫍>*:D`hftғi˰#!&#${ӡ+i^rxwn!"؎Sq(dX"ھk3BcI.ǘ+)(><Rg3GajJY(,]iNLfsI31pZ* &"9OF K&DWYPq"%Rpđ'PSyf]3 lPCVTgÒ빶ͮnEujQ괏Qƾ"I)`_]?fOdeECN2p+]?WO_΂}ᗟF_ TI iwHy*b? Z QU:QζLs%;T 4`>U0a,T[fĹV܅?/.D$#~on"c'Q|ivQosIqwoJi`֖W.r_䯦3;emV tdD/ak3n>~e-NCc/`\E, v"2 yޟμY '?5mk[0ldQj 7,8j0ޘ!T蕲Vz0ٱ&os+`B-6pI;`[$dd~ Az#/$5kBdh%\sE_ J'ޥ`jxtԔsɕ4 PԈܽ!3*qY(H1XJ`Mx'3T*KYa+:Mcfc8sLctԮ]:v;NxޫYh?P 1Ԫ.dZ:.$CсZQ,l I1SxgX}8=9hzmrsy;nXUz"#5[Q  품ٶ6"[cһ #%mS4HbkbdM&O h5 A;ݨiX=gT-0:/ڶr,v79 HEs&I.D s&>IW::وvusڋtvhHvࣞJŽDmA!:z)&Xp!Nh =oA1e 9eےР uaBRVfGDBNu nM!T/\!1J$msd1YײDz*If!mTuRשj(A+OR0<, ;:M(%TlA(և_+.WnXxshdџ^,/_W9-Y{~Ptz9=?0wK_E}+c^iJWWJػM86(n]! uZh,b'XH K~4'[}@n~=ʍ1܅ގW-Ջb꿖W.5rܬ(yW $4&>^H9-b 8q4㪡XHVJ8. qPVb-Xi)at\(O"3ߛ h;7/|z䋥vPX?CSN~lzm Ww57neZ\z'!")U)N>XQ8PNXː*)eP ̓q `z ‚ y(֕qUsFVP JNFN| zس!(i{flV#ŵO Q##,3Bi"J /EAtRuxD&?셠1TQCC~*ZS Sj(H*4Ψ YAD֚cf\yAafrO@PD3ZѾ(JDr^@@JJJ@!@󡤮Dp# %ؐ<o(\` KT O4G/$RkvA}1ZNɃY-MxWjëDr^̾q e@xm*dU,p)&UfLP!bdKݖ?8$x NY{j Ɠ ϳH$rm?gweXܐ5&v./O_>=r~Aq?}2gdu T}=y+ǯ~\ymTD@pWj% S$f@,{!Id{*ԝ rr1rfo|0aKmK]wBӒn[bDOGЌ2uqVp^?1Gr ^)P4^`Uǝ G ݄"h"uL`/gN߆Yc+w!f5~[sܖH*X L".{%,^3#$39PB̓xur )sTSR+r K1F18=jTjp{D#ѡRAK1@T7Np(Ss ?#/g5Arll_  p8*qTlUVm$W˱vz5'NχfL"&qdR,s$7.H,@ᖍm0h8sJe,8f,s9{%cdiԔ,-- .祏Z ̜P3.^+9>;TesyO̻#01qzJ+&k{͝Aᑡ*:^cVm{vlT~@ !p.פ Nz vPJ!}ɤdRY 3Q&8FҜzVvD4ll#gi<(  qT{gRP29¥W( %=BqTh=[G);Y(d9pC:VYX ҁ^Z((Gam$7EidO ؽ$_nfc!ː䙝߯ؒK[l5^Zn7Y*.& Qq Zj,؂9+@z㩷+855ޮ2N{<*: L? dDz4 {G#$TD(KJ"-/uҜiFߑyQ;oy4q1p L .ZRXC NRkPNKP_@'W enFsj©A&fJ0*px|aFt4ͯ 2ֺ6Z@*_6ó bg5C[_k 3̠a:: 1&lX1\حH/TRd3qF`]QAwbtp6A8]J"01vzW:7<1B*tfcwKSԔRLHo}*m@_+.mmCg7v6R Mx#L3c4U C(* -zdPPe"*Qsㄈ "?8*؊4·S O|APJ{0@ҧn` _J0>M˷=N = щîکjȹO$ۖ%kaʀN-$3-%Vr3J*X3]P&Q3&LRP5 I$H%ٸs:4Nm,QCK8'Gb9zeT3A_$^AlkJVCi1]f1-~-)= :s⾔DѴLfw(+wzϽ^Οa$#H|doR[2 τyfBr Ve?gr& e"N@)6:[l9HOKjoڈ.:Fko+:(?NeOi(Oz, y2_a#1gh5mQg˔Dgܚ ,/%ސ\[sŤ|MgF18v}iFqcľ$rJ\IE毛9Y誵p=jMnnwwNyu:GU*t5"i .۹`䪯rkv q8mA$P=4_7n@B0 |k&P2&n'4>?uwssk&P:&auÎfp،$6z.SBpqA} BHCE<8gpkƄ*"'$0qVJTBnCSAPP++(%K0Z-wv湁|yMɱ_Qki~6K_.?#B?՟9{gҲuKhXg/=b4UF4]ýRw2͵TI&&9& hM(͔ߟ^k泧/Ҭ.dkۺodRE}7`i$ ߣ7xlG<~L7-2ڑ WH\&h3\R"[)D QM3{,6n>fw`Y q`2 707wNjP(38rG\inVT;NFJΉ>pY9ȥՑT vt#z&_V`˞_P%4K{CVJj_Lz 6.YTB ǕMoy(W^vMVb:AљAؾrUlӎ&ٗe)5W\6znf}^OmW~>}<>8)nOd; 4Z}{xVH.Z̎?=8ǻt$ w-; N/aStLd8u-/(*eaV6<bh-`H{%C*>UCˆ3T\P*QjLcMSe֛MuT`Xg&l$d~x t EHPrNs*g\sl0UK*u9Dp~9[8Tժg܍}^w=[ڐʬwH 7ykvQ> B<ﬞra>4ZU!>eכXѭw-5uk7{kn[盻?}0$mGˋ_ҴX^|CL%9C-*+-+8+ӿ-cy4כ L-{Sr@#;_%jXE-rQq76lx_\fɼsUxwsovmTORsnAѫ#(*\&0?D-CdBb>|Be1uͿVqRi S)%zr*OTz(仸ݳ.?>KrLߕ[w+wǝ,ACk_9svVr=|6[v Xr.'QtЎN͊YukZ >6Gw+-HwWX1: A{ ̴".e3uZP+XR;kcQ %) Қ6~5c6<=Mso =@73Eu]{ì77(u*Uua0YׄFWz *e:j:E6*)cB3)@eLfrؕ.JSzћ/=a*G*6Wp:qOX*OtGF,*/)DNj͗Y4GPF_>q#C#~72hCSs/!95("mS0-^P^`YУDi~*#ލǛK彤]r ݍw{7._E Wm|0wXVyN VS3% \ʬ^1/DV;.F:Tjd.l3z r=uc E;' qs* &*^ 1' (&(kFTdZ:NȾfH~*H "$z?O/(ZkuϹG|ߠ б뙛O|YR_l. rGM&uhb3P"5P ɃѠ%@ E<DžZxbs<{⹈̑-6SjK[l |0ǃq$)T]kz2E$3d5"Ő8)42JH&TA k` c^e_kڸd3zSIH%ܦO_sz}ZEUިK3M|DSPW*|jusFxBݜ Ի7ʵ$iǟ^kҎw#f&mLT c i!'Dz+]P6oZ[rx=68{wD <텁WuO/[cz.@CLEs1b]e$yt/2y+LZ)HFF\=?ᣃcy/QF@*J֜7cm!Tk tvի$yS^r>Uy>Ղ+bkKa;PyXRМ[6K#-ᐃ˃"Vi6yfױ4>B miqp˦2; VonP4G]̅&7ܰXZҜ)Y⣗(YSp'VeUjit$\1m򀡡_سO_'gE/y"9qYH;I&b:ln0]˺IUMpu{Ϊi) SRLte(70W61htLSB$7EaіʂcX”?wD󦦺-`V Sr dJ?.P;`A'9wQ@,QlLJus@ⶃ怸 3ΉYjͥRtYˢyۅ?<dqBY ,lA)wNPbZ *SVTRIGyE#5mEY I`, H@0-+&F-lLc0D*< xgA83Ħ ewZJMLiŔqdUf + ]JI`0&<3&.SyiΕV@FW *x\q kU¥) b"ͧ.۷ c}o #3n) gJxW2*gJ fI `]pWYˣ ~5lii`0 qUp–+ ' ]8E#='{}yfw ̷Y!ZG|ρjQmJ K h08䬴Ƴ4GC}IPX+wN]'UM!x)4{H  @)c놨7exݘc 8ΗwMΏaM󊾁./X[W+z/FiB)ahƣ*cњν0}fr讛h?Nr] 5k67A  }:d;OEǁm/pB/ޙQQ+>>UcЊhh" b8vNmGx _j:Z.DcK& :X|meNRBL==]bhm.DK÷jWE3ī0Y25JV5%=D E^R8M':&k a+ #0_Q/@L*&qo5؅}#͒}ej&K#t%,—_.- .% >*ԦJ5,H3}2)ռvݝpu?r >|jB[;6Zf?@Vponj|/.旀3Q{aL"7b+`Vyyǖg1C\\}ly_)D .ClXy 2F_f^O~:?T%FSQ"`p,(3U<"QxpE_&4>ZQ}_>j{6NT%ހ,apsޯ5҇5~xwk ^O> iniQ``x׻Dpyobge<{!̗32f=KOW߾H{}Bҩ\R‹KViqY/+B-)Ӳ\[[F=OUNѦJs8:IV4 X9P9EV6MYBsU/K$DQM:I2pL]'5"I#caJ#fL!}9Ϊ7yu#n1apKMsԥڨ1&_hY9#xV93I6$|E'Ɉ3:Ԝ( HǛF$''[J%D8iш<)2A{HRfgc䅊$8?0Ivn|3dK/ExKwnG}_.Vüs_^" S0MBHeh^JAiGH?Qp8pIWP=zCD0C4(|\/41/NWѓ&"(e NIxlTH9ڡg'$p]ekyPky?zjqty}g NݳS4!xڣH! zlI|(Oç"]}T5/O;HbsYo1:DӃyx/J\FTjn;zZ$WjoQ#L*`Ot:\^_I9q'ۻŸt >ނovN#!}3_48Jf^|dKx}J O&u58%5,pSsJst_M.TiIƱ7۬'=%A''"ƛA? zoh3) mǹj[6UCVH`TSg/pg44/SsN{8ɽ{3^n۞ ޑ6"@o7B(|󾼨*82Va>[K  "{E$,bFE~O0Tp)?5+@$*^u2JkWYB>OrEV&EDm\CwPpOGmՔwNv? A &) ׺G AU/9 ] (GZTl 5'vpbK}6E/N9i͠W\V,,jMaⰐ !WG",bgB ~]off.:3nK$y,< |3nj=_|8~~j ~yӅY']ϟ>hrqU3w`o|{c],ĵwM?W(xZѮFiB UnnU2Xw?ᅴ~Kն߽m?)lH+;P*)/VQFJ\b*]1[ZE6|E!0ĕ ӊ)Hɐ%4ʠ DQɍHp !T&kق*Lˊ 5` e#*bQEAhdaT^1*/0+N1VuBDҔa PqHfawr0H[N3!5fZPC@G6XcAk-tLwDT;檂hbCTK+Kx+S*)  AP E nqV;!@ฤfB҂ +f4 :6`)wNPIUie_ǎJ`Ō!Fpx+vAeL|e+ޱ̯w.|u'ƿnzr__ݣq= v5\)Y1U^  R+ϊ+WqP V"Q2[ eNqµ)מtR5yoP מv}vV0`WAkH&\{AkEïrB1K6Rp WVi (XR(RV2Q0pPvޫh/Iɴ~NO&)JO&)gS/seS$ĹiʼneֽbbB/_S g/&T_Ls HaFT+Z4ex!,cZ1'„ )/&J_LQPp/_LŰ_L:[&iҘhZk .no|x-.YDˡ^onvw_-%]2CK {i:?"51MF..Au5395rVVj)*B*T U;W0ʲu-\wi}vlԩ#g0v ALG]IR0+#e8b 컎312V¶ &JhCIU:F"hRUZ`ʚެ=ce[s$Lj"iV ly~s?Vo/*ٗ/>>5|>Y ~KU+岪fA~Ի`{[{VўXJ&_.,<Ɔ@Qة0_.OQ8ϻow쑅pIcI*j5Rx0ĊSdRUȂV%x(q8;d8PQ+ >kIe6Yz l]Yvf}Z]|o\PDg=rLq´_apAP-ۨr|}֒!q>kCf|w}ZQo]:'Z[:༣tt5 FakrG L#yZ-V̭|;Gί^X˕?`nv.dV-JbEX|7{ȷc7s˥[of4W) [Q0Ǡ9LԨ\/TdߺZ۳-&:F׈`-<5u!!g.A2FbDtBhc[xSCCB\Dd[7([ D'pV-=Ir|,ԺE3 blwϭcn41Xt4'dZr"$S>Ʒg0z -&:F.n9mRɺ(4n1$E4H8}.uhn41XtOEH [ 9s )FF3gFB1h$:cnSJneM[ 9s )G[o݄ET!&:F0_8mR^ɺ7n1$E4H oݔRcn41Xyh<\[x-SkʠňS(}Bv"50F rt΁n#S]I2ʱ9_6&pVʜ[U$Swyx9/&(AkC39[9g+&0ԵKUF=9'&P,'16`4炜E.OG7jDdrFr,bL Ƙy=c9A8jz1fN|-ǘsL Ƙ99ƜcQ5Av̜25E9ƜcQ5As6jB1EYPcH 9U$hz1fɨ1c LŘ<ǘs9&H6o1$1c i1bvG1泈1+FbJ:c buJc1渚 )cc9S4xz1fMr9ǘ#kE1f s1GֵK3s1GURƘU1c JŘ9>!ǘ!ӫcHc9UD=!a#xq1sTMDM.ac1SQ3荐zA|[^vޕq,e//~/C}Y/ ɇu@i1r[=I#:!Yo8425iWerͮM7p5(A"}Q1`~!LjdPpEny$"NhEh"rV9atE°YCr[g4HEINM:$/AQOVꑷ_dzBIgi;>A,qUSy>O ,S*F}}٫`u) |ɿW ]ɉ6:Q`YB:DE#X\aҎ`154:`" B32 hz- D%ST,pha{$BIP-!|z턲NBT!L ,aQ 灂'aZ`伴 $B8&l4u/v! 0O, !t< dM @J@Ɂw@76Fnd’ %)~AFm5u5 xhYGI,8)# c8"A:6D!^7wzP@+s|?Z G/_OJ:K irf . !Joutq]~ &T0.!S%E V{$0jg@o!B  /zޙa4yWwv=QՓ$ 7W_L&Rϐ7Taq2̉QdӃNhTuw3S &]tT]o鲙sYAW;e/t(Uy "rz_e~yفh6/*f12uށ8/2ʿof^\mM3y15f_:KUq)֝)OۼqM b[JbJIp]y |9 L%ہy1W{l;!O,5"I67co'/A76,cǰ"l>eq$~[P 5ճU; T߳]91]Y#)~0僉'a^ӻ DfhH`'I!s%-TjɤȝYъPZ;oB5s ˆ.Š 2Quͫo~y/5 IV'Ȭ&d\Y@*\`^nP5Kg!䛢H_˭rx/`}ϥ]w_Lz`$~ߣ8]89 x&f1lp7Td;k!,IYw]VQ̩K\9Ffڌ%7Z~  8qb2/4hƅf 0ìPYֲZV'8I݁:BoĵjL0)h Bx7 pn3%5c?8&FhR}Z=QW=h_ߤb)[΢)bf^ҝ 5K7:[&]v#ߍO>/s#S4][OٟWco>*HB8iGkQ@*8uQac0 )&鋳l<^]&U$ ׫EzxЃw]fcpF/SRJ_&W6*+eWA_vd1j-Gf ̪aʝP9MGeoPO2-:,`!u6s0§j[e ii_5a hL6_R߫sG'SwU}n?)$Q~M5T(Ey3U80k)houB!JJ"!ZQ&$q#T\>s3Sgo̹B'Y}[b3(>*cEzFs`XV*d.Si4e;MqdNuyI@/k8.U|sP/>t&#lT•$E-"!bi:JRQJ%J7&@äԋä~[ )i0.,̋j}ꍌaKݕ{wOSw]،|j\;I?|+%N~1 dQLyPtQӳ^E>SRE0""~`Y;ʃ7IB\̓eU|s;_ۀpz%ޥ5W^M"s޹h.\z!%,[^jegy`Q Z![Si輳؟Ptyk`agKOj%8S wS ϙp=6wA2IZ9uzIs eFi%܂ SlKC ~2 7Lrx ErĨ  =6b6KOe-a?CS#Ie*vt]i-RA&?Z*EX2;]gK[{!xQB3,'x$D4O}2{33BZ8o/09 ̟ܽe41ju~-иDTTV4|%pq9 >Ӟy'ʶl @L?ٹY5~#&Iŭ\w;ܜvJ$bdDdwHm0}~QI8پ!?ZfD@K0W)6-n??-y^vvܒ-Wpmv,Qmvn8M+]zfo |-v5يUZЃ́#g>™봠W(nH yL~iMh&jU953a4ĠƞQ:m^sZk&p70{삓~q5א^ZM*U P͍9flɚT:w"<^y6QH烓^/DOf_\5>(=:{[o3-F uzjVӅkA6+2>ZMhE0_r?RIE,͵v>gڛ\s4 gϗM^LZBY4v2`zȚ=O݇FljX_/Jxp[Xaϳso2աJ|ѸȐDG*TtRgKifѸmjVr}A']R>%vVǿ44H a~'a]z/)Mye7I!(y'(*P0F-a yaƓ3.M COxsVpG!퓱7'!|dkn4E#Q}Q0A/E_h\=D"nGrl"GZ j- eRnp漷ZKsh,!^[ 1@T?ff|UTC -(0aF=Ki*`a 2ȊhRiDń#h+Cض@[kWa6]"K r˳xF<.!j(;rãUz1BޠyucsuznWYύnޔ\{W@OxކIcb~`wIo|`yY=;؆Q1Ғţ&R΂0U=yjv+f&JS]}fn<- 12t?|NO-/I_ÂZRCSQt~]Ňڰ<Vj`"VK5zfNm5tXi86pc*~Z^I^B4'R"Z6Hr%ә\SXio~XNj&ߛʰo|TbP |l)y#Sxls1gCOQMH'4 FBe+N`z3enU`rO߉:#]OZg Pk0էո!7т("FڎXsr=%ҵ!J ] o[1bax!UTI"y=gN1R0 cN\k`d q2=\E*8n\žmGط|@յЮ26;ӯRXHROL2"s%+37.wwp?P>pKV(n8EZw&fƊ&VE΄਱)&qFxbDqLc .LX3rqB~sk}K^k(n:9CxI[MC*5clO7Qܫ^+G-r ~o> ]s-2n*UQվZoSEQ/*vM-̯+Y7ιPtgI%j)cNF)Qᶊyĥ(ک fs#%STIR|Ca{ED6"r)0Oiħudv-7P@lcDzBĢVH&bObObObگi=n"kOI Zp&"edGGQψ#R8nҶVqHGV)z"Lk7U[)i&D@ZQFR9|CznmɿҧQcjl'e'_@УXG_FDYHjf9H݌0sjEs8 1ˌOCi !zXm:{#{yBlG= E(X3'Qϩr!dj (sBk4*PieB)xMH31Rg=eC$^be90̠8SDK: u>CqmF8-*c|ش b )05~.WNM7sܼf;P`1߿"^n| Pjbo~<8u ""A8)A/ \r8˫0V }3䲧9@/_`  'i7o1 VU6\$p4vW1W-&̤z`R8kb6tF渜D Y ymDJ2)2e߆k}T"/؜ZNpEEaÆXb1s@{V PZ=^C,=xp1U /H4p&1LKWV;ET/AO쫂 3J~b& ̫KuScypă 9E-륑I߸@ 1"Յ5CA hj|hIL7Hj927e$jk]#p5}>4W :_Ÿ@i_e^+msoUM4\{: Mݼ󕬫mmZk(J9H20\{`y ^o>,ьHf^;=_N|ujY Il4X1 ?vgEONbGRSʸzAiͮ#AQӖޕ 4&=JصB]HI) ڞV[* C}կ+&hK?< Yky*Tp?~Q8e0B< (!)+QtjʥVb &e|݉jqԝ|s^>^k 1uTK"mJqI3ZX ܯHK+:z$IC73H=EatT8#; vkEx].>th@|zo*;͓O}VZ T`60 .> APO <ݷ 7u!5A um_/?vieW Ğ &1{\)~0_̾p?[ϗ]7@y+tk縛Rq+L{HZa֬c@kI j8C)g FS$q,.OݾIrٯg*T_j.`$L*QJwqj$*m?M xdN4>C{E1X*`#KmF+2j1%qR}PN[UnVe6+!rUK`Qf'mwd ǚo D1g{֢RÅbSa;GZ9)+P[:J=rܖ領`>N^DPDy_I=ZHZP3\hXԫ@P!`KOu26 #}J$H)Ҏ9NTJc_RYm" /%RrѺ\r!EdMbBG@J4AHecj[#3Y !"|9̊Y8lsM%kuvm#OzzWTyQ?Od:L&l6OM`z{ #IMHp}< U. >L'p} &;s1\&te!Oָ,I(''=9eh"LPA 9ʇ`ȇ17-kߐ3w5Mʫ2$]'jKcK22X ~4]|I&F%q7b+vzN-ԞVߧB'4)PNujf0OtѬc';:V{?{3{94.oڐHL^.l5绩w&b' luu~.NɍGVjp O֢R`y?" Q$,\Cn6_k!cf&Q-\+P,U2OLB)Y|l}q"CGEJ ͶtWaCj`lV!# #`0sj@"LYmK1#_v;rĿ B>#ibR:5 II-%Rg08i2Rq#&Tj`{!Z֏&MKX"62")"V  X $oF;IM#a2ft GN/u8RXE0;r9S}_푁J z?Vex僧SL 0/, OTưA\@-ꧺ^C0<-p6Y< dY=ܭ;u`t;Mt(<\u2AH\;|Ruض8 ??hx24t/ ~6IU1͗[]0m0g{_H:nC,ZO^GJ7Qp,QA ԝt{<+дԶtXYOIğy 8*~DV1Sw*1^ -j/4V5!r=PKrNvz1&prڙ̦Cu`'Tƛ]棙 f=M'DEos/>2˜HLk 1E1zzBS䶒|Y%vC$@PpRS̘R % p$s ~`yiGN ω%_c{7] FT0 ';i RJZo~yBQ~\HK8$ /l"r\gHR^$/$&6٧Q <(,B"dHY eɗtDTD\IScZbclڢB"2$ܯ#˛? )2I7gf9\$$d}q龆 +-Hު\jwmHeoko4\wYUVvn?T7[1e98_!E`HIXbi4@#no󹽙WnyHЉ6DG(UE "}"gNj|lr}#]|geH9y9[R]s$F6n04?}{c{1RIIeYvEy!GU<0FbrԻg30cz.7aw qĞdL ъuGz:~*_,F*Jp<I&~jT(Ϧ~z2\p&e Ь_$—ՁzIY!RipPtpjJ5e/N|S:(ɓyJ\X{ݯ:N78X@QLǙ{n0RNsw$ .d 8Ğf68q[?y0Z|u oᓱы\;,̊I|dDf%{}k,L;F ]9NIo bC:X_fJ7 - [g\ p{Ih.Z(L_~`A3?h7>Q@,wltldtC'ȱu+A8JPm_WrY۔8z?Ќ]$i zg.k_5Q̦]Zrppo~oW,jjC6`_wl55Z#2S=Y+̨٭I3T}\}J_!3ߎv]1C2u )7LTlfop$ b"*OV%H[iw=5wSeZmh-S-,pE++N\Z78kٽN:QHGzu}: N1:[7uox2FY4jN߇qPnzŸ9.rs3]21~q>/cY8~77wc_7Á.,QhEίkҥߒ)SR@3#]+E l!t Ӈc@Q2B$ ՒT\C:RIv17rШrUz8vתU,F7}Dj:`X=ͅO.5 FcJ8TJ Q9bIr(SѬʳ(w]wϑœ){qQ;b Z% ͒-9+OewJHcs 1U\Cbrs.SBSg'&W&.Q,7qhcf~q"UdNha˜ c1l.8a2 tǟ?۬Ydydzg5#S$ ƼRUڎ3|z[F%YByD=G֧vJI( G9GeX:}11l9dO[c`FKYJ? 'ٝZ߹q|!LYdAI#g_n>; {?e _e _6M؀@k(K HjO%BF_GY8J*mi4[|{3y-mV7U0 * V>FWˏpf4sn ":;9;< rjD RN hb55-jouxuUiSŷ mv_]#hu&z FS0/ZV][ɀ:*|"45ׄ.mr'Njo/U^2ۋb{<3a*ԼL]͵QTUYgt H֒@u|p^+ùW{WoI+X|X9ig2ԤŠ2~r]\W6IW~n,㭞/K_O --nTh*ػ??d1(JWE#鳿_]Z5,D \|48'BoOIS?!:)3BQenLIabr_>}¦Ұv7'XO9{$b`!͕,j!(j$Q:8TbDN¹RO0-/- C!ΕD$rc)JWb1ʨ%;ڔlI"dڰ ej-Q@7$kn p5r"sTkL?ñx{[}?Meϸl`Hk:,x' ےi壖tϟ;Cc2\w[? ̫UYUR{SWynjRS_Os3>}mRۇ@ ,@ЄuLϝz&$f;tr(0(~:\p;. вpVˎX6fJНs7]K;~C?!S}t> ^G=H\ܞWB!#tGh?(t׼r &QWr')Lq"A_OTծR`D]+D4su.S9mf+0Ω0}$ECuDGq{=~\ yí/dF{=MV/qQ N$=룇fˤ+ ~#;25S5x_IM BOB)|e9 L0!L ՞w<6S83!K]b*f iI58ƅi*"l =y4cg v\c4[}. 2h.?I [7=îL3J*ƵV(?՞wxbNΛSdJ7"h᪬B.$ܗPdb܌Xq-Zb7y+wb}]Bkh#pЎqƅ~bKh)XY 0+AZFrM%‚蠴@k8(S&กQHDk*xN753q\K8}5IsM@4czWx^`.6 !CO#W}gق~4sݸpteSndOFu]2&nu.C.ا"k86>~Xj!\>VD΢.&q^}*eHWwIbQfx8|9)i[rA\ɿ{QbM!?NyL5gEҋ#Sd=C&лm+#dlY DbhM eg 4N U%A>z;WUGO!F* 26A7 ,쾜c"=zx vݚ-Yl. `gI9[>#3H`}o23t2әqn.GVp6A"mukbw56۠Fc5[|<@0**MQM`̼K!BW<&4T pY'%;4!ː|6&%Qsq8vQx`~\> %ݠ:SGO|ȥOgf^q/'io@ᭉ>< %3d'lvHن=*\&{P͠&顰^XLZ.N-y`*cJ :+[fâvVh x*<(⎸xriX|=`o=C}}5jC1 `}$oKX9Ey\aOS]|jM۴:9;(Zpܔ)qS2/O/gj[ %Yc9hZ*F:x ԮM,|w F}~n@Bi$:S7M*l>aGe(n{ݞޞɛX晌۩1sKM\$ʤANSID+X:Q'G&$3MHf .E&@LU”k< {™[C_(գV""X f9c9'|Y; ͕F=x-fJ@(O+ b\"PsBJƴFc`4"ߏM˅̓G^~YDj*"]릳9(KJF}Xp|=_tv&^o܀>_NO ړwĤmHFp=q3UQbԮu GEM.wc{s./ۻy?OSk=g"&cftο18-9g _CiPY s87?B?>:h_cf}*vSvDL`дPvyv-\^ a3\zʻ!dxPÙ>nQ7r~s6Sw>yr%@LNb *EqI"Bvss\A_΂Ogg?#G½p'pSzP.%_^}H \CG/0}}#ǝ}mx9N|k͇n!ܜ</P^{Cчi77RI=L//o2/OOYdebt3Tٵ4 E)L2۰g_q7d4X)w Es+X)laj+=KBwtÝɡ͢S3o昃._|~{)xm~ݱH$yzǾߖAL҃7?nDx(æX]im+2[D\t|5OS+R MKS7Դ4u%; /,{^א#ZLIC*|&ZB[ѠL)TVSfɰ7$#+r2q?Jr_1w dž\q^+K[ѓs=`4'ԲwZcrVL;f;Yvݒʒy& ̉В S{PX#jqv]vpFkA;}h :R;0ػAQm2~2mڅ=wfX|9B-OgB.@3 ōPuѸa,ՐVX g :Tfg7av{c2ӖW_<`i!EM`3>4 Ag&0h64ZVo (1o¬\On=r 9V)EB`R̺J*ZRA+oagڹf4;Kp$AE6DSg7ǛVx?d2VYCe&#?nQnzyVc7cd_/&?|%2P: 1!NƠvWO=֞PupluDoL]rbs~0lAͶ0cIN=a;yjױMgi 6)طGnޞ4;埃i<#=񼮅ٺA-9YpT)ȁ䜍^_΄eDLJpmMR"a@=L.]"TJ BL-%˔[Yb(QvS54&\al“d|lR&.3U4BpI`NF4 4]`=9D`K5YlBF G[^ϕoCEhԺ1ά.gbkS%dA5yn@b$l}F˴N_ӆǷ6#^ |8 ȞZ\scdOt]'{[}` '=֑urH/{]&xl˲yB8ȴ oStr`e/r ޳a}\<Nt3dcm :oOweFْTfz):J`_exhgh|ҏn K_:e6m+( ^Uc[bqKFI}dٖx_E`XtsK$U^U*h;/7 n7u?(qS@M 45PR%N^#*Ӫ")\H*j-r]S7ԥ?u1d.'"O=洫SQJdueh$P@)Ϧ(ov2WZQВ{*S(E{ǮtC@fآݐ}<찗|CV~gv;k_NgJ1ۜOVza+=.J2Y)B›ܜI)ޯK> O>٫swri߈Q ػ);mWZ{\_?Ss0Sf0cOANij {<28cz1dk @=V7;=-74X` !9#ΖYgxD=,lgظq`:9wUtTAo9_|Ѓ>l c`< pBZR΀}79V+DXhiZ_>}LtT^,hz3Ufbk!kP. )"XF ѨlId>_Oox˿)$ ݼQꞦH9'I$4mpu^Bmm}5D%ORh>_GfS=Tl77Θ= VjلJ(㕩AT9k.jchsc٤cIh޿.N&Nl 8oB):O:9\E1eIEV.*eljtT"DRՅ)L"Ȏ䍥 滿6L1zn|s@C^ba,-/ցU]qMeEŶ,/'nw1$䎶kXyz(ߢNE_ފ~d~gӋ\~L| ]1ȼyk4=plwU Ic+%[X~?YRG~ Kճ*͘ŽR.+.~vz6[N2iOWcGaIo^kߨb;Ѱsq:7p JotOJd,`xU0aZ}nJ8 B!KJ-o$ #~I?B⤄ۓP^@ 4("WcJ؏h!i"Tu0E' xSbdMFfm4c 2WLJ.ʘwI%ehl k=?d5Lq-1,%|t{w{ڜٻƍdW`)22E 9SB[-;!Qܒ1`Nj}Ӌ߲K7yWtt'&o6W耽įW&GԢXk*d-VVBG\EeGXW1S҄ b,ouH1۸&M'lD#=ppLyx.\ FCOcc<ԙˏMu{d`?eea```2vL!*LВ:Vuʪ3xt|D:dOIyH9`66hLns'Em8c( I[N.,+(񧐺ج"(-W~MRܼ;=QhPzF:3>i~SF6@Lhȫ^|ig4~yqAF,ׂF@nqx\߸ia}HP!/Oqpe7Fm>ֱZ΃2e8@N9CR5(\遉|s4~j2.݆&` n `5yO_~yv\.p mSwJjuʫNs/`pRk*̍,p(5rDًQ>?UH v\dsl!f%cr`Xy 5'F7$ȜMg,p彄;.W:iG%ԡb ņN~ֆg:Hs||.쫯HbF+XH?s~[+X/RD>ng,js E l`c)^llIYyyF`}[7d#U+ܒ"ko1aF\Pq-%XP[qšyg1 D 3r0J x͕aa+CEI)xlFNJ8 RhJĦ(b(:aN+: xv4ck&>''eɡ'j~ )M2o8YW=bNح]][7˿3kۛ+7߬/6/<~-J A7V+ /w7Xw95xxKs? icl~nᔪϺ|So%=e_6\z\D;^jmep>1Rv5O D$Ya9gñ|zGƶ웵wl&~U>^-oX_'ŷQN|+G#S+褀ӣ{}%$ *bǶZk/FGH)mW0Շ7W fq{Y:0k$+1=\^~*W3sio`xDX +ICvae^ W"t))$[DƂH[KWXXVHV)GuZ#j DR9K\QJc,82LS.FQFK,I,koZ4(,(z(QeJBf C^Y-7^&MeR^ %5%Sb󜨥@DԲ9a?DO =Gnֳ@-lfjҝ2lmᠣd(S78,WJWx]wЯkQWnO%) ), .E᫧Gp2z* OBTFwvtxx_i|Bց͢D9u0.uݣi̡LM9RtGqK3} 53nMfn߸k7ϥ.8v)OftO\zcv^1ϖ`zl+o=C}M h3i{: 0u;)_$A 3Xkzb13ack$c. ̜&J/0ӆ2+VZrUεx'S(nl睓$LuT9)#Tcg%MEUB"`4iIcUJvB67bY(T2v*iQ$XULaTUI,iE6p]m6w(3<l#ldGO5Մ/c7R|i%wi0:\7;81KmuZc8F&c6"]N$pܝmRD'$'p9ՈsN6t UtdA){ 6K6X 䌒3(!A4jx%#ZK*IKXwޮ^{#OFAw顆u nrx.dR URUR,o*bxWht$~3YC}@6657ʞC60`37`զUSϪR283in!$ApP~&P˓R"0Y?t"j~da3Mt;ޖ[qd+}^G8icc!X-*YVZ m9d%;0wBvSٓ dpcT|Q |>NJc=Bb"+Gp@m]%t 1Xlw/!7@|w:xzqЧklHac@wّt:91y==Αdõh=P6Cܭx%zv)a(Yh=qr,pל3MVйzbݦ3Zz6Oҫ3-9nwnzztw/I D?:b~nt G`^~JMeGFσ27Ndd"}4|so2;^LD~ M}g/!ELT YjTlHZuv^UK%>t`5a}:nD_|b8xa%}ՄF ձe6PmfUs:"? m0`bqv! l.Yr3ƴ%!+uy ~cv]xw$[V 6ޑٺɭX=n6MF-7K$+b~ݤ> ˃Fv;6mơ[~GvCBr͒)"|o=vA~#EҋPѴ[~QivCBr,SBVm*~W9+_鹿,n/_I;JL-g'̜?}B Q[ē2?#TCL03[w\'q\dd$&m2N۶x e)(|vy B9s//l~Aã DC9dIY)Z/LiaH7V/d+tw{Θn|k*z5E($lல-I .RP=2Azj[_aHp͖9J xIQ*2UA<&xA#c\iOQdk({[Zns)oNr, x#ý5$)Sʀ;SyHjG1bGOf-(Vj 1'#dbj1Zs: &\vJU:6+Z9^(u‚ ,>JPD 7 "uվD#cnA5:~L0aYTNരѭΦiI5FZ`98;kGVˢ<ELhMK%˹E!o,9R :(8'wy[+_ hBJ_"&e]qu$w~43F>Ӱ; /FKߴz8:`9U« 6JjjUi 092v0jbEgyxz4VkkDN28Xy^)ZQ<ԯƸ6]kʦ(~gl~c?K{6Ych%`2c_gq&̝ [9 @Ҙ2KzN/bF0QHqIN㍙2w{9Æ.ӯ. ޞK?S\V^0M`lc$1Zzk[nr{Uۼ?h)I^P'#aKk+rWcU-ٝ8TbmM%, qqS-եɣ;f9Z/?7svNO>~0ݟ/ 4睮e5( (]-Uҝ5&›t|HR"qGʜ,*u IZ#( YGӜE9.sOQ3R8Zk~ 2eK0 C:t\m00E odv;r #a5|?ʻIhCP|Qڀ#w/qдz9FPPtI"0GVPԁiUzH7ķB~ӽ%*jU_-ba'g%BrmX(FPuQ` V_xxF9ԠL],ƌ{*X/ ]Jk];?:@cL=E̚rf1֙Ct XԢY25q^+ :ScJa=&eXҪ !̀Dx4E-&DSʠ{ Q=s\SAD/~ 7E ]wpQU8; [%@E^_áƖ5*{|'N0GCzqX Yl1j=K m%h,k@d>8EGMlf9gniGA)%3&ANFX'D?Fܔ X HV(6}ڭI{5hb:ȽV7?+I0κ1X|a,u̙HW1x`aw0ibU=OcII`$kM3rF8MсxB늜:apvE x%@a5ZJpN QWZJWwD$+lCi`g2}:9[s|rN:<_ .]dhP1:%%2`B;xgAg+‡guЊ>-C}&(Gzs g&l:'b n+N[-,"6]WJcՐۖ"U&?I픒)9p Sf|6N IWycdž?U>5&Pw e>Z̩ə3ha0å %`ep)()s"/!Ѯ9T[k2B1Y] BE.48x4@!5OAHt`1aFiμ<|\Ga x 'x-g>6&fǑ21yPޏI)ec>t1Vm2F j )+5Ǩ+_*!X6Ve;0` v}J,뱱 *\2'Jw;acO,$cc(إ" U;9AҔQ`MأWK6ZzYb;y,;w,;xtTGF:4{+ؕ$|1M(NKV2EDi $x[Vû>=<4܁#wJYn@(Cr¸x>J^P ~cGEk.Rd_3 KJt,Q)\SRWXm%⹃HHCzǜ46FڱNTUOaT$*PvRC +Y\\u+=ґ;PtY*Θ2%`?:ɕ^syP>S}; UE7nLO+z௝C}Obž ؂(x4MdX&) X^v)bMabՠl~>U"r {Ȇ0M';?R!럏KꏛS}ӟwgg]_˷}ӏe×_s~|3L_p_v}7~q3tg~?Oo>!~wOg?{۶?=ɾ7rYR%9#YJ)EgCR$;Ό {{~{ptS?Qk~t< 1uN;:ZɄ#HEcx?/Mo;{yON*Dcg- vmP6k:ql̛.LOl%ސ;/秮%>=ogݾo'_5l&fzy~f *><L#%2:?Y?쿝}^g15q50oj쟍/4=X\7?;<˸ptupOtΆm^Fɟ_3H6?(^':sy_TpoI9/lYlWs7xz+2rTÆ6wՀq mͦY~Wu/pf@ d Eˍpi6F_:͜p >ne\$ǭlǭDDj{2#g 5[%fTX`GS(ȉ"$P"kYȗ"D 5A@d.|]$? |DHc`#*GA4O8+cevL֎dyhmC+%^QP"(E>{yl*8cxH Ǫ UU_8 Jf*Qq"l\scH W8JJF h,#81^rl<SI8EX 5p{{ea1BHo,Pdq9wfZ b&Rcxh WJcRv}8AaR ^4zU,,}LwYbJ`UXէ z'Q\4Ƣ4" =6i \¡ U Vm|eQiǂ?ֹ&@$z'MwqR="% . B0z,R cPHi/)]iu +0Bhx7ъ| 緈%uF+Œhb&6oP#e=t}J[Xi+mcvB)L)</81[m!8ʃMuk5da8Xd|SobrHp|eOzio+UѪ \%y}v/ڎE۱h;__t1K>& *x !o33!Ҏt6\ ?áDj URRͼ\> =XgęFDxG/)U>u:೾Oiqcg&MD7{ͭyh -=ovhdu9DErRsHjzJ`(x@C51}Ն*W 締|2XDN+V%Z.i`3soh* TaɅ7"0Xڤk8)4ӐZRϭPemR~;VCjPZ5Z'Ze3 D D D ģ̮K ~JWW c{uNv >w7w;[bonE w_rpi~1cT[)Larn^1BCnjG!dě?~z@{2Ct2=t6H7$0Dzˑ"%q+HRC4'q!; wTaTk1q 8Ț@ *C%82s8%yc2s3,P|^C g(̴L :R!zE(%%9,BQLhKćmA+0ApȦl^ߒʟKb-$Kb-$_ }VDߘr-0iE&m՜U)i!ǜ>d{*#0 &f BShk)pK||MiWR~MtiMH iӠv(M҄J"gP\HUE t x0<2o HHbeGjF(+YoF q:{{Z.X.|BֻbyJLo:PuroI,roI,˽N U› t^ LR&oKX3BpBAoM=8&5 Dɔ?9P05 xsm$0F qh#a>ԝI~$I~$_&äuU`ZnDL[5!lEoͤb~4^eF\5>{w"9BHy:4utI-x%aF@KQ!@cQ2Ki\8Jxȗ%o֦ZME T =)0! xR+F%`$L)%!5T9+!&/*E80 =ݴB(sVB@XFٻ58TE%trP=BUlX<[ۺ}yS7XL[)Eq"nVT!PSj2q$]+?.m]Ga" qց+B&[BF<[20yE]y`̍uÏkr~Z0!Xj&M4kO_8UjxÓG7^ TO>27z53X]Qj;Jx+Qmb~aRlRyzX ⺤䣔,.&RRT$yy%%uqsFBKJ:cI'6I/ *I}K(vKQuS @WD I֢/0>YBr*>E^~.۽>{F)幺/.2  c4Q}=Ʈ/嘉W(Ɗ,Y!Ț$O^p%fߒ|2흝ƴ-3dvw)!NvS3يc9uWG߿}TcM$w¨Kzb%FaKx`hϻ%:z`Jp';GMs1'eF7?섬T* /'FfusQ::(/)IZ$I@-q*MsVr"^Bb[k3{53GXOq-%>F »164QhSMeMŵjSZl*Mm^V$zMD*#(VpZk>ZN*D{4 {=mu5HN$z ]Ipi>2謬 7f{a?Alp`A ^پ71uxY9?pna۲&}[׿{) |7pjgH0ֈiEb6@O1e\*O2xL9P,y&1&x۳PX% /˟O{3z'z(b4lzQ$z츑$'t' ON%ƹ!fp<|#q }j~tIxar}73Ɉ}`݇L̃Mct .9}{{gw{g3g=xag?}8{&/лGw?u<(٨5?:uds_'vpjL&GB/yрƎ ywSgwR$Ҙ4;=3 }0C{|3}ߌOZkY|z9i@xU|yxfO^0qf3~6~]mS9+rw\IG*dRP!{nJ58C0 $ߣ6cන!CSymuG|`8= {^^]u3/< P o~M_}uH8q=ӝt15 (2 #]etyJ>}_\jK:w#;nGT׵Vڛ~NkO7~bb:g1k} kuu@OTBA7q{2s;tLmG=?}A_ӃOe|Lqܖ^y{X$֧lE-`lI;+'~_4Ӌ?n޺WGrQx}pI?źs>dϣ{[o_nZ ;T<*H%B]Y9qs[h6)a*O?NuuEΉd laol^Tm %C>mcڶYTc;];g& ̵M.7™\uB%KBCtLD͍KL YԆr@$Jzɥe{T얲ym gENp64Nt S8)~FoW/A1WV`)|[ReS9&MC9TW8 8R.x=Y!^۴)?JV"0!޾oH%(#Aޏ[^{rkKdSy8\X\{W,ITpkThC[cH/*F:[c[CA]qI9\iyyՓx8 z{r vc ݔLi,Ow,P%ͣ]qf PōEl(Y+vKΌ7n)T E?Δ(Hz b-ȝڹʶ#Vxi'BKmٽk{#:{tӤiRi|9&MIuӤiRw7M:EiS4ČC *T?ƙRQZ9S*XS9m 2S8SeVG.b3^9(,vn;诏#k#3cBa1:AgvTKEll<!F.6HX\ *}mNu:1ϝtP:頻Ay‰^|m/hTArUB*+KT*p{H+HT"@{qk*?̜u[>dpꆵn}tY9=ʷ #\wcDb>Ke9yZ ȎDNr˽EF26!:>2st]cꖈ#uKt]D-uK]5Dw" 9*]~t_ ȕ6cuUTr_b%FR%rҙ߳Ns:Alɲ()RqۏTaR5w|d#cNXdfwQ1 ځs^s\8) I5v_nHҫGc>5; #{z媞5( { `GN7Jܷ:߿}_^l?H)EdsfS?w)ex탣CѮ"峚敷Gļ($u3}ۭԉ?gmRbFz/슰?\WW{n6aYsxS xɉ_=57m?8LU'Zֺ1Y^i FeڳšF˜7Bgc,OSh!2J,41=)x{0 `qDx@`p⮑Jg\K0 $ji\dU9%0#O1΀k,x?I+1E7 $=CS7 $uICqs4<|ڪJBO&E;"b JF rM ( _YٴՋ_PLG+!=\76x=wO9.MӺZBRIgh|YIW_D" 7r;3/3"IVhtʣ,­v2m}qfK‰}+{Qlh"7 %`=ۼkk/^+)ӭ5;%L[}3 AVڻTWY2f#< BJm)(B R,6)$S-#C|03=wQq4Ѥ6dq|L*.B+|wLU & J2"lq!Fcx\6wyRJRF9:mO#ŕ KAFVEDRY,^O\ 2R6NXGUXGUXGUXG5:6/ (_A8g"+TϘp{xtbrMױLZDr)Ⱦ-e{7ٜzxџnK[sQv4!mx-,2F"z㶨P|ޛA)PeD".g*)DR|QI@c2 *c2.i{%)(`Bl]6/AfXξV">8J5̗<Mr] GIBTxY!hAL 3Lh#0g,E)9)ҵgj1%㗃! zȴ7UB*-d.}yDm/$B(*ż * SeGGR5\EABf<i)$#8m,y0 BK@]V7A-.W^JVdմT PQAuzj \ц g"AU:Qx<)'G<12ANX`0v5=#*AXKbgEg-jUi(D7TQv3& ^ ѓ%r[0 VvYs8G\ 6$a4f!{/Q/b xw6)ReBā忔: :]eňz)`v^zD n:OGiY)C6Ĥx(KERؐ tHEi!(Y 8p`z&ͥNjtre5¤*pêrE@V< \BT2Du5%>EFrW$ ! pH&&D{a)#,b%*EDx%Y*SvoQ4Burzp:"3Mq&FIYaEU AfH8_LH"eGSvExI1oQLuIi+}=\ bh R7cP+&XH\&*,LsUFkSoͶCW* 0/T( $15`\d  #(^+(y!SKδBk$멜@ UTՅ76bBX,6dٷXMe$^jFM( nj dHaE$&2)VR~DX ,&Ba@ְXP"atM sf\CؼT2 !sWLNpNYYA\`!l!(i-"'m*)g)MWN'l#-X#o%U[+1hqqq']Fl˘.XPY 4 ۄJ]ԓd1T2|k`NP&! Wvdjk"Zt&|R\jdW/wʔU !ZRSH) ` Z&mõRtFʤ , MŁPŌD$e;?5%M6.vȤzGNI!=J$ڛT)ه'cbM 9bmQIS"+ xr`jKa)5ȃ1KYu8#Rqswԅ0{^0.ɡkㅃ"i3A o mۺ(ƓjSΨI 6N#1~/@Vaw'3OR+QU[XUk[oۺ(@`$3FA b _dh[J܌-\X¤%z$i}jAhp4XL·m]QWLni@z/:M @Fƨ9g,uI(OV#TԶS$~  $(H>:َ8!MO>[.>PkTVoޘ x < p*+l$k0 8TNNdA=| ̰.>Ps!0pՃ! Hz! AvPu$)6*8PI`D9,pdkT*ItXw "DuA200F0 b~}2 `-NбJ %~C[PonYO%3{ܶE%eg8`J >.60 v! ʻsoHR-q4쇡-"* c85[XEw;- i1u? fݓĵT=`G5lݑz@Ɛ'HV ,HphL2IVx(AzFz0LD ὖhMe8 #uj=;㌎\q> L< ҆F=`]=\Gp8 иhhaOR?pnǧwR,It9e.<|E%diK +|OۛT=ww+ S~}$*U_ٔTŌWꡛ/ ޓiK:{B521p`_s*VzvY wͣ_|Nq'xy3wi38h3`Mg7" O& (vW%8ϕa$TIzFObw-]Vŕ!FGyvU_hgvޗ[pp iٝU- Q \ND>Ys&ضS7ٚC/d㔙˗)+߸pV nwj6^6c!d%Y$~2\3XDFު? @Ht:c=YFO%;$nsZ$Bb>s9*sT*TΛJaA9QʙsHPB8'D%jn!'C{*K>%1ʜf. >ߝ5)}{Y}ĺYCGC01:u2\Gj'!`XdV9,HL ]"Lrsm)eĤITS ?Ibpo[: q,"`WK;,b>f~@0}16@бTp@͡A}ͥ1QJ.ެA*p>85QN<NSJqUM>f!laS9x*WNL=$ܾwBOp6ÁS7Pģ* SQX/&© S!#^s0KA!axevJi.3#^|E4lm~jd,\ʂT)({;u/ 琵C(Ab xI Fx$`i)b33}QfpPj~كtցiA աH\ր3B.f˶V5Ѽs8R\J^**#Sbd X)~lj+|em 0'n WkGJmzZ}cvGMҪ;!eV{;9 %bW Y8?!mQ+%ZpR5glO1J45 Gz;udWsagN˩OX7ѩZŭÕ F[:!)B8͕m] f TU{c8Jo{d;B[ݽ+5M$|gfocָ{p{`3w7&o@T G-, ?%oV(>RSLde*A8*ámTF CZK][$h>(ø&7Dd'*Q25`c,ୢZOeOcFY\& sSlht"RIyY,$#հrXRG*#$ZLj4g8%pߘ1Sy?[_jlsBVZQ]XNG˔ VF: &tF *8fc?Ar)@gA(d$IIXLm(aw$wRŻM稬|O )M 頪Ctʼnr)*1‘଴ܨHi B]N-WspLxJM)#`w!j!$ xbTm[6%liP)}ɮMLk9cQ]*0bJ5@`y/ZՁgUmtYQ_MPnFysxS*Pe!;]]gn[Ŝ giud7}K]_TsG_rW#!07: aIvَ1 ^}Ny3o?_'{v7jiONc탔kTlCb:ZaYqr5Ԃ|ګf,ڀ%Za* ^Ívs9S,Wy}-z[UHfs,Qƃ+L='ճ2Xdix$R#Ydpd ]kLgdMl`Z`P!i=` AeV+UNW# )8-68Rp!uQ!4CYGZ;^A5Ʒ40 -9o*6\<+Gz1v.ثUz7 {^"b!b✞ Z1;rҮAkm2jcbpCs[t/m==4ژAX롍UqmAxEFeH{Z5r'@)dxӼbmqGzgX|:l1 דcUaqsA6V'.볗x99YQx\UV0[M> T7ɢ<2O_{i1Lά: -pOvx Rnn!9T^aJ7NX#ҭ]X}]KMyҭ|(xGv Yz=h#ҭ]X}E\&݊*ݎ!9)cSٯ?)+"~sS {=߄VQx\LOd1N0ܳڭ/>{-_\G@t ۽}#lJXC:0pR1p~zAƩz1b> L58Z}͇C& WѬ|d/ش B/l[A "FbȲZ^)7`Ǭ%HBPo%^Ҏ©Ej}Ixl7T}}dfիG.Օ['h)]5|״b\=p}s,(ٺ^ |v GnrPs%*]mn+֧oGp -:@'jT$^%$эqw#d8$i=5I3Kb" r 4mߛRP05Gt݀$V#÷(hS1,yu6Z7 uvUG{m~FP݆B8.va%;á'Y\,GZmlm.cBVM7[koMnZ uy5ڄ勒#d%ԕ &Tٙb JXx]yClUuyTS;L*ԇ1l>u宻Qn?;)'𱴉_?J; vΩ4įN=Vwa8d톻 iH*ᨪuJ9꘩7jj!eCXtYQaq4vRHßN{]v"́gɤ<є/6pEND/Cx>_[zd)/,cJ GMӊF%CSxf,[Е&'ֿaB#cs8c+'S&:9K>4DO~c،gks.OtGgR h\Żf)/@_{O%,3wjL\_śMKOh {*ۢ@JMC`K2eY{.g!m_V>))ևS;9tr9ʡ]!,gJ1* 8NcN6٥`X=6e#ŬwmI_Qm!ϣ{ʇlM.uv]rb)ˡHe;7C90 LK~wQʄjk/ɟgolͺxQ].&+ge6VuRJJ͚..wQH.BW>^<ގ@-pAjLnpIN/Ye4,Ij4SbH.z=z2.i骮*iG%BD^"/t-5om瓎D]=Q(iau8ǜ)DQiʃШ H%2ʲ~6Z9eaz,ǙĚgQRt7ęްBP))J{H"eKQ4sE1ёl۠u9Ocv \滪bܐ(ENӻƎ=ۜ]05(?c T {D ǐe!Ee`瓮'>R_dre_y14bhbĊI^  :LłqGY@EfQ1KNB.b 5Q" 5T/PfhN@&Gn;3Tqnn2, t F"}se9gYO>D-,'Δ{Ж#K*}o1:Zyc 1&?5~ -E=>Vkᥦu0h+*- ΋dC*gv׷wu 1ڄNt ڃrB|/1!cdmE?@.sXfNM9( ݞ7vpWJɦw{0M<S22iFͱC;S+GAP~KH/k!O4;2d"˷~C?#Ol?#Oy_)Ogf2" "dafFo{h̼3IyW՚Ushq:*1 ЖT"ՊK#sCꡮJ {]j(D&tV!XSƎ8أ]^+ 7JVNvXE#2n,EDId#( 4f&qE苊 g,~hL ros@~OXx^ ɉGbH`Cʓ/=aJ۱9} g=w5Ms?}$رⓞ0t[r'hSfH;11ʺ\Mz%{ D,Y@VTf5_:t;h H]}ZiZkW>'6<)M1O)|\0l$ il~QYBR$ dqUH*a1᜸Z521٥qð,tz`FUTti0EV$l>NM1Hr"K#IK$ HXF (Gd9TKƄQ^ڥQX5z%#رn!r#`Dwl[':jIؙARlaJgC8ncUء*1C1M0|b T"v8a{LCS YzaD`G)^m_͇?{J pjTp}CJ{JOhOO.ͱ]-ɥw)=9w; ءyߜ;C#PVfDF%A 0Ki*U"ZQA(Nҩ0=QD`+QIF1 H3A,'f^IAl(!g*$~1a%LqbJ3!<):1gxFvW/ڼ47^4?Kx)^M5MݵŇ6.mV/|bҧjh5>^&G2M{QW4c4C st/aoŀ| SG 0cJ҄՝`-A iN8'.P0oe'q{-("˾*/P iJa_nǫ:+%qܡY )ҭGv~znęڰ;TI Ae4JbCD%EcaWЃRI'ָ= hz%]xekOG;_.t \} >N:%F^Տ$=irTN0Gr֛9['1eW,O2@ʺY8H--n(Z "# e+]X%VѠb{mH p 4) pAx nU`usKsLu`n0JIWJSp 4F^TJ8m&9Z^(AUIA_t%''3r2ЧWOuR_5V@xbCG Ъbfęt>$($4A&YQmXTBDVr0˂{E#ـODW最߻|zPR!?K~2%ޫ`; dtE cwMF`\vn2j U|ؕuELL2E2BB9{mf;\f#J =qD}Dyz5ԍd%lUl/W@.i,'vS.ױs|EHxiP,BEr=E(1Zk+Ô:4*eKL@-'1_& ݦibz u9 hLjq-jxYԼl.%g&$1 ḧ́XD3$ZrVb&4%qH6o "Z=G#؇]'f)q7یKrpd׉ ť{?2y+y%ݝJ'>m c}D!sTE74j]ϝ>ԣ t^DiV¡+K֥l ڎ*%E#WN`bSJK-ަvgvTj_ob/eѵ"n[Йl vXNA:P2 nTG/zp*zHkmN`CP0fD_s%xWJ^]g: sbʀN "Rbog=LsQG,ר97P8+.*S2Z,ZzӛryޥxWFL4  7?<3A5HnxI AeѾ1!B"_659(Wph*&Yl\E,t`LCIvQRƙDAU 22|>r貢 w|X!D!0bhCd2,˜4iӐShHdI&Jf))E hTF(i 6Xz !3_304xczme|t|qY3rv=}^cc߽~7?߽}ɷ޵0mcIt&̦ۤM&Ih ʒ"yl&}HiɦLɱڝ:"9O}|EOǧ_/(`ƟvϺqj~ӿYziz)S 2,\~Z`o{O2s3)Ng)yn6 ͛r/ߵ"waVA% -Z>Dsޛ5IZYӦNM[ՀLj-o>Zo݅'Ѭ?g'nRZ'SV6}_:52rhO!-m(SSrn UW/k û^:Ai|>u'45CzԪ 1J§n;'x-?>O٣m@odi0g\yA'U|OBWz ϥ]hp>,/AtJykNj̀7u!KW?> Fiᅵ2^_~~_/5J_>Q^Jj{f*\q?r$)@| T(K*13K9R0[sg51si cfd.-sSI)=څ4F3Qudcq~6;Ujz}T*f&} >ZP;)VDztr]Y>8jU_P3 ^YIg>1Bzb _ O*% Z)42v'goK[SO4k𥍯='L@rav29Q@?!y*uήfw/s8չU>͑{oI 2 e^m\1Ⱦw*n]_ÐyL,v zCx ;`'OqݡOpB_|*-p}-8A˴R'!Tw\zKs\護8L9sE%ɥRJ˜]P_oƵ& (QAsD-b ȁ~Sx ~)G=yK̙&׆)amӒI~I_Mn!'Ucv]~)s0X6"?@)1T/= o$owi#[*!qPžhR;:G.V( }Ԥט4'?ISgf6.07ptv>8 wR:p;xL0rk(D0!p Vym^_r=g:O $rZ$A A>@ rNy ޥ oa]05~ ڮPR6‰GUY+pBVY8_΁BHHp7kRI-R OW۷^F?tCϿ޳cP}pO&w~y`10 Z3?iZ년Ϸ6&hP>v͜ǻ$P\>~6$'ӇT^T8HTo$?^/qQ}i3lxu xݥuV&{7 <,e<'؜@sȩ^X奍_,$zK̥$C7 Q.~*8&Wŋ7Eb\-Ycпe?6n !{o&{~z);,J_:D֏`av[Aғw`%!#()U:x֡ i=T7r,6+1SU$9#D&$ZN"elJKf)|Z,۪YCf/K12aQ%#jR֙MPii]ߺx>$SDt#*m`_B#N|+H@ɯ>Ě Hx6 d i`]",,0lغH\%:^(~H%x R~H99WR~ RsR.{r{0^\*W(zUw9P+mX>b _o&hra"/]_οE6@1{UH@ _Ua!^t*.DO%P5O+ri ΋BAU(]^OJ׷$:]>QCKS23er0B)|p)'4?֊*`&[hxL86 VN)rYg IDy*OBr;_ +d7UxgcUӓ">̗ߺ^÷8 RhXGc]N(603}9U̅X O1a䯪rhr >YZeOp ̎A5on}[9 kI%5U(ozI'rIt!'CNQGLhT)M8C !x Z H4* <" V*m6]do6s0aD0M0B:;aB2wIj(/ݢ [f ;DiGd  Bqìf;)DIk/Kmﴪ/F~7ZXVRkcw)q'[HzrHV (pCL$ {j%9J6AyT4Q( GY@٠w9:0"d Mzp:}}>8Mg jѥkDT wYMkԶ@Rdow%7E+yB+yB+yB+"Zi"6{*9 썉77b0OWsII(ؙ5@GiYկ}93TyuRz+ctYi#"b.c{Ά #ϾdCE4{h#z[AD_ɒT/fX; B 3H za`Fb3_t'yڈ4|eѢ1!f(NQ )>B <` Z[ BڛJjI2~[җڶl&mJ9+R1ԃ61By)R@":WH+t-SzKe`AĂ5CCa ʀ!QִĨTkHh+ `CdXǭq\"`/"ãȄp$ƴw60ߑbͱVQAx$X(Fb!k#7;Rc*AE+ z4aL`I=,3=Q)"@]qC|]?aH2,IU6I"[mCD&QkB% ND~q$«i|&p@YӐ:`$B\ܒ1ιh1BB'dm-In57|LbTMQ1|-kÒ#p|a.)?[XFX0\cmx"̷6`GVaT0 zkKƭd k^Vkl1V59ٺy_wOb=MITn)Y0@DJler:!)Z(rkwm:rMN)-6+#y7)>66͖Xl] b H좖 )[rPIFtbyҰ5sh(F2 ^Q$rkmFE99ؖIaO gl6/;0¶Ȳ#˞LSԒ$`b[-+XULdh直&r;wX^ DFth.z3&ʈFkRM CԲxm@iV3@)]jQƸڨ5N msJ.9SMկ *Nǹ1Da8.'h֮7HjD"K !Őe<Mڏ'׭n|e[dHy zfsyXh`1 He8 RmNCk3xkњQJ(~b͎9.QRwǨ^ӾLVഒY¨.eNi ^4MTcˡ=#3Refp}˳Q{yx~⊸ ##xF$ݸzWjDeq:p]`eD7=Z稦s>ue$K<߿vV:Zu;=d;WWu*sq$U^^X;"Y:Ih{&jj6pj) lnKݝoq6F $U@" HH! . 'ψ'XcB aCo7!tx>䅀->7 oxBءt퇓%yՍȿpUKZv(۳bKt Sb<<}`/ 0Mr[QP9S߷pښCZ7q=A~&]7x M6|y]7g\$9Y ٬ ~~:ͯz|TsU%n! >DvӇI n$.a! I-k$S?->>cf8w}~Y|tՉ q'Us?u4U\-sX)#}{FmFppd|-+h=p7 xع~yZh^-u:?.'n_8Z79f6 '4A'6#D+Qf4x)5|YVK«jڝ^q׸zjhקx6~?!TŚnz{!-Ԟkc&)T:]-2b:48H3rB_2mXBBYE̬6I2pB#4G7)U +Z\\m46@%DjK:WPurahBa42eV"uXq~qM0QlU6f-% @*c y .3I MDm fTnIUWZ 8PiRXH,pjOzr$9(o2dq=֘w),32J3bΔTT|ael$-)F \~AZX[>B?^ jZiӝ+QNPcɌX@V;͏rnPN*r(DV;4CJ3xVnJ9%Eݪp^fE-JB\c:KM -:d5~qpg^xJ3U$)/%qBdH5yex3ÞfmQR&n' 'ggJƿh okr\o 7;4- ^2'a 3_ɼ3ҘejVKw/y1f$J:bp6|ԯFlF^Ʊ}1'$H$Zo&si”TI@4$LӔzD檥.Mib4%+ɩ![;4[{`gĜZ|ChWF(i#a{qSzɳ2ߤ0x[6 Ex§6 TD_8=kzN]s H iz*Z-ÜHHrr M*]DNeaxd\Cw]K&Ld ‘>zF5=e[i_ ' nq( #Ԫ05IhXlSIay*P%DIaUȁkYc5zq+ Uw5 Nt=G ۍ6 %ͳ9;[x~7''Xsਮ틈׭0ɷw nC4WhI,!hBԊm >.OM"Sa DRHA6aDjLd,:I`@1)jKa&LpS3wG,SqvK. ׍. 2l: n$I Q*aj)iʥ*UVA52i$#1Uub( +I[w6>\I"]3̓X6B-}3˅x:3NH!Wa|k\>:}թ˟Lq>=hg/sƓq{ocT"g '3;xx?>?]03n>w^keU 8 `lNq~pjf'{CC kes ^k)<K>ּ%<<@S& 儃JIT0,yA?9̰  p]؅ Mb*@=7&6^I,& hmYIbB xh%LVv+'_[O#xrYa?9H⇺jxF6rx گW氈[a]P.Řt0IY-90IQB QIbݒAP #ƚC{v=mI,%M4=%?E[W&oZ HU$e~E# ^kj[uk "ybMirhIjh>L(?/K8QR΋;;xF֪AV7W.TŴbvN :sW\]ڮ[-ӷ`dZ xdUšs\>D9 5Z]jz P Խ!Qsld3n(Tv TX6- s7"BMT{+!O=KwWޥE_\`|I-*\WQ\b#k]O!m>u${҃ 7E ?{TYs$7Zk>} ҁh.D[RB9f殉ΨZҹb@:Y5}ψ=#SØAAU<]ψJ${FT{zF N6#SH `ҞqThn:}|O5hҨtph=TFʘ.(\4{EXy1!US\t\lڇ2l%2m.ك5QC64[׸ r>K^C %IpmDV 3kQAoEj !J z~(&/`I:'|TR:sAKc̳So{>|Y[4~Hťk:abD5Z 2b)uf># PI 4܀ b&FHb0aSP4!&@T%HA*}j(9iE,_. 6bK [׏~M%!,>*Li[PQ)t,EGtRA=rDkn'9r~(3Ǽ@;?X`&i_6}yZf4繝3~|"v"gUX_N7g~x4~|;? 7NNZ|R"*qwΥż`bJͤ&N?9 *rTտCy`/`FM3Mssczv̱+mѿEw<;~__5dM=Mp~ "Wh#ō1I/k߸7}|S}!*q{$* pkRSԄk" "mI%t|̰&\/ʚ"H¹D׃YA2gfš+-8=Pff1k0yFx5qMo\b2f~@#ʣ$L&B"#ɤ"U"&RW2(IhX˸Dqɏ_ ז\ᄎ^?| L߇fx4Э6|%FE4& h-5)wo.C.=qŗKv TK..Ob+Rb+N*{г93\:U̝AFhdFmV >H̼ƃ6HiɄxRԯ˗jU/gȳRKp}Vm:lXĻZ+_ɚCsftolyןl,@ן.mdK֮o66KԼo+Ⱦ.bq-iZD!<`<|kZ1_isÍ'f4>l"ln|z2:fKi5ϵo( $g*=S' TL&Rd̪x!w VlFC6ME C]kY77rv} f.2Y|xBW)]v[csdA KHQƒe,"y9"(BEff6 %kBM0:6\Ǫ`cG `?L*6Hͻ>?bK3F@">$ b,b}-ەYIЀW3(%#˒m8OoA ѐ~O-sPb-=2c Lm$1dB ڨJg7K3ACMCo(l98/T(6ON5klU .@-B6AǬx2QX-͓2B5(p0b))ٔN/IlƤ[ fDBX`"uDPi6bۙ4P*dJj: uzLqlf6y.@ i%2W79rpkeXZ-H(Xb ws:G%IhFzY\f~݇%9)9Dvk 3'Vq7~pjntdp -4dm?}cy}2fg5QzSgMe fpi)=1euԊc#Iȿ%jh7v9ht^Ϸb"iƐq=^]WX{r_=!Kx5p]!=bQ)g>f:oҒFx$Gӱ̫tFukixUJU[ud)hbǙAǐSPAN"\" uTkF\W1E*hQF"3iyE7u9ΩowB!\?e@!~;ΚB r3909k,;zr* |1G#r ё7& [s'4 S r6kRC[dB*v&!ev̼Rx 4=Vbw@p3gz2ԩ.b hFْqM1W,QL 1HI0k'S.fT[^B祙L6^HG %'4%<1OL !-'N8Fb#sheɐ%iѥM?] #-T%y13mШ)3WrTCU-ƔsZHS9k~_̧1n7 xUB n2- i# ]#U1sj1s[Z{eL+:JQ9x:K Ȩ&YIcdK SINT?f,Lu^R׃AzF}K*Cгwo Žgۧgl[#VD8GkQZYUm+$}sY;zΒdV#dYk2JJzee/Pk d*EVҎ>ZV(6ɾFOXJtݪ:?uxz݃&o !YO$6Lw7_KjZ pLՐs=sOLM=~ZG _Ψq kM\MumÓo*5bs`b-UJ\#Tpu\3O`GkA58[CDԊjǴޯ*UhC_( %}rf'cs-U?T84Wu'D7y9zG?C'=mןmtTn2tv&*># N~Od #;(={}גOKK SKlo^3yZ˥[vtc׽S [=SwTM? NNZfjV t9T-~w-GYa8B|Yc'αi;(ȻOj!,4oM.1;uhDwB zhPﻁ;ܺ|aufFfHaUk#=9Lz:uN>\|Vw]k΁ӫ@W7hp@ג]}8j@Hqs~OεG 16>#|\z+03cL[}|rf;:v=҈ vF'6 EM_ 3IݨI Ϥ&!%&q-BUGQ{?f/A~_K*=rί.UB:vCcG^K98>j uv-c[_VI!±7W sYh.yc(ʡ_0+vt~*݆jxAT4,0^L۝#Y2feeJT &DC*<:׵w6#g{A-32N7r(ˎfmEkemc]` $f3u<Ĝ4tYX;)]hy.=A;;%3?h u,Ye{_Bx=Nj2W{RR'[P+1zN ڂU1SG>k5M=6AWcQ|{ }7hD9Z)Z='9 |zú\rwzkϒzqTNjm<SgD'O.$ы(&Op;VN˪qqq#s˶/[S7;9Ӷ疳m$Usl{4Ur鉓3'gbb 4nC^|2҈sZOg2Gw=9p&-4شޏ~XI)4FZbB UӂX0<2跖FEW.d]TeE4,R̢DIgT!d'm9ư 0)L2&H1`)[% tWTʙ胋dapZRr `P{E2!IPRm"ytSb9R}?$ v,YEgK%\\Ve&-)zr"K8bEvkNҼjcCo3::WoH~{R1deI, nB_RB-|Nh}sva/Q,:.E<_Vzw,ʍku rMhI"^*$%^ϧ./`W IՊ]0P E^ \WػC#8{P \=ղNhYGc?gp5Gx3?)ߔL9Tێt>m6t]sVlߝs'ߦ9t}B :d[La2)z.s<0P M4-G twvۑX ЖA+/e9Ǣ"g9YGHfcAjۀ% l ҦI6x<^0pÏnf>$4V 1Ba*mZ+FߍA2h[@ބcdG]GAøA#=f~k{~8zFr\JMuM[$d^IOGbOԆ< | 9hp&ӛ!UҥԺҫM)Z! [~wX&JMOd)r)d~>e|Kd_ߟ.t|NϪwOUᬊD|KRRx+ >$ +UL^Fb]XޑM%լTXlcϻe&)tf6Zve +w*&]:[fNЇpeq3Kr4׋嘓Xtq2 Z%0!3I@x{#s"%tT<߱eޥQ<e+LS9}ON&w %[|lYgU[\U\rN+"ziiKkE(``p^)¸yKvׁ:)쮃Gfk.O@wiDV+X*ylYY0`S4$/ :W9Wqil67"nǂ߇kV vfاHP&A ,J45Nv]j]T[\0o@[ȡ$'XpX4HB'AڂUb&N?t!P%`Wq"!J\oRR#9Ju;I)J&~h rRDJцU/X 8˚{h~I1T.3ݻ_-"@r-ofa|`QYs.TA**Au>)QՆs$F2J M:\(0(u OTLR& V>_m]Z,&.ߩ(;U-!],}RᏱ`hӮj TXPE4:zoTZg~hqKC"+D! J!qPS뼳2a'q'PQ9>,v7n @2`y1L3*E{A?M()iWJVuPt0`̈́oM&B~ ;Fj%`Э0TiJ;`(7{K,^h[G b?f42v(nDD^oyP`M9C ;S10eYPj0F Z`hPH('T;e1ۍBE QG ӄ/ 1 " :!NJ:yp5_?V3B(-Y҅TQpnrs6,^}o|ͫD 7oDvd:Zܿ?M!_? -H߭!d`nX%~vpAqPq7NcYP qן^q($ƌVtގt>k־[:|om RF -ұ,D<:pB ccPRz Nh߻M CD/ eqLhRiOVE;e}֯AfgCC'/| FއdwHd͇ lx6 ?dP28}F?s0x(=tgoĔdL( Tf`Baj(fiL ciM^oE .B;Ym4s/4MF/q%D&Udct:ꀶ.'WaJi|?M}.>~fM|[!`%$q)PQ޳ J`ј>8 xT[rGzzAP\"Ʇ2=~ߎƸ@_1^\|N*thsyHwe1(Vr:tI8]@'U# J·YCgdZmk!|Js^j[+E!B^%f8f%QTKd VX4a{̃`A\ZW/ 2C>Y1,pJx5_YaP+O~9;@<<%cW0+EB0 t$G!45;2iJwW{3a+JUXyݖ4?LW^"0^/?Й<3?h^BS^"+D!4z Nio!B/KSF)#R.rHwVYS4D, b-e υLrOA28Y#VkPQNX~s+~VRLD VZ(#rIMВJB'Tk$ H:rN%<2ˑFtikJ)8B0s9A{[8މs`WNe@H9df뺫DMCwZtW@`M G7PUvTutGr#%zA{8mL Ÿa îqwf3jq+&&`O,I<鳔'}V͓{BHBYzK| Cɵ*8@x)l)$րF j&?[pJ~?pі˰R~uɻ |7Z$.GCzs+~.>N[K4SA> ʵ饅Zox8;|dsyw$xs"tL^ Jı&~܆mFKry#Q<1ȍ-qKiTk Xŭa(ecf/00IY {µ ,ڞXi4h<%$*rةn,0Zg1ҖJ ,R|ؗsJpr0\rqmBDL 7)OK:&9Xւ3e9y- F8:4v!Uh} u)G1e唺T KQPtпrj'iBKj?^̹{;`)e&(1 h=QJvm-6Ce7P}94W@F*M /FΔP7&RG>=T0&.@mhUR"PEJXfzLmVkp8[ ZjI=5q2d&nTgNu(٩nqctAlj˽MŪt>Gn$hzVEs6|RSsްr#: }t4kڧsgfJeXoçjN5W$#䩘jfv/3eFImE=#v1&m~S(o^N`l@ycKH=^j>9W4}a^"GF9l@?z0\|ۺw\hsg7ֽ[b(rY n GE0,w Q^Ͼ;古E;щxN_4H7y{9A0{ +/5kyBA1%8Ȩ7$:6T6Y, F 2 Tw{9-wB]Tv blijd8D= gEr:eM9G]56vqYFA[5Av`؂M/_{=n5Yѹi"{?l? 7m|V76^ln\P㒟naB >fcNOtEɝvr)|/FD{Y^Tx;jdQ?c6,\`2##ma 1*3!Sf OV>]?,i4Qg`};E'n\Ϸt 0wF<{ qw\sNc?x.$,OCjANaڧc4x@d)xhkPI~6$wr$TP;/gEq"ո8Zb#f)(j~I^?-r$TjfZ윮aX3P]I_*]!lFנTƾʈ蝫=k[t/RE3S0juYg;f ԕ n%;R4PW*(bZ}5KoU`džI"4A2[rVŞ#h)spFfRJ۠W;2.˔Ұmd{Z*ZhLT*[ј J'Vm㭺Nݺ䗘}M*o7S.>cW9~n'wF{n+mnVihs9luNp{w?Ь{;U%D+#OjWۛG~luBQ&_okK@ZMPa s1>,RmutPpK {/Cs@H9h/Cғ#-!dz.'"W;ECOR~-h5YDzعmS;7=Otzٮ)V]׭'Mڔ}Z 3UuSvu",a>`7_-0%G`j@`J[uebD[f^^&mc E Ti1%IübE% KdZ3LJm82aJ 4VI !W (+퓤a,Ŭw4,w!unH%g="Pi,3&6\ZgT)LbIx⌉ Zp^V0ʂ4L;ޅt }$S"NE=31hgdYE)пMs ﯘ`F%K~.z0[/}6Jzc3ls7|3:q팎kgt\ `sE%eV`DlbWK,3B8(C3Ø=XCN%rvKs? }CnyXCA T?6;'f27/o TJ<[4Ej³z3_ꓛqNZ!b8dikfa蟰 \Xfe*26A|+֤|zv6wq(Ga~=c B6D@ixg /\>O8 Nm'>y'켜!nyN$>}65/sٴneͼJՁYs&7ۛoJUD4?_7aВ}]~_q%>)m 98XȠH%:N0V4MJ4~ QjbIyP7m)wޥU}XY #R Nj$BW j'wIι9qJD΢b1HQF4a4Ոa9x. RЎ *Ѷ;d<5Di%0,lT=qR|t22^ WߓWw^tx"T,<Âz?f||A;u%>[!MY5+^ui:诺} 3L yv(FyO/LO=EV3Qjڳ܁O9|];Mvv4ir1Q1pT% ?RrqYʴ&q*I"F #D)z;}QPЗ7V_ʼn(!9R)5<&qĄ#sK3ʍDX'GqS! <(Y$b4iKW3ƤK0Mp,Ҕ*j9y&=u-dVq7d( stX^p;gyh:̂PHVf0  TJI!I:ϩA&Kb8 YBpmt,D$Oyi` !k@Zj |jyu6|(~MGϔړֆY%"ID˔Q$(&V8NIxSB-劖a5Da W)-G =a8CH̨Ye1S,l-ʼnQ1 ,q o!İ' Kp83}T ߥ<_ez6-lYm'ݎjhS_vrW0Bow \jF81o\;[)ADDGE_%%?!o7SXVM02Ua30O 4# EX O=_b*01,GͤsLtrW@agU;xrja.W~%+ ΀khՔ^$U9a XD"q/+tv@HH+VNhE-~U]=Mmdq0|pn6ܚ0;+:YjrTY5LQ?iyVjr>Za@mT;n`乃2Q^*;NA|fp@<e1l7T! 6KiP!g[(c 5i|(P)wB|i`d8uV}0|tSЊ9Ok!)8U$6IJ|II"䓪C=;@)EkRv>ܛ\ŧEfj{Ƒ_1nO^ Ivf~[z!;tlwf߯hr[)厒 AڲX|"YU,*_Xv}?4.gET&Ɵ .{$hgz6/hj&Mtmb)c6CYr-tWBJ2{lp-/5m\%9k/^sYxiL{J"N  Im24.ú1n/'Il˗,d{rGuu/KE褴Tj s1gp(9l̆}Iy ՝?Af$F20%I\$9ȰLK%Ӥ#9W%|=Xw.2PNj='h&SK8gfG=p,e>ɄB ncAl0з=g*!=i!(hMq<.{L00XPX_—!TqP;;ZLnolZt)[BY 9b m3@FZΣBƺ*ۍiѼ>OIFW0$H@@㋉c8%}_@2Ņf%($gIJh8VyRƎ ėO#|6΀ ne୥%*/T"s*"e4icGNcV*ىQ n5^Ր_þh_lp% *B3wIAUdaӌ#=c΄K$4x2Y &Scx1W[!_*UTgk2}9dG=y 9:7S6RkEOgѺC U')%K94W:D\15)";"q!:g t+1CH󛘎8HrQs!p6k%]zoČE 1c(i4ם_;Z'ˮVH3I8^}pDQ_`S0[FM]wq1\pZ*,6O 9)0yʬB, !r9[4I @HK2 s4Ф/n4 J@"1yQCXNJyM@~p8M6ݟ\Nn̬_OH"3WbV.RX`\r0oFKT0ZԤaSb 5O@K'pA;M>,V+{Bnni3Xr'Xٴy}OмnS%X_&T*KߙKY9<* fE)&\v!22ՙ具 6¡WaP,D*Z\^wukfnխ5rݚNK$]%gr($ϊuR0dh0xUNd^*fӡ thuZ-|խSE迪sPK4M1)v93Μ;+;y( c2R`(e91! 0L᧔J4y,y,W=UORƻQf>>>} ZXuFsqQM:Հ =Kd60sPn"-|&;="JA2TdxH[P2KDAY8ͧnvnRq;g+$R6D.-()ͳ}+~p~ݍMNNIDkn&v\9n}wdgw?ܥwab;dsjqpߞܿ#ILN(8f{+^g0FnGFR'^qvőO8 q|);Dq4W[P;PZFo@ py(՛oGoR.(|PSXR)\!u AJv;ۥՊ$$ ZSoFDTq@3MǪj ~5"D8. th%kZHY6ގ`#U kDeq<~)A`b))xp9h(I HBF4H5V=U/eL4(OERbTκI盬]/֗mh↼現}% @S]{aQ+(y7zi6{90yFTukLm @Njv)׍q DH=rX"X` }.f@^qw fK33p;u"vZ^N 9XO6ïꃰqg ǧYnh-WG3ΞFee:bZ n50d̋wOdXHV{z=V? p-^3A ; #a1Y<rKW:O38 yB"!iu RчPSCp'8dؕ"g[7M*, !Lx6Y Л=}u0;È}!hݗ e&U< i ّ:`p;R F8ဎ}x"|426 +wn^8h(-}"CK_Դ6-a{6[qc>;bĊ3J\:>J#Q톨n ޅtbrG_Njj բig5S”aZƌ@*TRbL#cTʊT3NHi+ɌWC*9L \Fb(Qӏæ`M>NGSB(]~>:GxJ1? վ#¥>:J% .(.=+ʗ* ZqJ=#!B?_'sp->_ jZ %n$4ѝ*I^^SCkD$W[(A*Գ(uu5Aieag!noXAg\q98j}0~׆4e$hK |:|f>(Ax(r:"B~t咩N'^ ^? Pwb=x}49@c lg 6KFQZI !肐;D)F{ lmYxon}֮#f`3Fubތn}#?Ou7zuN`ϗ0ue'*}' ]'mWZV0Jx# lqņoc&}/ aT}8b74ı ޑg:S(Һ34}ZK1X!p"RXoBL3 6 +b_l6nýfP()18lj1bhur녥a$葙^HOpJ0ya6DQNuNSb|4N +U!4Ǿ6 "«S# s a(Kww9jl-&~p~*jy9[CZnĈ6aTPy׌ Z.+ NE".RSEE~83 (lyNLNŨt"8Z>mfDUc ufPn ƕd]<*#Z(8D('>UYp i btp7MNcP /$o&enTm?K}M82&2>g˛HF)^\4ONefe~2_ܾ(FO-3)}{e/^g够ݛm5 }{(O (j5]0ӥ0IWk{,]U>Kގ uҕ(jxb(PW*(*nj@Q'=I QP]"װo#J@.ԕ dpmۯ\7rwIE7͊u=i$BGϠ:DQ<,/3F9`셰 ó(*]Iܫ Н$ Lqi۰"KN}MN1ثaӈ}$?暢m8+N de]8KY~zh I~֫54,$.o~tr[0uiޙ9LۛX Ovv_v.'ۙJ^$~q\j!2\2;C9VNktJNQ|\&Y%Q\ƛ8;njۮf2f9}|vTW_sIh`A,~&x \氚+ߺˬ:Ĕ2|^ɨg28xmsiu)MRv>m"GW+> Tsp#Gu:~#->KMW?t Q/[PMW;Ĺ$ޘybk2:$quy*"~^efc/:tlޚuanIRHhmO <|k]TƼA2k9qXx!X^f_uدbOב&cu M)U 1(E!*"spź.1Żj].%C"/&]QlKH?׏E8I}hĪr͌l(c nI;{ 'dؖ1քC &c $n~k9d%A1st.հxэ1èyљD+SbfE%uGJ,L?JOܒ ]F /A:2|o\> tw1g-=?CD íf5 pK"tl3~i)=G0+h\eB2?Lwt k{Q5)acĥ5Nm%wXuMR,gښƑ_QeϜ]:_R凭$;5ufkRS.mMd+N2[OCeJ&%P$%V*3I, h~}Aw8GVuѱl8A_R"74.hkʼnJױuHYx)s2D#E,Yf=5,;M'÷ooN/{9VG Q &?ՠ >6&7s3j5Jx>x5B>։%XE\マ~ q {9dm-{|̣puZ:1J_Kȃ z Ɲ Dy׵|ar?|Q 2,VXD\gF\7ޝjJX7kՇ^5hHsͭ|mJuAml/.}W\ `];eס7^ZoYdĦE( ' h+1!˭؊rTT3v)Z?^ϛN\Q<ykG%U"Z}kH]%N;&VpD%oƛVߨ֨b\Q8#q=[]f1We#w錍_3)};J*j_Tj.0{q%Wi^6ֽ-Q 9kOCVpѬA %ROZ|-ԏ(%Ax A+^ ]w<h3GvB)֋;f:B~ٟuTЕ+Ω ihWN$AOx#wUT)yLNe!xq6P_U H`B]SGm;AU𯁿f җJ$l)}B:(`w L{1|8r{})&>~%c)$̀+RN)VRw݊U ]X2x)]WPZ¹䅺#/H%oi]ٛk>_#S\ܘ1370|1`|o`f%ބI?OsO7eKF%D]2٥a_\Zҋ]Ms_¤_`O+Ev60w2Y;4Du{$ߜNO#ǵGJidn]( Akb*izlx^怗9e9`4 23S"A3X-!1)(\3!htF; 0J6]Bp{o+ T1Ժ e.CuY &,B aʲT" a,q"\(j0.chk:Q#k:y ֜O& EH}߽]"j|tmſ~x6΂RYi9=3pflʰB7|FSαx (wF C Y왳4@kvhUUbbR\=aD* Ôkj꼢_0hULrf-)PܿUk bQX @\6N.G&cV,y DA~W遻$X"}o_@Ղ&~/3]U"5wg iPp|uqd{wFfc/߹]x Kh L`0Wۃb FDAi؟]X29=Eu)8Bj'Jy|H '/a0 '+0fcɠ]DݥFsN 7.=Vv05ow<^G> [҃[o:U|Pwf kȵ>}VGX%Nv+$0 9 Vb/bܥ\0eg =Nq^ZӶ뉪x{\up7ηw+ x3ye\^݁T~d>;o* #ys|xNG烏T6fFdJN4ꚣ͕eeDlvy@}J->;I~nV20O+7eZ+ٵ޹6G`[{s_xϯFثO};olIM~ ?߆/d?)090{u ֋"LUh攺G"Wq 8> 8> HBjLPoKoc7  G&# {?jݩQ JeͰe#%B 2@8EgWYD2Dh%*y匬p;8 o3qgMסYk$:U,U% q,7<>PdAd!0h5{^] gM`,ej!"hws0;LG:dr- O).[dH2ɹPgd^ݢLDFބÒt߃&.dFᢿ7ff4L,~<ʫz{]hQRC,1!%JXՈ_k+.Ű(HY㊅ t)uF~z?/Xo{ w $3;>6_\gSxǝE:M̂TԂbgz(Gȷ(*ϝ Ovc;Zi4Dy;R(I!D!Yt TVSj&RFR65ՅsCd˗.MµsVLL2횼.G Xk]wlF#Q PȌ{|".ҴSXze{3<_s315 gD[˛o@;QF+sw!%;jCBE %dlr{3M%O䞽y=.6Ps! -֌6 !;$кfWɃcٸ ›&"DŽŒܦ[x}w.܏q710 ` r6j27CFgZ6"Czg~\ťu &ܥË, ,%A %w7XΪ]Em˵`k{} +PWW'_%8E|C7mkXn!k)p[ "K"Qsr@M 2')/{qyV{9/\֏Ȝ|(Yl3rg;c '5 o!Nn*)z"i a^s~)-[BHSyd5f֓OtՅR;x K&ڀl_%Z . 01 d@[H D\ CmJyaí<}F 4JG:oXVrL,QjF)n\,<*FH}%mGecޥYlE[֭/_ 1),Y)8`p T$Np^s?5g0- s׮eʽv_n1Z操#8刈>{ZA%"MkVړ2Ӝ '2,8䧧k}= .CY8Gc䧓?&?2%%C%i %^$d1Iryqz~ rJu}tI%SYUa'0О忾jRrc(hJLO 1Ubeӌ KN*[J=h nwVOBhȚrk3 /^3{xKNHql,ׁb8 40SK؆H!q_> nC^JqR^,ۺm[x/GrP} 浮3uǃG uY80T4,~~"-NFZH\ou51&(  5{[FuCUޮf/"Jٶ/"ΦQ^{Dt1[g/_:Wk*\gq3B^w_4~tXFEU{xtmrvȟcb3!Q1\RZb<{ºR*2쐸5>HbƿKw! ߝ|@i 9 G4bौosHŹk˙]X8-lw6{`ڴ]s˟T{Riqe)< &lyK07_3vϬQ`' 딛a=ϰ~fhut)ƿ]')#ȧٷ1y6v?#bd?D'F!ڸit6_e_[KWCp5qVpDupvV4ֽ*S 7S);x\+ZoGN|['U淴W~SkSB׏~7ٵ@kń-SkE;;MN8׳:mkmJv;pfIt_if{Wg^?tF䩼j<`F=m#y3ۃ/YWg}\&q3@6ĿA^g7ӊ6AhaO5a-U &c(ɝkݙPo"Hv>?m0dŪǯLhaw ֆ}d(jq 24i jLnҰacػ>?^$nw|/|v~ WLlXyY?77o߼}y47߾:HPϳg~|{߼gyo]/m?v4ү/d0ez}>m;?WOAl4#۹wi7;Yۼtr̎ɇ =3 Vݞ~z@䳛C􀺇g5qO%4%վX])peB_8yϖ,gӿ;I弥<߱ Ŀ~f3:92]too].d= `R[r0ׁUʯ$oR]/>Q7ܼyO>u6eu膝g鯀d_0D t;A_C K?ty~rF=<{ԏ7~8@e~tlKZ$F Bރvq2Y};34cN]۷>TN7ni9GGr[ȿ`+H=~BAg+K2'FW-LлDdkb? # i-1ˆ;lbMo&llă dV0ԄT[_~hY 'ڜ7OIFzߠ TƜ .yL/ ^߂#BbNw}g9*3{9-竗?8PY. y< A'ph% NYEĀmcgs#+F-72;V 6-t'; Xe=`4!B䴏T?iPw meT̆ۦPHq<#WזƭSPcTcÀRP5Z N+FH8f*VKAV`_㵻~v$:b$VR*rԖ3@$-PX.jgX\x;#L`&F? 07qmtу| "J^T%'Wљ^_|v,Y7 Cs5@r{H@"% TqhsŒoܿ۬3؄b*{QoX\dMg ZWFaؘmeUju/TN_xF=U&VȞC!Q8~ SHjI:I5E$ ^'Ge}g#=1H`vk.i%)ڼǓ b'N>nqT_:D@uM{hOV{II 4*·xZ699eJC>XQBBcl PfQhh1*&1i[W@F/,[}=( P'Дb˒>A 1h RELS) %j{t JҚ6(~C^r_?umJAI}HM7Aܙ#$vB:v4 DǞ]MKB;׿uv4e :-Rq \S||tʈ{-<}_Lm۩Ê nOSwQE4[a<r2NJrVBm?8IBqŁ9)4c \QF/Pj:N8,b4f-yyɐ_P!5G .n T! 쟷zwsTJ QC {?`6AqojlFf;2iHa b_|L2HliXEB[7z݀bh,BCDdQnT"a!RB@QPl(xrDߚyՈArf)5vl+ôkJ&^1l >BTIl-m//qYH߿:[~:[~U\,0B`Sm*J"IDL# f$ҡv2X $U᪝?ZZLW$V_g101%F]^[r r r r iE+1dBC-Hj],L!a&#RPh vuW QOy[4[bjT[Gܛ L&?%OnaQT^Bgo^ތo"b\3"/|&lJ:V.mrZz *L58}#駃lxz=*"A'w7ll(FۤF1۲ >tZ?ktGv:,?EcC<nu$gtS#x>֜V8Su r-HpF^,xve.[_'w @-޳PS^~ZIZAOX~0-z(J#Jʏ`k N WN;p%(A"F-!WX[ ӊvT1\фjC9T,lj|حZ6}qvKI|rKor{nT|K[˵qwG(+aLe[u(۲]@hJ 5',;sWK"ԵOvmzZ65Q{YDdz= }[!Hc8]@ĐLa4#P p+ɉEԮ Ja]\!b N*c2Tm*-#Fz{Of.)mU˞xÞޓct6O)A8G۞,@x=F=Yzi_%ƽ2uku.SeBȞzJyb,vT25H_0\DB[j+sSYa cžm{'CoB3fw4߻wCRqHqoX#2־^v@>uNb8a2vĮ!'I"4cO4Qr}]:mH̘yNG% wSm0`pF#Ue =oD@LsW/T+iYK/JZ>hj=$rLx4ȏE{.bH.!X[Kd_Zef?Y?!=I?J ~yP|~SxeҀxm7yH/@!SĔD\N({;DĖu($> 4ޛ0.`OkH^ɧ/ 8?dGS,ylgxg_]OrvɷwOn?oד|[֢s 6V1ZX8cm>r 49¯i}<-B*6C}[)6c+Tze@Rs۱V.w<:8;lxRڿp*Ңu;wό@kC%hT9]BI}*c[dG(7zz@G1'u߲ZbH3]R,|+a'*ItWZ7_lriBXX:ƽ|0G.Qk1DwsC`O$UW_/6%>yFșр1%pjf2-Fm&ǩnWOFșǔdqK |s¹2p;AM;ളZw~H[Y<>5KQw?s;n'km$ W-gy-V;@9>؅7n?;Qq&6ԂF& 6 _ OEk~0W*2'첞FW8 @KFlbŪ20Ѷ, t;S,=h­8?gf z|rd=uqqnz.0{usfq^٘G10ya㓾-s@x*PliU]/opgseh.*ϰ48%Ùd9J$wuCv^Kgk*z=RA'_?zx!gџ`Hw:tϬOexѲ~䄉 GD/"2 Q[?WR;` pS⹛itœp26%d{$0nIEjC2pHFvl6DlGGo?*#hš xK{ۅ8 KXAD$/1 ^Y$-N|ip̿{ܝ~q>MhNj L;tVrѕfqu~~Ө3 bGqxDQ?-x.{#DzOtJ]i4\(pg .G``O.Ami.?.QYK/xڢBR(!P/w})b8ID ׭3Lr) z<$+ w$4QpQ1MPq(|z[T8AQ|8nk_)j.[N-o8IsY {[1v6cxX(EC)vq}Ul$f59s> m(:,ԋpL,8EQC6L5 ?D\4hpX4q!CvdAH2~zBK#UC~בz(H/hϚѶ"nk ׏5E'KVu_%e1}s~la JVrgi+ I }5LIP)MοȼE X Un&rdDY c*E!eJE)8+BD@XVZ@akYxJ"l/~/u/+w."wބ߂p@Dz;%$۶:1ҵס `ȑK4YvEY` t39j6nLׂBe|i$L)9>$]oW+0P y8뷉"_#SOEq`s+v%tzFQTifD5P |Jѥx*iCj8!j8}3!c7+9$on4j}bUH)NF;^YN& Ja SV2s e}'rV!'\f AL0"2xONf7Gല}G 69D+_5ȁۉ5^Bͩ Ed$Ye(TU"׀M!kiEe]Y|\Ze0fkex㣳vJ` VIqP9,yRұHfPxJ 5&V3΋bͧD+';`#.Vo N@I^Q2"/wqHD:vXIr` ѡҨ3xFcܱ_ \q u8~ZG}Ũ%j+KJ(jRgv9)d*uEr:kguz/u C#*Bˁk i*G4|Mi;v ) Zg4+tBA >^CJTLpMFEȁ v"F=#:&F{/6hd#8r RaI}`Ji85RҲiyE d({|Y , K#^InIzXκU[ Mn/a 1HѝCHzװ\JWEV-&⩶ Z"c]XNК )A t>EVdow+IuV־d n@!Jqb';ڡ;WS>?iQ}tKؘFl {ټ7lJ(HZ̏k9j(22 sYS-Zj|&ytV=%/J]3:Af{)yN,dfY12c}¬PR#"c+)9#‰^q |Hm+E^Xz'h/d'<5-˳R`qHYagug/9sA/[jzULE=}.£#<.%|s Lb%[|{37]-c/~˖s>)fRB2-EV9nff2NJݩg`_?zWX*b,c@ k"qz{w1($O>MLl3;@5}lVyɢ ;mӶAr޸.P׍ssF KR!ʲ %w"jPR%! &HWMݷo뮽E:izzU}{"VN0ңv lBAܼ`#<y <1JNIaTƲX )"LV40.r^*lG\hP(XM(ꕪޅt5ΒJ(>&VÏ{|[ɪ 4üLC)Hwz=e2[H[%S,Lni#a9F4K@|OK?`m=u$zV]Pu @BK] _UfCj@QW!oQӃیD9;0!ܑRtQ׻]A qy>8BnHHuP N[]O$μQ1cOQ.P\;տڠ'h̉DS~r9T<+L4{1?7>͗}~q{|G^V±=Ǜ-$ 7C ;~a-^A}cq 31,$c(_6"rq;_{#K>/_fRuJgVʹMe_2<ԏ.CP?6Po(>t4a@sp^qlQ!-ԅܠ*sRɿ\>LOu[=Лɇ:#uBt4uEBMݚ=a^9o6SoƏÜSMںm<6:h@ D89򖧤V^}p^'d3I6yƙd!$[w}OgǙ|F"PkhGhQ[LdnF~I0dTǠ?q 45/te|= ʟ'n"B@m|,c)H>j;?2(r Hc@4%$@\rs` PPVPev(1 +B?~+*)!ڗk쾼F}]v.BaPe VS+˜H3+,IZUQI+m#In4<"/~X`gƸg^vB6ǺV$)D,$abU_е"FJ?+(,}V%-uo}4i컧U]OJeKoJ+|#yVdLG6  J`gy jqj ÁVKb& b4~E.R~jt(%)*>Z>{M˛CF6f +V%3.&1F_iKt!A*\{Ӂh5b]Z!^)l0uބ)Q GuKrZ+rFi.HD'$)H#fAR "@$8 ZpbNNB5*?^8{叫^wRD{0Z)Ak րi,Zp8e\9ehIs݌<#VD]FO%xS낙CtL+|!T3") ',K`CrZ OS t> K6TD&1ňc_d bl`[@i!GDHX0Q [`NdZBÔ[b'c(AQH5Asɽ`(sX Ӕ3"M|.GصsLU^7X^I̹RK8{f8y'_ZMgL EMhTLqz "RpHئt12tPA 􊰄K2$8W3Tmr \wނJ= >*Ae1e8` w$yh`F` $VC[P/UC}XSW! 0/6Wq|(QI2|q eJnj\ 'qAU+C\)]ZRLֽz;SCҍ*0(/VI{0z2 rAb<=bitN˂aFƦscM\(p|d4]֛ xX!hVaN 5 0QiFq#j<D&fڛ [wS9`(l0\,*4VtފSZj Xa)`2P6BK˛k 6A}b;ZcfO"x{tUʭ<*299Sr6=stt"ڨ 42& K(ES%x_rfmz BF +:c6닰AUw٫S~7`Sru_W_1`# 1u.ñ$Y+H2(FQE_ip;լߍUV8%|Gfe9ĘQ)HU0LD^R. L1 4`+x*ooY|GLqnhgѭ`@jeu Y.Si73?Z></ϕw]i=[w!%- j!h ` PzWY} D1gq[Mu5JP?Pu}2!TTQ;TV;;等6䂌N'>8g"XI,DcPyږP=pl7>y/7{rWqg^yyA㷙[1#3ٻdd U tbܨbi5ڂ5E޳Q?ŋv!sVsͺZrx0Ș(Q5V!*аqGf_?|hmI`ڴI${.ۃ^1ѤLXGwo 6/"5";]lNa|`QTzjJVM@o/諶ەX { ֿߗLVeY_Wx/ IGehZ-ocBx|Omhy"]˺@v"V:T4רrJhW*VRNUh3&USZS.n<i*UݠBC ][ N0 }Da@61/,(SKZE{# DU*c`j|D-c2[mnmpJ¥L7JLAml5*ơ-1B+z,A1F%5#liGŻ 5 H6}*n#T(x8ڧ&ڒDik^=\#mq"w8iX 'JCGB [ `ø/F*qo됹qoqຬ+ƽFՒkPnG&50L\[#ĘՔq0-uʏDCLQMr]#~(V!hХ"80v:bwHvN)n wAIo(H%9ҝeI|Rf ޝDcurs192wrNjDCʅ 7{yt2sGGj"^W_3@4;]}Y*WgLD(*BS""IjYi/"5_5e>-D+’H*ޫ4TyoI(1ݍg/ 4ifӞ^~eje>nϪ5|lT핋1@њI8ny*B[^1cit*褕tRGKXCK%|B$1V@ȥbTX ^JvLeGy4#Z 0V]qG-QnKnP !._'@hҷ<]/HjXy&TgrCӜmk]E¼7 Ii^-r"pՃ+^5 B6q #p s]IhXa[L0_l_WT켲z!8uTk6!V#VN$wRx4fZe4.ů7z| ZJfHf*C4xS낙 gZ@Vْ+? ?{۶ /ECwEHSri%',u1-S$%RđI>3;;;;;iSPbP#LX ۔H-B R8a[l PA?RPY/vyjWGjD(]0BǜqITZO %0a8ɍ@$:"::@S\d%y6Oޣ?r_!{CٺߺjIDEI(M"n^VTe?;=27Fѭͥa'6p`a~Oo\̕<U<q:up9{C;6Z_;NTȲh:Kc@xY _ْC[>9G^N&VS/=@A?:{(\H`eDʬ~F2|ˈ{ Hâ%rí8/6mYhID7f]Yx.ʍP.~ڹBf,3P2nMXh&eql=: G >ݖaE0Kp?f7٨UnxξSW/_-mk瞔{_bټW$8Uͅ6X'jt `Y#1e(;! OS&d 7VF-Z,nŤ+D%# EK+̋臾63֘s)aE9/ሦGؐ$DG,i1RhaѱcciS`O_Yָ$ZʛivF'l56轣9Z=U6sh9#ܲi6z8ΨIXNFCV6y% Osv>[לV$.l0n˾@Q*AG  &"DE-K)Y{RR%%xhm$=]2)uE5.-mT) } Gs[\;-?'<{~]}cWmyy[n m-+Z`&qh SJ0&  XCE4Qa8%$n<[o78:-xQ* ^\tCk{O6{h@G!Wyugʥ.b1m|=!$Ҽ=a&F.,; 2A-ҺlIe6iuB#xgM_GMJ*~6yǟ3im,0lo._Gu|n \+7K?w4'Ůzڝ,Cb:Q0'2qݮ%>@ԂC"ۖX%x|xڰJ:šm[eseD읋׌&CYYޙYN6vPiJκ4/ 6ս3=OJ+f wԃP~ڙFiGkCPؿULvy<09&UTt}Tuv8z0=[IApq;F,b *u4S0},Y7 έ#M e'"-+ H#hjNJ%qLb-MARCKa%A3X$w0QPhpFnkx*@^=~QzjXKz9*螇4N$Me,#i<5ڔ "l@(8SD[j)@_Lk犫5(]5 .L$e1'ng 8bESMM,@F/!qޒCRi>q1E.:3 }dHc_qyw}M÷Tr^ih%R6/']abƳuxUVl lI}xT';y[e"Ϟ붖Գc15=2ͫI̟jxNK>ABio79RLjбE^){OlK}; *|> %{CjY{顧;(Gpe9XdΎkt1V$ &H3loǟ-H |_nHynV rzO; uVQߍv4dңj:Bm7^;q5 DO0լkI^[ %W fiکj0Lb-BoB n)%B]U@*iϼq ?R>n &u& FqDci iSQbSʄC¨ ;XE{L&(rtgW|]`GLt66 Ӽ`zg׷vS+Rʴʖ2+?h֥Zc/{~ ̵歁i1>: @Dyn/uwn+{]Cj$kSdqAU)UoCy~+_yW+O݃qv!P 78WR_QeJ""Q"OuWWuW=%ZՑkcկwW=/ y2`7K)!HIl{=lN:&z),IJi1VRna (DmIT!xt}v=.X 7`瓚! T/փ,ih1Mꠦ0T?ndbk/7 wAxup9b"A(Ύ=JhђT8 -5%NÖTRIb+H9."igqUsU8Z8wE؛"_&6F8^/E}{>$\%=l.5|v_EBmx8[(>O푈HڵdC1bY{_|2026נEQ.lo!fm.Q&BYsE6UL5 80y 8C\̠-QVyA0N<B\n#ZE-^k 3}I` ̜3&64$ujqjtY Lo G%ۀI5 yT\N 낱Yb \T¦a*%wRGOURL Hʐ^PЯFQh5 J3b_P*4h ֓€'XP E|o`LNQ7]PV8XPЍQN}`s x\^06x%JtT"4{QP-&,8d0;Lz4n9Ӏ E3G"1Ѳ_9EӒ AvH̺h*U6YRhДW l1 cWiP+nz_n8x5:DHt4;>/l0{_M.BDI H'}6z4Bˤ{˿"DO Dv"(Y_@/ǟO6x@:$h<btWs*>W?+;Mn> q+TQdO 4>v:(Ur Ǣď?s)*W$tK]^G:JzȊ4O_ a 1+qܠ7gk&l!-'ؿO_ķհr!pk֣-9ALqC9gy/6Zbș.2TkΩg x̣B=X> !J:f+c@h*MpmT yL pNK)&(vCdlpo"%nux[ûz#bwVjW`` @[ϥΔO%^7y^N*NF+%BXGN%(JT R4l 3]W"R1H?|,2pC.E/.Qv3 7U\R!U2X AdyG2M#/kDP\,+|VKgAtw Q3F ޠAf~fRSZӬXR6{"qt$wRt8 ^LY_4UCΡ#w%Ρ]!0'POy:1mٿf}Kg IRP3^UJcPAK,ɅRP`fGFm^i2/0fVLPI3M{8zB(8M6̰nMX/9Iu$;~v\(S3XI'\`A 7(x(5C2PZUp ARI&ASQڤF xlK2QpJ@#H9o d1yプ/wH:Z``F+vLt9u(fJ9R&HJ/ZS$")s@46 Q&dkv}g =m8)';3!~}uU-=iآVeWj1sJ5(&ӜutݸtpEVɤT*TDq(-P\z0 '.f fooo5&'F:jZUZE 5JEF{RhYPۢaVM.cM @Y e$C!m"&N;Zj6m|?_5Vӵqce#*UtӢ-qxUI_9&E,8[)r$hupBH[pVjO%[7&T+e%N)deyj}G5Zwu]y!)=zsQ:*uf:dC\$pk{x`rRZYۑI0jqI- H$2 ,"-'bϩ8M2?&iW4gß}1_2֖!IoִrR7uiqEf'^W8$=\|7=xМ]Up CG"f[G {-z3?X ua%F`KNGA=Ry*U!{K6p]Ӯ^]K Y)X_ɥY*cgJ8iJ hh_K,̞k(pM5J%c4ʹ,zXಯBUJ02]TNljj%ĐC'`iM!Z*属t.}]3|[tw3(Yͧk1^ȩB=u4tտv:I!58 uhF=/<>ٿ27C7(WݎM6Ŝ~yO>nnN&;ttn+|̤3+:h*ˈ 63ˢJp;vTI͍WL)Y" %x LhYBQ7wRr<҆r R*hi;}+9&_[ cXب/@a#$JU9}5]ĹW >huMeW!-%ժg/{t |4={hw:ōļ.qsti$T6>IaQ$LK fVPH U5T-)7-D &耻_gab=u}W&CumlU: @L׻$<uaiOl֧9s&$^,&<-m\1vl?̕d4Ih墔˵T;8"Ĕ9tk>^l:D< _JLIzY»)Ԥ5SI֧%Tr#Os )dfBdF: 엒s5Њ|`5Ѓ }f^#5"Xlk#6鲇uH^YL/V+?ju;}7x_ܣ4Tvp gR1.N-`k?R"{xF$jCH9Ԁ%"L2ӄ,Ip\h/=j񬨺RSQABC$&;;]^ %0zȣy')bg=dkXؼyB*@ tsYk[`?ꏳ:#@o JU${Yw/7.JhK]$g?Ђ 0cTC6\-w/A-(nX\rF~\قXtnu ¤Dۈ>晍̌ɹJfwo|ӛY=HIǿQFej/8}!v']\9T|=R}2y PeXw<٬6^ ;DmWƟRQM_@*cJÿwJ"=q㧗cD|]ۈz?=}|1$a\1C4$P2H6}0v5rd$&s?{Wq <=S2d>^X%VlXjr8ͽ`Xawg" nᮓ1q5cp!Y<+P7wFkGί^+sʲ.8'{8G9Qݕ;70?(}33u\c+9>7hT\z]M.,,lBvR/o`46 &&^MƏ ByOZ Д> )"h!"\ScRJ\>=7j ݌>]G 1I Mr5p4+;cxxM89,eD)]KtN9ye}xfxOʊV Zw1+i<]A"o |'y'9ҠNO_p]{!F)x:;ݙKh>d!N9X LNТG.lɀ i7=7=5+Hb#E)6RTc#[^DGV_E7&V7堕B#[U=Pׄ#۞oV׆stL)BleP!{Gb4ʚsc STr$j:Xpx^˷s< &۟箞֜)yIwђr&1ӆpNc h ZEdMgM`\SK \erN&92r"VA(qԥd< 44rOQt&Bռ %.L^'cޥNy + 6: N<"q?Jïu`TH,hEG1q{Gs=l4 9qAڀ&Ho;4]`)"6#DO$daJ|~zUx=\&80Z&W vm\(Z [ĻɷtcmOoG](%Qۉyݾ;lMyMD=AnΒwPLR:gҍ0gX5-7*JQ&L>&<2e `."Pu(\cېԵ#-r Mdւh2 Lk1J8QpPc$v DFfq[zV+dGVj95M笈v5HY+Yy es0]u!Lh bӞ y^F8r8倉XGhؽ/gs0@P`NtC4]}&}]>ze|wU_S9;>^I),`" Y`%Vk7\{d\<\_sNÉ[0Hi\1aK:<Gͷ{iSrf *`c@ 2@Xc(ŎP 1rX+O=d ۅ_v<_ogҿ (z/j pa$^&zG񻗏߽~>-o4lp`"$HQ-3?;Pql!A>=';#n߹3 ՘J?>?< ^7 ?c"d95y I?d-R#0{NЁst:p_F{ dbh6AK\)Ꮣ_ܠm+3 c' W'uIԀlIPS0y0we2tbاg{7v I{Bj{ tJ [ 1Ȭ0qb r0RHt`ʻc`nõ,lE8v!z:ϖ'3hŕygެ<);F G5:Ͻb/b͎5ΗAtB:gKK Bac7\$UW}/D6E16d1@b4oG&(52-9L_H5G=߂թz*ޔyï/f.J0~aoR:/F} Ć2)6Hrt>mI> H+ 5sv˯4YP AJ^J~PCqL\PCsQ@ իASc>7o ned%YULZD0S$`T kY4e@H`c <Y@@u@pW;h^ 38PϙCS)x?b{}sjj W$&P)$J;yS+㶛է  <߬\P{l~9?#MksQ]6H. )=_[S|0sŽq_AUcگY塻~AiǓ 6@wCƖk!G΄x!"-y:{0/>`k$mwΐXl1C" \̐Ⱥx ͐XphD=IvsU7ίhej=*{]/β1iv9 &xH7\;߶3 Aqp^暣ej18z]^x=‘t- XjX `ݵۣ6.=ؠg*s:PO{疠*: ]u&v`P?T O8ySl%$;y$CLT:F =g;"ߠoF'Dz_}B.t3\G gЙ~iV9} ba/~U6 VS24˞`& bW|%[< MQh8K=PZ`{5kJ [mBR=GTbHy MmrfZ+@JͭZ3R)"r>bEF3'JU&zf#(T0̲:hw0d*TSa83 iNN$95'im"++iEgtg/OTjV2pAV# ¦4 'Bx)pXzI؁*v1G*lc A 3sc*Hf("&ϣXZ޿5q'IP.c>v@sBy 0[t1"X^+x$]X?| kyhfZ& .dP>?'O_'gqȏkW']l&wc{X:)o^ (jgm(Y땈^,ܷx)t>ҏĚ$٪J<:*&P'yalӟaʤ.]LlFµX_h I*JBE57+g*E;pqlF>Ēp4}'y\!!:)~+wrA&[;S~0#?Ŗ}3ϕ;L)I~bAvql:ԃڶr#-r3g==~ wx쐙W<$B7-jD`aǛOg]_d!%קּL걁ud| ˡE!X`Zd%Q [NտXB0 ‹=c~55Ypü:K/V>:5o'J |Z7fԎK9p1|9!wKb k ᦆl*A1B H(m8E#/= j6]`&gY&Q"*ԗN_i^ϱ/CM T KH#*r44iT!h\2(?W BV$gD*KFbSav>P52͠"|j,AEATjwVExFZ#J$0[NEq^Y:ů.3sZ-7]s_qb.' Μp>H[^ Z;µ:(Tuw\pR"PB@#lgB;b>D塊H? -o -N讇5΂yÅ#H׃5vN~b1;/Z۝uݯKڽq:"Ёqk)Yvr޹)%̰jC@{v2qtJUX9zտ/β1iV9I Cݼڸm7Vk\tb.,=5=>}+MR*w + 1"s:AǢӰKW{q-;A7N`MpT&8 !ޕq$V+C`;kiB^10nIod]-f6w?s;]_TF&} j X,ݐ%1(jB[U ɏŮz?Zs9ڋ+/Dyt7^o' OD9§6U}Hy0GVqՕ* ֆrem;gZ*JvLS:PJc3;Zv>V| u_?p=5J22bZ>}'I;9H%- NlMC\C477DDpQ.~wo.l  U럙oΩ["-< F#0y%$ǀWԗR!"# Mj.MZS"nZ}JYJ/yTjs]?,M#]UkBajonjiXc8E2 f H%V[LbV.zuF ݖm{p H D' $q׼`:~9*l){^a[&_MU3\1wOJ~}"%.2 [EzK1mKşk4o.A0^u$&-Q7J|˥qVuܚvr㜍adyQy+A;Zܝcp^:fP)鯣G3gcT:jX{TPY|\6Y0fپ|o$/,Լ{gbr"in"IXUU5Cעw֫lRD)o@b!UUf~) kw?ZpR1FomcI{$ [C<ԼRy+4.E$j G0eKb{B'k!R$pGagDLj4Q_;S/nf crDKS#1TXsv.;m(J~l=>,Ɂ=AѾ$}m69u2s ~jЩ6q:3źܕ"cogBYpW?}v+!݀ M]52T#P)\uWTl%?vXJM-UĖ_ 28Y2d 3e9E&J7e oz!6hI.X&Dnz:j~A^߿yho'rQL9.~1yxi&pSސ?y *GwU*8tŃe7:PC]x!vѤY4-1N$_(k4D8lD!EqSq,PTc,G%p2ó$E.B D PJfLaƚy2Hs(!y$:i"Z4^t)\M̭ y̋p5:Aйx1M_U\\M 0NY!;͌t) `Bq Ϲ%td/UΕ*8"8[ SkE0Q a:4oa6nTs{2֘GpӀ{ h@ 3faK~Hh %.EaY"Rhjg,9Վ[bU}$[䞫cF@=µPW7u-;e5-dZT}Y۝֦ Au:apE Ǡ,֬%ֆTȃC@ L]ٔv.P ȮcHt!'dl'.kSb[-x$Đ`"1t`O4̓5(pkĄ+{N[HOn Y|'A;'l۳QpE9RUbs Ton V( 0rq\ĎnX]J;vUvG(X_ No8"+}4=y.k 쏫i>'Xbxcna='<7<Ɵ&mpJEߣ(N`. sF0J / AF)j52nߒ1V[2^E"BNBT J)DuQbGХlIeڷH65Ƥ̘͏~?ğ1NIvI~~GI@߮S{QTG5^7@*q:L 4:sIqes$@ Kz =4WMYz@>g/Rh|³%\@;J]4ys]NB:=Xur6Zk붒]"Ĵ"aJ O뺩ɒ:\ީ>Q?6tG$rrXq_)Tp8 Lܠ}\% oA(d-F;#6 ur s池LάN%c&%b+J(尭O9ь1u eJ#+t&!elP(Q FLZnE[Oո2>gZS8H`0OjXrZNUZ 5E'Fcnp5T369g7քPQRZ +ko* =Fq1IE%%M6f r]f, Kr;t73\pƞŏ 껹s-4M(ᶱi=-n23ܹk+S's ݑz5n 0aӃkdϨt1&HȾ96rka_g6w1>n#%ğ<{[[heB, z3ʓYdњbTvrdV0ݿ;fl{(k${Q~t^;jlyQiJ\7RF):A?vUo;#1KbB%(k.hi[yș?%xd[;h'_!Qˡ0/m@#{~?8k6e/./ś=J6,;yUdog/SKm=ɮ`K+ >X*V=x`^[&>u{<7t ͡oXQ] "M@ښSǞ^1P`\O f6{7<|hyh3&)[i88o+="=}Hk6ȷɼ9mHt;E+>1b,\chcol@;ڞě2 G4cIew5, EG?@z"w#j6Haސkg{y*)ȶF[* ٞqIEpɻc'p}7MfcZsľGc_bFFN!IۛQ(-SUF45}BI4j6ZyRk1pTJ+gYoK? /y}ܦ1SDsQ:˜1@҄Z\Z*1F$sAx"AT1CDC۹|đ?Ƌ/]'ubuQz؇K>W,+ ..(ιh!L5R4֬YQvE LuY#d( E 03"P`DpوtǤ9iNz66up !Jc)Ģ(2P&; EXp$&X}Q.P!M`aQ@H{'dj%XM-yI8FN%$iˀGdw.#X.%6 R*U>JCA??vKRN@<G%:H -/sTY:iSw 5vm^Dd+sZoO}/Kԗw_v'7M}y4ia[Oc*]?~' xoJ 6_ ԱqZqP4qEL▋Ձ3(鱹&q\lVjQp UbK &_ߟ6TTM'(ow0zwӝ!pٍWӋx1ӭH_ep#۳O3Q>|WT{&bO{mWh(kϪPVR1\WВUa#GyUcϯn~k㙎وG&(oDVdB^R _\՜p{!V6}j!Rgl:^CMD=\4ZmCx~WKtYᅺu8GL7}WYQڈjg:P(5AEX6Q#wY(ŀJ*L<-(;#\Ws\g偭{^^ƒk8}L4F>KjX%z`ٝdz ToVlLRDJyS0'#%%5!K,I[o8-ژ WKmPB  j-nq;K>ł%`h<1o0vm0e 6[VW@ɜ ҡ L$˾CJKSf7SdXa|y ^- J޵cO*R3K Op E &5,8'( UJW67@G|c4WlkyEz1=Nba@"v/dIPFp pC5pdPJ6%ꒋףXI9jHNvTHp̐ fH3Nޢb[*nFrz꬟)z6ޥN2Nšt4ƹIzw1Zw\I"rx8Pfi&d4'wQ'_pĈV(1gG4g|- SC#&'ESE״/k|+o+ %|Yɷ=6۹\\Gp3m,g|7'g'ZsQeדw Z]/KTKP?xEb3}5WkxU9{ ;l v.Csa5 ǔjĘԔe8OxJ\z1=ǞtFb0k^3Ĺ$HㆳD@Mp',%#mNBSyDf XϽW뎦aU^)8w6w!; W:ԅ` IS#Na؄ V(pB@jn6Dz\zROsYi,ҺKL$y$ kt kh2b4 )(AYwнLUɗ  m!VcZ56V#>$WVfmQ*0TQK|,d (9%Qu ށKZ@2XePw2T ,i굘L&d2Xԩ9 -^{9^;91 jQ)>qt-q rHI,v(=ٸsR3~00~n]*!N˫tgg0a!pm Hȩxs iH3)L 7WPx=BCda\?B(nofŞo}wiw˛.[*e@$q1O|p,#yx`~GLY 첹h-x+)lI% 6%N ճ}CQCE Oxn;l"*4ұ6e{* Ϻ"*ҊZXCFSay*kt^4uL<VI3\L~Wfp UK6Zn)5GC3櫆6+ؖ}/NU(w3'2GMM8b1,Ѡr>Y#ʐﺁ!CWc%F$J\=)U}gzH_!$9~1Ɔ<,<"?Cl3Rhimj8]Uuuuu]vTyz<A8t&&15,)h2,c)sЮTbh+[U ]:+i/bc{lhk#jmM%>OqjS8 eI=SB kJN5^pNo+W [{܂9'}|=m6_ͪdxK>x|x(䢻<%r//cRi@ʑ?:<٦9{5i(Kk K}[V!q0tIV!!/N9ZɆ4Up9b&$np]:TX{{=+~F{ӝ3ݸtx7^Tӝkb2:U ўH9YJ K9WH# i[<!:Na'Ynx#8(TNx)0_BU:/kB!b=BkU: oMVBHu! n o5,̦F44e<1,QjRIQfbAEk!0IM[0zwPu~{sbvT7]'jz iC}`H/Q^3PD,c JԚHvc1Hv2JA_E!zmF.drJ XE&;E"ɒRA\qN" &/([oVsǂ2DT֗sTi TjIu XjqFuhTw1JȽq3[|fݰ>wFrq;6EU9M@Gә=byؿ,_.5 z}n6)|?+]>ҶBoݯӅYq <#_"7 ~9-/g`M˭zو>M阮$䙋hL :F֣M4sn%:=[E4H|݄{nNU[gN;=[E4HyX7u-/ vGx퐱 -읛^O!knt!C+y{Ђ { .}y9}`.cȐ+ZT·1X81l3Ӭ.>}nQoxV2&N TSvm)Ө&V]~WCT2C?EGQn DKsn[zQE;cOfrH;̜oyYnd4gY\WCXswdW!D.e`sYsͯZBcIG6gOhxe~6$<{/b 0^P_ (xv!5հ#f]oBH]vIvu럮/F21 Ty&.O\]_\䀬ڢ&U&w7_OtX ?Hzy&1&ls (sskn7yerڐ޼Gj54>&)$ lM6ZN4. T:eSasBrJ7.F`%xvzarJcTOS)`S] ߶ Alah`kj!N-zli*,"Ys=[',"a8kopKfsM^?+V!}ZME!%Ո3^f3lo0Pk ~womn,O>WuElol{.^l_ J">`4fE[k7SK*"~|wC$KNRR>xaf9;$b_a[)1)ǘgc&4V Gc[ ic).\Nކk*w _# .qw׎UR9N*aݥWݞ=E }c!Q~_,$d3i#jz_P"`c j)Ǫ[(ݡh {jі ?{-A~O-lj)ItyY+Nrx,Q}ѷ*'~޳:r+E]RѴ~1AV78ո>{Li)}9SE1k◽m•[7 R%gZyLn>JceW ƦS޶Kk4;Yͦsduouc%Jԍ+Q7^g7)1"2AlKXc$B4TQ$ 2FaA9N$Rχףk[;/z}O= h[@޶;,/nB84 H>ت 9]tl8K|Dm*Q*;B$8cI"L?TۏTf`NMOzVB!zVhJVZ>uhlUDvߍ˻߭:>8 +FEϗDWij;+^mǾ=yW9м(6VV۞v5b>Zbo2DkOל6PEp ufT %R4X`ib;De1:oТm+MՐ dQ͈n\/|#'Vc 7V&oLXqe 0HyVptiC҆K"b(GqB%BWQ:PPF൅Y`y*ib\JC51q$T [16NR/TSLhgLȓ] Z ~ K;fM[{:Xhi%fLD!j <"ra_ L]>K`."P%\"R(%ꊹ>Fl}(Uj $*(DXB#yZRʛT=,íjC͒1ԵDnri̓o7٨Mi`[m2/ӣ~s{ZZۇo\ :0f>\_=d>H*[~+ "gFOo.(\[竮;r"b N^M'4&hΧϏ/PIDô}-sxX@I0' $[>[R0`DJ7q)ddbCDG1L& '-S$hE FBͦc$Un,3dX!bYҘ2l<ͱ*f@-1+N!edH+5C:RHeĉ">TAY  L"#̲,C4!&$ psu A} -1f;m>{A@p5$;ON5'|[Dᘌԓr:u P7_]3Kpv t:R-MgD=<σ-Еr ~5l|}4-绉M6AZ Yn■!Cgw*&6*P@9'?Ai~PJ(}jՈ-`SD jDr4gAi,Ifh"|j&S5eD&-0nUM h:_jbVG'xv':֓aj} @T'990Mv@  EV+-#-B;\ qsu[il;ll1HhEbW!"͟ju;[?@LCOf+Ǎ'?hyu^в$ރ1ÝI6<+- ~q2|~<;dnvi:|^2mis4'ø%3Ksn[Q-uj.\ӝs8Ca2QpN̥j+D0A! BQȹiI0J;l?6mVQ=:y{58Du,<( m߀\*}v.~44"UX(蓅=rWߪɲ*"*>]aBbXjn߸\AXa$@->xJOE/9EO]%8VwN^1d)ûA"0jɫ1kI4n^1w*0ٺVoH G_on~[a_|U8J$ݕB)ywmy 1quǼAXrĐXxdΙwL@לS Q5KC ҂պ2u[3a.wڎչ7m_*8Jxlth 5sU%^̳L)J3JjxńELG jeOHXe(b >OVM9+`{sbvT7]>ՂpLzB0ӄ~mAfHPi-RLƱAFH+4YU(ReI*ŝ-D589j$2PGcL Y\oylǶn5T&B.̑-!Xu} 1œk[;#(:BKUA(/z ^+4} OJ*RǷ?wFieȩҸ9 Q"˚kU[> cG'wn5nE5?`/'0v;ٺ.n+Wr=޼mZM?oL`GPߗFL,`wM /`'/V DC40S&QĤJAr/TЄ إ`6|k1k݇^;巃bzpcό|pc J“gÕpv֜a yZ X `7|VRu^7pQY636 덽j[wUt<_vxO+WfF$Xt9(1/3/ ,7=_.f֛{Vbc9Bd,RD,x)#DŽQmLfBƸD7*ᯉj)^T*?{WHn/EN24Ov\afǺxmǒg& ߯ؒ%YddGS"Y&!43Xu7xhPý>,gǘ{{q~c\"Hzrv1ztw+J"~@ A"0ʗ63-mF Wl]mX(K[9˳G TY?K Ose\M|9L9y-O~pXkpp ^v2_Sʣ/~4ǻ.7«+~v@W<{X #| B:ﻮ[` b*}e2gw_ncRT³vP}ٶ!V#_@m7:D0ҺG[ +) )YY/ԂJO+/}Xj2?\{?y]?NB,;<;bϯ!&|&A˃|i1wh>x%49Tzz&TXPҾ4KgS`m.E)TRW9Bq2E2;-:1&&ǣeB5(T"b34W&NKp4G %z_%tfHA5‹Iu n4B@=/H28mZ(N{,˜YI(D$ZbO˶ZhVG@koU ǺqֆHA:AzcV2kd˩x5aѦ.'Ck-י 5(t94&#QcF pwM1Қy>H#ӡ$Yn]ĂA~OqU.Z񆆯4-m/I#$ AcmCAdzԛ _Rb3~^A@l)LIM*4)?mk|mo澰˩݄ݼYrRN0ͅвI)U;Pl"daf \}׉ʻl8m-5|\ߝ5QT60NCU׷_"wΘ16wi'4Y|B "OQ&%9~fBUWm71'1. = uk93J3\8FlΜʽ!Κq!Stj=fZC(麗W]v\n Z7PSͣ_4㭽ҴE\ѻ6@[!!>c>{~1ZѾڎLy*{|pS451zTzΕkmG9a ǤVPz  a˜jgF#Xcp߂ɛ#fAW{mYIw (]62h`Or5< lv٬Zpu:m/AUIc8ϮԊ rWx #bA(mwL.uMlX_jrz6LwdR5[ˮ4cRAiҕ&hрMI} mظ泑B -ψ$5v$J<Μ02ZX (GC2:٪RDt&p4Im RT%%-IՖBScczcSz rkR2^p2i1oRi1&sQXrt49CQpdz#pL. Vg\ GqyO?+Bh$S?~o7&s>P5-P-5@ xO=uezR`k#:} BmU8ZuLkZ"XO׵f "PU /9Xv|_׶QÈŤqk dar&C4{UPNXe/zI׍wths87'T.uHM+N bQYS OfjS}e,Fs F*Xcd*+TyM.1'7etAuDKIbO(TA)EwR\9ޔ1Ҵn]sO կf3 ʼnȂ 9|y,[Δqm@E*')ȧ<hcjoo2 _bsxvxtWIpedCgi%Z+ٝ3Eo&Fު%ԇ> @CqGF54M:>?4Z-JIBZd./ }]y]}1^CsL@*6#rc`|Q"bO5X6^SȨ9ޛ{ F*inx) *;b|Q{!7Jp 9~>/i0$ (~7tekKN+%S i4e% N8С|Kw޸🰝~A& Azӯ/@,2SO>{Tv %EneVLs ˈ38&3b1&IQ0,E љ5 ۲!4}$l=,&pV!{ۗOCcͲ|dy+:_^My6`޻H͌)zjː >\:89,Xn͈6T&'q <22[r9J@KM.1\[nFJ;-'>@28?olX\T ep)ܸ|ZH8![/몬ՓaVH|[~߽?A@A`Ў0:֗/IP Nvxbtj!`P)+]鍝$0Nպy0bcK'ks|#=ZiFBRbt2"{rsKoytcj̚;F#"Ϣ=Fl5!]3 rlXS-%Q.<%ƒ#-IՖ32m=9\ѻ6B[JmxJ0<2#yqz4'0_Tj!=d aF4_V84j{t۹pGHÁIZOp쥩PhМI=s˥nv:zZuSp$Qw(kŊۧo#gDdJ:q62t)3F tߒ U]ҡ2AIx~RjDaðn_}CR'u|RK*7M, -vП>M,Cf-֕L [*m\_]WLڣ6Je?jH} *=xRF:j0=JjLJSnHSkM; JWʩ"E#Ż :֮[*HֆqmbS lˡ]+kzĘNgxyޭ y&ZߦʞXw0ح)1)m=ԩ/vkڰ7nmI!|ln=YcSDM|ϦsBlfK+Jtmi WѯQڕnpz(LP0. dB^*ΩS"BE42Hj9hT$/`ZV 8/k%p¡;flSpA e%LI 3,ubH3K¶,k-$ӧN## =mz0 zm)a"tvN.)H-ç9Zo |ĥTOF&aD =}u&n e7_I6He!q( K - Qww?헯eW~h:p`wK8GW.ҼeZWb=/7o5|=^_]Q?k\:;*tl'y0Z"7!SW_*O3ogQ|~{dI2ϞI4=P\ ŚA+Dj@:'E\gJ'$2[w^͞&E= rsY(1edk+@aUF 1UW1 oQPhg7WEV+ԧ 8!+40ힰ ̓hV@Vt*BS> o=a. e[hY~mw*{ 4L+)qe[I'\ .D+-kW;bmIn(TB6Nn4 nl AF#.g7Vy0(ATS ekC*- O -.Q#lU!Qmlx~e5,A@W<אKP, "!;zxyZ9ɼL0itqLr2vM7_.3JWE}{*-_sh\MC>=ϯ~ Tдz <Mn<5V5k74\ѿ='wFQgOq>=_ 5~ ҪVPGoo]^$re\zrk{nхGQtY_Wbw*uakwK7Sz246| Rhw ?קus:#+BcNCEr))ÑFs,,1UvRnЊ> g!>X 3d,R F6$ W1@HSe,E BLj&815"jFb`TF%OH5⩵ҥQb9Ÿ0!^T+}zďͳ &a2z|x P4aQ9!M#נʞ!Mڇ~UkZx9",OMt$n '[ʹԅD9 0NZ +5sվ}oY倎>C8A Ao*R&d5+ifoVJ w4840;)O(Nj<CM~9FTT&\ -bin8Bd>xG= ;aoց(ÏW[u8U afptqhmՂj9 S̴`R3?6:ǠB#71pIcB F06yBN[JlJۻMT{k!Y`ܪXrXdB3bRd^-ZV{5cZfa ly6؈` 8J}NGՙ!B-m 9rdPA3I4*&@$jK#Ra@T$p Bxv)D3H bͮ|ᇡ_5fX4(ewOmx8ΟgcL]ӎ%Aw{59,w[m[%?I^wK`}S f bJm:Md|ݠ#^xrAQ-Llmnh<- !#U6 %>Kp{$av1DsIz7@EQ>Y>);*5?l;Z3@ ,_N:}oλz3le6>r9b8qoտgV~YɳS;`x%DLdY mKJ`,ՉH-Ɨ"᣿5ϻ&߯u}nm67}<|$}1βa?s4u8mpӆ|p9?aa-ks?6/_Uף@3dei279;LL8jnJ Wj*y PST+5wƻxM GgUq@M84>hI jO} %7^:A@x5JT5]CВ\n:"󁠗99/)z-1 0\^k JXDꘫ4$IqфR˭duAMCrRJϒBk 'Bi]UhɥPVZo}f^2E$q*Ѽ~JڞqhFp[(4VRrxnc@Ph"6$~tQa,MhP|\arwH~:}q0˯M- MUJlm\S~29=DOnt4xA'y>xF. Fuv2xqNFa~ESO]2_}2=i|GYߧ4Qz5y:la<3E).ּgT? Q= i#pRDNu0RJ$=MD41Xӄ5:Yqy>Mo5y%IiLqӓ|k0/Ÿ zRJߟ<EYuJ!D4ej6+2CD" !J!Ϩ2u."+rP]kq4YS@?'k3c:tꗖ(Mj,kPy i2}E&䃡[D B Y)7i&mvP}vtSEQJ.sQ6%Xb`-rVkAHv7B"DqH ь8bt4ItHƁMi?!Njf\kK!NL*FJ'2m3SJT F;!zP5ϾLq9HYA$A,|VC)qE[5Ņ,%ۢG^p~ S3ΝU:>,fB_-; 6 TWV@%Ry 7^^Ά n0DRsjDUof9MEM#P W I ~(&8lER %Ou{y0XY$qG$f8 |Ɩ0X|1hw4֔Fi&.LKXu2*bvh\]5 mEta>OUbf>F K[0_ܲkUn<6P 3>p)"3RJ$os3ͭn<|qf m@Vu"=0DVHrx Ӎ4ޫRecΌ0z/@ޜgH1V# 0FO՟d"toz052h{kZSfN~JѲW0@#*Z ѾBmSN{B=<0 #V%ӤlX-7Ez$vBc+[ ]VsK|ϻ,43S*Y"bHy2v#_P!Vmx=@"=9Mn"Mb/*q&PD5{(m8o2U$DCd,w}tH:$l~-k$H֛s2@~Z:bS,ά㩌4: b h$ZA"ikܴY=?ҙpaoq|R-d)0RHʛ/5&(gag~ַ_OPW՗;?YGZN;><5`ޓ*5V-;m9F j5[/E\iERg6JG,"$ڼ+H|3J%Ad7n?^teߜ7~oog6?fciC\yiAX}tdp?Ac5d|\c8NˇLMJu}v5x;7Ҩ.<њ0D91s)u>WuRrCf,[K"o$>RrRhH 8 $Ʋts-|Ln׾Oo/x_=8z5l⧋us.o***mX2( 1޵q$Bf䮾U=N89'FdABCA&)iH!gl˰uw*k@٤Z W *]t(b:~J痟+8Lm`L )a&RN* l6<{ۘWwX F1G ""kY@Xp(B|R(ۘGLEu:~[mi!Um[3s'sZp0hw `؞]~Dlk2`^$J39pg} oUvv=DL]oӼ$v8s-{ t6%No91BFs~œI.ٖvsN.?Gߎ#|?tg7d'n֝[O[S[Jv#nw6N//2Z=hZ2*39cՅ@Kq4?LC~-Vn+5YY[r :'Hn\^..,h*uoIH~k7N<4ĠȠ A^oHd(.` ^GD"l28@ǥI9DdgQ$+N2U;x ufAhslݞ.[xj޴[-[hRkMAǧL;b8VHlim !_.SFǫsl\]GM@Þвz`}Oͼ v?e8'%y)feA<^ &" ZCA ( 1]]Mm=Ύ~,;j4..NWk62ii3x|}F}[t=MnME'&.e; ƑC2Gs^@eMFE 5A%/eZtd^qy<#c3z>f]_ȧ-}f]'z=ZǨ0=Mf};O*ވ&ӥy[su)v8ip) }q!fhq6&Ҥ~\ԏM+9)e Fi#Ҳ̴D9@~Ef,:fpt(:Qu݅(ӻ6t&]-0C` Q'6MB1Z097G!$ eFh=@ qI+.4Sk :Q' :l{j\_zx(i~;HUJ5f\f .F|8+LɅ*?Yb_vPLQ5'J>#lM2Kћ|Hӑ|8wY[IYZ!NxB N) X9rg},EY4Y'd4{/{kKo2:׆"n/ 9~&,E2\h ޔ5^g80KE/}9^|ꨑ{]hFo=;}FmْXyz׬JoqbJRu WZ~z?sd3h-WZ]I)seUt*dfmsTVB>]&8UV04޺tݱ+kSuVY_pErg_Rr[;uAfcK Z }+FPR_hą`boKڧ-*3 l6nR_,hze^P j%A%\SP/* h% { @ 63_\(bZ֠S4pJq95R!pS >y׵j5o雗/]hU*VhU*V mZ3 V̅]H0F7Ec<"u*Y/=&txzoܶGOÏCB7VZwoo;N_FzsfyrhO bPݿ4џ/?9t7ggDcg9s&\`=_l"A$e?=:?;dH }4 cAoajT2C r%K .Yr'էEdjgwl_-~UܿV_- jAGuY w)U #]e\t'*v?hC&*A6ޞKǜa)FHsHDԢ!$Ȟuu2`DCW.TI򈡒DJ;0 ApR>1IK+c:E.r;fI 6Īc{Zb.d {@o,d$ dAʊ<C D'ǔr &P rr+[b:QB]'Yl57e=a+ǒ`'f%.`FR,͒6L!!$f6$CPkzj+ Uf0t@nh2,F# ŵƂ:8pA@)С3vޯ%J'a_~;P^37UXnFv Ji3mAĤ#dԉ|A2%4ⴃK'71.(ɻz;KOgHVKw66k/y/9.ieAl " vHYvF@j0Xpl9* ":L:g-;L18 QY)r!P(AR]ymnvm0MKOuQK*`ZϴJHH[-=]ñiP#;2cȶJPgZ%–1i=J%Eh:ŋ-ZmVh6-i6$NWp FdrZnxIrF\0*5^i 5b&{۠%-ɖʩKSų(OAE(R, -֨-&`L?7brSQ"YLd[PPj̜KIz8P&IauL.'\{ؖLf&k5FM3iNAV.,vvqIxiիΠ_NbB|;pF!.X*kwK] *M Y,'ɲ%i;A@V86 ڨ_nR ]Od儍\L(d"uT>@aDmd:""%2b!ɓ$-YFBCeVNSL46zF)7`I: )GgOYJmmK:'K 3a=|)@:*;gS@23Na (UF99 <XC`ZH!O<–lC >L0g} -b22#%9BFr:)IY92[Z +4;ɤ8cekkٺMp(~ [X k(j-G[ĬceLPwަA-yIww{gy?dlv#sy2kkGb8CtJD=e??Lg!"`*qy"6i>άj ڡE*)Z;Ѳ;R3i[B`}mH1-,%G']Da)?3 ̩YK,8a4 -7akkHB&nyTPpN-dHzDC2D[=@àSdTjݩcF(^oTa=7H,)enV>|9Mo2ֲK"rduUpsq<. ہ* r>fٔjݝ\- Kn-ЅLIKp)?;rPT{Spd-8 j0m[^6 wG"$EWQԂAƃ"*%lHA׎DԆp ϔaGbgJdqm`Q{km&;Czlo``31$0 j;ESݤ"d5JYXbwNչ[hMiF'#8ΫcU4ҒPie 5R)<]XOȎ*UAd4F1+q9]V{$iL|9lMEO`oUZ#-b5]^=9e[/~tC_Fc+%%}mQe9OȺwX}c?| fn%/5(BXAk4WySEZYv L2 p89 s"TOxB!p  cqompr5[Aëª|ŻpEYT~xT\3Th$Y ɭ,xkIs\ SFW M_!KY2Vw&ϻE;d݉} m} 'AhK8'<v37g7*?n㝳+k#2?Ͷ("c:**] y5uC{fIt7 cެcͺaykD!6ή͋;Uk ֔c4>"Sl0S}Toĺfy "GWY+c;6 ,wM!PuхQL0}(8WFFAs(01%qkl7-":\`2w܀MWުj0kF$7̖zK1Q+zJ[mk.^ HRéտWN<>_2r2#c!u-C>%&@`URPܧkr2 Д_᧲ͥs<e GKe˄Vs#<1VT`ޞ@pƼm?}w,Q;6'!e <)9>A`%Ҕ(==Ö2-rMj8]rd~԰C{G ˑQW.G r}D;L"@{= &_ܪ!M5ySZD\38 oCMb"uv[CAcD:/k#Dc%PlL"#4T"%1RRyfj':9 t9sҢi'14E]@yBo4IȩG0zXf :yxoxXckH]ݸ"o<_MXL>n@vBV0]'+*WĊr?-BɊzr_&)~&j Q-1y(O9ۃ#LqjHQ}3RJg%unOQvݭ^K ~}vbk#@kH/啩ySyoo7I^ls=o2ܢ6c7Wc19V὎||v _atXPbXQ1H;r  qAm!Mep)m"*޼A %3čwCh8V{Sc>g tuC,3HcK]L壞NXsl~/D96Av_^J45"N}0j78Ni`rm?\Fq!%xr ln,^)D5'%wrQhisb[hO0R.h)\ fցA[)_CdOKpkT\-!1R,Tg,<)&M-NƲS._%cˠ-W|.{MjtJv7^Al*v6pRBX|ȡJ !O(>y)5NT``0)NJYKL( <2C0CR`iJ"ޭR\HK%+ϮM {岄,鏙%WC(J,!IiœQbc58Rp/FVF2ij-Ruf lzC]>ےhrKvXz)&k>:X9(WS*=4Gt1۔ԁ41L+TD*A Ä>QS Mb7bPB# *cu|;`!.=]ڥ@8@f o90 !&.\}L;Ώ9Qhob糗{ $.Z}>9} 9Gp,w~M{ HBp{0-, Ws$9;59(VذXSˈQì9H$"r12NvLeS =-۸ SHFO9q[LJqd|$SF"9gUp QyKWwW<5_s$3&b䍉?gl3z ^e7} yKR")A܀Es~n` "b(<7z8H&r UZrQwЫo;SX׺vwn|Xf06>n. k5|6K{s#7({BM?\qcL%+jQsq;e5$qu 4ӟh-;R & TSD,ϯKOntÝ {jVEOZzW70 =|τ-Ok]e]CWggtQs#bzF#F}e(FE1փ7#n!oپyw&'al{Ĺg?QJ3}w% wآgc;{#JQ 5NDJG9EMR%F3,I"tR%.f}`%|`> uXIX@=uއj4+^VN'W`ޒDu2D ޱN<#w TMișjkѤvws:'0GPsbf͵J\\ohK!K^_B/xG(>ҏa+H^a)ⲥ՞!.]gl@7x3v,"Sp]C:'DoqNrJSU)L%?ϔ$$̗ 1HJ(Jifb2kL}렒B!Uva⫱j<<8  \dž >~NJcsx αFdԕcSDIRFy)VƧZgג3vnW%ˀq )ރT_)XJ"`cĨ䌔k,$x""0,n.eYN5 szm*EEd;v6c}zs""l`9d&3H$29htʺYG8ͮ⃳}h5,ZU2 Guy52˙`?Nᄇ+L䪶VB wIkS#Gv" 9li l;Ǖ@,!·HR(Ƒ/[C#Q", !98V"]xض,p$kVsɞD|cEh]w0%`9&'`ϯh(G*Cw=j%st~ip &]9sa8(hေ1Xȡ}<؊.؊J.R G_,:l>ͱ%$UM20nc;ѩg”RR1$uKzZ5y RZ]?rJR5VM͇48覌"N{>4aӄQH5Ȥ<)Eg&4%jKwQB, 55hBlN&lx&c=ʦOupFJ&fy$څ7yD9ەsԔSucŸLx"\۴}!Zj&4| 1HG+RfP1‚θŞ@䕹&+ Gn4/q#-0YTi" %90$ n̴R4 N.9gpm^l)KFF=pRuSb@xHf&iȎ5NTId쩚p[fWHIZ" -Z's$رƩ( th Q%qc'Ex(iv+>+ >~;ʮD֓Lt4lPԋ_Է_ې1CsvP73c2 -i>{R?;"^Ϟ;пdd[R,Tw߀qL8f F #ζ(獃{6([ig"bn n%r3uK֨"$e0T!R$$cZBn":-3Evt0LJw|Ņqd\ʹ=}sq'Wa~X*'5Z]%pçUؽ)w8e@J)# p:~;bPn9 !ՉT)hb '] f@1T0+PpςBʵUm]15Zc|FbW=ٳM,ngqS,ʳc,p)|G@85syT8BJyI.< د_׋ѥqZB׃X hp'Zbm68uE? g|$&?5$$IAfJ"] Cy*A,8T8 yUSx197(R!C/GhЌs u-䮑-NK `Ro8L$LZukrpL, ; P$sMOEZHj' \]V8v|Jצ.ձf6 <9}~̧np/W<ܯ/bwݷϿ(]0-nv_1apMp38z?^-e2-8-E .+gQSs8\ F[Zvݛ  ..Ə\妊 :Bj]Z)NikR3qoc1($0z,p*7 5*Țy 77 (Z+:G90&F{BKQT%?Eφ~hhU &9:`\YPfVP! H;+[XXHP`о&wIbۛm|@v7zfvsޱF!@[ڰ$ׇrÁ=DZ(7Pp׀ӓo?t֗37/9g mXL沋U/GG;/?ңpe HDj=M~u Nk],GB)0kN^nsFn5.ΑhKޭ9Vwͭ> *,\}+S܏{[7oߺq6|k- ^zmC N~zEѽG&v|\}W"A%8Bsݕ ޿PdSl D<2 "X+KIJQ$*C&Py"'wwT7Sn=+g< ; r%x%;ҵ?Ǐ(: FѻMMa3=̂3oz x*n<(J}vtq|WGP%utf̑Z̨G|@2?(g[bG 8}y\.F@t 8q|798ģ"D<棡Hޣ( JZPrmNPPtѡ *V`M⁀u7}p`Ԅ c_kKכ?|;hˊ1<>uGy-{Iٍ\hw/yc|VqnX~.=~c+XCh-uúqyh֭.)Sw*xǚά[Dք|"zL@\Ԉ 85yj4?fwaN'E$(I`iW@ ̈́?<@:"xlZZVF7*"­,^LVUKGm&\|xtO[$zz|[QU>fm1Թ=<l"(^_]atIDX`@Uc7WCt2?ގn<.t>i&o.j^kܟaP -8=i7L;o2ֈlyǯڏRqZ[K 0ݱZ Pi-f Qibf/F9*rݨ<!^쭉AI¿ 2ߧ)Iwk C&,giSu# mw7?W%7 mχB*_׃wR\Sw!Qquo٭۬ڋKL?{FYI0HL0/0xM+_dg&X_R֥mj'ıZWbX_)N.(B{rE7:p7}^Iݲ˝⋞]Lj4*w0dN)¼ I:0сRM& OE-`a#gd 93HGhOOI+UEq_vqM"DƟRDUq_< U¹QI 6ߡ>$Y{ $L q**91mr/Lz>x݆T@tWNqSnψhC&%LyIז% JWs]ItER_T]hvn =!PZNp`x,#h$Q4iňwHLi\}׮Rψ:I 8@-0XzØ8E!:cWpX98{=''ra 猋Q{'&اah@ e?|L >0(`^ZBǔy$D!HkH"X `'8]n݁{/k@܃8h[񳿾װw}x)TuR oo͛ b ~nQ4 Ui[Lc!D%V( [Tmk,>UGT]*3Qoۀs%p AALc+ 53/8D50] %Pq`!Nqћ0 DZ)e܁ۀ)+ Ƅhĩ 0apTP>=4`f9 MQcs*R .ecPiY ESc,&DQuiL;->JĂWĭ2>nD;J|(2]d ޿ 41L6=\ "Sx~}ԪBpuܢ{_,5e*/S m闏>Zէh] Kk:a9Uġ Z-}tzV4ܔ )x\08O&RKrPWGV@enʡ*bX"ǑX#P]7W.]诚g4 Ʃξ9؆a FXfLƳc2V'OͲ.LN=)R7p& 0v.1yMzrkRG.돉TRo*JD?IV"R NW. Ep1@;pHp~i1#iAhTDSJYLຠ0ޡÛש~5iv;_~ ~yO \4+R4 9),3W[GƒQ_MoY71vFp4m0;^sm.0ZoQ^Q"`HysI/j1:~kMHwLvSwGb ~f_*T[+cIyЛrw_߯H<۽,ה]YqĚ; ǼEۺ^<OuGOe_Ď$;"r}kEO 11Zi9'OJ':\DdMt [)9S:Fv³O"iK9ڭ ED?Oih1hR rDt6> riDs[hLI uǾ*~=Ae@z5 ԟ|Pw\s9"ۉ]'n$tm|[%oJB > Å^nBU1V%qǨ!UZb_f@k_q:ղYq]WORJFNygρ+@ o;d j>]f/I$*y& 1H&b+9qI(QI.D]&K lS;N'cP ZpE (N8?ԁOqǛɇYaTLJz3Y4);$Z~?x͕+hLZ]6stɁ>MRjlŅ@DuTŪ wS&ϋZ=8IMZV ߞ]|ѳ镹X }VXVɋgXBǓϫ2s4y9yyN%L:aVpS)3BkDnRM& Oy`\q`AMv! yzr+88i?u?dA*+D/(@07[ VkꭣQt([)F-xS -^I顾;)IԱԴL9$jLk9}|6e2Z&ok0"9hJX<4R[G,m\*7IKȁ*NiLZ9"Dp%&#RF9sRr=l(hDYL%0\_}\aQC߶uwز>1#-_ZhU%Ҫ?#1TԊ>xh`/>\Ŭ_uM1ŨOEST`r}s}OdtӆTH e* Ǻs}qK*[ ɻ~r}ZPbXk42pٷG# LR6Z栴ÃbQ$FۻZU7Z2MGn~Ļu.Wo.%kXBH{ ɻ~v:_}ig]-Cub2^,ŀ`YIBծb]/SݦdDy`?6Xx"Τvx&'fJp⩔ vq@VS0vsmwQ3C=Lcʟ ˲CI~jCYrT W[DεR Axt*( p'b(̩5tNW~_:̓U`ePVTd4 _n%P`.D;>Si^oQ|;~S!T&G2uJsn(XY~8 c c‘Y,Y7,w@blA-f|*e=!? KF [:DhK5o^aFbIkYoGA¯Lci-`Nޞ %k}fW =|=ER&CG!33RK6ÂT')4n [N~r'a,.|{6 ػƍdWd}}Ap. ,a7/kGI\SM2u4`XYUky[g6ax:s'~ {,/ے[-.OpSZ>159`5. .]Lt0~qk *L/8*4W/oʼnBTV`J婴(0/% ?=37gk{MsM+)mXf `elʉu5:Q8=׭?GK}P3Ƥ1-`p(rQ-/Z;Yu oV jxF^Ԑu& g5B@tH ?9in ÉJ .CZy,]n=96Yrϴ," RD9)`Ũ:AUHL)VTi&qK*XQpl/BX$أ /b9*ds[AP-Ic.Fscke1WI)%Đ(G 6_ 8"+ Fr `@iTnZU mDZ˭ ynz\Ȼccp=DG&<m3!%si)˩e\ie 4v(0=lt oCv1Uw~0asl/ꁻ[N;LTH)@Zoț?|>\ܲ{ ~_{xADdgꐿ o_^!拕ixWr&§8+ ,T3V8;+d~냀d!@/_31p%j8/\øZ#0"-,myZ;PT>C{w<;x }/y />psw /N&tN8ѹHKsayU~'%]Đe~:\ZҔB~U.=|" MpbieuG0^,#.Jb )6~G(L9MQ(HjΐJO\6D)I%{DAA)Lxݔ0$ƼGw]Kg<(09A8!4_\^糋\}*h/Vn1xR=D4)H`F4#>@l˅f)Y:-k8c~cp;JԚ8dE0}'Y4j;#z &jYpi"_(g[c-$3g1aUf ~w WHXO:_H|F¿b"SM9CXYV\...NF D;üaƛ'Dtvfl凫8_`|S=k!F5SfQ܋W㧋Qʬy)C^SDstB Y;El5g; { =l ˫{zHŖD'?ߟ"oBFHlٓPA}*-1Em IifZ{b =ko}:CrR@(wiStHz>^'CH b&*^9\lx4AT"ōVZ&q1, vcR9,q.YrTa1hczDPϥBc{#(^s@hW߹J/%O^4+Ak_p9+"ֺ ZBXRp&TQR81C&D{I*̨Q yk%.RtO;($a+h-_u  %6fhΛu{Gݔ1 [v0koBj'i٣8TjzT/q5KܣZq翶JNIxu}:McaxّLPt mCMxAjmR15/H3$|&ʦ< %[[ bL3x%fAĬktHBr=ܦ(rDJv4erRrDEmCc}{R*= [F&Htµ5 5.SO %>Rl /^j򚵈Op4VHޅaGzkz}޶!^9>Mc'EQ=4(5@(+}941,+ʭP6F>^5))D頞+q X"D񘏄a@.jq.P?u4+ΏB[ ;y ς/?/3;5/:5R%zncHW2Ri+Gt! /Idz+N&Jc|*_v"c⿯"Xprxi9\/!b&P`V9ϴV0pqkL&@>NsǖTHHR)I\9eTp* >ntv?9ɹ-$seJ'v`:r S!snjݱ B=72xmW>%Xn 9e 9-D&4ܺ6-.Z 8 O 1=kËceL4?ӎSF; x-%|"T4WEBy**`‘(`Lf^q}k"-e5kI93*X-(KǶ5 Hݡsk{8x7˭c{]XȻBH$3̜˘V"SN yrȩɱv]v;C, I=ZfsU23j4JȮHYieQ@VrDt\)4"l ̆M]3Y6)JbщztD7r\3AoW#k=M{P\9Cg_k{<}yY66}hrYwO/5dBg P,e2Í9voP*'TMD5-s|Zj(L0Yqr ?gl;Wlˬ!R-Q) tIgf 9?N]aJiWmB/~@e] 詧%5ts^L֚; |%%)I2,3}G1wwp\ s9dQ+H~u psҭaJ,LՅs [T1$On nkIP<%H/d֠^ަ (T[*tкG^ZQUe F$Kݒ3^\Y?lm4P#19 oʵ"u5iUQ)2 #llF+ .d^s/jdb}Zr >jiiTAc1k'MNιSr9[hsv }?N;G<m0g8ۢpT-Q:^f@94ZZiB8g7|#y'*޿7;#D6biO9 BG|x8>|:u Ci3 Y)s C^rMy?<ߗa ui31g`H8SVdR)%`P:1_,.cb^aϮNj0d@PBB8 8bbO0t<Ӏ#t=hfyn: 7 SAtu| c 6GEB?0{ 58a]M>|h \a6+~>Y 1tĘ!JI܋%p2Pw>_YotM!z;ga~0uBO!)=r]o :R~U#LrEXם_j~90/!߹-f"\h\?{WcmJPEzdb ΤIf&/3~uNHL]'/,BYU_.ZK'׺"@7Oy# elh E_<بJ 2McmXJCDc[cT?f{nNV]VWXCxe/[ ۯˮȝ1ƒgX}VlL7r`i L!ai_iS+hKb$j k+HTS,QK[jZBvZn9f~3g|Y5d`_&5T6 IHz*MG<rƞSHc_حFhYa4UWQJP 0SW||Ngł ;u&u@W m;`l6?3{bdZR 3RSA%C@\jXCE.¾GJ.V,^StWqdv-jq)I>u< c0u*eeQ! JohARA8bQSWڜ47IEJ ,UtA:K+U,ovmIw# ?\na˃2,yZU\ߖ6Vj89,yH =gM}DuFu!-U+=pӐőˋa^.inŁTقҙ[zR`j\;U]_x=x 0ߐkr#S)sm.UKՎj1U1".oLPG<J)1T8s p[љ[hc4dRsYꗟc|o^vX !aQȮfHš~7wPr"}A:+]h)~Aj2J1(t08VpnP1h\cvss%:=vqmrJIB4Ӡ75_=tc%vcރg.a pKe sFiug MgUuT+UB|*~2+ur0-6&>qo!!4 AQJ&8eWu2ԐC1P7cf 8RnC%Wun DrsvZJ9Lt᥋g04PBuP ׹1`M:anB\w`r!o؈oCQ<^V$qQRR9̽ǎyZ$I 9`ቀ!3\jZg]ͣo_'󙟀:>G p lUpwW}d-Gw+ʎ*eK7URzٽL"%mn(TZBg^'?e: F ƹ;V9ܲ3wT!F%T!:dsl0@΁lE(d@:`Ӟ@܀"0AkRm\9,<uA^x G$ ')OحӪ5 > @K_RLBuۖUp@3`+CF'1+0f3Иz: i'4 <,S4߷$B7%]Wu{=+`oe$#Zmֹ :I1( bwsFI :)>;|8xw,I7ʘvwZn|=hyv/Rn%H=5/_/G&Q6~Yvt%Ϻ9 Xxdzzf9x=!(:K{kT<~T̙v '<{M3Pdƥ߷Py&OK&N)4Vڕ*g3ؽ͜u:}*^>IEOJen;n5Rxl?2O♷Nv_JI9arҼƀ58_|7~0ĝZ6DG0!F@W fF(?pܵ,HZsQ>0P;[X]v_4fsj8X\vo)gjN^KZ=HCi)DMa-99VP>5qc{5ܕaY<8ID9s/ޞA/xK\DqLQ^o)S̐c1kWH%FE' O*?7ٻ, nvJ"hNc(Cp&6O&uV5 ?xbļTv1y'ste#9et˨Em,#>m?npƶڣv79Au8H6zyf2Vs:zJ{K^w;y^zdZ9&MUIHK+@ֱgA7xPuȄ x&(w0s/1GR 1PT=7oBB̔%9lP8Ag1@wE*9eϝX8へQ)FV9"&@oG; $R g?T<536q<7dK8e:ṏSſ~qd5 -NuuE[JNiG})`8OAF28ׯsA}}A2QW^9ͷW_wOˋ;c5lKڨnִl<;zͯ>r?"%·f}T:M#V[SŸ0 ?S:aR(͓j[*It֢qOa)#nok$z`g̣Ÿ0%³I3T=ͫz8MxLSFX!< ^yI=m9m3r$*i>{HV/ Ro0&[, t O)#,^.^7 q.nqzDž?e‡pd4 իgi'(6qi !.>SxY-^yle.Eq  h`LItѼUY 4j&]xBxM`6oGG-򘥹Ÿ0 B^ t&-ז7`'Byu B)91ж2e#iJ}۟2 j]:,=!JvXp?@x@4asTد6i9b7QfcŸ2sZR&2'Q57yV)ຊᡢWfH kx{ϰ;_8Z+ _7Ogis ݰ;s A0@U}~--|~`DIs(0ӼZ,B5dh#"1(TIJϬCcv~u`96q\w-"H]V\{.`[طӑ3ي)a~]2v0̾\z:; /><?|9Mٛ_rxΩeota?Y:wxcoEg:go_znkϲSx4嫳:ma~0̛׿9== 7C[tLúXl}f$ht?`$s|\?&QjIͫUg /CNG7i$y$c/=zhڣ=:<~չ:cYиx/ƹo/]GW{t] ?bXqYV&kUU!)1J(|FJtGwߞmjEkw2He2u176ͯ ;idফ^+_RZWm4*;k8yYB"8Q<U cr]bhE? Y7EHQZ9BfDP4%Fz>L:y&ȣ,t{[]v_P9 2&#Y]0ӗO$}'oiY; TivEٌjL!?=N^F5b0iyo? fx-H2ucqQB;i$Gh[o|1[6"E$Xmӌ2T#iF8$Yl[q=.IUGrkLZpi7^~rAYy=Re$87L##M>eNTs ʤǞb.Xy23>XXg5~ 3.7Fɂ2"^%t|,ߏ?GnCrFQ>v_62X'h.bi 9ɛhNon%y6>yR|~?1_Iҗٞ1}r1 i,q5^W؟ƒj1}ED&?wKe8+(~͎QB)j;Uϫ#ߨe^wuRᑮüw I-yZr໎T6zbLN뇔hֹ6?*YVR`zQ̾Ys WKWH՚׹B$Z"14[ճ*'ʴ*_,[yMp5-^z^bJj"} M >uR<\+*/XSb0 Tz_"]՚E^^?+crhTk/W\t6\|9D1g^$"VR UUgK8|MUPf#nt 9Uw7A㌤)pq9k1c4zsZluGͩh7ikx3m_s o,(u]֡Hs?+{w!O(ܛ̉EtlJz^hY'鼗?Yj~k7ZZ ki@H{;f&f(u`w*=xWHbr1i[H{_üjVWªxwjZte;Z~!*}DU.Q-> y A|{Gŭ;kxKPi?-aECYbXEnt7Op?EpNi-߹Ɛ)[Sl;~VKoXlB-"eÖm-?Iwxx&;Y$J1]7uoM.iWܰx9΢44A͛OzZ<1C$@(2DSF eiqBmO:8]&+˜Sxɴ\a$0|*= Hb pA;<ۧ=b;ʛqZq_*vْd;exQ':zy5(2@' e^Yh)]H֝O(;_Y֝ۧqƊ=Qs֌ReH4łQᑏUVR P|K86E%G(hm "4iKԁ:zK4T(Ӂ,39΂"F+'״)Lqps;-z-PmٵΖ5f_\5\^V^X,\jMۈ:_ &ZFP4c\&5BncH(KfzP, .cH "CrǗ+kj@@ =n0IJ@xֆ9+q]K%𭦝.uDXtx؟Ecz!*dYq|ݟ88RF Ā1"FȽp6N[+hVԛHiga4­[dR0FhĥFMLCn)dܚF|B23 &4OdðW>^ ;a|\9I&t0jVFwk/vr uށʜvD P D1 ^8~"Ӗ3aBK5x?DxM%22˹s{/B+,)C14 g(Ss]AB4i!fPpH\PxfУ 1 `iR8' kF o(a8śS;kn܍W> V:}:+1cSdrYF\1p7&tB|ŀ9e֣b{at¥`:Ɵw&Ւ4s6*2WG},c&E*iZ+A`CclS S< rbذиp -YOTkmQ\7*{Z {DY6 go5 R!Gl=g2{ʃWIi#E V@=FQWl5𾋭H ;yPf:E%wg%A=Sݒ"k%TۦX!8MįDph}p:hm& lU8oDZlQ_A*&J5Im.TBv\[mQow.2*~›dPP]jlF(;*8 q pZnB"QS^eUB5J%I%P{[rqY;IE"88k)G~<."Y\EGƽ#BPcy]x؁x8WZaF H8'<\!wZhQyX@ģf,ss5Rsm]DY+2@"Kt/T `EiC1VLBL&o:&h9_ |׶{kѵQٵs_V(h).DZ+65=T|T}NQSDﰗ.SU~qZӁe` ?Fp.L j! cq!b)1W C`uJ2#rXsШȬ NSP1PJH]QҴH/- c!(,182\+26f@b V5RF, 5!@!'qmcɈtTh1R`!8 |HyM5"@ Di k6t(y4  %лc!K!X#%.A @ND`VEʠ8>J`L3" C3 uit]h6F,0`HơXN035HN& ,ԍ xS{ [ɅTrO!@fπ!!" (Btc8JoJ$FSIu"Q1&h̍A7(*QK) 1.*2k 9^#؁ର00DD3PUd^2%qS,X, 6yʁΑ {1V^$F!4w 1 6*3+/;q8]JT6/VB UQUh~d3}[ -ZMjh=2ڪl5WBK0Ъl`WCKGVe[ZE[UJhͮZ=*[ ʋҏ*TC?.]C0KߪQ-EM՚:ANrcTRJ8Ń$?oR*,Tf 3p=/0Ar-krDd'=@s{JGxmGO^*y}ӛoJ viYNsF.e4!`w|I RFJ Yb]hpWo6prJNTwNLdEfL()pS`ȹ9Ă>ݟxtj )8$ $h/fJ">Q&g4lN/wrPXQurlU+H0;^00y6mAUQm[0FZ`hnNb)XAR22`NB'!103K  x0 ue\I\}f =;MeP4chVKLo_,0*;Kx:*>GRP 4էx#ʡ)q4LƲi4&oP)rlɷ:nnK)G9"Aт&Xv!imx!*H%Fإ[xKS'Y˒NsM7$5`s;!)P0w.*-`[D h'-X@2Z椓1Y݅* Pq''0p0܃j˺G]3 =wq<@=4e'3?Oeý[vi2㩯I~k@QEz5:|E9uE?G/Z@zb8In,ZB@/Ch7Lé^QSj1A_?C'te_z*SsY@{e ?{/{v++Ƃ ^\HTZ:na6+`V޵sm jHذԊw혈kuw 5뿳#DH`82O-w:G^Dє9V1Zim"my:$Zj*w8 >3Ɯ5ZnMd^jXjDՈPw'*z̊wgB'° 7/jHJfb&zGtsFUoEɘ.}:ӥ /|% VL \{ ֩,bR0SCE]gu<Çm4j%,9](+GXTqE )рbX Kz.o4>&U! :]zSdŰ~l' w'}JK#s{ _3k.\bhj0} X9xTkGGZϚOuA/c(Tsv:֌xٺ i[e^t>?ZwP(q?z>YGw ituWgh`JP[ϠA;GL( >7a=7z};n4|֋ƱT,wP~8}xěAQGy:xQ?zvvQ|'c3n7,y<>?w=_&v:~t _fY>Y.V }㹕}kb +ۣϏ{F[zj&cvOL?YgKltLs3N3N b*X`N?YY2EJna%;.U%PmK/NM^nvk+m hp'5-UhޟOy2Dž@Xr)F%  ryr7lz&|=*{at:Uә{Z 8ç]k=ZbQ֓0ֺKi?M[Fr$X[Y4\Qʷn`vVj@EmM:~ӔP.ərogIT$}_np9,]sVJT'QKkOkFXo򐵊 n8t9dS?ݹUJ4wgx+a)D,n?l=;?l43zzDxe{ _ڽOp\NV*7w^0.>^f| ]N:csp1vXL.pr/QþMIԡOυ.}u-x' -[u;mv3t>:pvtq9 u`*_tM,A>^^Lwޤ |\viJp0 [eɴٴ|:ٲĤnNJﶿ7O᪖1>5Yvdϭx6WnZM`nG73!N uߎ+kzI7އOφnd,F냝s:CSCRRJ-|={ɻFn,;RG@6Is-@ƖKn}tuؔXU%$ߏ|ׇoԷii~ TojmckxP8zw@r_I_zZ_,]Ӎm3l-uGz:WJof.{iv R뇻wu4po~1m?@ԲQ駕xhHԭ1nXCou-)uUU :^Tͳ֏Uz$#q#'u7 ]۳k+jNp‰ ۾6OI^'<,'WpnNj3 hdMm\W^ WNu}*j.πy}-W{e&ηG Q+mĖ#o p-ǙkD-1B FM#C\VFkZyΑp!a^Wxlx!\MGOiͮMpp2W/BR+g 6yyw{}is;Ydmmy!vuJH)ci^ &TIX#eTyZ1&BG31iul;s Z9=|xg(?K|F]/_yb c SUځ!(r XYi'>.W :&3kmBJي+<8H!ڼ9=A]5ib8d'oo}2 կ4i%2,~]Up0#fg)dP%c gz-iTArN"z-yqOh>IF9,*)sNB㱬"՛! Orz@x`1U\9x[¤H=O|N]acvL|pC+Au3Haݬ]DOzuC|F-L4֡7b<ڠ=O|N]O玉Eq%Rp& @~Y=;K\ONSм)lD/H"-\X=_0 w%`㢠 ADbΛYcFA?dWnd縲pc:F~ՠG5K=]]c.;~l}<1}!˭wk I8M}M dL~md-@Vz5|z dv~0ƌ#(nO/cP p=F%SK"bFɄbqF_YneRb7 s%p0(FAxAQ[΄ Y/נ\yƷe? Xm@D5JH|N_F1SY]w)+ 3afq ] VtWw"]XF*IYze k ֧(y$HF}4nTY x\d0E`,";BZF<em1oRgh.sĨQ@Ce8ViSIAݽ"AB9#°-;`gp1a?J7 uA6bQɀ܃:v=,^b 8i6 R$ .{D@^Gd`[=P3*n>%GL"+?PoJ -w`SqtN ,rWhѭ$ &/(])40pg40itA28#Y!/\TiRR$qk{'Sn9FcŤG+Ќ2}/*/nZͿa[=Z&ćꮍi#ҷo_̒K7y6]P QqK BCc>w !ϹUGXCG 慐 8*Xe ,>f0jiHm!=F1Bw*31oB qnHH,x*t#=&zq|Fu޶r+Q\e 5NY&[ޗ''XiAM)zuƿY?LSb 4M/EBd=krX<ֵAVjCwx P??룬& i]~VeH1e F gLCa:lr@ӗ'x]kojMSi7mkL:bB#JiR_0M" X 0Ш>"$늮Tu*NSEik'Nz6MIϦ)4%=׃;N"wZE`-S$VZ$VdE$*"5Bm< I]O4O@Ɣz_|aC20 $ 9r i)%2*aYj0XPp%Dg> n$D'.;Mb@4 >,D,r^8b oYɢNV\E.pD"|-'c7;ajQ +fv- _q"\ͬٳX2MڡޚԊ9XӮ&ל?m\󚗘\`s FT޾n](jlnm@K1A>/~1 8j^ۂȂc\a !ec[>W$Er•~#F#ٟn햎9z1MOb88F"MR>,l:w=)|K_]_a,ƤGIH MIE}(&Q@}C^Rgx^r2 OLt˳'w7~%F+/C+/GIڣ?!Sp.}|I_V+9xr]$.;G+G|Rj IK1@\:)I\}ND$$]pg΢E<էk3(wpk8ɺ7Kk:H1ԂV [0GrqzQ[>\ =\ϜEKxJ[_G .H\pI) $Xg$\jB+m@ɬ֢;8VX.'uݯhpTႾ%jZJwp}a# yѹh]XGxLLm-fHh+ JJ%b8.Eb)1$J<Ky Ͼ}(XkSBA.RroCg;#pрdAYPp 1jkHŨ&G!hqZ:ZעS]t) 1Q)9X $pBCiRȱ4 %VN8ռO7hU ^a9ж3GhOwB0#md#Ҭܿ^doQCDo@ś|r>fDt|c̐}ͺ|hoo\?QD]PT7[H„ z :H>H(Kd 嘞/w j}|k8 ~aB!t?pwb9"X=|*cP8#[t-_ЗX-[ǟ =%{pEar.nJRt@ٮMi : :Vt~t-_֛բxI2(+p.(S5Å]|,ZSca_) J1a1NnB5m[+ݺϜEKxJ[Y)<ګ蹗Az̴!&D:RXbJ+pS),z{>H{T>tpsJ`惲dGW_ s|*+$ $*nb{2}M$I9;*K-s[ Yw[ W)}hJdBRq"$muudA=A=yA3EOpzth:?0:TJe`9lV0K!ܟ{Y]Z1;gFsz]sx?sW0]W.!y:O0k܁AE AoD+ODK>4I(rZ.1.QIWLXz(^1UKξvɛ=Hډ6S$ _k;{3l=<kowEmGC S5&pܴӴшzrۮq)^ߗ_shi8FIc 9~&WapEH"BH#l9ѶQqZxxu9̫8 ش +o908?'uψZ)}^ -L/4j}~wꍣICg㤌k@[p_ uz{~|B*C2^{uA_E| ;J|7Qm2vPCt6,=>>Xvz/y*hTXL{`}4e938`uyR97 s2 AD? ^T[7_vnv;H#s0û甋?/#&=_?W'lC43O2$*#At4+W>NW56¼3Wš& V>$+9yߌ?sْ8ct/ԿILӋh~ ,'X'? fYLO䱗NMND/ CLJ3pk7dkN3P .tߪЁŐ;y]6G7vp:nn]ݎsv |T@OjrX)ELx23 VLQwEޥ6#^9ɴ:'eju*2,_z?DWYwF84y$yJKP@= q)L $8-F1oWyq+T=YY.8rϔ|Ky.˚N@Fgݬ1d ,k:M66_}i&d.IQJ8-C:RKmsjrPr"1)wxlEPs_Gs}`*xwQцv|sc,&dCByMMQE)[<\T|IjE4;טQ 0``xL䙏2q,O_ <~~cҨ&RFI#Ooށ=59ˊ3;f1 r!6تH[7a(JF?QYqj2+<Ѳ8't_]!z^y nxT6X~f*i K>ER I"($ݪXsQ2(ZHCM M`= t(q A(DxO4F="S >ۍ4ǃM(PjfՏE~;00fǗ1Q$jWi6I;k#JAMc&0ʆT)‘T b$*!' N0B߭Y ÜYPbLHCT5 <Ƙ2YB,RM:VPDJJ]Mm\?]]ͦ3o%Ǡ I¹:aԎni9&]N#,u7|iJ#F8^HA&j$)8 ӣ+gZuRiU0P|/P仚|n%|["p$cAv1`l08.L4:!:n3;Tvxu)G'2E;G'I;xKP_7[Ǯ.sBG;4NO̖@Ûs9"6Lu9TGfu䛗Qc ${%wnl!>t8fUj+ofaš9kԎL+˥YoX8~i۷yI <_sGlvdqV96 wd x먧뛡] @[r+Q_Iry(Zm_"YY)٥V^)wkP[es'=J(NdIE091BTր'CAJK`#v+@S#aIrC,Z[vk/@C$Z LV'fRXBkfZK(6+w;Q!Q<!`XO:zӨ$9HjYH>w  JhQ;4 {?RwJoхl-Hd؀9$bq9*.Az_rGBCPI/wp<,4DhGknjYPQn͢X&G;B7~gW}Ѭ1K[4JC>8Z*-j::e>94k2U!24N&PܪZ'r@nUUY8E2&_B+Asf>>:* X[ sCi8VQZg, 2(-) RX"EН[YJa89$FHOmD\aQ,L7 O Iza!wa)͉b0D7N"0< rknqc6}~̤s,8MeH3X"ЩgyF=m׷mp4lIXrllVsȷ^{op6ꉊFx%w ȉ]0AYeDGkԔ+S@pZ.mN>bLB)V4"ܝ;_b1۠AOxwMjeSߌNr*@f8i'\ϛ>OC ,Z,+!N|D<[jWH)! D*]qX0ˉжD 6zMFId={D+^KfNs+@LT?דx x^'o]Pe#/MP]VU3~FN{-/Ih W XXrl)lx)7v52FYp՟cl"4{,2 BDpM' _7?덞DgsvI/_>M<qzA|oVk"e<|![8~`nidpq!JN< S}|w߫I?Q1i.Fi+$]"Y֞\B>\<8P<~pB7mԲy.P b#BJw3n1rik.R{mr>6H e)"0=l̅l=ViT-0iVk0^E`8=e(on  ÖuD;?]}^^ɋBr,}P9/]b`}0 O}F\5s(`3Zj\9SYR8@wx#nE4V(z(]~eTxМ.Jav,l)N"*5L0yssB[E`W%t#b$Qj"KFxE 5!" U{B>hVz%U&\%DԨrM 1ٛ 7V8{H>X 4G;OTQfR*JpLi\'0X21{8d@:pJ4 !FKx;OO@}n<E'7!{:@ JwmqX4R6~ `6SW[VKn }OuR.ExIlb!d{t^c: 4TyN,'7IIf#xdr+?ҽaQm[PSmcw֩m. $|ޙ-Id!UG= hD Z[ j#x_;sch蕇o,z7vyv~gwffr~m_>~eLl^iF3o>̱ͧ̐cS{Ibn>a H(y󼯁8vbġN`_h}MO!1a)LSfȲyp+ +'@Í (#C2&ͧ̐cDܛOkͧ̐c`J<_fy2,_f?VIk|w ⱻicl5iW/f?,8ocrl3F(|:H=CrQoVVRFnOέ#{3ў?^Z_-lY` ":%XH#i*x9H"WHm&̮@P\rj!щTH73 ʠlOo+j>QЧo]=:%}xL\h+yX@祃}Z G?8w=8'oN3E(&/ǤF \S/~/V(bGbI-#G$m~1J0IΧ@@K*q%:ߥMo8nDC֚YƼcƭgLN,JJT'ou[X_bX鞕QFF4 z-L рDUBK5;*#iؠkMfKcC#;&>/|Tϲ=]8 uB { ͒{M_rw4qBLq6bK?gK xmP*LiEck0ҠZnZ N`_K:jX)Y0xq,ދ{C LKBY >@ﰾ}zӗic` :!6w|e5¸91㋫U,tS}W?hAd 7>a)s 쯴U!@Zo, cd-"҇fv~5u*j:'BS.F'K-e|#)s O>r-ZKByo_ju1loN_/pL28q=̉[D"RXuNwC K¾ w1n`i9'}-th~x·r`׮)ߝ3 (^:r\:Iޓ3$ipjfJGΡ(GSr>яMERt)(9-~G/pWiyɣ(Vs |NO=Je/9ZJok^yp9i,%-8Kd=jL'!Cu-5Y$ijꮳ!BMW7"BTiS5S&ED]JЂ[%9ƈHFQLx&D̡RAnɢ-/Ü4`Ap5XDmOA NT'$NA\Uvp77.܄ŪZ 5Ӷ{/n|UT#.f{om7#v` uJ&?yfcĂ@spe^GWLb>,Nɭ@S{2rjKq25$)maҿÛK`qwcZNKƻUaX# )>HH>&aP.y)# 3Öt ppVVhʯ M>S.j<hʯ H>P%Ob{[xl8c[d e+.t&|,A*Jy7MiJIyxJMcB4|,MI%=Itg$1h !vOaD?a[F)|y9s&|S(K\+'>ry`EI'zsS7ZIbʺ,,>j䃛3}-C0i+'!ÕĴ`~y&;hi(Q5ww!I8 UvZ#M2 -l fi ޠh-'ÖÔRQ^7g +%hY/f3J)f49/NqF#>6HyGCzԂ-/.Hn]Q!4Z}t.48z}*7`3] Am=5X}pu $6/&G|ISfM)-3\=9ݝiY JPʜϏ nB mCOК2ShT^ABm xOd&a)Vrvݪ&[}ׯƜRx=깧&z(FXqTAl 1qMVp dj3ڿ +hyuA=Uyw R׻sftud2sZn -\ jW yRӀ]4ƈ<3{Z3`ps!|{XW3?~uEdGcR)g2sN`.j*D7";5 =n՚cJfeeg iKUr8oxUkT,էDı̱Z2ӽh|m8?CgrJgh_JQg*7F) xV [IbB,EA`|:aDwQLPVݽ"_ʭZZ%!U7ݽsj=_Qp|4έT*{+gQt gU_( GBQp#]WJp>>~ι^q眹T" 5#Ǔ#2"K0IW^Y9j˃Jgg:T@๷.1 zioT&e;_er1-f&g1Z"-TdÏW@ R=V/u&&5ywU$׮:\3h5ڮ y}뽔_14-x6{zFֺF娍ʼ+3Yr1/ BN%PO-2-dDP"++ 2Dj.yatBfiLB^xID0Lm+y3nJEv=>IƜBj+9ZjV>޽ܔv>D0_=јH22eVkVx n_?f>f\_Wߤ-}}?~ ؽ Ȯ^J`Ws9sjHCH0#"3Q807eꯋs[⬻x{V{PK뫖YNH o2 ų]d ~'1׋{PrO߈(bDcN˺qBm*$,`O Uѫ@@)3"790H߯W߂ o-Lw̺d[&tJmJHz%-H^jIדS%b`5a& *Bϊ$.K"M^Ivcґ`0208DF0d1 ܓD u ԒbA2 [C( v!wQLNz.yģ&G z* #zT,t*%ims0,#޳$lpb R;Zlp=;2; S2{k TJWlTD}@'fz}$#T$5C`À0%VPppdTbWY # ܸp#:Z3m-J'f(-%Pب1FH{40˄6aIP4\]4€%c(M `zi6RS" 9b,nLV Ӷ^aD=%t@c,=c-HNL`; 328CSD~yZ@'Xs} Gε Qr&aLM)D Ԋq@ =%y˗˔ΐٛ`#dd\{Q2]u'8TMp*(e]*  1r:ra,&Iц^7?BM~G"8rFLi<~n6"t>  iT8"`c4Rǭ^:VhlѦB8JjS{p)q&-a *[G;C:A e[EƎLQHĝ \sb @ U<@0H}o[]Z BNhFލhd\`ypי`pƆFb:,4($,g ՘z8E"J[D8FKB< ș|C6 Sgf4tcΌԃȻ-Hu 7,=8 X&l~zu$| ?(RDoǝ SR jpz㬏)?W=?, >g+w7zônkao#Tb#ID Ma X O *&E* [ 9"w w r? 6#‚ qc"coU42ArANRaɜ"I#x߼{k;nٺIcQkv\|؁c&W&V"j+ x, ffߢ͂r$v8#v$-ȒWGiq q<^_֢0vMB|Y׳01o|c-nWbWź]!#p_= rjWo ({]Xc>~`?_P!>5{`ny<ÄփYń . J\q|ƪY5~nweqF_KJF}J/T)KR\zt={3҃"8~ht7KRRaW9&n+r"`٬MŝoZz̯#&Mcv[i17@99‚^9AbG4\{廋l6 rF(`G 2{.Iё':_A‘Uƻu:7ý=yXYx!˝ /s"5Bnt;OPb6̀X70HYNš0{ptL@&o?'z0wtrhܧsx>́"X@'Mxnț9ȥsO_O)w]]3L#S"Ǡ3=s;N҂ `1O2J_nX{L{wO?ЃZ>8;=z1"zF:sƙ@pdg9W㵯{MW-,F.mT𕶚TAExX}o򶫂cx5RAj:a0iq?Mտ^QL,W 4HXkoVi ^тY k9>ZQ+*:dSrQi-VC5|B1/B`{BZH֢{Wi@s8_xc tt|Y|+@;8'78LSV s-ھICkO dvHn& C;,L+%VK}u.|[?jQ3DVpGwURUx$'#c,4wnyw3n؋k,~ԝCdtG%hBHWӧ[ɇ7aR6T+^o{\vkU4i _kÆWA4I߮ࡋ.h(rI"ߢלo/E],p*0yL1XpA,pi>=?1S \?|:!/_Mn{}x]H]`zGeaX6"6Vn6" +8n,;gX\o{חx9׽'DuRZ'syx0tў n"SZ'ۗ1XMӖ1]?10j&IqNR0>bVXiQiV=w >qBDpKO5ua/wNh}1ģ?vg>v3 V[m>)oy/ok B ]qQ/o'~:0_x{8] hJ yVa3>-}ˋ 9o@_mkﺚTSc˲w ?fYw5Wt ҭ)t(j}}KJ.87΢E<[TD huS kqiQ;X[A`GU)kW<f&iL8Isic<2ID0˪~JY+X}Y{)RRnm J1Nh$ `NB Ǥ`Q%xaXuR/.zSZbMLS[$gDdӄPd&Li ;0iX#Z!$_D_dd:{^|N/V ')Bdc+f`<7퍽\ }RE\ph񊨎W"szb3y r '+5.G4z'*ƔBEA3V:Y]/aapn1 (UYI"pHAE]/> #h0uXe ;g]S>5B.D$׸ rMYT-}APxVYą]x % hˤp߀Dx R rac4מhZ _[\mg"j08_kOQ[)9S:Fv>_ei]׷t+^htC~,ZSg#&8[!YS8F#6M9pҭtҭh{9R\3cWXg7ȩ\RDŕgTy\)*I%RRH]PUm|# qWs`Yմ 8Y)oŬCsݭ1/iEG.D__]Z~BP E#*T]1Le3NcC#P_͑&ejxZC)DpikP$lDƻBG11!㬔GI.Wh1ޥCKfP @*e TC$FSy.Jfk⊄!#Lap3 ~ έE/tx+^YĊ+m5: TQo\Wczjt7n!i =j /|בG9f;-55 8 1=s;N"Iy0țH7$݉:AwMH/ b òX(̞LHwݝs g%yF3>0& #crc$]Vj,ď1j_0V95F5OnY VG .EWpW/ {5=Fpf^78c+<9S~pQ|ς 2 f]>1[(0DπBU!BG2y_VY;UnYOߨtn흝RV\QB՚ N]EM7!FlW5T9b=m,ÅiwFei.[N*Lٻ7$W~Ye}Z^̬Ƽa)ѢMJ"QͬʣJcM%"#⋈!ECpH+8H#,,'sbےr"O†/5˕Fj52sP2S3 kz 3MEIRW fᥭyK K^Y OJ*.;䘠b[cUl݂Z|Tl}7qOٶO]~AW??3gNĿSZGwH!e5r l;)m؟yw$b$L@!0QvC rpK '>-H1$gufW? R D4zo$3)+R1B!Z(锍کpAm!ao# yP>tM7ن8Jŕm4Yy#11y 1@P9#^ڗxY476br0E55JPA¯B_q'k]]ŨOhYn,`儷[@>HN<|,\$"~uKP|q~eocx6_tOJݏ]򋧎7,Z :%xaa%[o?*ם:S`Ku R aV5RQX=(0[n[vek{<,z}ܴZ X`sD?b!3*x@ľ ?]/ +ٗ;_; 8GKq–t}0_cz ~' wV+h4~i n_>X|_O]?ߗʸ1oհD 9j<3q4`RZL%oч#UM3b V̈ k_%H[#L' aa$MX:/"q]f3zJ:Jwͦ;)kCGN‘6 A %'G=:r@d([1P((= w(`V*uZ{9 &#?xF"ILMG}*aRC P㝂4]Ke1&h5iARN' CͬKf bɣ[%] 3yU|owjOEo߯.2|ϛnHO˰~jgk[n?H}Sm^ET~{۞8v6UѦ~ݾ.~}6]uz'=XKM<3w>E) $y&5fruK j$>c6?T|uKhaΧhҜ`!%H+r-x$Ԃ`;=ǒ0u+ Qf>:}S`AjUIe `rI`W?.'lr噒@W<[`@ Dx%WXQ%f$_ղJ.f-_/-?@gN7ET 'A\bq:G#s ԛ%r ЙDS~% f$8`<\%XCӖkXD**lRڱQZ@)e@HT ["Lqmig\8IkPyCaԌϷw>/:!~*(D9 Wnp<_1XE1cb܎_ȣNɧH o+/V a=CJB!AQ)#}(L7:ONAH 3GiBw) |H1YbeH\ʁ>}&iV;=BK/.gpf1}Q{W& QR1v9:v׺/Q@)!;9NUWE*`فfY; KX5Ǫ--L'Kvmk@Or;mX]ogTWfIy^]hZ]dBچb&0N)j,):z^+ !UsEG+鬑:(@a٨*E;~ų/VNYP/lnOf6fmUBIUX6D1EC7L/o.1e%O0lC%6j,TZcz1p+=x&'T&jc:hHGCs Qe }nC2 ^tk&M aZ /uJ|e1=d̃WD :njE0wܱ+|L9TX n0sa[܆mݽWZ WC#.gaZU$iPR+[]LHv%ɟ*yAx%&hkZmJ>L_>X|_Olb~y\iS[S (lYu8 aʔm)dHa'"*|ةim {ҝl3ŏ ^_KcקUfxo)a#;ˑ AO"?ҹ*4}H¢JzӁPa1a͆RZ M$|^~j Ak(mT8}7&ʔa`ߜȒz%<%PW({1geA v\FAYC.qX.gA Kft&u^_mCRO 7*X}6PY"m/=4nŁ9A5joyf1 ɊicϢ6ڳ[bz?[Oy0>+;kX]}gLB޼ muE^b]B)ߊSI$FʯA,rF{!lQ -ظ# rL sJk O@<7 tiN%Sd{y|6ۻj# ־7 N Q`]3@&×bk]uQiB3|4L4o@Tko an'ZA>Y-pZ18A A>cmz??k~?9`FٔPMy؆h~|HLsDkX[=0wC@Τ)E V'5?oFޢCr&$pKb"2t.[IEbkA($b$Jk7ƱT4ʨlti C(AVpnV歡IbM3!9g!LsID#?#`^RA*Ɇ"Lyp8:7`H>J\ ʜyX7-}%2OwO^ }aLNGrg:h‹07C׷oyz ~ׯw~A4#={ YXL3&t [ ۰'\m7[6X`0xm #KKr{Cݜ.J̬J1qq!#2?|aTA)z.Cɿ|ǃGU v?j|{ut܆c#o^ s )&w|[v`WX{~>MvQ믗@(l?NO/asHh#TT-mvC6-ʭ^㢺x~/Crk N &Ɗ凟%/?SprƩ?|7nJZKdQ{.ss!]AeB)R\AxꏳxF-G稉R^Ҹ6H02~Je9wฒDAHIGY:,=AEw+􎨬8ymsYӼ&?^4g0Gx.;McqfibSIد!ahN8RnL:)48`ny+$cHn)*s{O!!qmk 4sjڭ}DCۂnKOinSH_\DG9:9#R ity ' UjqsJTg¾VQiSw6RG ޳ -ut N)lw/Pg V[A,،lΠ{Bhwd,A+y] CҖyi8ͺК-͂:+W*`mrK=9y51.}c`NׂCW*5sQw1%`Naիt{SFJNY%C7`nbMJ n ߬]SmjTM{Bxʃ>XrzHI>im7=OZ)m4n{>#~N=~|i&߯#a[JO=^!aKƧF_awLYm~xGُ'gkݠgOlF?_P?urAoӳrz҇!̬_O.o\/_>cXO?>('n E׏?ʯ\Uwfm.&\dGRX$bv SЍl4l^ 4 VUIf㲒,"MdIʳ#?γ2,M,~ ΜX긇k]x9^uK0 0\_رF{AP!}΄6?ȫϯvy{I#aܕی >L/_<ύfN_np~ӱWe%~X35k"ڋP-͟z6±Bͬܠvn,u$6+D8l<Nq&좻~cH;aݾ$&k>z{aG7ճb]Պu.1ahԔaX)uBб0-+>7 d6c\☍2GݱP1CctcꙞ J_1^ΕG}1#Ύ5o^ɏٺ"! `u6QiyׇaE"za%) HyߎzU57<)g[5]~w8O&ϾS>aW]oݷ8I_Tj"$<˩!{m M1³Xp!fAb$X^_aJJ4PbxWiZYGj> N9#r̊%dyX 83(G'TFwKuɵ8̿PLJW8ZhI AxSo@%=ꅌDj[G'R@ i@0z"" E(O]$a =cH^+XAAsyo MB9L!d_W7PM|I.#tLÞdFvBKh #WUUqQPϔߖSB8c;*\xGqOn/> 7cjyz  ջӺWfMovk'^o~o˛gJ{H_!iؑ> AEׁ͗ѧ͘"$A[=8;!0tP3UTUջg VF?{r9v9O'cWnX .l[Uh֩q` TqW7sܿLNjCK4ksHjh2A+0Ѡˆ-ew{S{|MrH\#3?4wb!Z'N'?XR'+;;v25 ג8\]q @,NV3$ysnWYcY/iCSڅ@ϗ^Y4Wm7es.4`a1!E t4&^pEO'c^UkOo;ffJg;rme)Ջ WJ'|q6s+>.vfvvtY񘖕X8Tr8gwJno>7*/I bf[4E{]3B[{F^[u{ݿ>}-r/qI38'/rw{=ʞ r~!r7XTH-*(i1d 4@vhi9;h%wpRSF i^cWH6$;rq؊I."AR"$֔Csfŋ=`+MI2dF 0)q"'!Q}vjIJf)fj!HO'GcP1x >hP@(eJTĸr;MDHtdb& d(! ', . lrBLt' ޡ r (!h!wuؤHM(Z@5G"Va 0r! y"7,4Ԡ% Cd P*!\"LEuH#7KLT{(Q-ZW0g'_9LJR5 )}T yүQJ$"̘iHCTݧkBrtsO!让z>IW-I)@IHCTݧ#ORKxN$ϝPjI(KyMO:^\I%v FR =F9J>j‘N*bQ\f.D@EtEak:?j&/ c@Q!hI` %FJ{.ZPjc1 WTJR$A0$k$c`Bx4i*P86jF؎|Z[NJ:.cŜRƜŅh:'|2d`?6( !q&P`WIm"%O#~jWjK"Pb(4i`Y6$TCp1D2 TF%Pe b8՜e*Q(́D9$IH>B>-cuE]hH=#VuP%*(Q,eoO .@A2n#u6f]lU->{cּ;>7eEϿnoWq;6N슍@po֫"(4PP:ZAtcEVknuz! BܴP_hHiQ9s0 6op<.s\MZg[i KU`P\-\yh15΢i`zo5sv㐣U z:6=F-I$Id16)I1S{ l"g~QMf4pP'I>)E5 N\ 4Rfh zM̍:OS|37U就.4EИNp4qsÍ5)MhBVR>k۸hnVNsTz >yw09FD3@AͷU <"蠬]#Ŧ"_yN2V4^F$Ӎ RPFVֈXTp0=ج?}nЂP&cBW5T»424"d L=7 \eS/j./0eRkTun;dČ V5a)}z,p(bb*sEeP(Muj,=/CaOzuKG!Ɉoy;¡@FSB%nPoA(Sw U<j$^z@| bK=Ǥ S|p ;Ji,`y!Z=}- Qݫu:Td"p1ٟ*'\ta5TMDUjUoƫJNW5MMl#w{6]3Vq馍:m3j"i $S371fZP VQ&vhߎz<N>dΡ;c~\ϡ8F^Ȟj327?vo qz5$bj2$@TE=J &$b9* 5i [:Ȍ?Xi d*W$%\˖nlkW?z靡\o/Wdl mv3;[ŏ3[֮;Nm]]>qU$ms} ɝUee6`yy^_~8϶ˍϫ[\&v{&_(Ako|w-9%qH_0lKwyBAx}9#&|/ů1s@&zB=Y&,6KA9݊!$5F$ T5=VR𵿵t-||.RHrD-m>nMG{@]l9ca@;>\L.vK~ (A=-ɪ:dŠ3͓:F;I؛~+2 [5?yw\̶(1<٩*8tI[{ɠ? EfБb'9|9R~8@gTޥ|g/ \ޫgǗ?>_D/WW.e?kˢ4Y˂FiG rA 0Pisk\L*b@-w7yBl[9Hf@CRa}t,FGz If)y=? ei Wdؽ,`Rʎ )l1o}LnVѭL>~ZO}RFX! ogcrZDp}z#{5^zovZvb~{FYs9D`3rah asy%IS e2vڴyyfiwqkmFE۞5I~XdEpr6^vdrt`[ݒ%lbmvX׳ II#?UW遟/2ɶſ?> uɇ"WtE?i5ķ\%m\fg08'#-d:[J妿DI͞zp{g EAmNg-`)J "g?p8M$1G*}6Em/Ƞ?& )qb˪WileW׷B<[YUnO/w}Էm怺LO;Frܚd?o7̐nfGho.d$Gt+QM-8frKqHF橯@2ql|Xӏwc_y_FA[L>}c2a>y y-xI`H!ΘE-OWT ةru&Zj!Vm5ԐZ<4T#Rad>j5l,*x m&2v}>$jj i?]Ϗe\}&5՚Paےu4 Db~a6v be A'.0ĠƳ4dokB2nLE5k)ܩ&t\5!yAD熂P62ᥳŌ3A2&1d҅FB1vJA5h;KxI1yN;K/`HPK ` r\)DIAB 6@kW70CY˒să3tV@\p4 UNyeb/hB> D4 |):(05)TUZ!?C܇' wbӌx- 4dS:!ů1 Ny3`D-X*sZWܐ`ڳFnq_tg82S|nH)jC[\67cA,:00 /Q}Nu NuS4R+(.jMH>JVKi^zas=:F mxNxomEz 5Sjt5)CZOzxOMI`Ǧ U>)Kq34eZsCb9|^M[aKC*]}P/tղ:ppIHBh[_{An}T5՜RӚÈ2/.IfDq_„hш^E8z2,^6%ؗRF؈+\5#MjN6T&1nRD1T'/Q}N5椥G2ԪԱtGc-}u59ii)$q T[3_ L6 F\o2WT .NKcl]uZDVTN4},;ʃ@[FK5A;i1jiZOXЊXRNT%F!*',#Ceb)X$TY.8/tɬ_UY<MD%Ղ1^[i7MsQTʃRjҴZ*uǭi*0ÖUTsU['-=F-MVEh "1][-r.IJ)x+pZ9f%x:ϕ BagRՃw%sC\jtBVӋIv B))ZTک[}!ﵪ7iIl / O(A$JXH st1S;i5 j0[I `.ETEF1O"2~%9W% tܕϏ{1>YMJYGU4Kj@=꣟Y7TuA侣u1]ެ[Dօq)C>|f%ny.ϋ2ɆvػPMT=8؟t~3[X9>D* (jSXZ˒'+|#&]k|Rn&`}cN lXI{UxéisQ)WW ک1 %UWUE)Rvj5AʇD])l6Sˌf}yM,.|W3Jl+76tY\ʂjOAh:Ļit_F~! I!똶zDt^IgGƖD6YŌn8[|x^'&l9~|uuuu]J\ݽh .3.w0ogÆ%)}Eq-Қ0 S)Gܜ cՃ'm<,X(*aƒs2b5lF) P)710^ ! /}^9.wH ϖ@pW~/+^jV]Yx'WQ3x6{3R;vܗH^l6T(ٻIn+yWpA28^,a}`Ukn#[2fVdgdުΌ8qv邒ā-&,4gPpNeA^5p;&3n`R1W1l.A>' )WPev%~,"[䬐r+O>@ܲk(eEOwqk6$@Gq2":|49I'D@O%Tqf /MMx}UYb",pB..0<\!Z:;I%.mas?qB lgX͡ie S.碮4ehH0Ȩ &@wn72?R DVʸRfm!UJ mEY.hrU4VT$IZڇ579[vOPyDdoۇ>4n60S3DX>Kϕ2<7җ71z$ vˌИ~#dg^myG)jG`%U |#(gG\eYz*7'-o55%HTjnm̨ȚAe7bY'c[5}:C^8ES8OGev][*!IFLv0Ll/6Of!)ĩ&qe7EdT B6,d-(~`v!/NJe6@qh5Ujo8O:aZnUk9Pk9dZ#i-S;uJ`ǁ,%+%<&exvфفȕ .c]:2n/IM?:fμd5drR@5M,s.χ_m7>9Ǐo^mn+>Wa|n]vsy5+~K]7rG7]cc~s #NΒCx^@N)Z1VqUyX4BNXIHDZVS6iTíݳZ^E'0!湵'H/*ڧGwW/ve3sB{#S;<{XNE)5GPp $rH6?:$l'C@*FXM(4x?IAs( uEh_?Hk*rD(VVZdK&O u: g7GAYPp܎|?#xᢦ4_b1k&R8 /'՜^K^d}gZ~"*&l,:IEasOUO5jI Ug㨪3bUi*˕y.˪ hYe2cIl#t 8`!{ݹ=d4Hp,1ܯPk]LJFϋTa.6s,m A*2*<{0B\b\rRD5b(M׬TIP83\'Zc *-1ikLFK|5"+ϣ5rQ2J|ReFo).pyi\G{^%;[f)0ҭ]#ھZ[MHVV*DҿY<]+q$ JNFyoIyϞ|?pʩGvl=RiX݁ $YCzs9emjq.<Nws;2EIйZ̈́%-FAB{8u & CUHz9?k\AJCl%6Dxm(ԍR& {'$w;l%4~4 ZHۤY"D4Yݍ48}\J/S8Dd@搢O1 hjG^!}zr*6c qe+dLQ*dD5j}A(. 9TqI,p 8nD͵>< `7mm8 3pH qr.DxB%~'od-({a5FBˌS b{-kճhKaV;4`KnVm:Qe8 oĨ ?Le8} 2<C^8E8WC}žz/5gR1N7bی/Sqemppu~v3%dT B6,d-nF`v!/9e|BDVN@c-}+|`շ]I˗x`Yě-:Pmc5#+K/5yn@2[}SWojqh$ǻ ;TW,.-끴OZ )9ȈH[.Q䔦_tHZ# C߅!VrX\5:(3s3ApJ/n'wV#j$t2(}nom:{S?~j.n=}Z*~X!oN1|y=o i5?$fa<|&\4>,z> E#xe-X@+g4l +e+,2'ɲՑo:E2vްn~} `?rtt >΋y1H' !K MVX6Cϵ ֩qY¥UD,.Y"1Tj‚^]Jp}Z2Ȩo rqV:7Eeàԕβ<6"L|0~1MD|֓Ͱ@}]+%zN*ll z(ԓϟz$T\4ucGw9OM?C̵n.?e m<DX/*TEta)GW1c`tQA"aepyIG,p 82hf8`>W}˝Л7pYI.ɹreL#I~m8[΀:`8ܯdxٷ J2zNd%q+ -N_fy)^h?F1VhZJ Hətۦ8!95=UJmdYa:V+fԌR.bZh ЗRm0ZKOY}۵Z[W^4K9&9Yzۮ(Y'ؚrk5++K/&p*a6S c5.?mmjk, =9m|~j9e-_L*{o|[lPu.%rg <̼2SJ;PH,MO-^?Ս̛c qӕå:әN믿NAn~7u-ֵƼ܋ee4!)wzQ\餬g`)ӓGHYETP;I_;MŮӷH>5?t??-^tq`Ydm&WeK>l+ 4mZvj*|e%T0*[+k5YVnkG4ëE4jҰv[;q, xa?h^Yz, xaw/E4jn4X*'Co `ΞvVᕥҰٓ|=i͞vV3e4L)ynSՒKÊ |Z,^Yz, T kCFl(9Q9eTnWRof6q%ԽVA:!sU%Ը + pG\ RV l.YLiʪ,@]{4Kҵ"W#1GՕˀ ? )wynpBRwUs:n3v-Uv*R^mZ7HMā#pΠĻIW)%$ !j$W[u܏t&3qD;Z?{WF/w"/Eh\$3`f2fhVI~ldb[R (XX,VB[f-Z|!Q,!Z!5m[)5PJ;4(xJ^B_f$hvWKΆqsڶQQEIG*+th):ڴSm):HO_^!e= Q DsKڌM#DHrsIXRb>P1K%2H 2'2YJ0U{0ZVD(OlV 92:1ih(!ReiXFͬ ). A2 ` PvCY!@9/A;t& MGDE c(%GWe" <{Dևd,MQiBcJ'"VT٢jx>C6Hoa79REJAOn2_D͢y}7t__ " xBxnr".0MD(CM?Z4V >'ъX[0|&\\u8kuL3%4O*.آ`Xfdl8gZpV\p/Ħ'R"|ǂ2uT4\gp[IK4Qd\T&XM;B$)3vb[f6Vڪm3m% nTڭ.RSw*픧)Kn[y]TbX~QPZiji׊v#"PTH1295}o!Tjj05ڛ-3eG.CaQg+귳yۈ T96Ls{3Xmqqoqrm 0v`r{ zƞwR4DY@x)iꀯd͉ DÖ^_Z(k>&HyV]@OI`c) %g?*}?aq&uz@^xvF@roOjjy"(sdM,sRׂa~-Fmpؐ- $vHQEekasQe4}sD۫$fٮQ-(Sj5y ׳ĢEVThȈ%܀:i2-;~0)Q ;jwNT]?oWtDS}N5(ݟRJPJiKKRO稾|B5m2RVXxLV4TG)i(-gLCiAP iQ : PP@)4T`.OO\πR+wJV[$_DOVrT-氖LC:2r?^HR'n|`noǟWEMXX)2Ʒٷa^'VK(O!A$%ީW}Z")\>:ib PZ%ٷT RyJ1,*7S&ScbģnC>΢T41xwrIq[fQt:='0i:ggO&jND3%,ܵ w:Gˇ]+8$侘(d}W \Y7^~.ۯ~6rz7n>~- E !.jGE;:i?4ۈ)|.fӨ>oG8\ޝ L 3%qš:nde?,~[&XSm&q:~bLOl^M~TR7Jhv{=y"zr_J1Cɨ׀͉IS+z"}Y9]T;9xROᚑBMo4L\x9%MQrJ'MQ$2'טjB8:f$,fVYƼ8j(5*sҒR=xJ^oK9"SI^hD:H .CeY[GYRMjTF#n>ϩ1wXD1+rw=v`xf9]`Q"_2/ͯnޜ\/$| gEޜoo&vD𩿹^r4 D[>%ieB,rG7~fC}̋ngqv!U-9/Oߧׇ O>쯃E/_.!9Ƥ䇖\|7>$7DҴD5CKtbW[]|?'EFe&$;|tn/w"-b̒3}翌oGcP9dzʉS}9qT"3 0,0x'sr1r&o!ɏKw0.^|[vER{vu&GY׎Ӄ< 7/:aQ7T ?g?=Ѩ2 K/V'2~JSDžb\<钹z;fuw3|.cnpƥI!f,BF0SJT\'gyoοd/>rruɋbJ=_34kyQ{ʐ,Zmӆ0)HIp[ȃFI֩DL'wR}j$ԉ(#7KFz7xe"R7cQ8!Z!ӽ_ԉ,h` ;u5_+uzt9ҭvO'GOI[()᤹H@)xx*\n߆t:'^+A=F3ܐkeƣbm+ySFu`HOD/Nߗ"ӽ9>MO0ID[x _*j{$^%@#UEp( Jr+ʡOi9=^/UԴIߞC:M|.BKY|NsH"9mn:*D_E 3:q)x B08wD@k8 Zgc/\39 *'j= n|v]jPN#F$gP4ĮQ{Dr$heu؈LRM1ɇr7ʂV?Y<`&wEE'SHv(jDU;dҤ_^S"#Y䀥CKDQn]6 qv^I(i8ID( %Ё $l?wXDL)u)JgO31Ȥ*hU|+9n2C#"( He,QP̞SWbr cM.?umnn:8w Wju L4sBꂧuبxD ,X36\YbOe!d NxP̂5$E vt  14 i)M\>vEm ()]4P1/M$m%A:NVBw%S;Tv\X74VRI2\_P{_fTнF K촗Qx{UQ>ZpYIODωh[F ,?git ϴOldF)KsYE J9jB:e%* (n)4#Yl{Q?㬭_Rvj2.Z"wHAe$J8dHuE2QЌqdHu‘k;WYTo{ө]Չ(DًX>8#01KU*͌Hq$ĝFr@!Y$Q"N9%&NJ&RNgLEhw {pQ0D[kx?_ަ[‚ c K8χ0 <=mQi%ǰ8&Sœ$qFv#c7C~KX[9(Ba`[,ƺK&fe8P7K.W/|kXmx<`{:׳hx 2dV42X1 ؟wh{sE{Q8^!B{[>~ij=w0.JGY3p&$/[Xs=Lj 6!?[>.xrLr_@gpfR?7\ss>3'(wUJ2by^3ʛtkt|>8L2vჰdql,0NU IXYuw[?y0-Zӊqޢ w<5MnÄ6O;m}\'฽M hm4?%jXq߈1U+բ<.c=c1~,>ľ/8*8tSmD+I>ZTv _H:V,oXzY[n>4k;C6+ȱ[N'V]L˺PdGZZM-&^U?ݺ8$8Hq1Uߖ+Z6MH) gߩsޓ937LLha9C(|O~38zЃ\,e{ M4h#ƌΪg ʼX \hTo`5}Lm>yL'EO=|A>ն7}7Fٟn~k Y녊P\Kn{D*Awv w}KBXD@i0=}Z=8癗'pEI| A'gb'*ާwBnVz!=(G+%thtziIuHeGn[[r5o(.=SB-I4V0#NC]ѓ!ґ!rlj iO?(v%GB?q BxXn NxPMW#2>R:(2ZS(*:iSmSdZ\'"9)Zʹ[lňQ͊FZ|YNhFT{-Ze-n/#-^oh\>3ӭK2,Xx _*\zp$eT[Ge&]lLPLPcsMw]dཌྷY}&bS3R9Ts-[j"ɾ#˭] Rqrv Y. NmBܡyZ3w>?4y*'/SU[oNX^ﮯku2|o)٦R=iy=F 3B3}p"^k K݃y([mqmE {gy>|~|#+I[qy3WO$-OB$H]Uh"ir(DzJ։RKIIT7R)$RZBHd&I{R>ޢoҞR%VI"ѥP-FEBToDҌ=ѥ®{RgvKaL꫑IN ȟξ%C9Z0c/t .ͦPE{ "b4# )ӑ3D@V?Uͷ Am" èhO85vBRq(dYSz|t߷wΞ8mAI QkWѰԏG8Z&R ɑ^ȫwEy]Bۧqu^g] goOD5,=) ).E} ؘBx$Q;h溻 %MљL85$21IIb0C2%v+R(FUR q|Y qnk^;1PcN B`Z28Y*UHEž״D12rQXg,4K G  0DD)Sd=i)eMXfJ3}1!Jlґ83]<̋'ǐrdU-Xt io*MaJqf&aQFTcmEq#38'y=8{p^'ϰf=8 [lEڶm7"m|a]VtGj&b#?^qZ{ ]!\ɗ?|xʖC^/ߐY{}Ƀ1BxTCgp!|Zy%XK{Ɍ?=xƒע}4_|2mScUWxlZ>>Xvb)%䱱X0L~ɗNc(3k{ۆ.Q=V;QtbGgۊ#F=m oϧ#@X`5/QeW`> VOonSu!墇32ҽds)?v=d9%W~O%q3SNTSZ)^w.S<ݺhDiZHͧP7sZr4&o虆,:?J$?]U=<ڗyb@z9u6*O7iQJ0|SzLKՍf5Bt_ƯS_}uF 359O5;( (2٢_)K>w.FyxaJl&V @sE5p3북s a~Ozsiɹ̃O? dBNN< 7dK ew]BEJY9Z,gU.9߳GoY';C`J֧l%Q;e !OLR]&IN-. ֡2-9n[C4SZ=֎MkLA:Z?vdnM_pnM OML)BdumݒںPYז\22yq%ʓ=Hxבe5y\r. H澲Q/@S} raoƞ,> ki93r\\ Vԧ6F_9Tőgt`&Y0*g&[έ;fx~ `&vkա>{67V] P5d!Ѭ)ۈQ*5,#  wU"'3_/ -!,%v}pq/}$;hg(4}5b"{ƞ@A9󯓧'm}0=_Pi/OwsiE/V=J~2{#dϸ셳W@8 W_.8˧a૽8p5 }u̓rf&x|9 NPͩy+ٵ6P8+V؜Dq fj&>Q*!jrA}MCao3JO`r.KfjKh(=ERRJ-J . 9 3JO~P(e 9Ք3JO9\+>١ZirFIc7/e( 9ՌcG)!n(%Z|T/PJJs)+U8Q=UDJݢՌQvFI-zR8D EO+QʅJys7T 'RH,v{*V3JO3 ҂jM'R*PJe~D^7TSic7"_K9vCiN5#:abt\sy{ 7T3~֥'Rָscmdknd[{I,FD'O-/l\!jmm|.޻Y;M?lh>|~|nu4]/u%m}(5 JZB DXd|-H(Jv)&Kţ[Y* O?2VۓsWL=Omw~1 + ݑ5Iza=|ݢpD*S4'U#ï:e$cd~eWy/a$8"=la߉H[:hwqqҞjws,"".5kV"&g*լ *,T6UGl"S:)!ې"P"E6$2ņX D\#cfH۝Kw厥^MNTc'Qc:~!4#5P˷b3 Q1UG(43O!,V8D0ɑhXI&&1!ld ͵>Z۝4!ηO^x2>><u #$(Rrs["L/v@_ރi拕D>.bU}).TjwT!<\YW8G+Eu`<5)ILKa?-a,;@2Pdg<@`l΃>jPB5- -`Vc*rG^st2etYWm9F\88)o_ dO<_2n!ahMrı `abyX9⺣Hgdܽ9T̑aJg&#J /d$o hsVoMxwm͑L%ulqߪ}ZSJb/S*׊29qR G@vj)5h41JV}^)%t-N={w71NS.u$cQzf/Y4[{b3\RUtyg|-6ٻnW_1+q7vFsc2Ӗ/fO!ߨ%z_^Gn˾_M~Q΁l61xX-Pbubsۢ]=,o9Gj><4C>)4Nc8mH}D떊AIFu;wt*h- "к1C>)44NܵnOºb2uRcݎyBѬ[z Vuc|Sr)g \gT+㓈S` yD)Bp0h?=ȉ;cTpC//:i:$zXeo﬩J^Ը=ٟonbǫ-#~ßgm l$: >|?=qHOQ'6V]n?7Ez҂i ;Sab#;W7,O/gů3xc" <쮀  a Ec'*EK5g-7tǗ{IoDD<,uvF,BΪm_s kXvySge =2ٰwo~ɖo7O 5fY{k40]"/ri OmNa[VTI)0\6J<]4+ҏ68p*.UlpK- 㔢Y#*!hp$Ey4HLI!P0EߖkI ܡ (ٜe"g~v)EWiע3Fy8"\J6 udQPE3]ߧnF1tcB((iھVWK ȋA}I X3|H Lqx>ݽ`^bQNI3?jA%_vw$ON.ؠmeԓ`"F ^V⊅$bf-Xa/y As4 wqXR3|q·$K.dZ,%*lVR}g)Y{%Ӱfi%8ғa9oMb@?;\3k}mNV3uU RFL!)㪰%gIȈq-I 1t/1boIPn\fQ2X/YY>SeSw^zx;˕̱l1vv{X]9;ju3P.G*w~S'fUv=[U6č;=IýO3?] Ȫќޚ~jx=S#G-ۉ[{߼{p t;O & Vw[X#FRWsmk7!T:Q3DyQmFgtX|rW>FjDoD7vU&Ja;GPJAQ cGW!hpz 9M

|oaV*mq}e_)10w}ٗ~f_}gesf_`Wsy扗ynE$ RdйiLΑR lVh~{;ʫbi%CogV3yd2ܔݪlݦw*?5;H' t7SH(BlVRR6cXڪ\ ¹ K`]dBYMl jsJ󢍄"!OJXiJATՖ o)j ؝U-b .ekڻv Ok†#;)+NF#9=Ip} 5to=hgxj]QvnzݧѿT9 T 1:Vm>56?Hm=Aff[Bp|j/F/UQU>)}UrTrGۤEB#1  .(ak~V7=ƧeA~`|SXvUpgWUCNۜK5{B>O07Tc0+v AA ̐! 4#z(<ˋ|Nm^(x3da/CXawB3k- F4/<ːA#p.JtT g62Kt%5 ^CوG++"A֌s`nqЩ'6B+CJap#p?gG涬0%M=vUwU^azz/I5{?z)3!a?l K:82UI V|1+>p9a  K}]oK"㕶uS+QK!i-投Kw4&+ x灧S ҹBrZ;?^ye2F3DQ-x B9A˧>^jH[yFH2n@@Og;Aޏk { 5!N'H5>+L#Uj!@.`Zm(!2I`'YB#9#U^LChJ&"0 U݋ph$1܉29Jh>uNp^rz9qdy#'P~*.1ƖKao[?v{t*~–ꋊ&zo[b@1BR*ьۺ5J .u!4q-N[:eYм0rKvM 8$o1Zb[e-պ3zw.#Ҋ虴9H[iNqKBЎ--8^s?{G^UBryW A*j1r[ZNS!gMvy9,RLOjsNiP8PEYJ\w@mPw/NQ@(J(+r̥T 2hJȐɂsYY89SW%PP>.K/¾ "MS((74nP.{{!y0eꇒ((S?}*+A__Kzm#R]~zV<;SW~ZW *Ah~/X%?#x]~iꓯIt 爩f)Mεq*+;ߞ}zF8sy1W_Woguw'#\n`ps}R =oxQ/_wK5t{?"8C ,rU0lNT^d%S lIj!tFFReJKۜ)`NmW5 %qp/$C[;M!bn)k$h*K93:+$&̐ss%r,%J-1G*C! Đ.$Zcٲ䁷۲-}@7WiYj[E{>aܝ>$@ j_Hdδ |I F{Mk=Gۆ~w)rY^ MugM-ʅtRj fقD`l}0 f=8*f3{_*r ެ y>Cԙ ҅Je=˜Tb3DcǠ82Ov3/{Q##p޳jT"g*pi3}CCʫΖ0A& |!p$3A D@tf&()#)f$97} P\޾pb0T1m [A/BD6)u Ios3͑ ݶzawz @q 3 ~'-bcԪ+p9xl.=y%xKm8"%En7VlObX%JDLʉ5skq[W$W31ΨULhpt:8AC" ?(` (&(WAg{4 (FԢpJ0O..#F^F@!o(jSx7lR2]srz-0S!Gp#3h#A9WE@(^5o0k9pk BK$W{ mSO[c$[pkZ/F_pr=\7cD` 8NB'[C1箾`Z q2CgkNZS/l%@'dw@FO}] `5.:}K;:K:ReF!cZRS@&>H9uNOy$gl|HԜ%KqһnW_yq?{ƭ0/HKfY[{68/`W[V;Y`MՒHot 3VX*?7^HmL2u`n?/zOgfK \rӡWn6|;n:o.ޭu^US֚w'tnu C S+C@:mx.oy7-KVB^8DK$0ႩÂ^2\27T 21I ɥ|Nf02YevԒ>meOgq|O-¤Eɒ'%F}MatӉ`36d4߬zxu{6A~e$osx܏bD$Z~i)IYןwr xG%+ZnE4k;kȧKRU9=e۾<G$DR|He(DD>$%UG5eb<4+?'z`Nlqwg68ۧw|_9W@ j-{CVe,HڐeɆ1L+vgOtqQ}+0.(Pc6H\"r y54zšmXq - :\rpK,!)H08愥j*,c݀CXD7˼+L׭i_gdV,c s>d߃ld '^#ҀVF&(4gZOR'G0ֿ>ihhh l> b!EJ2LD@N5BTAH@*F2&Y5 D&?;y4aڤ6CoӾ}JOXu$WeG'U X]@ߧVfy3Kx^%zz 5W+/K(dw1Q($I" c+'m ,01! $c+I(gtrnDûg 74|]+3F(*DLH(=u De9eكn #<.p@_u[dU-iψۣHzsnԉ t 6܌[]<>wUO*% v>C Dz⾳~z+_JP Sf 32Rsl\+tcGh8us-]xq 5@MEm&5]y6,OvZB ZN/XU8UzO{|b3p|ςݾ#RDMIToKe7T+Pch߱j"ɹM-tɯI|;XU>&m2oEy[Zz\vm2v1Ho~o,-'=ʼn¸)E+n')-w<1Rs|g>䒏zs{Ad%nS&lu,AzBZMmb?|SBR%kN!`2RM;:QwZtNZSuZ] ?;upAbb%DODsTADnսB4Gb[G + x—K- y^{H+]'U#twu-.AJ)0]wT(Ӳ:D5}{(]F {+5YRw_\KSM/kIvS !gxr7Ks[|f>U_)Dk(cPQ}SFzRz4ԩIlo6Mvo%1RLQ]9 BɐhgМ3i$Pb1 8A`߱~SSNMJ~N[e3K Nip0 I(Dh…'F80 0)zNCHT7Rc Q#ê@0k VƁD%9^֖5.K:=yW@1Y&P9 vqPQD%14Q5;ָܐ(T۰>6&J9Б꥔00powXɁy=JqYmpچS 蠄nh˥9&5zz 6s:5-X/#8'KZ$1)ΞUnY==˽5TtV)m)df1'P;T ͔rkݿ7,C&z=9:pt:ُ`Lw\.4EHvwMsTwV6Ujn=a=d\ICq!ߺ7/ AeR{od_.50 yxH ]e:6W5"/gP8=Z *gŹi5% IyG5Z_gAW6mŻu 5_/i@ лBC:8WKVB^8DK▘.f9)97a#_"d/ %AV&nx(XejY@AQsI|/PzJiYj<\y(iRƗr@)e!4QJ,՚2>Q:-,5p^5J1C) E/PzJiYjjDءd' ½@)v(ͥV (jRiR*'X Boinm= >'rU*p 34)䨢^Lߐ+e$'NwTҗ[NE( n(]80¨{rw5 f *T}FQh rԄQ$mNhF}:Xi9C?W`);mv^G,*3.r @,CI!hZzOM TǿTE WBP(i҉YJ&  (^BD8Mccʦ@IUAI8汢Hٻ޸q,W "lGaf:F`ӛ)UrUw}PUTGN{#Ŏj"b8|e++SDBc)2C+!J)TYHP3B4Xh 5BfCM\A K9wmg)cQNN mµ,?m>7m>[5}4zHnd|3}=&]eFM5esybY6o*.Lo7/smYwMosަOo', 8 C)3tWm~ $&J[5}42q@G77d3pxNOA}"Yzņ ]m+qټשyNa3fYe3IB5ᵎ4pxhJ2;ƚfS2i1ՎqV ,@Sbr=L2ՔeAu:ľUVk hXZ{N 2haK9vb"J I(!1Uh. XD#M)by,Mgi<q3Nȶ>Ìܤ~ϰ:ޤ?ϼ `;yƻ6 SFB$kƣxthF |1ΎqSINZ.Z>n;7Hɺ}@c/K  /PÛ\$7͞MȖe'%O(`"/ .AhU!~9p5ጋ+4/șPp~ǝͰ IAϩ2=v\jB<O']@,3ƌ>)\@eB( 7u)^if92NT\qE`>ҷJU8Nh.(vQÅ_2 v60N *@q@2fӓ:!.SUjaLGb5Igɿ3Bl1D—̋Y #rNr; +½n#/nVϹ", ?3[]˝i+vM^nH8=u%/{_(*j hKzp(:;wUPR#׫MLQFJ̈5"BA|Od" G%LY- >+{0 یy3sƂn+h-rrD}Yx.H\tO۞!9BW|֜z]H U)ԣiN(W$\#;hO&ɲ1v T0\HŌ_%Rp]2^""UVcYLܺKWB$kLF8 ëT|x#(RBݠvޗ+@)dp"H#Ti*fQy6wQ{ W!10eR;f#[fg[EBh=VVѦM4kj-fBkcXOW0ec\Kۙ3|KJF#QevFJB^̴Jԓ,Ҍ.05G Pyv!Ϣl̞IؐfDi\~<{s_޿ iwڣ$ЍDž>B>"'A&K}_&RQ>POksrN:ms27<' !V4rJ9#($bZ22E,#J4a#\uz&$-/3P&q^c_ŝk)J( Ex" bTKr2Ab:\@ \8' f{_%Jc6"#AW}6>ut}" -#lB٤D*|{0WTñMtx2Z}=k~w(e9{fóxl[(kAiM{n7=/SE4Ek&Lφ^W7˧n hI`4Ons6Ddqo~E@\:J`]wY n<>Ӵ ]GE_OL{ba뷣հk`A߈CN3)ӓvwz'm1ps}(,II DΌW:u00Z.޿p)Bdxn/P"-(׹CI! m7蟽wnGP`ی7ݷ \ɯk7:(*۲Mřhٲ%$\Iڟr6Xb\E"i wmɸ%82xU UDi?jv۝)2OLL<~lN͗ҥ{wgoq?5oQr:liRTg<53wqk_·asz}Yw;h1w%`pNؾ̜^O.zҩB2kdN |zFﻝh0vG>}l㫣˱I:}~W$&!}>}f3y ZYZ{~*pvU/Qo.@﷣i$Ne.S"ɝ3JNj7n|uqG͗ó$Y̝;ʩoG?M͇Wvۣvy:ߵMzx4.@.9{rDMi%( 7xH{ ^"ڱo(ݴq{2E:&_;^moo{[H~p4K^MW<`i̞ݝE0,vZu׳w~2/`IL~I?ø{6F ?Pz^C7ji?~20ΦYcVІWKK{w?yچM:`rO{ 4Gu3 l~HXSww=˨K)uoxrͷҖZ79'G0%[2^mS^z\&-] Ԗq?: ܰ״̔!{=g\!hU0M=xȃl<37KTx.JX鴽I&gY+3sUcfpu,==*ÅJi$`c] 5 a.@5&)2y:DYX X=-YԆ nbUѻ7svW9d;-A lȔ&4FdjOd**`\* }4Y2,YS+e izYj?FmL7LjILc(HC+ vJ6Uv/]ղ[ձP=&M.BʼlV_d qUDP4W.fXp͌3k hL9mQdQ%v .gycr/<9's[&`'!]KB{9L9)DzWJ Z;`kBMVmf OSjR2?)I$ѡƚx4q\a8B2({SƁV$d $l\Zo[m׋x割'%FsG`U4p:QiEH!)r[ks-U bwO2c\@ R&PFFDf&=n0R8'yxc IB24ADZ֡a!:ss>l'*B3GWh{(0WөTKEUs."9G2Y !a9G$uOuUQVm(S/63JgDH0Q]%;'GqPLBO$S㐘")w6'!LȞL%scb[̛ 'd6,}X,[':+42Te6m 5*%`!BZHPWyaJ$J;W⏣IU# ZT=<û2z,tD[;T6k~f<XnS^[EKdg/Lߏ7ѝ˶g8DI=ދ i՘*:ΏW7c 8ʬ"}CӴ-`h<<} OKa)"jYQ*<]1&#lʆ42hD.H`iVc`*N^@/WqQLiCl͛sC",!FXB}ad J= %w, Ьdfkgi?%Ǯ<~0*|t&Q"'O7q:n$*фO? M0;bvy`$#[>uADO>zшۃht.ƝN{u @Iawŀ?e&.e%X9rd9l~Ŗl[^l(Jrebq+k^J#pDCy7ӚU qcnͧc|c& wt]nFuehPcFG~~ݟDsq<sq<W%ot aF8  pH,t(Xu, c)B <1c l6-\,Zt1Xw`>/Gpy9`~+~̯=ʫ@c(IR>cp:VE[E:(*WHtHX҃+<33E%!Td(KksxDqvZ&IEUotN6ULA#m- ($w۶JD~w uhͨX ޅ[֞ //C^ۡfX ~hCx0jck-mzuWąHÆXզPE,QnL #Z 5cLR8Ts{.2a D\!/I. I" 2Q4)g"Jjn~~(f8[)Bk]{q\9.&I&BbZ Y"F,=Uvn#!09EhgDIӃ{iC4WR_T  :Պ'4^^ "C6zRhyAvrN ! ڡA |aqN#B;U{NjcBƛY:p47"C62H +b; j- >`Մ[{Y!\莻e_3{ ఽZ^k5*)Ht)kG[K;D*96`T]`iUаٌg.χg}%Y־Tcƥ\f"QeĈRﮍog]ipf?[VJ'4 0.p;0As#H>s{K9 R+#Ð?CMJɔ1_J?e~Z{T\wߧW% ޳}vq=˪`,pek_~ Y:h$ qO]{yT4n*:G2Wk@\a=V(Qr%\jSNZ<`=-};-ތ+uqo(J&MThd"Fn$1 d_\DG[6lm8ڲqՖ: t"W(ni̔#dQ KEbJ3QJ)E!]Hu(9sUKZ[?irCµ$bS 덽AxIERr% T1GAhHjVF1,WXF,-u;ĸ}v<0)y~ .4Y<ԼrzY!O|{IO] nQ~ex ad $Eg[=̣f ^%N2/bAؑ [*BV:=-e1P$H::~|N}!5yqFaķ8Eplvt\aSK.+pF? T"j#Yo0 Zgs<-=9'%PKkR*j` 3i|(`;kz? R&a"Ӛׁ(&B5ܭ}3VdΌ)RD\__{]3бwT75Ћ@5cL}R*r'"~Tʐ 6#s<,,$,h9Yh%Y w۹4+B)o+Mֹjמ|N` t%=̉A%s󻋷oan#*FB|P`ҍv0RiSeq Ňrzq~i/<2fes-xC'5P{Ά/[ŵo^/K̛>Ds,wZ>*B#z2-r^ǻMƱ^HnxX} 0y;e+ѕbwN$P];'?V1ʑ}81cΕH!37tLx4?6r~p7viF~N+Wڻq#atVu_Yk%j(Z`5ՏowY$?|O>Qȯ G<Hv PW(rV˙*H/krHw$ A|2͍ٹ͵Fr0%"RU[5/3A*_Dz4lw1(PoxArq<-bp$oVO+ݓĔ/hs U)𫡽= J md\hV:E!z+ m)#t؝1 rD09uI1Ɨ!0Md0ۅ'}]Dʄ!I{}]1ͥQ+0YpuNz iOct(%fҧAꕶ1zNmZEYo00GS>Ek9 YKR8$SX=SHK8-9@Th"-*0nYfO.z2Z›i QJv^wf͘*gއ2ϥ3&0ŐdG٦*A\mu%uQJ+Rb yW‘1x )GF\Prz9ggA\Ν^NPH H0br< AފQ0A9VMԮ,gqzC3jr lvf=j>$ w >W%e←<OR oI*d3|{)Qxa71ps08jyT--u# y֌Bs;]?0jܛ$I2RmL4kvSv`78y:\2ϝ Lh$Dw%GH+o^hToGh8Ac)z o=tӶڥ3.gx"k$BYuQ:=|$W.]dT9@T-hv|saӄ% z1q[@g̑\+N]ۮl>YugQj"yLn}u 6/Ϟ]*Ӝ'K_ڻixRZwRcRӈm,4m~.2Myî4ݚ$RJFR"XIX<#v9RPv mK٦›:S>K@g )ԈRq,lBerSb"9| 5"dVɪ5K( s3 U1N4M1>,«9|MlD3yAmͪ} Rc>LthÃ:laMaZ %&^"!>DHuDR U Z mϨhYٟxuZY8d??j8j8j8j㸪7%#V9F\H)ÅacNADlM?|,ee:d~@eCl?*Wj߬ VeprJ=S !*\8ugjܸ_ҧ*xiJ|:Yˎ)j&Ѣvl;|p^8T"{5;xhO3Vq9IJd"{*QITZgEb6^$gv"14so \G2so5FM/0͋[M ![ b!B/z[)Y/ZK mcd.]h)_FSr>j^rd=p i^u; 0HMUm.$/F${_Nߧn\NjK?7<ܘ6ȕ}3ü{3d/^<5VL*yyT N ^xN8 %fٖ7/&V%Q3/PM]T;ָ5q{B)jYy|BT}- w |VZDtkڋ;t"Ϫ|Z_?-# {!F$.v\L𕯰ڠ$< hk/ iUbLeTv b. D<ΎivMmo :A "5>j$%VRR^x]ׂ ,NN7hF "~|>2)Fڡ{- Gp bd`%1%Œ)| ! v'փ@cu{gM]dnG꜠H!l:GF%2%T#¹[i$ AJI5¤[QUb|}Kc AOhͺbrՐYjn l偛K%I@. |9X"RW kmHhVDeX]Zx_鬒dOGZ/[s/߷41ZiWZᛙv!pMq[: : m igvN7_\HRQ1HFi#RM\7#s@EE4i?:~:I5Tf|S&wa aQ],ct:1N?˸#=(zr.LЁ8ukA%ZRpR$qs%yJA۟2vAW9g9k/&%dkQEiލ~ֲr*BK`z7}F,Ɨ.SE7uu#rqYqꠚ]ΥLCvJuljZ3 %0Mz1ݤG9htDPcVkR.2ªIhb-tüRSyY]u!T9+,ݮ7FFp @ Hµ GKH-e./X#˄ #@+qZ(HUDE,-U2yIy4WI$5$"LW1:)G,ȲmVj6 /49r~r^]%epe!;Ƴ>lz;OH,~k!:ŸфELa`:\3lIyo?!Omw$+tN>2hQfYƀ!Wxd̓; n]1 uZ޹dӛ-%ؽN&Zv/0qKތ$^yhDwd"zx[6);go٩{W߆TC*z*\ )+rөsU(xUGJm/۩ϭͭR{?~ٿxwmnz1yiO{âٸejZu")R ^XEFmL[RH  СԀf0n||{>몎]3 +!Թ9FFkzvZ10YAEXepDl0 $ )ʄAjH/(0#@<:sVQ2J D? @ysˮ t (]#-dqEys: ʙY c;^|c/UsOm X^6,3Sǿ緯şw?y}eiiRnu3?x:Mdc+? D˃:MۭW :m;={i69JZ9Vj EݼU!.ǫ%U iݫwSZNWWyEg6}E{D\uMezo`ʓN;rchr3{uL>g |E yꎅ'wjKhoz7];wkA4Fvͻ5hwkBr=ڦ C9-mm0k}#UN=W vȶBj7G)@g`*p ͵i@#4a?vCآAN'Fx ]";lI+[u)ϕU0 "?_}Y`Rc|ˋ]έIg,]ɼM$(:P>LY#Q~3zSN3d]\h +#+Mui6Sp>&"sTkx&Nj5*^/6$Cy豇Fx! E7 H}Ψ4b{T+Xߺg~W`R D8E) %l'~Pd!I U GH$';ZS((Fjk%}qP׺& dGkc;ݩeteSVJysٰ\* x%M#70A(=5::g[9!4J<|#Q87X_ `nWEJz_5Ȓ"8Ϭ ^bfkI#ep4Gt]ܚr2/Dj(T]U.U^.3y`Y Go1;m&q0|NqP]0hTeм]@8gt:ė^azu{7}?4W)[&[VRnyWh;\52HŁ8M7c;@p~ϲ^QG Q^}bŐ!*بNK6n{)vŹ5@&U'=U*% hRAڤ*sFOդ d["o1Y,ԭyci3pL8ĈRJvp Ϊܾ0Ѩӟ^$M W_,[,~ջr(ͥSF7d@p`Ҁ`v>g!1՟9HzH$އC ^HPRx_IV7@/4FW%K$< X¬{Ch:nmms]~'*y5Z1J@A,c`Bx" X AԼFEizL>FH?v{(Յvl׫)4(z=yF6ӯ}gQ_=;ze0N6Jv׬xA͕Q7s5+(e AX\=Pݟj#L4UҸJ@εWɨ)ݙh+\.YZkZT0Xi2Ǹ ̿hzM{iG V:<)Ԡ:T; {GFUd:o}7(-=iA]P@(i8 i ?khm')EOvM }6w1Y_ߓQsEV'OHce"vW~"#=SyvÝ.f sW3m v; `z񽦓u x}i{gg!rNCBWbp"pk04#AN}vEAZMT$Popcmл/b h|BXE$hb * kGjG!\l~;o1.S-A~;@Pl,u0hV#I EjWkj&_@pOL^Wqwnz?唞(9_jI|?*7-`j70Wzl5.nuOxf1y;n]L8=+WjV>8kpn)CUGT GN;lޡt䁁4 Vaz|pB)%/ڻe+ikUz5*x[g bf;<:23DRrٯ)T?q0ۏGt?OGt?o۞8ጫ D[O2aVǧ kg Cj2):0-ۥ/*.}yo@X_4\Us4-g,\Ҹ4EI.,Ft +qV8W]Y6 `"*ωO6/faH<0d7WkeL*Ry#6;W Qq]̢R '?jȤFGHp%L\DTs08Gj/xB^КTBV>(bDܚԏ'/]\5|+DrF/TcK ni;ݟ Oz mnYZw y"Z"Sufv3 ]%ԁe CNֻVFiU#Du_ۊ֜#4< h"} 1~;ŴS8*kߨ1#lҥiPG[$u=G Yo I"vYlO?!"CxxF;3kΎ,ۋRNx9rzMŻK꣍64דf%Ŷ$@Y;cdUϯ^,4ВZƒVٱ;Gq9Œw@Tj!}L_ԍA@]NAA#94Rߊx؊:sN0he t ,('6&pkAk,е,¸ !Է5Z(d 'wpϞ҄NubS)U@!Gp4P թ}D-HvAQړjm %#rdP!kJCdndҁ0b^ER@|`^PZx _ʃ;wtߣ挘Zm|μBB^Ȕn]NBQN"Ȩ9S.mAĴn[O4W !!\DdJRV> y}[N{ +?Lv~5;P".p5H\I/NaY+vmqVs7yOx(+xkT< t}녫î;+PXꅣxi0"Gy>/8CCf-ScLqP*<7.KǬްsZ϶1M fR[deB-D``Bn6;©x6L$Z:QR;u x Œʶb]E(;R&.yV䋤 ۼwj-F,gU dǯ}iZi)Ӏ3v〓!#g6\~8bs^H?$Xw|Є-(="}y~fƢ=\_%5'(ӷGW leɴ7}d䲇F 2(ZA۔EmTrDC@aFtqؒD1  䮒eRlArDLi*/a[YFY{ȬU)Ry}Վ<*k̹MNE(!9wX_o*gOꗦ_&T-Rg}8oJ) L]S J+5wɹ]>~dd/?]_5=Ma $W*&O"H$9*Vp:aJ$~|jvRt+$*#5*VHˑå1j?G}i,Af‡2ggcDC@-<LZaa1>~55 76e2U;}0IUߤ k] ȩ{kz0dMyR _fiκK*m4:TzY66h(:Zyik<|i3Z2Pj)횄6%[(^wI<¹$;S!^ lF%GfnT`h\4s3!kt$Xj-J tqk+*#Hdֽͭlz(A*PVEIF6s/)'rE.ӆbS5 '(OY1 sOhB j'甏KJi%s 3ak y"_m&݃.oןS+؟مgry,_.kpn) 섨:H bA'3O.`g.iy~rLRNΉ[rr4'N̵NM޷&RQ2:$쭑xwwYK6&󿮟FqQߦvd\X-웇;?y>L>Or?Ln߿{vJ?}Tݴ_CK3Mh~DcH\3d7lu*{SPdluf4fzGd ~8ydwKYgD)5Yi/@vnɓg?'S3"s hX{4h<^lkզq$RKD8T"k8H!Ԡ$ʿq2L.ӮLZ1ote X4*SɖMh~{ |T>$iTSBKV<)%B3ўh7aCͻZWڬEF.0F&(lB#AhhI,~ oxoo<}ӭL҆X$mw^:-X8.|,池3Пo~wC)'|fW.ooO'O')Yn/OA㝹Hn,q #qE.0{g?,$k>(,#d{q+ /F:@ᐥ{=w MN G{u-Q0h@ycPPaQ N1'Z!'#ռL{SGӸSĸAAH:^P5yRnf)THA3ljU7,Υ)7$D+A3Iu籞0h$:D k Z{>]$ ̝ubj~l_풔 9ywbBSW2 cɳ?DɾڡN^&7HVӟ7EiL3U#FRZ險JP]J ƒ6p”ý'| A6YxIh[Grf-+s³hҠRZ#I(Q78$lIs5py7x8|`x6a+.oPpA4#:"A:1h$ q09RpD)ĥh' ̨ =S6X(4`FjM*Vܸwe w^{C7]"ra2>ֵPaޕFr#" G~Xb`lCĚ*5  9z嫇txj=Q>xh~@sjQ^GWOd1QFHA5n>56VO3r;?V Y8RlΗĨIx0x0xRp^ {|J6+j} uiݟ]s-( W"tRX5„PټG& BA}ݹHE"]DAcUeK9}.)ɼm) 9C "HtZ:,EeFU;4i±,ZqRT qd3'ۦ:]g*.z8+*X,wToܮpKR㬔0%[?SZ1ﶩF__"J_JZwMiǨۦ^­4i-UOSC;.{zCn==K&v,k3'&'ARLNΎnlZ'[,`$ 8 ?1쯅ӪN W6 ŧvoZ+ZZԈ-a(Z}>K?$y̫v}lC }ӥ(CY~z3ލ? i/_AT2"a?@}عJΣmWO z`jg-\D5B9VSYג1U0rJ#:JʀdRgZ+%k_m9GXQoEjEjGuށ;Ua뒔p!]â򋒀$XWHUrpƲΑ"PGM^m@Pv97Lou~e wv~~\%?Ԉ?o p}_A CMW?ݿ'& B矾{bV_?67$L  \LpbKy'Jr`zq-}+0}2PyA0p9/]wK%T nٞ,=ndq'+}t&n,0k@[=/ }ce?Z ;7o,vwliG7@dŀbF(!]ͼ ־9{FKCP>VR9HIUܽ_,M);uܸm =2D)ކrL'05Ɋ|u*NE >*QD5*s(Ir[m3J:Nd uڗTU#Zժ)5QW{tX'a7!^/mS'\[onnQǏv54}U7W٭?y)qr_xXϪ:Eyza>ׁ[~qF<5?Bo良s˖*WUSOhM$; i T~B+6avuB@ڻEz>, 7,Bvw9x\ĘNwx)ϻ7Dz>, 76h};bn\'hzhh倢?Ŭ$=A\qh@ծ. k,ո5jÇT_ ǧT%iJ|4ѵ4JOro+bznt-n01%QkGIjӄ8C *z5aL8!-J) ?triaṙ@zfǠYcް3ċwm6)_p22J Vnjx˶RN㬔 F 7rg \ uҋҸ\7 ޳Պ^FF|39V+wEpɶ+ )'i 0̺lmg{\љU'WKRB)Yأ< \'}Z: x5"e#b'yM|E`0J3TB7HV%X[T̳`ArZ1pkNA5bsr\ZEb6d%Dw|j6|@%u[_Gnxn_nPփ8=8/@ _DNtɃ1JxzyA~1#Ƶ4Z)[I JUeYծ㔦B!Th2lamDnI4ϒ!EpT ư5 *Ŧm[ԞRV qB \+FF0&K Z}>4|(엛zžwW[?u{j҇yxu1~ۜuM215I@Sd2Q je5(Rfy'F#TDkTN X,K餲v"8linY;F6 \|&N652m* J*dͭ_S kq>ԤBO$f|D$iE>8@??͸)ь+#Ym`_ 3C3n|Q!yIM,-ЌWL} > xu{_"lcHC7z^|DxPрb9yP͐H;7=Fd"'s3hbxXC|,KFJ;0MFv$N(|}OE g&˰r B.H!bmݩfeww8ݗ؝j[t,W1Nߑ:6᪗j2!Act$XhMi8Hݼrec:ߑ&Ѽ[DwBpͲ)M>n9x\ĘNwx)p-'#һa!_v)!<5F3.`lΨN 2:ܞ )px}x\0,AjO-x׸5d#&j/@}~ ԙ  zfNpb !\ a;whlӣ;d휤QIF<<vNӟPЃ[=梂@zy4&%/S% j9 q<,^# lΎǼa/1<5S'nj ~%'0gψ S̡-=?O>F_ & X+pKoA.P,QITr x.Y%-TUҝ=*QlT33Y)JL~֮|6K 1ZؑA 8!ҝ ͷYx >UOtޤ#2/+_j=\fƹmJ RiӔ"SUԢ))Յj m )aҜ*;}/ϳ_Pͤ“e[;mJ[͡ 1>-ʒB9.R\|ob {eu@Ȋq)* uItŮpd咩N 6_ִ·fuVqZbPoo^it ?e|Ӑ'SLf͏]֧gY$^6@] 9Fi6gwwśtHMMY/Z3[*l֕$L#2%-V׎B5ZtJ|4 }:}/Nd`ڕ<[k˩6Pcpg fBqmO mJW+oWHPjrIJ&\egʩ4i.OdN;bK K_Jᥡ&ʯ+˴Y nN0d0Gdz_XCBHDz sNb@ҢgT hEaw9oK^a0%U- 'SoIQ_3^OϪ^ b,bV־?ԤQQeiJu@%%X^ΙE*Mraȭ4*Vt M)[l (?oaKV%F ɰі}!eS8a̠ϲ)KKaԑa$uO-w;T׶n #LB;QSHF"@ s}d`>\=>wn6l{3@)E="OH!L^~ QQuM*lK̮V=a}-NKݺ栽iݢ713 )LY$+]1>I)8uccq5zzIYHdj B]= ˨c'hWd4AQN|Oj|4BUPVJA eYI'wyyX2!OҦ?!H{WlK60{W$\( ) '= EccQ%CBxKΡ rE$9m%Z'߿Uz'Vb}wᏰ|KnK0-L/??;I NmpoNbj~X8-x92Ue ֕DH]h]z۴5g82nn7y OެKe_6,&CLDt&3&.mШ\% RJcJ#Iu R*ҕtuS ]-t.Wl+y9];l|FkX@"àg1:,P>RS> 㠤`NPbURWw`vbf\z8vsEF_;_ ٨%q,^8*4}3 g~st\|x*[fva#fBd罒5l$*[ؿW%Еĥ8^*lhb{0ҫ N=b$TC5o\ߒ@VXqﶩW+l+V*f IiKT84!T j{[7Tc4J|kB n%Qm˚p~%UUҊJ~ D}Q> 2ܦZFR-gƾ)?-jt^|( qm P&nKhq߭rjRŇ';;pκ`~x?pа@0b9b tqqCDʚ*+Jrٻ6$rی{HpIv?F?Nd%r߯zD gh&⯪փW[[bqi)xI?En\ʘ)`h[pZP\y}ŵ*@5z2-}c7So:^:ȝCB&y7?_?|TBpNL_¨k֘C)3A߳2d@WJ sBr- #{*:mo]O"z 5h)eFgm4 :{씁@3U*gSvzƹDA؂XW9@iEV!_E~Xereg8j}eh[jnegg?CNU=c|?iig\OY|hKH] |ܦ[R6׶CFIIcC6EEamR!aAdxJBȰV}nOGnHSkug_>EG;rT{'vJ B'p>yG|DIycyLQ4AAWvn~D=|Btvb] ;ٓ~`҄L0vHR,sagFyjFvl GanAc tbwlW@#fZO'6NNsƪ8^_T#[|Ɏʡ?.qƚyKx zo7aЍ Ǚm*ArdGlXTJ(u:0G++pxa]%@޺;yw!D4$rN*iĒ}rl+&58!pVf "`dvM("@t{ʼnnǚޱK#aw;=ݷQI":P&*h!9gJeР;]G"mK' +e4*svNh"scu {FTX_E]vSIwW+?Zv[&sN{-Ϯ(V$IÄw+ٰHhsd ,gByFwh{ܖk֘Cfq"PSv7Biglxy_ !n*7gG |*AǧdPBk.y֯5kLg ;ٕ@$LlK uu]ʺ=sޛ_]tv 5\uQObB!z ۠eʁJFKghErM"!r2UMYC^ujlJr tyͶ=@*ɥ'j00f FVtԍޙ 7rn5"-;.qėETa!h!a_JMNQ`;֣o*hA7CJFLWt/\Frv2 1amhu )rM,ʼ7EQk֘ C ȑT!÷! _hq;Q8T_cx1bVr4'V9kC|(% {?ʈ-%wdO*ڎck^B-ZLAc+Э>㞺1v 4֋mv]c̝QJ,] 4Xb/a,qܩKh[s+Jz"N$"[$e^P, ŗ-KtU2kVRf;V@ v$r:lf;UVb[T l] TDcdrXR@uѩs1A]ϨjK½_&~uviGN9]9ȶ\$&QK+idI zo7r3t iWX[3P橿>$¸JmGn+PzEKrUEZe*^RZ ( B*h_*%mOEޡ.Qj4E# it:H2)ܕ#h ZiP-jqbU?ΤRZcomH*iZG-_޸~sо$V >jgv!?b]d~ή_W\MNNN< }m4#C2Tbф*{#Y$ 1S!F1ɚ,qS麟,5ǢrDp^κ.t㕿4GlYY^{?kή^E?fuV+5&e{>}/X{XI<dm7[o(%ʨZ>.I>.jCkVmAZ?]^ i zx]%i5 5j?݋~vߺ7?ٷϞV5 }}֪~;&-̟Y`X,&07RN`8qmG+o3R Gy)eHamaZaJ4F`H bco[aZLJ/I`'i/O5ZR5ZtB%c@m-/0>RX;0;ȋA f@" kÎ{4z{{tcﴊ#ݥX_7GL X֋#/mF*vM-n-/8>RbwqZ<3Z~S|Wgşn%Wٟ_cEڝ\bu%Wv q$%w[G.E[Y\BEa} \jaw <$xkR.x:󁘪J:UCw0ЅNkN;:Cn/v ֛^W@\rwbAzHbZ"=k?z.bVv)Lf@]l7b v%  Il ,Ad5( 0$ ĺӪL?3] `A)wnC"9=S.Y}o+_fߴML˻6ws?Óo`7|>m)}nM,|QbJi|UQ77Lo{pz6Ns֞l#,7@b'B>_]_Yz<8,ӧSڍ^id8X_H5l\҄"jzbvZ-]y0E֏oAY}4PfGJspvPYCVM Lj[ʹgNܿu⬀>{~aGU1fSfhzpJ{{7A:N;NRyg^{3G1BH#j{R8}I0PŒyUbp|F3k֘6P]v2$GG I: 2Nhr0PdN?}YXP_7BC,5s*m*գ9yUq'ZhGd*I7O@Vj<\)R.Cڐ5b?ߴޏ} vSSX)TzjAv%tyEAGltH_f /pm0|}/o]7@`Eᯞ̾X]~'7^p=zv}ϯNLҎFƤ[CKwΦ'dyص\[Y!߷6&#zo/fowem$G b;b`6b=aמюE^D"<$@JEV#A<2++*aG.`7 xu:;DXtx0'OwҊ^ꊦ~9@lLM`BI}|2HWdhN8 wǰq9e` G-lJHvIDKQZQjF;i?7r.xhr/ʪ$$xq" X7>-^75?1ɧoI~f J1lໝ.v~R_R"nFq^l<_&? )53^jX KX Fe!ku^z NT&eO (V}GR(˅ ]W$61dzS~Mk ܃1LnGj9Uxw{s˲S oVq [WHapFۀVA<ćaRqI[t~8J{q4ϗ7<[  Ujy%7kYAIg\@>CyRpmYTp;.K$/P~`^Vϔ/n?~_`;ͿƓ`aP&ռ^%~O|9kD vHO]՟ѕ2#c5'` ޲Uiާ⠉ T 'rJst;涤Egj˶Kλm юSGo4n[VA`HίZU{zw{]3]wfl>-OViyӗt{L򗨽޷a6x1I?? 0 0 0 հ*T7hR`ʥӈ0vA AMS9Q(LFJ(v=XDp'R@׃v0MT Dr&T-#>F2V]bJX>9g(ݽsQa,iR\j%S~TsJ}_.ZmTya'mu@ j\4r<&KՓ^}&A I*8zim$((w!7csn|6:4(YڀT]h%xI2 dwK]\T|wFd=&tr imaq p4}ͩ ws>F+ FZpcuZ㐊XT_\u0O9?53O33BxKT),m<@b>e¾^{dƴGQ_n-cziM8g>): q$_DZ򌺔=`KO \|xJz5x~ƃLu j%Z_mk dQ2O"[M%C\S*ϕ$KHO(Io&V%j-k Z"Y\S<[fhk8T<,~]>4$pij|0ʹH(b _&Mݎa)lT]b/HqהaĊ%W_>OGp\tCGa̒DŽ]b&-zBF͉ZH`D0ʍA1BcLAH-&MR[!뭨hU&,̐!w^>2D)ck9q*2ǜ1MC~Sh5FВv=VO#ARHu3w wTˏ!tcwJ#"G"׵bwL-aRE.3jfr~Ij|5TtSυÎ_no:d52k+kkQ15*Er3~\0W8l\_-wa_t!{I)ԺJVj:X`+C9g Z!4Z"ȕA ai%LXG%PNިE2בcRJ.eݵ ŏr|$\Kb~ ~G涴@1JUD/Ʈ ^aE&TsΆk9S7ž[$SHkw)b7M23ቺD2wm)u6[x[t n2$JtðK0Z-ssK ٣hd]ig.=%?}gѧoҽ>Acc_LhRMvmO*kxfF9Ҳ½peBJm 1sU>`$1ZDcVQ?S=c<Czd1PeSx˗]yY~eԂҶm $MR}W~mKD̃'Vr*$Ĉ&TjIC ! ~,QV2y@:K>EՀ+! {'oW+\O/HBR# *vNaFu4[Sΐ/;3 '^MЯWWVe z5dœŬ'@g}%&阢gsGPLlquFXC08u&U<_ɫ&jaUZ@i=1٠,Jw1,: 4)1HG ԨAa pB3i֗ V( ,߈VY% HD FS>Z !V Z ۫y6Brz-T15MѷVZr .{UOCcdu{D$qJF u! ˄\RbVV3@H`mYC(%@i8vI$fxfH@ե `֔`pYD@)xf5SsHd4IWl& S@sYUYb㮗 &(ʓ$f+AZMlunS4F{mz߬^Z725p Oyq Ⅰ`Ap-p]$;$"ɃP% `iV3akP~jpϓ)E 8E)h@`{|E/JNjS/_,Sr oTwt供MZ!|c:IeފI~f`+|uQO=6f c,z v(gQtiKz ՜AHRȏw\q[ʼn#"BsٽY[>g<"[J^sBV'~H]''جlxvONyq[c6qNhБw_?F #f"ϚQlm!Puirf=ܻPaԟp:Od?s4I}D3 ?dC.Dtbھ #SLW&w6yf7(J/6E BҚvfpNӈV Լ^3ӑ9*6+s `Ҹ>ԨʚYXN*=2Nڬs2^Yh]^Es7 rф6)93^،h;%zvn:BS:2Q²L̞ UL7 XjkUu`ZPQ(朖^PM~wCWm@1«u h{j5ˌk&K6{љ9%{x{ iUYG0#;cY!\Ӟ*Sϋ/:9KuTTt\2}f)DsO4&hNwv( J*$o$o$o$oêU ;)6m\9NSI#Z8SRɭ3[IaPDz ;4QaoGf?#/Hn8+~]srR9eqy`0q&r$2 7J$3)YODZynSuc\ 7Fŭ5b%{]y$#λT͈+Ju"Zhg]ׂZb5z`OjZf&/m{OoZ.FXtRXwQAitCV7ͻqO[z?}r+ĄfbvXeXx z87,|AF[a58:u(L@z'*  R^nxL{gn兄hs;Z(`ZOu[0QjRZ/`n{]D?oHFH5!b#ЕhsRq񱧣]yq}lY虪-j;kN&._olo݀lJd-tOXK9!Sj1wiCњv]ܞxbDW, 2*>n0GI/G6<㸻 ڷQFE(>*#cCF="/I$ve~(2TH98ֹ39Ѵ3 t*m<Bw'Gs*"`F03KyO`=BQfȤS9)-;zz8! !։N5.ǰtI.͚#9})34&!9Wz:gQ[; 'q~>YM9@ $CknfR~Tڋ,Gj#ܑQG eUC[.˪(s2qҤ/&$}U1YsIkvCHqy6X U9ʧ4Gjp;CZpgB TKi X$?ʣlWmr݃k ,HhMOOaҽ,_ M_mǼmʃzj! s!iٝs#{z ,wko] o~t]7[@J}1"CWT3 HRv7[yeyhH D?"5ZiX(*2\nͱP WH/A}?܆>B2>ǭTm_vm0Ը;IpJ[Gi6غ)ᗟԥ_H!3eI7=3`'`c5mڦvuvgTr&[FS!8@]nh!96x Vrz;KihF!G8D{,LrJLr`ݙž]x&P-[|iϫN  bI|DbUh7)]G J ZZZz/ovOp o'$'BF4B=dP*qdCXÕw4p?g/_:l6 ڶ`'I-9yiMÞ7LZ&q+\ fhϴ,4JWZaB]ژG3QtgZLY$ hޡI#aTj|y(X6x]N\տW]Xʯ&PYw3Ջ9+ڲ;9kB|_//}ںu?WO (EG=oڔɏɫi9uATQ8EGZȉ7ѽ۔H4L@Ke2R[_AG=貥޵rc>>h[=&0zK6n)-LKf/?6#)Ɠd&!D玺Vv]Stys?uHOdžt; Z'[bODz?Ҋ/F/yc .b_[,.-bSkR6{2KJA֧cBZ09%:*=J)i@O =5oCo ('V/#at9 Tr=)c֯.g Z4IIf; V'M] xu}CzAo|9:龙!BNC.'2Mx*mzP]o儜Ԟ]xߊt$Wd7WRHq[\hӧdgnȨB(ۧn4AkOO'Djx&BѨ)'zg;(Bue`L(ifZeK*4F%dbK>Cpm?2!Ua FPKAFWc{>y{̴P,vB8巏ħ N _;ZHF*:.ǃ LU$7 Q@Db fL_MF #$&'#{aTIdc4ł }[^& MWO7UV6 U+mx8fB9F9CVĊ9ڐJ0 Z*FdLBVIz8uu\G)ՓC}D,G.]Ev7 tݤ]|Avjqז($KP[)Us3da0ŠB&RUsANiY(KlW+VFV5wyY w0kttf|Y77h_~̋fU'm=3?ބqϾ\\1sB~d~}ax\<Ń޺wkOuuOY1+ReyLOG f~i\v7[yhyQL+ ͈5(J[}ն:DVzԭVuh|tWmQ6`Vz0ΔmGp5} t\tm82fqEX譪ebZc/]^Ar)y.>-icȂdoSAq_ŕB.W&sr +Ӵɜ`jȝ6<97r^W0k3⽍ #Wc9b#G^XP4}q)H{wIݹ4ȁ=}̗.<,/𨑱xBPR5Qۖ8"MRBb5Ȋl;TuR9WjCU@ !Rf~}z P{ ^ʙ2$^4&:㺛QV5V ^{QHvȽmV_XἯ{Q(Z)ʀ7"Nbd5hiB)&"~ג{ ǹpBeMi\@ żw:egCGoYTSĞ6l`o}'pO÷/>Bi7ڽ5An ˝ {62ϒxC[n`?|ՒUo;Ff/J"Bߒjd3՟6b!;B Sd9i(Dsjc4&Ǭw?,̸{gEa~M rk2\f3tKYzXvN1܎pê ?\Ϻq>%9D1JEߋ#Q $ik܍:w#js@9ۍ26qZ(o~|/zg8z*ܭٱ<}>),۬40|smQ "/;ϯv3e+(uX΢_pR0r-8SrsoV҂U շۧgzȑ_;Hك. ^زG?Ւdb/l[f*X|si-hyسvhchׇԄVr\U<&c.ǡZsDi%4ƻVpRiŀAXS Z)Jb15!(VSMYmmJ!JCF%<|re>aԲqjy8l="IM(!J8R6\.Gn,A W*څOX[r2lM/ošE#Mw0KJPݶ %\Ovhrg! oq{פsRwW) %R Q'^㧣āǤ{*k y%!FPr^_8˟,|/ P3NxhO啄N₂&KY,Lϯռ J͇O#d2]ݕ8eWPMjn7/k iF ڳiS y2ƥXwfm߀/BUn>eb N |L{좁Њ8 ]۶ d>m^Pxc%'_ 5I BSPH@ R/ ckACWia6F$rbJ9 ̆6e~'vxYcDyCGG3dRۥ8m :,~#cBelB4$Φe? z-k}Vf2@OָՇ|rZ_r&ӊ˻J9:u|§蓈q/I8VKɝK4$ E( 3HrՊyT 7I$?PܔJФ#sL:`9@2ZA*" z7H%B원PBeܲeO˲ϙatu.:pu`ZRƔFxR=_OMfަfQzmxVڶ֙hY}B{AΙtvcwaWcYm1J;pr=hg2?]g7ocBgy3">Mx,~Q^~k&}=FPY39 ^sem ౉һF;pMں4Fυ0ƭ%ZJv]s-Pv*B!ZRU=|ކn }UT(k@կd㯸NeZhp*hQ~scD{%2+O`NhdU+Ҟw90m[a4 2h&x]\dDF{ٍv?*Eo aI`k)}8A0I_$ / (UIbt^y\{TȓV# R寬eCUUX2?f./W7oE_1E8 rL.D_pc@z btl*483]zLR穪'LJ3 vWo[Wvgy+w,7ia]c :aɤ˘g b)xu[0Wd]`~ aLhT=nxԭSt\̅e.M ݍc\iN\׎IQL}7/=2Ziq%F}]Vr\plŜ:rBJe,KuԞ3RdqXR svoץeV~2^VJD& ?,R]m-3C-wfEUqCqE%ohoXQؚn/BY7A,y]I+6sҕ\qeȜM %PI2bLI*!4ER JbW-:A}JBN/EʗR=}D.Ɉ1++R̿: ک{WSjk2YWeˡ." z祝+h)ɤ{+ޠ\~*bW7g'0UZTW5LoNKO &>^^Gg')-fΥLnszO>_ N &dٟ%WU9=/O^ .Ӄ#]jOܠK(䂛t%1J0f4!J֥{׽TI$ t uC/щ۠@CwnT]RJ57F8{)~%\ULur1B)00@Tqj'@ ]|bd85aT =0b/MqJ:J5dz\+:fRq"(_o*<DpfLz`*hg㠂㚗2pF<,snѷ&Z_Cpŵܴ0ˉ@L 8-ﳍoGS~}j3\ϴ q ++*WM \zR8-R4 T u?ׯ_ [%QҌW*1j$ii4ЄzZ.Jh)Zc9ìkg<iZTr\/Ԅ0l_+eU,ȶ/4q[aӧT̳fT6˞}fiOK(Â)9U{AXl+ֺvP]=m6%NTdAf;W|Y(?8 m5zQz=kVF5ю^ύ@S8&9SLsw?>ʱIۋ3kP9`{yO>Q oT]3hN=ZQ#UIon+%:qj//S?O&GiNʐ+PĎ2LrȩZN1\`۶?șR8EHZcDI0GO.UTHYX0$"9ÐS 0tLDMٻƍ$(1|J&~ILnR)Jjf cfW% `sc‘Pj@2C+# n21LV1~d]=V] ʘ'z! tq J:-@JbWU,$HEJ f;~ 2$ c`&[JX2Weᾪ ۿwMNbTi6t@جb'$v&b so2CEDx*H.֡Z4Jxw4l(viH^K !ʗ1u\#Czpq' 聕c&>D9cvفXD=ՅD/79A!MTs7搾 DbNxlڲx$_}~/Yz˰rwOpYb$x5 ==G:pSفll5eI)!RvqbAG {ˮL;6')1,8 Щ@! גBh|x/10;R;|rN`ޠ# 'It5%uV*vZ^TZdmACNCvCJW;#E RuFz} =Coj*X:{!*ҭ9] )v#-5"7 r#7 `C ;+^*4c0Ib/AsAŽB.wXD]iḦi)3"3 qҭ3r(G"B¡ZK=Vf''=Vm x.„;.ɝTGn2zv49D ]Vz͹[1n D{/<>~k>Ϧho)xnu$/?!GhHY\Nsww@ks7 zCEK%l{iq*,I=*}gw~ةA6Ylw.2jS9#nSp/8{Q b|*Wq®,a"T#a[qƻ8$2(!TǴ:]Sjj*vPX=/Lu,ޖlUpNX|XeQ:X2(1lVzƅ_Ӕ;RqQY=/b%ɒse)6Qn?f%=(Eڭj$!߸fGߣh -v;>nh %j>$,pv@)h\ RD'w.^[~Dև|"ڐ)(Sd%oqr,~NB*'m: :91 C8]3'X0$`zG+MV{£D!6-]{vmx_!?B)J90h9Sm)d5lu8Y:,p- ދ'{!! !d<:0-k¤BH-g0÷0љ¸qTídVOawq P k򆖆p&Oє qAADR;]8- COyC9C#CA" n}䬌 3C. >8ud2^WP/.f|/Y6ײgw pW"Dr&keCkQ0G/zk V|C=>5 kՃj^c+~Z*졋F! TKegpmwu>Jξ? }zIlEˎ͖lH$ɔSf_)Zm,c)|ڞ }cp5{UA!4?T3:AG6P-ivuAۜ]jb4Q x\^dޘTN~)61ձ7/ZwlifC2d\3JS c,iuA5dX RZ0Aa=ZyUx;s^zjtMx"ӨxyNDoՂbR^Qj! %>^կ~K?NEQuQYm ;~qPJM=Y'`H =>p ?#ўOa;`?+5g`Ǝhj)I@kn3Af0v/3 [-6! (: aKu;\\Sݪשni:-TRݺp O^[F;hx5)"m叏rܑ#@pY@Pn򡺁Cos-zPLlsss$QRtO:a&94Y({vY!t\t)@R^--BPcF~>Q%>;{U! 047S5Ȏ>lC8'Z 2Ƌ{iX*xQ9V;Oq|A{@Ry{<l7kdzwlKġ1NQN-i^Z.+Ag3/^zss!=DsShEUFj}8ՄZ(BX')'uo-|Eϑ`P%-ԅFWg)>r^ьmX 1ڛ2~\‰&M&rbw GLY}`' $@iJ瓇.|w>fuY&"} #K=|D~ !j|IPf/~?!Nl\ !!cM9djS5}gj>u>l AYl*J oK{2ij|LN2\R0FXLZ1'#E¿C?dCVo2!7AcD03 ?>/1K?/n{nr>0i"|^E*zW{7c`TJ˼̣\ nP) H1!, ʁ;fbX:܁>br{?]9Zٿ1gWf~SxK0Es/kUTWQ_Օ~]dk=C,LH@hp1%#X1葀+hJ܁$8w|X]+Ass gT)r$P)E3k1 ᑥS~ʹbkh؅ϬCSuDg'm 'A/O8?3Sn`ޑ=%X)9$CIRQK S1UDw^ 4fO8hqibo "hrceY2h4JGD3V ~n9An(A;J_c# `X]iˏW H.`Jqzc,8R"JRE[O "嘀!i%":AAl1zɡ| Ѝ`8Fmڭ:-. I:miTVJN) f%Yau(O K<8X~7% jK^y$wo596= Mס\:c˽N{$pijR-u - d-KAkyeYIdqIk)%]{+Yaa,9ٵ$87Wx)$<'#[Λ W>u;6r#5feش# t G3-9<Q, s$"T &xjU$Y<DDpVګ,n,dL !XYIф ~ 8jn?S LD:PN).9Oy]`|l=y!^FeK9q) wS[VQ Ach~T /ej|eQᖫFemlLZ*Ϫ2[}[9J/hRIo kkA{o\Dd-I [.);Fv}6xs|4=Tև|"zLQ$ӡy)Gk^p/Mot2Q@^ r K6Z_[}NNºe)C10F4d 1O<(BqTuPpX 3># %HoDʖ^xu|^ɄY ̙o^Ur,X OM[E /.?ݭ\}g$mQPAkS*K3)>-#@}XKn \Xsmfa-q%[onhpF#Pqa54:.L!]eڴIeV- 6MON@hvH6+8WW$Rx!%pN ;t<њpA:n'7B#:;63E^sBD4HU!8 o>Bc^cՙ-F%PU{)Z",&n*~̑4™9SL1Ͱ87C3vwQnCvmks:~1PǒlFu[3@&cP 9^f'Y:cy9,9dKӤꨁQs0ޟs%9Q:ToTlzuNx2\{8xxq"Dt'?+B)3dž-DI,1&R*р$P#PUL;2bJfCA.](PsZPscEi<*8#VRro4zꃒ:>hH`c7j3BI,h:g-̈i5@aQT~§޸Saoŕ4T[ZXxթ¸ .2 9/'=x({ΙR+e);螣VѝHׄDƻt N鰩)p:]EеLGQ[D58h|T& [Uv]݆_u_}sO%^`1C74J(7m(f2c_EdZk9Nd+З\R.Ot-ө1,JMw FxR6*JgUϘ*BGHaR+LnK9L e^Vz*2\hDF{LxjPQ%m+0J+u  ȅRb;5d ȆLK,xnJs/^Y\!FnILKMhΑJ񎓘Cx‰\Qww)CO,a2HUZiiС3̔= k5n*;}*.V}VO9["EXI]>|nuuMFCsu8a,.dtK;gVJ baګ)ӄjBN/n{UR ƫ0;]? iBD[Ť2~bPLKƷr=i-R^NQ<=2Dէ7V9?^/2zi DP`5WGvȧ`.P߀g2*Qn8vmYwiN3~DƤ:ch[7`OYi5~d?;Bƹg $ :1#_smXB0^bt Hv^ڏ']`$Li aDu\]ne la8Yr3I%DLHIsnS]Jg1ޛװ rQoӓϿONDkb:F]m:vbQC錡_nW%Ui!d}"pg[M mK}=Dd OppPZ jwYrS)'qV00Vm5iN`!fm"F 3"%Bo,gBXiƆJ|u6#ܻwƆBQjkj:BYG!w 48$v`Ӿ|Pxy*rxfLQx;Ī(ޠK/M~bLwřcbMߙv"NjeٍMKƬE~|5mzrC([ZL%ޒN驭/yb_> JNXP rB 6d8fL^KN3x2qFஅCI >]cC\‚ ^@N,ʡJ*k)Ti"Zޚ65-MsU=-4+mwۇccEQ4JA&#rwj)9BMnß?vӺR\2cfv ZLitn  JckYˬCzUVIEB:l6LՂ+Zo&1 #ֹZ{Ĥy%wLYZkE&R.>E9DRb`괇̑Bqd+(IRetreH\}Lr+Wn @0mia  JsA( 2 6_JT9 ( C3ALAИֽRCx30WG Ck&7L`8W[HTʏqj+r OrHQxRL":. -!ypF$%|q@ƑI-|&)xJ$,\x^9DX'R: >. O `ޑ0,I{Š% *(($HAEm|&&57,nxаю/WO|g>O#BbO~޾yӭ嫽( i>~gX'qsXK r8 JiŰ+0|16 䕁3Z$qFл3WY o3-G>ߝաVO%7//Ȇ[d+(, ~X˭gj_<o,T)g [|~k .uwg~y/ǥxwl ˿s?{b oo@'g7`Ap%g/~!/8DB$HMoAv O &Ur7#<)L c Nr&H]yqM@"Q <VYCӳíoI]E0j.!=j1IFk LΛ (DC"34ҝi%U  H\;^qmw!K=8t'C7 vu*tGY)h 5@MscZtg}"p>[g?r;c!H!]i>!?5ߟt{g?^?^^ݮ>xǫ#\-BJF-G˧Mԡ:&6SV]^:[n |L޻W/>G:E:; sDEyDtU/;WS+6cfJ<9ImCб +YtN_`Cҹ:n_ڽܦc#—}={>knq;uZoL~7Z}u^ZGtaULn.ΆZDҭjDөxXHB\+iz /ҳB_ &P:2?AyүNϽd5^/W;"G]rc;s 0,pp?}TkBQɂ@5TU\NFȘLO_-ui2 |nfO!ʃrmAn?wC4A{b:go7kR~|rXQ^myqˏ /c8CǐBEgFϯ RŊgϱ,t:qJzV"K(rUѐ&!q`;ǖ ?^*+a0 O𚬢KARq7bpw?<ϯ.oo mfp,ŒP< NXH^Nсʼ++;z U>xDf覺Ø("u;ܧv6!!'.˔&4(d,[PaDLKiפ&0Rg;%)˳a>w5Zkg Bb14XXg|q2|LH^uY36Ȧds̅vK%4j ;)otUԃ6S{fbny#$g1#xYw<ŏC ts aUoUDIGcb;FQōÓX uLgCF!km>UR9Ĵr!u3/Sv cx ԥlM ($1v=H;. > G0d;Oɛ :ČL>]h}ѣTO;iR 2Tx}otrT,%c,g߿LZxGׇW%zfKQ~?__Ycf|e=36)Mim]?Z5BD~*cQ0Tk'tlڎNˊiGMK>lZ-Ǻz`HH9G R#1,yqoq| 4pZSxI6EJ2ʁ /3k~TM҈4MCDZ"41QQ8) `ugP]?"RDc5G$.uq'sD갇 ƞ#tǨu7/zhιDٻFr$W,f< ޚtgvH^mjIE%%NݤJUݨJe_0GhKjU%=) V g-ҟZq `H:P2Xj[2"[p6,oҧhpD=1x|qݽvQeY=닫/N_-}2&1_-](= dpqA _rwz;m=C')8=dvtI'R)&JC3GK=kfvt9ªw+ [d hR2y Ht V+M Oң8~/Ž cWEe y9EF q[1+a!I=}V!! ! uv(%S.V+K21D4 Sܫ)%RVDcJ0QGS˱(Np \ܙO{rh 5?Y,'>"Ay)) !(?b: I^˧y$mEC;ֲGÐ_( >#'q>|J._%f6zE}[q|}|@9gBDŽf$Hh7*o:MpBMC+Dpt`opFVؠ RbD-M)IؒKڛXuyW 7oP0[a 3Ƙ=$ 110?\>PSaql)p:XIĤtd`睱{AN.;.z8+#ؚy9%b=G>\P9C 1ԇڎK̎ h|Y>&wGӆx>q>!Jx˩d)@C*I]p%@B1ȑR[ʱ^m-vR U׸V L,3yJ%Kc<__%t%?kbƓ?4a/§?BwO <-1\ dB$ۣ?B#Ԍ.\eOn B"_ͱ$XݥJK6R뺢KpT J@:m>n7TzؔVGGD%B C}hTc32ц,3I -WMAvUacQhMl<$JrFyS(xOx5zE+tSFUcn] ,rҦbTңh6 Qod(ϳ]EҮӴbF36[ a8D%bmOhB8iUaV>̲{\ά/hި㻟Gae.z([k~|kMA11(ec7rEr]ut `ĉK6L.H 9>VGLa uP`aXۯL/M/B.QBɫ!&ػy-_1+3}tf`If`+d.Y5ts0kޫ8XLx} @3lk\j*+TX{+*U.L\ 7yan\#xݿyc`4),l`c {#'bDfC fqŞO~^%u,]9zw1L9 F Jk$( TY]x(U_l.?H 6MIaRK= ^c"k24bYXQdmOE-1P+rF-f%pb!S,|5?~+?.X!{ޖҁ֔Ö:ʜD0E/0Li!? L2dQ妮"IUKG*pr-x=#5;4Ƥnp N):>R8㧐㒢(1PIz8OpNd3L"g'qᒶyDZ@{Մ@)~3.ʙz,Dg'bS.'wƽn2u>Oxt&M3wR5cX/ e Bk*N jpPD$bp+5[l_f~4<?*rEW>xx2k/x͞3G[ZFvg~x io&eR CQ+%@"q8B(ԡɚ[c vE\5UڬA'BG%4vCaBY'^k\DZ&2؀TpptBt{SJMV螂ei`h~Gmg&mo\m? ~bȚ~@sM7 k(%YtD aOZMIř{kmڶǾ,ds _TW7JVpǼ4NS#Qrᥲ#D#B;P˗BڗU{!9x)~{;zΣVolGJ)\X0%T;ÁI[>»!BCۅK"S#Am]pA67\П&]vՂxP\t=´…Iڅwṁ9Whc?ˣ)ܟG${t~uٌ=~)ZG# ){w cG/OU;Vu3 DfBf[Np_[Y@ 8مnbqzPi{_&4,γužYi\" Qia ]q1&;00 ePyZݵϫ6P'o"m+jcj) i,ڠ Mwew_n]hL[4Q9j0MS$ijXYcO75; ۅm2f w5 _[]d=,]`KRNPn)[#>*En᧟Ef 5/,!7419RVpI &`TD9Z1D%TAX˻ d󷹻f-n36fAI| YN@=Ln_2vRS<(1eK &!5xwSvB"Q3;qo0v/#焌WZ U*fdBFi{c' 'w?]Iw "mOQy~/D}leJ<{nX4C2z2o/:~;T鯕''~:+JMn?L!{:}Qڳ@/jt8w?7+|M,*9#0-Zܽ6e >QtQ~҄_lǩ4DtY7"(3$GmyƷ.JR} j=a9L`.9T)i{S㜟K{RO)hll"8-JGDD3Nw; M2+lu(2{(71obixSR[ +dL){^sCMPł"'$%@>HjPMBS/wqԑNnBrg0eFw /߸?\_K"ho-lڐw6_#C>1\P֣}8mQrz8j Nqf$t6gdE"4i#;o1zt05d['UٰD!ճa{V6)X4*9畮u$TuGG@W- UBG,M"N#͐{$proR+jͅ{e^Xא':'j4F0dBy|QJ iVo9߃4#BD$@NH~1MmLj8i! c%g NH NoT ;P!Q@, Ɔ2up2&V 54͒R|hB68eQm K!*nzgzu`pmFi؀p܁o*ni2:1!h $Jo @<WH fay~]#P*[<䄉:]٤*tiy\ǃS''7H>?VjZϮ$1ݴʧ]Q)9Q D2n*="e_=z/GLꑫW !cZF|))בE-ѳ҄dZ[X,2%B6ZہZa|C?LsN0݇MRu܋"bx\F0 zqI>}{ݛ2}zniη*yo!4cA?~E{?):Gd= g}2y6n&+Bxψ{3f*S![}Wx\3xY0րgb6%PG^z8hAFGF $)#-уLte9!=Mc!qȄkbzr`drZDkŵRcBh:b7;fG^ ek3fm̀s^Sv ; -u@; 4#lȜ̞b\scIoZ;OOF!P`Tdb!"xьu@FWn4cVv8ޭݸ;j0]@=Bkwt4 8EU2jKIA$\DSЂZ0Iؑ)5}~wߏFX_/ʟ亊Fo3%q8ձC1.qV?HuXc-$@g+˽={޿*,~XU,V&Whrxe( Op5űj6u;k>v4u9߿3/q}'[`(t}!srhoW'ޟhfm HY̮c,7TWaֱNW;b֦-:4c\m1rIJ穖uȬꐫ;jlDZH{[ӃKq7]6j߆3$xnp)n0JEA\P^3&p0{eH Ȏ !X0.ڻX-f<Ԍ~<,|0qF q"Fa[E#Xp=.AFW0&JHO'Ai9iC~`m[Q]A*<ᒍPRJSmaU|z%DN iAc.mh6 E*+E׺`:dM e>ZhO|Q.v;ҐtJ=u3^Gnmy:mhbݎ}I)HƢo3Ӻu!/\E[T%kĹe݄\Y9uk˃i;Fv,"̺huBC^S$SeyN弡TY&7Bq]*(ޯv~~iEtHF%1߅|a~Wz2wwJUtbrv =>77" gWL03Nϕd=!l87Q~`:8(GXH]X[Hª J8ehBWiA5d6#50gl篾w-7|1u X0L60?OMPW=T{ JvX5w+w[j˷P4n'\ Qy=7l;{=__Sx{_|qɕԡcF2o, 8@^0|%f{!ur$g=!( *>_@+ŧ0E^ɩXx:5cM@;"ERG!1ÇifG|_Ң{F!,۶G4c֤;h߮xFi@m'=rWrn F'n=Rp_3[$C {'bFAGk܋^!ewM¼3=짽޿-H?{f[Pi025D J`i~;=_nzrN/ʽ ,iP'p-ݗ_WB6!rgIDH53ѐ>0f}KeHp"mnPX hMTL-~~Ko[2|~W@1Qk(F9tOQz /)֌eJXU1F"^:Ž/~*M,%V3ΐYb!.`p2jIEݦ`n@ xY NLX~֠ h* <10V-iջ^p(L"$׸&nl$/'Xgd@pˈAEH g*Ml8*>٢y گPҏ^SI(͙|)*XO6tocҮr׊M') S95ܖaz<^%-!}y{Juɢ97 U^pީ>uo6S \QOdS(! )(a*:N/ѻ…~;+v)g}9KK# O#n{K{ɒ`_!q+s\C]>aLIRiFüZt*#IFwm i)+HS=坯Y*Zz ?)IHVq'vxg;w}>/Ȥԓ;~>rifQL+;es\V,߀ v1h`d`EaAhf@S| ʂ8洷@RTɏi>ߗ _L'4} }qvɲo bU?X?ML+zX⻇;rΌj"fE9(/Djx/Dv{ M5z[Q/S2Q/SZD>-@h 4Md@i(! ȋ!g2|moFnZos"OZ5Ple?jaFa嘕qJ-pNS][wۆwLgnWfs97s8 T&d'xQh#%nDDofP͂bB8 pShLLl5@S4"HO^idaror.D͜ 5@s{CuVa%CƩvbt΃w `NZ)!JLvĶa2ge$a (R("W !Tj\B׀k>eepwoܯ ^K DO_4Z>=H#G/<~&%??2o0͗bg)N(~__\>vɩnnD.ڒ sۓOo$QӶ4WJ-o_/eBqJiɌѽNoP8HZZy|Q)1z/<*8\o mnܕԈrBe*^Be*^V 5AsN:P񇲄[LN[I@ǵ,/{ƍ K/99Ҹl'Jjl^NJ`lmtHI忟)Y#1B.AnX5$5x~h4qggo_ ws\C-(qd3ԦrI@55Fi9voZ3Y?tpr|o9a0+|y1ڬ0qڧRHP#Ha1BȁɍB4?y{KyUvoMՂ'CȽ帇>|y, '?߾?\|O/۞I )w p~54tu9_}g>7 a wِ $t=$skk{xn~A%EZulsF|-Mw60MWӸ5L2i% EH\A WͽDq'hKZt{Lxcn*xŕj_+ M rBt٥@( Vq`P@qI@K>Bkq?5uy?:-O fq}[+}qwĨGLu2RhOa{ c9)j+Ԛ¸wF=l0ΖlS'beJsp{\lq}ws@'*p +(}b3M̓Y9 +\X8ӻE3IЪWi\}yIjvR]e/ZvjMCZ( Snʲt뒈QwFabkGVQq0taEbsɃ9cWP2CXѼyÚ݆ dVg`4h^h='6uJ[ۤ9A&pwMϷpgLMh4txz~V/OgFf)N< :yt򴩓wo0VRٺ@y5׌RV*%,X^sNGpnJ PFPn dj^=:6t2 'A w[[C%j6os2~:}ƗCE*:6/2/UEQ1%TyB&* F)?^UJItK,Dǟ\]n!{a.c#1y3$S4B3SgmaV(R oqu$:A\^Q/nѢ`Uӎۀ߯`+ >cTiѡ#5y;:ʍĀbBXrun辰n]$^tts^v F^+#oљJONj12e*m(D1Z.W<;+d{P2E=<3i̖$LzB!vjǵDq=)e{N14avyK5[Dl`Ú$YQj:gH}qJ:HtAt-}pH^Q:p3lT]3 qOR6F!+CxVI>J X`QER jU]Kh Sz9ڬ(hCY9+C ɡ#zXYF@t\$-o,6tdK@!jlQR/J1ú,Fg8 k^]*ELAɻ820 (k;(qo H.{\-(Bb.'sC T*L-$⭭*YyɹWO.č  JI-]EJ[HZxC 9j_M|8RI a+q]>}u7y;}um!1|)Q|> .?S0 ==q"w|lNۻp\(߿eox`:% c{ a hN#^M곛Na~=??|τۻ+suҿn// (^WB*4TKuQ!8z1@bjT.U|~.u@:@qlN[F> -/VLOO]H;v/$~k#,~]N`u7L)hǚJnDa 3ZUQ34Xib~6[D7ZG$ra6:=Nca[J<uEК6*Nw(r9w! @=4t=9K(K jzЃYvBTMr;JӉa&oŋ`c;|6ҡ:;?/hnQ? 0ijAepc:ve9aJ ,UW"GjT ="T8N/ g\%5\3칫J>=qNt!Ćَ{) >M"5_Ƙ>3Tq( {.]'ۨ mXOgv {h({H{;Xem0"=q2]p ; VRָ87Yٷ] us&ftՒsdޥpx Y^8\L{hWۮބ {Mm"BQ `1#=M̀wsAO0./}-nb~M?TWn隸,3OK/yj$eP3풚^P{5A*3'_޺djP]4lݠWLōb]pXqF@N LC! sRhGn RK&D_Ǿ(ajŃ=pv6nyyS3wuS?o!y5mٲ8P\,_1ѳfF?? vh62:3"4 #P@&͍30ۇ#rRQ8l-4Pha6PǞkIJǺ1| 4[Md,Y!KiȒu̒u쪠V^Wҗ^q*Xm)5Ӷ)c)u%'CYI,ПwsO_'77٥C'&,.7ԆN&'@[2[ld7t609sɍoIerBqМ*?~8 `Bi8՞W{jOqDxI9HduT)YjʂΡ*jT7:I*_PpJm|k͕Rj:Tq 1濁Lp^; ~I!8۸AB}j& MgO2F4"9`<oU@`8)p ji |jݵ[gދipKŨxavt ~bf}]f(C'ʑ\i _@i! YAu1xtؼlǐ&d4/y謚a[EV&EaABt3aPG/zZ >]>LGQhFvs6}AE2Î M_X\`:NQ_?w_Se /7G'G5Tp# k)&6$5(ӯOp-5͛>9%TpH6AYNT uV^ ֔5ޗʕ8XgG%; `U=_j W\cPQVp+4B%OSm d1V1LbrJ+% tgt]դ$6!JkYeU@D.aEBqeI3@K&ke*)$Q[NGXfQ*-HJ PTe+WזUF *˸7Z"qTiMU9^)U'aeɍUbcb*EDYQ} L+P)&Hٟgj6_aV.Ki_\WJ;/R w%)'($q@PRt;I F\!F!?D0wfr+t \HVhpYb@<`gadg)ʒe dTG>Ud@;I4O f 4K !(Or0˩UF.o9eۏ/Vt2w1wA_q !:UxT@W;j3j_ N"sн{̈́2d.u(ړ'\vݤ,= FIy5&tSBhPK_)R(C` bD 80dQĩf-+Rj&`!]CBTN-1^\9͒\Z.dʁ.gYcS`o(X,aJ3¾I,㜀Q0.&0&N ،0,::=_vOo%'><Ig)Fsc^@w`0xO`mi!TGsx/A0-@]XP-.9Hjq!Jˡ)g`}(C3Lz]^I>؃N(ðKjTcl3R$ġ$!.:$!ԑݓFIVilv qߝgf9VY&4kd'A:{HjEPD@jmL8m~s W S!?exfV ]c. '3,I{K(M<$HWX d2DaAqOX-upy2]@쯣i?N_+,Z>~;'~IB%J!qhJ2bh|orP_ n=L[O T 9BϥT;h {kog);ub-r[g) {п)&*՝Cgz#FU<:`>ZIsE>psQDF_0|+)2X\2#. ={x:{p]ΰWqanif7sZ9ū'KDAKy$]ߺpܗsyy2/H|>< 1;E9ך]9]-_ !iDP͜1$:Nl7V+z_/ik">Od6k8 Lif\3Ǔ B& /!s0'PP8- 'Ȍк~S W{ He_BнCPu_R͠#/~Ij!2 HAmҕF̔Z\>GC?,$˳Z͇|$awsa)Hd;CTp[NoUv5s r(5E{25ӿ,9G/4I=L[ 5®UL,N}+2wt 8"Ě AhMiۻ_ڴb[faܺ;Wk^@6ɫm+Z3<3֍`ɞLAݘ*En>oW˘Mwo..;YK䒢`l.iFSk ؼX˯QT$Ƭo_9|ǭV'ڒARe)(W e`q.A/*e`ǿH@t%8}8۰n-N~qRts_0e MZWkF3aJ[A~8.B|q3Րӏo6T5t0mk2% T֝f4N "{kj}mb(1l۷w٭&$TbSÁf$EJS2LհXxPJLk1 (G )=$P8|8 q[@Anna BOl~jHK{ $`A5s rwfPA۬X 6 C4n?[0Yg1bEt{tzHԭC=^9*G'ѽj2NPI~EӍPm;V2ѰB,O;F{L +CPʚ(t+)A8K?c"Zr`' 8؂A%R&BJ.é5 ҉JS7xĤYRHJ!{(uvi TX`Jۇ0%> Hu,3۱}P7IB ڧ(;+(O5622ZqZ6x" $G,[mmJY$PL/Â0t Ipa$0Ch4+Bg1Z; ʄ C$$EVE4xFywkJ5SEZ=&~8d.?$;t}׈vU>WQP% ˵ǿ%!T!\{4`?-ѧɭna "b\{~ ||Գ[>"Bg<.wn|wS ܦQp{Ε8gT1 gE=]XPp()PpãF5"&~.6 (h?'eZ)ze)ws9re&ȿ]XGY^̗ wCV?#R7o{p%|mh ~;{} w_α.?ݢ3nIB^Ȕ*unI4AD7jݛv?Y nmH *9S)#-e7W;› '_\&ڄ6T>o-b-ėӌoKW0T!̈}=3:*3bWh-L@[ ~ّx(-.e Ȭ^J+XӊעooQ􊅬bD)QyDirV2Dx$)w;}I3?*Z2iB{+V+qv{S07++>'wO%i;XU!+AlX%=,_8 >)єA$bQɚR_"Qݟc6$䅋hL 5xGnNM1,R/j6$䅋hLi}rƚv7)!NoQ&ڄ7A jW&7/b/uF\ #R;F֡1sUSڥ"Uy+dXrmgQH[+qgzc Bf`@ǎ旌–$[ 0KؐԌK4 #zo CDhk' \1C@d~@WU)g BϳzG&9֎U%YgP>:j'*jh,M_7;kݙ|ɻw0<4Yo=F@V5}4D'][r_C@͉>*)$lC_wqe)Au_j.#g)BxޮVt]_P "BTʖʓ5wL[3~a-J q9 C*5FD^&ڻX θ\νFd2E6\\|~I^z;Ȍ>όu#fRѾIWNEjO蘥Z8 ㆪݛGoao..8ː* 61(Sv*HE$TP R?@(KXϊ}ȗvHbjA~w/P!#V/,|{py2]@쯣i?N_<\$坻aog-?k?ŧZ[*kai-eְ޹5'Q9s+2mqZ{&);hAګذfp(Ν.{B{vrI5Y鑭X)^!j&= 2 3E]܃Ηb!A (ˀ #)I|%s&TJ$dB!1wN$FP͉Ābф'a$40㩦"QQrIdԷLLSHgKf2PH R4x 䈯.5&KL}Q 0K=r~ly'N K9 Xb gVX,6Bjʱ=-,6EFP mXD+v _ #CqPx\$x\V8f\a &YgZl^de\ z|$9T*5DY@(۩Tb[u@2lQYdDJ AfReqX1\bcJSRm`}2 SXBL,d밗p})֒՜N[\Bw@0r7WFfznxN37^|?һ%?>?Æ>@  <O??דlylwnt"(; x~xhI)Sbq3@b',}K8x)?ZE'45bsD`HK@:)S)& lBVQfSu. [ Z~刷jmwI Dq;rn@&t=b:"ӵb"qWӰ;纚i+Iɇzn!o< <=U0<ݴ=v& $i_5sw.c v܎-*2U*j;G/^`Jہ)eKl[&B\7HUlM=Ax׌1\)818zz o z84"#USr8뚣-YêI'U[;3Lzۢgb cHm^f(HN1#z[:I%H%m[߈X.iNIAt%#YƀZJ#"+ Fɹz&K#;j:k2r{(|Aތ1pi" .cMnUnw0"IJu2!"3%xFd"ˊHfbܞj9:!%j$A]"h@`<Ÿъ5Z y"^,mmaۓF I)IcB3UϸFڷ>tIצgzܙ0ZXofpU4ZyvUA-Sz;/omK}WԌNnz:\<5IUb4K`gygq<`<{@䫏?l+3!Z=˟%g76_Tϣm <<&Ӌ5cuxE٦.J$J\!u#C?И&#cո, &Kgh~̀wL"Iأ]t7j8(5^a+|pWS8j-`x.j- uQkU3SE> ]3lp^!9CKP߾*U7v U@r)BH~ M0W]a,wK7ȗ*CiRG?;GX-UG:_Ǚl 2ߟIץK']t-ސO4Z$sIlre$#5F) 41e8zsͽub:-< }/pnz[.HEIn !k+D"&N&y]&)0<Lɟ(ER 酱YN6EiV{!d`"B8фγ{0Jƌ&>= ˗ 8NJĸdqE4D,_R1#OԼm;( v#˜LbrNy̵BG30ڰ% ^JYI Z*$qMYZGe:d $iƂ8F)?l;M `Z>LZ#hפC_\Epe"?k@H~0@!QwӁ35BCXcˌfQ< Re BL?sRwF~e2bs*ZC Jjxb>p=Nq$bHɖ+ޡmi4gm< @d0ga9x݆kS 5O}DΨW5kq&8"mD̿6; U%K$(NN1UX~W TJQH-͐u,,FSp-8fD.39%EDW*ZoB ɬ!yPI0fЌI2Jk%80q -`|HY j-?}fkCCj<jj1v~$ 0 9Uv'`Sޝ\<|M*1eM?ͬ*:8qO;lp~v`yO>WR ٨ _\:u}Ioz^P?V$zB;^=ȁ^ؠA 5KeVjyfsD8q.(r=j<Я>dqIMԟ@?=ܨ?w. ; ܩv{s+_*)\pBd'0sP)Tfh=j0Y[*Xӣ:LTL~o/ѓXj̵D k28ꯋ.w.mFAZtD5Cj4s7HlEwR܇zxj믿I{ߤzzh!̞[4Zz8IEF&8髿IUP{g)b%,Oᴂ84o4iµ`a&B}2-m&RiH'f}FÓ )&-XݢQ֮lkud-7jc3`GYeA!ƪ<}uq=Є.| Ixs Q7U^4TYN%3{=6˘v\׭Qk9q~/J.bԅ`6Kd<+eآȈ̕Fqc,Re'T>g`5\v+4#Cq\H8u/632LU--G0DJ D}?FeipJRw&|֤I& JK92s؍t&R1ӌgae4+1S) rj |N,)F,JLJ+uBvlb} #)ɸ`L0:f2+$D" ["91X+0(7D}%^ܖ~Sc@%>^<,}ghlE;LjLCKfGi(!EfaIKD%Mia] ᝽4oV].psމ@`F}rW+i~~;T?G=''/Ոc!L,;{xŏumli 4o+̛7Val=o`{!?t\&|+Kw~+ݨ .(ڤ4stzm.׌/'fksqDZv|oxa`u^ۇqq= õzb@vF\ Oz;D)ƘD-IZ_D0@ 1,D3ua 11. 7]=\\86b^l\'`OÃcJP=MFfeGT9Z=MFH0WJK#E2D( +P ]BvXU>K.m>K3n)Zq%v8nq΅JYWz0JꦇpKI$F8\,JVDxSSfmd ɤJi8GyJn )A{XK bBGwzPPqI4Un4nR QE$tK )FPXh5!PZ\[6αĔQjK0La%W##p]GĒ4-n'Z9tE#481\DqG.RkD}C E+'x5wJfa-ܸ@#M_RSXƮ 蠄yk:#af>'3OsGByix#;X-HY +^v\9Lv0HAot0Ī0ppD{|Dү׹(|~1`Yrcp`Dj,M_"V3wA2^xrԞXb)tѧLI}0;ҟ.T;>;X!,7Ns )?ͤ;0 U!Biky8mUZg.|<Uuk*8R,s)a^p萳Jz*QrZ@GC V}7vbȾLՁL`@~I}3Kb/eGؿnTe}Vo>{k6Ro7 ąhka\S>=" )".7@.t_ 9ڴm'TJذXخ'TLKe4(}}^} 6n5QCŊP UyП rVW c?[曀pJ`_=]h.KgnPWAtz_:tfK1d{^:ٞmX!K-K*) b8Ga`?!b(u!KI.-.g7ysK|t>Zz=E*L.o-\,t^(2`*G Jꭺ_Ηp5×N, =}wr\gg^Fɇ1|n!oH0yH/ڈhy_MBv4uNpS3 `erbEJVEƵSA3m̵Jat91@zD B.Sg,eB3FeaQ 鳥)aF-Kƙ-AȋmښjI̓{dC"`A2n%CV!{nW3ZtP3̴NR8uVvE3jP+U$a).c(]Yo#9+^vgi> Su`Bo=(5eKu̢)J)<( ~ A8ie&wN_(V8J{P#"1Pu},(Va7`N%GέctgͷY@b,{+^9֭v2 |~4$~nqlXh7{z;s[\x4ef b{;lXa8 û0˿ɿyTOh_qh ,-+<@7,()8\OtSYAAf^Ӗ' @^y4~q5k9(|aVY^q,Tp(`5H*p(SH\` *  1A&VgK:2N(0$ r+NW&/',Q€&m8 ca.AvW W! DH$S7E~ NG <εrR?8xVj3{H-p jr4**M\qM :!J'I i!XĢ`ѭ̶k w_Z|+=NO͗e5˙ fr1]?-qP77sr}] lPqu}LoGa}͏/aE bX7TB5BP-L@CNȔM"O A./C󔢪+I4;H`2AӸxz7 8}R!P=vOqF|o\j0ܟOq`F~\1k~}v<HRG($A$]Q0H=f:x"3_Ž?.q^+~6ow5@sq@ ;ĚB]<\VdKV%3(,7@=Wt^H"6jݼT"N/Vl_5A+p-iU#2ܪ61}%3zq- ef2&dPi@6Et!o(KЎOj,̄4}&=CJ$BYȇn'So2#|1wxܮN{fh=•Dp ^g)M4*)&n盀b/SQ{يH]lg0G{tV`uTÜtzٌe"Szx_<5E?nsM_B u]@Q բu:4$󄷥@:`kwc]wx\=z] JF7}97¼a%╩ͣ]_)iy/Y9@HF $cDsND1\G&DŽ^xzܖԓ"8;uIFhwJy1TT+.tm,\9#3@@h}-Nu @Czђu1r1`킐{@똏U{ENXrCω7l3/2Y+IKOs,FZ `99sBaM=ҌC,qHE=ZvT 9OSn3]Rag6s ( S$D o$cBIƜK:Z g4 D<.@K:wHΩ{3?tq:ѝY0?>++Ң?\"woaQ0} C;M(sF; '3UT^_\׬\ 5Z 3ȕ~IrV@';oSv=]SU= #@kIyF!Vk6b%d,]IS/_%8JǏtrUs/7wἐy s zƧU#5wϥs$,7!I/ i9v! 1oƼ7? &puN*z[G{RXU"kԋ$ȎYZQxi ek)V4f¹XzQL>t"qm<6X=S ❫x&Px:0uhZLHO>52,njH}Gxr'v}\=:7̝Pn6Φ&b8;Y- &m&G&"laGm 6 ' u<MN<)lr;̟)hYmg^žXb꜑/=C~ v&k%=+QV[X 'iHB"ZGhD_WMΨb":uQEj <Ŭ3V{ڭ 287|CЅ)=Ё8S0zu/*ŭLu'f+%H883]~|Rcu $[ٷ"70%%2#pOF"xC$zr@)1jqNx0s91 F(xQ ?:?˘{rgxk~%j1ѻm 򱺼0qmQ6uXGoS<;13ˈ"LC`$Xmob1˕@!2_` ;ZŊz_o~Lė=ϸ<`,3;2"9q-1`TfV(/AF1%g8g0*p+婇"؇ uĹamƺ.H wTŅR} ΀EC]<*81)k G,PݻjZ8!]-mW۽&fYh !m/GY0mvi _W<^e>H1Vsf@̅fE/U:E19Q,0| =^qL@/$cQ^= qn+Ϧ__L9j3 }̧T"$mDIfN{Ina*ldm&E+: DdF!Ah,AATgNbV8KHbM*"a7d?V5ҘQq'G۫ $ya)ЫZÆag: I&~y`gaɓǘ(,H &--^q%(gsVqf%C3X .-?֫<lY1U1n8zW2&&.!n\΃,qȮ= 'co,>L'a m#T@PYJTV &  9g`=Hs4;p9* y'Q7BEf0V{%XDc0J,^ I0\:q a12xF?6>.f*FbM+g֭vR8F9 ~;SC ss;BymӞ @,0+@fŚ0 {G5_/_qh ,-0q |A yx3%R"){ cK!k)Ӓ4vʴԝ-{ʴT@0/BGdiX@TB4J`1n hp0c[vEx8#_2Qj0|Q~Erf( LAN2xWrb\^a c|#N=.cv 3lOq +FI-q>z4LO췻 KT» os`n`kAsI!g[IGI { %A%=͗-v3.Xs:/xq!ᄷWmnMS=V)iYYW|>"ͱ謘2\;z@mB=T.TQk)[bٺ4E&v]";?T5)0/Zwr22RQ6GOPa{S;kZ_,O*bVY9WFD\䒴_4b/Kvf$p gxgp AjFVI?{WF A/3Ҁ;.h2 #OhP{7%eڀMI/Ȍ8^]Hhj¡Tk8+&=tWp@t 0C)&Jr~]g4 ZkZMN9ㅄ+qꪰV<~*$҃կp/~DFH Fś,qo8NG-up%.dtH՗p%S֎wJ5(mV4z*跤RI faj5Qٚ=\=FTy.}k%W7|+d3 0I+=r*y* n[nɻw eBV5ñcX$JpsdiU1Ca;^5'&~7{SE|3ƹɐM4d37dL)0fҠ UNDD&;ZY%"JUO;F5N[-#H}zgz.@.~OMH_5p. 3(YV8=-gE܍No^MĈfTa<-+O^Z}ՊNiec. hmƪӤ<.R=c:HFOqSҭ +9Pn[!OֺvuSo2]DR{֠Rbh[. s&=ke9#Ժe|۔l{to 1DDoi#,@eT'Ӹ[\Axa'֨L 8֘ 6ƛu&'Dz|A+_ &ȒrwrCQ[|[>ZJ3o57Lt~Lqq%Tʞ/qW ŕHé8^E!MP4n((А DN+M(2kY cDYM"&Qne>B&Z>ںvK^£JxV yFmI;9̈&JzJBJC`|& C,ۀtEW%ca-M3.E F— w yJ˄f* -s^hQ ("c`7[H2rЉh/MZj-d!ZJkd?gŔ0@DT4 Hh#k0<*("h0})[oށe'worYgʲ_̛6argFh4Rfy [>R4p3AMø 0bKNmZӻ9QL+"ݓ\yO6N'ۜ>MQRpbjԓm9L;_j$WEpzF~K3RFnv{=?@XI ;IFEoR7/"$Q+vJ3Ҿ,ӲC8{˭֦1ڵoUv"b&rvv.v@\)ŀ<w7y{ լ1E!\|J tBj_!WoR0 jzYƳq|5~LQ\ڍc% 33K+T_Yu)bËn4 uYs5AB"w\I9pYp$GN0 )kKqBZ6A:3(YMm-d49v~:$}̞P(GC1l>hvk<ݎơ 3E $l$uHg!ؔL L2;%bco$͕Gǟx^HhAGѠײ;Z[%airKMD2A gECoEkMP%ϴGbC.sS 1BId\T>:P`Y⫣{ⴷjIL4JS5z:mRARLsTɧ`xRAd˯u>>} j!w6az\1DGT>hvw+LRfjOoyE \}z3? '*nCͨDQ9y)%CeucT8aƯmPܺ #RARxrVP_iKbuU7Kk hMr\3IZ;4)".MJSnp2 W*Rϧa>w bB8EtT!=Js)y=xǕ9-?ǧٻ7iD4{xoÈ+FRą|M=@<= n_i/M{FHzJQݔ5G|nz1L\ oޒ/z0KifiZsvIߪܖ=ݨ IonU ?TA^1Nrӣ$Z>}NdA@ A@nZvͲ'*#D*՟pdae6Q,N|~VxFD/{}d\>=݌'gv0wOd F9훣,^Vn<687yta i69Ij'e䶗k_>8|>˹4yfԅ>8Ҿ&a@N0?/c?-~-P5G{GCkp^{N  5=GMuTIpd`K픕hHƒ;Tj|2yBBl[Car=HX'G?_3]XQq&AP玁p̙MZsjgVD`fNR(%ӆj#iYߧ dkaש\3_Ed:rrV>ݑloÓG~Y3{9CفVJZ+2^ȉĭ/}(_F*'&a&H8O2BXY8A(L^F0g\qeQY2e4e*(Uy{$mw7{O RJex@'ЀCYʤ(86Z!.LD˥#ɋ8B|r Նa*LKӹ dȤp)ձ،(9~ ne>fJLJ$CZS)__  VĤ跠$x4-]dN: _aJJғ.m?7`璶X2Gq%E6˄3m ŀe1#Zv,}{=C1GOS&M$N3}0#(A!jRb@4$=#["i/I b$ΩJ 0TZ9AfOVq0R|(Wv5@ѼG@NT@@*k[WFxDDp3S&F a/2Cw)LJ6vn!x4"P2hrb)B*(TBxWe6܀.ؗ$8 55rwo]/%E9˱!Z*4<_IJR3aB~w銼%&VUPfqf?3:ݴYZ>>ϫ??dQUq|3' OƳV<\ _Ju~?y m#j]cI?RbVlpӵ,;e6B۱whe{Ow/CrЁ>HS. 8GizytrR}sGfD1Zvר(,U #V*O⬲.*qi>* /[2fG&ބ+R3$Wƴqqf|`뚻Ϣ|㥈){8G0ʼnzI]tbՂ]V(kg*wZQ 6+}"Sܿc SzA0Ha$lrccuir.z5Ȏ L$ =Ƥ!C'—إ&: w:Q`x1 BEpGc(5Zְ$;QUmnTy(<+ծ r&=;ڙ;z\\kW-t?E//&9(G_f CvLq#@GJ?">S?{5{Z_ũ~z^tp3(x4Kߏ_ 2MX-/8&+_XqLVHR+To#;1̋ U̠?O%]8ݍgfs#vDUhV7!=Cl`Be`nDd̜ &%+dHYɱC'֋*Lk]))X zٗ<}a Қ vB !ޱ{؂p9o2gӅܨz;b0ei7`. gUZJa&c]痗pK ŭn ɻ~n 6`\#\':dox v~l5Wpb 7GLIYYKJްY5 ij'ӼwmH_a]nC\);JŵyK6wD9_cHI#>D`f8b{Fw|ֵʅ*%21ǝ336<Ǚ@뭺'7#W>PmGPBf<;d6B%Hi;+W&-NrereYr^ة m:\K-GN>)z5-Up<|PA3i8xC`Q< {rݜjcbOKx@_3`e V[F%;%v=ӠX ;3RBASE#8B C&I>dZ:)GCFCɤjO@X\[gp'\Q01vn+уK>b0>TFSQaHw@0Z]Tvȴj`ߪtgKF]^׾,?RX /ͩ+ a$hMS,"q#8E&PV+…JgvJ䬚3z4'iiX'ZTXpSNBq3pDb7̻{tP-Q AX)SpF7$ K@GK+E{ 8Lk˙ _˩5cJܔ(YX B#(TQRaA`~/5cjW(XeZXZuT"(2F BŲKN ]K.JG FNеrG YUklZQQ0A;GL*cCrh(4iܙRU(<)MZӉHh؜򓲻lw"(Ffx$3 4v@4r $Z Ω`/߂CQe 9IZO >jS )>E7J0JS ktImbtR YQBP#8?ڪM).%_NdNj'i7O (p{73~B9 9G#JJ-QcVSyS6ÄClt+ 1=׭KyՠlC%RWi]ѫ#Դ5iprWj]e.44\~Q1(XI1^,noZյ71PdŠa>{~ V^{xos߾YPQYA˕ //M}~8Mݤi#~e2*KW˖|Fݱe *B>wO/|S%,Y)FEJq3g"c2lJz[.);GvT[FS[Y4j7IUKfthRBʹnz2(Ub\|*1|HN81aJVc]k(,\ R^M,#)b0,?12C9fT.f#閝#jLAV{Gp!׶9yšD1DY_yӏ)apYng¾[1+g {z'QtWԵacSz߲W5&-K5]N^{Ԏf_,ƲMZ{J#aw` $XûUjIyjzy\bTsWEjy(2&E鑑_VmjMl!brP!~#^q`o9J1l,[T@vH -Vj ]*3;f}hPCdc8nfR2y78홽}*p<,bku0:bW,Bc?}@9DJZk(J +a`Z42WZEĔ"̌f3d\۫Y8[\#Fgg5zR."/{˯X\P՝Qk,[оՏ~{ *@I_,Vޞxu?%e3{fu5;+q{q-faT1DHov*! 1\(>/`8f.)Fs49!E`f8[~\^ ry(`*z{'*=kۋp] ߫_y9ߟO^-ޜ7 P/vd*ŬY*ŬY=`_N{a3E/V-~rE@u9ͭC|ޥb,bѬjtpz/rIܾ! !5DZR'aW+̋]7Ym EKb6HVM~X,Dcs446pڸi"L%Ex-)h膑YC>ѻ0焋]/RIOg@%50F3`J S^JnAi Dfb`՚pMynEqTOWsID1DZȁbҟ)!KK{|фC:B.iQ m$q%4V~ȳ,Lջ8 'FUɧ( omAu=֧sGPGu_1HYNC, 6UPm6d]ͱ!&aLi1Qd5GϝB$hb@rG Oܕ`|MF)܂'>D'( Duft!Ϛo \:rE,=W>q0V0N !f}vQ3K||bL>YH.Fc|=#s{Vg1UV<}W57#K5\ɇh:iQ N2suWj-۷3^N]['8~s{o:!/}>h6ڪҒ)K7Ւ˩0QjE^P* ʙ .K -5N}aEAIV2 rVZ%A,,+tK4X E'JT<,6=zhC4L/NgjqI_>ؼd27pͧ<+ͧ<|j)Ob^|ͧ5I|j)Oصyyr D}*w*f]~0Ϸ_kSUj-|>_,+Ȃ@V?w ,5ksTxDz>jt_ޕV/fy&^|!h7n5r,U$~D|( .A4 :n,> x20 \: 0FM1n)y)WQ] i_j9E\h0r {YpZ)nV *0gc@)ԁJ {Y =Ek3(+.&8}zsdRAtNVe_;oZCS)T4p&4.j'Qؚ 0+Ҕ5(RiGEeù>Rr cʦԐI @3ፐ,s`{4Bc$F.)ѓq&:$}b%yd)iD f~b/=QސeJÊrDXyN|ߦơ JZFe PutP钊oQx^p*lql,gWh!j\mHlNl{ڀx;G (h止fBP+.qP)̐lC>@66q=1Mb ېhf HP6e<(J2lÍJlcV]SUj7p9mTLiCk*8nyhzpV޾Nٚ$LY$;i>np*/`xD_mXMDE@vAmDp)ɘ>t +mId2P o tIl0Em`҃ 9C *b3ݢ J*5p%j]o%2ph%o8 ܇€+9)f!!5*Θ t@jAkd漢@ @4.t ƿ81k j-P<2ZՁ8:HPQ]݄{"ɬ@lFvzRqu4Oؕy4Ճ`(.sESQ ߌ7\s4׶COy#xH7]%ÖF 8\ཽ(H/GBi*d485\FWeo{mफOZRӑQk]5<3[db.<" ?g7'3V{8f|3{yKuFO@TAF؏8 Ɲq/p!E",*"94VsK $IXj* dn㵟U./UR8/as)T3keֳWih:@ k 5)7#/:E e.ժ}χ9hD2]iyn; :҈8Q`h0NF]T(":s@_h|/=T˨b3K)FO)c?NuEg$(!htGjux%p?OB ZNBJTPӔ~1kݻ:A6գoœQh]uos=Bj]we՟c͈5/Αqu!mLp pf@aB.tnu美T^#.)P8Xx䧷[ ]F;4-2ToZyAsClte;b5xRd;!4DC04v)&iBJAFDJ#I5$IipQJ橆(Cv{C fY)bYA7,5˚LG eOC;${f0wIk yZp+V\Qaε=0N+z6\xЅs~"Z)3z]EVMaG!Z 4qDC+J'2]ˡ5w&O+ AN3  =RH!#Rd. @1@.'OsR\XNOHhC8Mq|)v_b$žG&6+Hw[hDqI%IyOW"IhS1ɢH+P1Ҝ9BNH~G):I}n4ps<$q1MYʫ`&tM̡y-oɚJ3LWDݒfZ^hNsyD}.:")XłeU'8X9ExY(})K}K*۟ETJ ;ΈF!RJiDlbD3UA;7ǽ<ռsUVm;.MPq?Q<Z?WUҋYGHsU1v<^CO׬UG-Z zÎpM5Cx )&愥U)RAYs<ͳ1SB$gg͆gz(u>?6ݳfU 5Ĭsb6Xu 2b8uQB9 } Ÿ kդ_r7M mnj:F8fkxmgyƼxKƯnEԯ^d _^wqu6\WmEod*&kPezS4A sӡ%i*r-VY™"r+򜶌y%~Ey4Bx]![2tC릙\%|/gޔ\OA2ȗdeփW伬/duIpš*Fȷ@hs -gB #i9N+Q )#҅EyuRsktW'Uҵ4BɌ8frƉICBTFF]qkWΕiv+n^JFsx[ %f7<\ԡtKIJ_]k+szSȧxm9J)طeaz\z]z1Sx$V5ag@U ]|U- *῞F"EmlLmv[IZM"ZzR-.n'#-sZݐZ%pf"}]஠Msae[@>8`h+ۑG|xD[Q֊ {d9Re5о4mu[:92g)3GX˶1^n$6WuY1n!m+V耶Q0vK^{%o=ZSzqih=ڜ#ΌIszb yf/zKxʶ]q$U-ʜ=" o [2Lk.{'jg?z,+%염y`y$;\\~𕦭۹ @\hײ-]}4 s꼈@Ys٥L'~)/-# r7ŦF{y [r;"zevK4^ژ7w/_ 2|-~ #72N?Q6yʦ;Ot)Mpy1 S)JS6M0Z0k41֥RHQl!sxRD7XtNJ&v2xoM= _7zInR oW7s˘MCk+,p>B}o;785:BjyhBAm? C3|?q㗜j q h QbДBIe,cri1Cup^;fk~2\gejjU3 r{.s;O5MFA@,nGwOt62?H‡z-J{޽}X'qқn>m*)Ž.T`!ٲ(@+:hpbmDO`sXbIFX\75'KJqb4R9]c]%(b6V, XmMg~(Tl)t&Y0TJU)mqMCl.Jh`Ms(j\%P "Ijp}A i"bq$ *Z4h#"%}A!H.[c ph̹aƕEvY>{u< #@pFCG4t*s+88v* <µa%hFJLM53`?FȄ3mW͹?{7ݒJmG1V Dd-UA`-`fGX?50 ־^P4Q{}εoQQF8@QApyنuz{1GKXkwOQ^es " 0ߚY?[(G4!si\{p \ShohPB*/^7_ѓI*7adQ` >~G^L73=[ԨB 2p{Npy= ݢ2pg 'tSa <_Pfb+>*\_E+DcW MpmKrig{x| ި&[ (Y'V#xZ1l>fm#Ҝ<[wKvqvveP|ZG8~Nh.h1.q!\ïkO t f|KkSd%I)1qUUan DQsdRxoxF!7U~yx^k!XlٛGSEk]\斝IA0&ل1To^cNJp켛e7 .6=PpnGO6\gyjPsF ,c:ԬQ%nӘ[_"oeG=(Q#K/Ƚj{3JZN9S6ը˙UkJ ߭!kI=)_rN(=רb|qGb}lwfe 3&g 6F;֩2Hn3L33 Z8$34%DIC49,ݦYz3]0#hr O@nUBz8O@Kj_q,'{H};\Q̉_>,V׉fu^:m~~~vQ»̖?!肖E>̅K?muC >]s.d+}.|,Xaj$­čdW2@>Gota/ߧeC@ÑO~zA 7rԑ ˯-Axq),*pdXcxĺL!=QȜFC^d(m>\%=J_FbhR^sX/GB]6%<Zֺ\Y6;)ޘlEľzǶwk 5ktAB|/shf_WƼ2`GP3IFH c#% (dQS\󧐶O}S>vv]|Y_8A:ܕ ? RT;^t* \&ȟtSW7ۅ4iz9@( Y-_p$ÖpIaسݤĝZ:A;?7 H)pJ*VQd($eVkXY,3$K &2rP- u!UY5Zjl;%j47F&+;yB&I33),)\q̒,DƓ4W{0 73~/w;=@+S0: nLM-L+$5Ɇ<#i,)7x qgq*0”穷qǫZӯ`m݌0CAdaGc櫾TZKRDjbLS)TS% hnhΚ:Vd*:#Oy7CwKɻ c>3Q_:NJ` o2(Wa9+}Jɕ+( n$GF]'Vh9S ,wRO:̢ nPRiΜRBIc37uRKC njI-)"s8L *2tqMc &`m"mqen-iPׁRhR~u;  \A[;u}ph^%SDag)}{[EYlJzkj+@|)P'n"7K\,ߣ|);żyhh~cZ򮩬xZx&EIa{@:vG+c\f *iX҆Tk^ڠxi^v㗈ikj˭(:nóTj{bV5*G6J\Z, ta9UtQڀmv TA4. N SʩЃ 5_wV Ճ@+>Pg51l z=QXC%0T+p/R;؁ӸH5+˨R~^P8)!\MQx8tTFG< #nB޸6•e : i:mTboޭy_wkB޸6)J#]5 twntZ۵wk~(лa!oDK6Oa*$!J>/wTd/ղRZUv@V}H ֪Lu __*9DI}Yia!zC`tJ*[B+iX\$JeF.l [5'b8gS” :@cl%]؇ [vk2 l[Pi䆕.q[kMIX B ibByA4Fi"l$y,#pHTE&UvSK;BU/-q鯀 .S~"nJ>f `U>z=@=qԭiCN|( D![F TLpfB@T"B vN1cC@N'K4d%ώ\fh% 'eI>IDXRzpUJ6cə٭ki.qou5e~7w=^;s3.7r?}4IX\SĮZuV&J|P3*)TRH%j =.O ?^ GxF+OJ;=ޠ1Qm@l dscoM>B2ysN_c=G:}OW_gGOž9q Ү R"ǜ:TO,IuP* &~};1UXb+K".@Mi2b{DY jE:yrl;KK>ȯPϭI -0n)VP!"\':8ڴs(4,y/pA' q(,&njUօS+V+ߌBQ@WK `%W#l@[܆x RI;WAn.\ox{d2ԋݲ69}~EMԴj"'qzVw}6Jv46-ÈG@\\ cYAc5:(r]k-4l! AGH</N[pqFiOܔ'G LviNE> ]F\ƶŽCwi{6.=i/yq*̄q9~9N+E|;Hv m*O;HNZui-p7T (pm+NJ ))c pg?DF 4 ~7\q$ ?| 2A^~ >]ޤf1Jbi3eo"8ʏ+?x݂)+m#ILtyD^2w l}X<[դd$%)RbUheXEDfDƑ7U5A8%WYEu"i8xjF.4Ӊ4qKAnx[hہ^5RŬK mG[`qUϋ%ymq[`r+ڴZRZsq 4x#rBD$O Pp :hQw 2CK֕^Żȣ}a)-hdqFH{ A/$DW!@iCÂ^9⁸ \eA^,.),${ܴA1xW|J Q߲o>?=0u,?it_ݾ:cdUV.#gOOSZMr_;vrs;7s>__M@(ZB|߄r(4Qg>꥚S !=礜LRnRo2집]u>8뒆Z[CT+JW\~ s\5tLxS-|%&*)V:2Rd8ۤ|G\ bM5KFY] >\KQ5a-%SRc &$K/L!J0~&u9LK*bdJ[,{J43'x?hy֌Y3QO\F`QD J4j&5ބ$V8]<)ʘJ=]qwہZ%x|k (G@|HR;3#ɤW$Ho&g|RVDa|]9I iHGic5>i1 Dq!z/\iZ"aT2^*׎Ӕ9!CNٕ]b0R؊ L.V"z%E/ u@ 1'mJ dOTz&Ꙣ =QuМ:j<и$N&⃋4_eJO8ά^Ii^H;?V[W`JH-3j-'0Yp\R5Q_%vGgFҩѫes\9hDK*vkaB4GVԠL]!V(ClHpk?T~Vx6 IBe$0؃uاVjZ=8I/W~5|'G[([g9)Qb&%< $c¢mN9w9:0JR2x|5#9Xjߟ[lWY'5o[.I㴗.HI&uě:o''V\Ri DJ e9^z;^LxiQ:8 T*D8N+ܹѕ RpPi{w';~q9r/CȮԋ79|Ș 6{!x,Yϳ彞C4_kI%$@p&g8ETA(ϔIQhqAہZ\>~x"8('YhHQpEbЎ&GH+%׊:J8`BD^ )( I*5hzڱ flD-Ľ4c&Jeh 9쀣9)DQ< -p`͢9㪗QPy ܢQ&%`[g[C5` lT4r'؜VVɘY +rklw9oV }f%r#(wKJxFO:Kdz = |<1ʭEDrt#}=^5]Jپvcu@\ SE{DWOp'\|r}>ЋQzUPWLPE4JRy\^قoh쉅]Z>~z\Ms}DSRMֻ&[.Yv ?7aYeq7MgA"7 ^$;Đ6,`t_ūYDߋeiy E\!s̒R ;<)<2jN3[đ4FȾ/9c[&Tkr`L={'kΗ[ pUIyVٳ$-4 d}QTNjf 2TJĄ*)Tğ7)d=PITk kɰN$JpznW 0L!̐(Y_dR犝4jG8e'yc֚N㞝V\bho0Aj)uRQ/GaFTKU)QߌD8J,#OzRg9[)Bnc%A[,l"|ZCd8N$1#A#mHN$ 7a0CSTsoZ}޻)l$ )j%oӆږm4܌wsO0 9 *q~s{{9z?dr? ο8|t w]>0Cb9΋_[wMR>iVَ_˨BPIE긢$/2U}1ިi|F#e%ƐV7mKuBOإpud%2J1` t%Isۡp%\']q)us}wo}S>Z/-H& A2[wVߞT^໫՛[0\s1|^vT ]N_6MX- z,xu{GF2,"ڋTWqk f"CwǢƒ&"ET+&^Vۭۏ?1=t{k{L=LcfY3NFf5U_~8 _N^9?D_ Ozؕ&&   򧅧1 a\pt!ί]IzID!ĐRꀂK夣d;@pV|i'g?rsWCL4z=`9= i@}ƉgMcBfHI6_ݩVɘY RӴN6M; }Bي97Icc YFyOzQlZ8S,6ϮP_;D$ vrywл;Ii d="j嫧: .>>J(UKO=$jU*eTAU(p}񬮮Ͻ3)t{WܰQg+i1ݱd.)/E9b$Aj/8aIDQH1\NC\nüs>B`Lh5!?xT詡F:.vԃc֡K!`OÏbnPD'TK܎jb#"x4m.bVX" ,3dhT prтjQMN2: je Vi0j'c` -TF sގj%@ C`ב IYCě4`%O˖-\j!9)ygD:K\"RIڢ}`ڠk$dNYy<][o#+_Ee/L`ə`sr;/Ir߷j٭;[< j*Ū%fA$>%Ÿ V":1VB+NzOdh )$<791cBF:D%W)U`B)SecFO I%؍LHL5|s`C01I_P8^҈38)QՅ.EUf0*%:]ضO7>y3A-\1?37 Vw"˧G)p|Szδ/_o_`R DY>->OW T_#oރ#p7Ky2c~w ϴh_=+K !ߖrJa@ %x1z,A> ή|$|QpŢN߇.RI1L^{z82`"*5JOQ}N5啽쟞ZCJ AF3ͩ@IC46 R&\%gɌNcVdG;(!\.\o\v{!y?|:[)f%`%,M&,A9E0qa%y2_(穃y"nysClث;dԡ ))\)I`kT%ax"0P"EKZl%`q- tew%fám%f5}6Sd񇷰FVmy偯0TG8׬kb__! DK+rx1LHm81Κ̸=HZ I%0A&Zf22% Hr+(d*'p2ʭU"6AÞ>Lr$'i'yΝ$s29K&3ig#T />ʹ Ac53Zvv޹ 8zU,IRIGݯn`pFD 0 +Ƹ־E mw:/;A!x_O TOEqfˀ*Y$ `~V58X9ͽV :Tj`6hJm J0y{G>)Θ_pӝp; [Q_=׭,Σai=`*B6tOkHŮEa7]qB^V $8ͧ+?^-ЪiUc%ѧ1ÕsAς0DjQt wב dWQ8J#!G=Fruֱh$2rMl$twKZ2DXjRԕa)z^O .lъz>8A_Kՠ"քhHjNe.qИJj(kkSCŬpr>2޶ ):` EOҬȀ^rݬ뮅TC>CǞi`/f ֍ճD`R]i߾r&jB)XWM eeF ʾuKEpz%hlQ7mK2 Qӊ E˛?J}?q(|~=NnI`I[Wrr^C__e9׈i$>17\4y4Yul3OHD`R#!DW-@O9`d M6[Ün`9$SւtXv>J56g8J~4O.= >IR0&YVl8|sG {y^p O^/`,}Oࢿ{Y/_*PYg;^~LR0C}wK{7{M{.H`eLΰcV+(aR@` sj\fH|Ϗޡ+l![4Eb0G`Ο+ռs4uA .hyVT"^Yy618PbF "m[a+do+Fty3=;׌7H8!Y9+M5c8hhRAZ٠z:0*1kUB2C`sH_lbNY EN`4.%& 3 ,q B2Jl}?dTf,UݍWu7^xUwSUu8 VX;iLJd:IF:'.W8C sreF$ |M>p1U4 ֏YJ,4Uui]Sw IC Y%a0@iSFv7Ω lͷW5_VDW fUOcȘmkmN_=㧹JùUh]}q,*"10[qk˳LؔF(4ZnL@?2c I%l]R0E3!q2+$mtS7)QM?ʧG *p˯^$ݼf+O]Ow> "o5 ؒyɁ'.SF+T31V'p mV)\ 4ƾH 6%^*8~U@c{T)gcU ZjA9 5j4Q5&Hs-2$n c@:t%@ȴHs&HVo gRv F$T S.9]]RFT}CN3i5vwOOv #TAh01pLD!F4vqL眪5 [6w4tBm Z4 2RqN:ԓPNMXj=VQШTGBmΞ&d{&,011yy"sy\f!HbRr;)u\ GJZ׃~`SZ5;;t`-bEl0z9_G7}B_3V -zvFh2W-T.b Uǵ14pFfh |h OsvԄfF3bT@b^>-p&|iŽuu0`K.R~4S S#`ߟQ!d㨟O0|VB!PCrHyiֆG~7(x-Xxs32g*EF ˅YN1a [QX5`ԒŢ=EDծU0[_RP-X^yD`U1 B*!Jp*Vͬ"wLhe-cﶗJ) X`J 1zjU)qu.tŘ8RX ZAH>שRzRg՘k$hw s`ki|yȂ3KV$?%嵦K+yAf Lxo?=3ޘju,EGe07QlPǵ-W}⃮)HD*QF+޽X11J(o쁹\\^l&{~?%+g(n}Čt9 jp{$Xw؏T$Co*'*}vYKR$:%,3}FDIvL;f%/d ʎ;ٲ8M?o_zƶA[{UljWWC<^ZDzlr6EKϫ*=kڝ7&xÀ3=k>o0\ 2'LHN\T-HYim KVRL1[w 1_A2p:gKFwՆ`yR X>zA,+x;n3N#ˑFi|]dL3R8ɕNiؒ@lXc^@RZvúN8۾ܥgc `P rEۻߵ1@|{1iEU> zr}O_fi]Ήgi7,u#~7kw;dToU*}w吕:gE j 3VIA(%Y,!An-`U%iѳo}]".(vuUkܧ ݻW[ȚgѧnnZjF}v>քNwїEGʛ Џ5aJ?II1֚WatQaj8z&T0&uk*mw/-!߭[ ԂXHą*@0]'+6p;ƕJ1l'+Šw+7.3 y|+A"~*Fէ:)WsZi*oLF8Vڨ!%ofQưrݶEhP]rT w !rDW:j%bݠdzjsbK'ըiٺZFpeH.zzǫ,l'" } cn襦F1^˳G7+~l>M <, 6^췋欫srO,i5^b哛WX(ѐ|"%Sr{Ȇv(أv DtbǨz $;_J4P5!!_/SBiFs[ ߨ)zjP8qt[l#y`_]\~6]zY}m .˱c&/$ O^䖓D'L hkU&TQO> &/'/NJ2C2_; gĢb2JX+ bV8zkb4S־Ÿ@lu S}N5\p]B^!ETRSJ ҂j*@)d)Bע\:3ߖ */l4P$ 1kaͫd<q^q&SM2 -4k2R*8FtA+-'{za7*rJht50XD4U8FP.(W0^\^l'wٴ7`-i[7;1tJ$PwV{O6Rt M뜈.+3UiFa&amJPTٻ6r$Wz٘-WP>f^zCX4މ )xXEhhd H$>$RQ7A5hk 4EFC+OmImaj6pE{Q2$p[qƢv+QΌ̋WW7!z}*?Ww~䥺CѰ4{̃A^|'ZtPtOx`4{ݎ |h~kM)k6z_ l>ᔻT>%}NkqQ05G(FK]kS"w2(P'Z頻}RnW<5g\rM(0һ=˿-JY@@4;x?eڱ8;G>^rNлÇkY0 i G“]ewn~СU6@R'!$\4&μcZy`$ϥPa9o??NTNu>>)jgm_t>Οfo+quh(<7P00tM{|bLJ~Qw5%k!IJ-^ui0\;w;~@i 3&B;`(.I`9 LUI@S/l V 9Jrt9-bLykh46$G5p~$46\(4+G5=z#p/dߨUor~nV7^̖RΤ983+T8S}NfܥV\fܤ֜qj#AAY}_skEZ.&j=NO?M?xފ뇏qo),\Bߖosi1_Ux!+WCFշm^UHDB(T9< 2L]e0uU:I'L/T,% |T\ͅsN8mI_LDjVD/|pBaJ{K)%S+y3&/Fݪͬ}3kjƥ>jNJ \Hљ&Ha^r 8M:'PQT^É攇? &^E(Gu6MY&"v$N.Y # c$8A `,J"))eK4vpK0{ͅMZ+I&IK_Fk`LQ&xɎ1H^sq1=[q6|vS0q&).j=LZm|8cTBm~uףl8qqj{@wgnYV3w3x{gWɖ)2/!ZlQWLM8ɮ.4!r)^#_Iv`h>k8F*zD=vڦPq&;Zhc;E`6I}$5'EuEeʡB21qER&ED54ZBۂ%-9Vy&'کCNMR_.Jͥ$)zS&:S +-cΤ JJEa#0Ќv0;\kO_#&"K)^Bʡé"AhbvsKZXy]y=/ҬT. 3evw"s"OR@_H6jIP_T4DgW^icCp,T[vb ϡ@uX6?Y=˧KM˗нh' H!:Ƒ#S+#)6:GJZX#u>?7$Ty2@eA:9([{9%8_KRqD`^RK`),\ G+uV'NU P@SJ!RScCI䒱?1tАɻ,Mũ/fZ`rHyI-rqݤFo}q$BVzhL :HC4"ޅdVIN|.u-+b)jZ6K uejT*pa)Yka}-fBQ8;/{;n:O{|E"Z͒E."L1ۗt̬7w>[ PLRLjP;/Pڃk74R޿9P0b(vs!G}c F%z^10_JfX^b̮֟[IIHnI88Rb,SFsP|RӶ! b(58K!gP_LJOTw ÚjO#qGҜnGN۵`y'Tj4v](d۴8w,RkihBaՐ)kT#(Ų>sϹ9d,Ԁb'έ Rt]5*Xwn;K[k~#Hq[7U(P }| \ډ)nd*acd@m7%.':"@s<ɹuqG= vA?bZҡ)ԡRnQò_eT 9z,؆;n ́YC57ajLb j퉚%~a~u% *$\TJsξoA5uݨU:93,'QR+qP V#>'FZt gjO08|J"cإxiy CyYpg@2 zVƇ?R,+1PGT-C`aPVc:0ppI϶1QF%Jg;zsKszū$P˼4J #hJԺ6ӆJZXCjy]7#e&抽29[eh L\`dԸMI :_!Sڼ@x?F( (yL*fjqD(ه{]L[d?NF3~7鑐QD|?/\1Oƈ烗%(mA~嵏sȕHokslXh>b >/_ 5>Nē !1Zr Q9ACeUyaKZXPP1:4 l=J5_e]u 떠\PNu=dQUD_scfխgюwZY^~?YSڹ;=On[#JϷ{T(lcFT{ :B,ߟ{G+- 4lNßv5ՠ#H@M|8;~:H/`4l? ƼcW<~~g>W{O-÷Mg )AP~w; xy|1Ǐ~idkwӯnʫdZԬ dk>(܍Lz!XkϽ/ f.|L3^[ո_=Flh bDO:1"@ut]6TǮ]MzѨv;`|wZVgbAsŻ5gjw)!~+3`dG. HԌ׶b0Dy}lyi3vRot(c܍Q܃ǛX:|q&}Tܡx0zOldW!Oi4 …}lYqbq\NYGMM {n6? A}Gt;`(8{5tks/D.,䍛h+GK A|G#t;`"nO$[MtgJ 5װlyװ\C]Xs/?7D'ŒڵC/%@bRNu>LFb OQa@m%/@ 1ݥT ,aJgzH_!ŀOk0yslRκ[G$Kqee-ɦH.hB$KQaA +Ľ1N JiF6G@5#:Xd:LIxERmT+U`sfWu~sk&nנ?\!V, 5h\QbMvD vDq]aV#d|ڡabXz(qYv2ہ \TDMp{oyy{.4'46#A!HJ j@. NfI q;!RyNc@-oi@sh։lT IҞugCJ\pޤN۪A)#͠Yy)%\1$k23%6Ps  ((ՌOGt;$'iAN8y AlTќH #8C-JQS[TKF䈈zhjSN:M`DwPHO1|jSgo咩W) o!dF{oie4@PЦ тo~;T-q$\2ߐՕү?#!:YAzIdUhTbN՝U"w"vPВRT&ڧ7 'R/W* M%'7eʌ {0<62_A \e{%@5jw^GP͹֤kZvg,BBgV~{,"ڭP 7 7ΡTk@.NϤz9slr`9F  gCd갠@ԑv:K!PKo8]I#hDq:@sŖ7,t⸔Ϡ8iȹ,*:LjlNa[ް.5@.s\]DaR-;c=sZ0|\1d[%r5}\͍vPc5VIm'FPlj園>QhUO8(B%*+츶aq6iOkO'dOkЭϝ+IZ/6~xsy{_ol}Mُw.^7[)>ꗛe?y0hh愠[cz/w{ڕދp{i,o?|߿MXء#NF%\RJ JFK-۳f)t@,BO{{_+aV>+epC Ǔ<j\Ekiu<3I1 Rv<]iց3L7349{AV0/VWB6)jp"#\cᗬzwؙDyw s:_~2|4g*{x*`~գtFHS,eU.<>;3<:;C,AJrpOz;gR~jkfP-FYW^؜/V._]2v5?\_t./m&Az[](;٩l9Aʃ $ُ7*x!|qvWoz(gws@Vzurׯ}*oX_>ٚ煏Չi]p;v1wBIYfFKG0\]d,iK,nq.XlDU MFoX8KvjQz}>Hƒ ܩ`&9WΑ!]i>q>œf~tAs,zq377?MW7gusow"V+THs߽c;<#D%:WnX=Iڢt.`u*w2 c8؀o|vDy8 rl\m*q%MyRAsN`Qkl 8IV+@>Zאl<'PlH_+ M#Ǚ -HBo`8B*19%FT\meD'Hi j ~؝bZo6FbZg< 86[3$$+ 0oYg'9ֳ&L;E"xP1?YF;,ծH`mIMvcW +DIs0f\#u7RK>]kʩD3Aק9ּ2)+MJ"!q^#A(aJM0j"4SeV15).!J'+ڏP1{Jm5"8өv(&@f98)#j#.P\BBR[.{AdrNt9gILh~N1Xu_J$'E:FgtWr@kL0lV$=j~MEu |*$N6%rTQW 4&6%Ȯj.9ۀ%Y&6ଝ Q[,`ԽXn8s-̠h!N{ Ϲ|$]L#\  3!DօYm@j$5FgPM{,$I!- Lu^Ni m=_=l 2Lue:b;u`A[KEj^9t ?b BV|9'Y-kXqJqιivVh,K<kkvU &l6Aڞ] "mz<ڨSYNԹr=L%w69"Ɯȍԧ#?J "l{"|kÇ-\tuޤgkq7KLӨL(iq (3z[SF'IO0W !G?h.8R8Q8uNu[mΡ"jQ×>'C'\d*l̓?}6±2tm.U\B6jѹTuaĔOhB0^A3X"km~/P9q1ߖ[ئWrW$J*2:\9@JhTb:&A Q'."|gog^^]]ŋ9:YDňEvw1S4X:zen:Zs[mS ;L#pD 2hc0WUa<x y&:ݦ@I'_>Zмʠt~vu Fﶄ8k侳{3$ W GJ$SdKT5pZQA"T&^yDs7vP+:>F-ą{U=}rUmS*0(wuxhSu sG'/+ҳB%E3܅JgKJ gaZv+8>w.4iYPaI$P;'™_9%Pl ^}F $ h#:tb#iyżdjt0H:! Z̷ayjdΕ#&69+ ]d_0#3$ʺmMԢF)ANVD ڒkB ǨH./bRb(t*fi: ɹ{R**~2'?^9#,p^# a !i PLJq'&@ rvoG]5'pYdXS\T( B+[TEb΋\pǐg"\#",BqB<197ej3vEdNYEؓ#I0UcV\aYh9~ԅ駏ۊz֗۴ZTrIEgޢ !w`v*ypRC)sء^=Zv`zWJ:ᕲ r`v]$z'8 ,eX"m]\P8[zu,D8/1$d؋m=)Ix}M1Vrxo& |Ɨ,kx,~((o>!kamhN`-mւYG|\ٕE6/ L."W C96߽]^TvyQo}?@ d,rA:qy)BPRJR;7÷PIsKh-rolEZ'6rdDɹ}F/\FUD)Ɛ2%jX Dla9CfrEkJ|$E(E}$yH&iM{,K#E%P}HBEWl,'eV;"OF&@EaPfc샹l3I[:=Lymۏw+:}0L#l 9ySzr5TL\;H[1hz@"#e-0E{fLۧd,)2Nʚ&i"*"Imuv=:RN[ o_E ŒL5[ď_/}4FZ3wZFk:dt  ,dl$k)M> z6KA5Z86^Uh6uCm= %O#c[-cR)93?c(%W>dhWM33Fywcu6Yw[Bβ)wEc$FϕA1]s6WmvcxWݞ]IS*lq#<>fce{$lth46ݠw_݆@3hS"SN;(oDVg wX {\);2p>zƼE0hB8M xcOW:Qr z.Rumsϳ&]8D[`i VkǸ|2(Y_Ya3_y>-{Z06v&vC!zB]bRok3F93*&uXڂH'wUmJU{urm9h>?(_Gr_ҚyOoi<7?~kb GD0`S yu{θaZ_\'VlXJ9ћԊԖvD}aw;u"JT'º`\p3O}rGMa%h [_)=o|fQLo,6 mv,홲eI$ns 0]&aHK|=>5UƔ, y Z뱴xZ#Jg̀(]8ŋpGÝΖvL S#KT#.P;_Z+8a:):f0 RZ9%H$x-\:*9STi#< |O9W.5_8V(v1" \ R~Ō2dSFL)1Qi/ kd|~z1^0(9**&!$dL()ջ2(4wm2r* qUR5}NA1AVcK?&Yrzc1#`Lil;f #[IJzZ&M)'}s1#wv-mczcZi]D*xOVnb-,}T.!eOl,f0LDŽU7Jq-cV;Ί7Zh7i'TWo܊?E%ڢ=ח?fcV\ L+EXPԦx[%0*osnc IL)16a˵Z}粬ZKhdՆN!+u-uXAkm~R٤ұBv%Q[?Xm!/UqI_ڥˤ}_{nuZJ"A_rԔchDW)]>M#Ѧzs$2cFC CX׺ M:oѻA t>wblƼ[D6B9D`J}9ݴEV_@3xm" n_uInC CqS/LJ;,$ JDd*X(XfmJ w`'iD)qկ)5*(m&.$ q9 yqao̯'&+׀\?j96Q$RFrE2Ma)u@2xY^Sd1"q"A  K@ŊXtg"sɈ6"Ҝ5,E.C 2cdan8HYb)wX1<$I&)Qޮf0'^(ܝ[Wxlvj"PH WAZ95]RڡToVLy)iF1 s|35K|Wj +,dre>y")p +l[1QxB| b{g$̈ ,j_ho,qb)wX18x=ZP;{Eu 9qC-&JIΑ0g\'8z'rP2MK-!v&(+TN}NvV #b StӅc$S < B5CRKAdǢ6F(VIo8rüԆ@ C1*[1ݥ'#H+ro\ 6P4Jʜ ka"_XQXk9 )O'pnVI}ԔZ}t 4'QtGMްr]D)K\qӲ.39=nIJ6r-YE6mCIQ4{"ZXC +hL3"@ wX V΂9KsqaM.n]?$*񺲽Kw)EjR:{;RR=\3d*sMw^(D( hne@TBS#Z[1fs̒hnr>|0(r.WEX!Ҟ9\iue(<*RzsGExJYV+>WFiګ|BXWI}Ԕ})NT&fPd`|{|AI}n40^WuJPZJ-FiZ'Yv"ZqRi(RR;Ri(Ųe#8 NCi)5 ؿ(UFFT˘3}iQR^OajMTb."^7E HuZOR1{sNs'8)ʁi%TA*9wYF.\z+ &q{ׯrGsSk@gUbRNRM97*AAw>sgS%WL^}g%,Je&c/ tԎ=SxzD6CP " ՗UVy0J9$x낑bx%IH Kрz޾YD)lXki1WDtW'K<.Nï|{S8TOsx5ySں,,4 -y#b8(SRrPnUSk7T~$,]p9tCK2Rj4Rf U 󏶼{nezZuwgS;KZ \Q^4l?].[{' Jk}?L^+F;}_owB5eyltR|w ، |67,O~oT^g8I?y87e)7zRVNZ R*ysw,5p-E~~G~~-bNb1e,i!kW|faeY;HQ=9{!%Z8fi{qBWpu1^q`a;&Qd&J ACXnj$NnJ1Af$n !F y9l K;-rߞth-m|OWZ4Ue9rwb:ʅr{sy{o銹D'Asgo[9!V'z1Z6&P `C*ybYe3JC%Nk'Z>6ֱ  T L>McN:IN$ Mŋ!S5ʷ _BC{ATYwǸc:m1 Бvuvr0;lD^UjmAf* -zGfU'[P ~#NrIŁ\BěwU=Kg+<8fJi3|Ƴ돞Nޞ~89dTwIOszn-Fbl³,7nbV(M'Jt<yU[X/'E=lxkToM8ZsB]et}$Ѝnlo^y`>Z|keiv*6:RzKNӷS nw#:*ݥ ZrAzBXRJv4`6Jvd].YjK~Ij(AOy,F=dv~A˨-U.mU+w=(چԬh!y[߿w`)q@Ņ.b`zXq/q'~I*#e{[7*\w x w{ y>)L@y)FGFy7Nw뫃}Fﶽ+ژw_ ݆@3h/Lun8Qn}uhݶ[KRy})L/9r`U)wؖ?O~vݔJJQ 2#;=~.wƾ+&a.\g !@_u85GPp>@NElK@z[ a$ `4&YQX=ڼQ bY8ߺYvb0ʐCR߻GÐ`EzϩBM&ϓW1D/QU?J 0"#Yp^ jcc 2% GgC<(\?d~gԊ5\G0A0aǿvץ%S!/$#L3aADdyB`ָ]|q8V59QNblr+C/_;ErN^yj@'PG=3zҧGbtc!wɵb[}jp6fOv.ZS4%WhӃq{7iJJԯ=T`2pP4{A $ιJg 3莰70o{  (RO Yƚ—6Y&K\8C`L8ݑ4"+?gp6 jA'IiːJs* <*,{9I" ;$h1#ZPn]HکB!wXI Dјin +8/Z!ҵ̟bq<&%wZY7Rt<~y_,{6C_ίxQXyrnUY=Yg~U4>\@Jju>#QiG:MhL"^&Ს2w}eZ$yl KYN(RCYt(̴9 ,QRlo}>E* 9M$ tGLu }~L?fnPEd2FYT f__1Rf25c+t+N(ǠnK-:ɼ^/ .>SWkJEu\ YQ(iaMekNvVdAN\-wDȀR5(* 5P\ɪ5ܛRV6'-4O{/.׋@VLmb4'kkA6'"~ra65툌7tL R.`6޺XT/EՈ`cQbRRBXw \#VB;pY&)[a?pafwJH3PRV7\W#EP4̫);E6v j9PCF[! e$*"5{66@o`6vF*C5fPξH_Jhdy ÈN weN+c8<Fv!(m^*nݙ/WE,AmO3zdf'oUū.Ud#R>~٬ YûGW}=}c65Bd)7t'!pjaĮ F*Ni'ЃAR *Y:(\YpLY1KR38PRTjk"՘ųu ,g1w©ح {TD$zbAjX +7"NFր* eI(0^PX!wn= E\&% ԁoi ?½&Z kZ{T$tj_U4@0#eoC'N&2)*fCrR|0Ǐr Gu>~+9mS̴̜4nG40zxitچ aж7.<ntCL1ͿMƣ rK@Qy_zB[6>P;;tք)8Wb7u?&O{Eaoh" QБ(1g z]͠Enq9|ɠ[Hj9; Kx4(E^j)4e(⧝+r.J eѲLBEFlv(} X5Һ\Sn\N5+JǸ\`Z gsLxvL N}Dp{-g~V/j+7v4wۘ"k̲%¶ۣO:wstcg7 !bd B5?۠m6w7M_Mdz4뺗Eo켗vw`۩մ1s|~Um3)>iM;4䍫hN6Pb:m4nJ̺%4ֺu!o\Et87NgnĨNjM!O?=uKhuBC޸&蔖:}iޛ[˯AȡWФw;{o$# n>ev&{rmjtohQ~WZfdeV/) ڝY-Ӱ\-rqD \ÐdVjAieSDiH&O7%s6hZ,jf:'P0w7.SspOjX@?>]?\Ü2#X3oGg?+O饾_뫠}5O:@ɖ:BfyHtCi-=Tg.-lhG~?wmuqAlTPbYMJ€Bi?Mf,nRD";OaS$Ju:KrƳ$Raw?~K "@!bRFsXɀm/s~Rq&b_]M Z./Be gaC#ܖe.4^e ZyL{ wO^@X=EPXh$n(zjP懶RЎ6UXv35Y1@:&5dU4Gb%R?&+ #ZzBfQ!j+8: 5NB-Toe[侓QJA[@t.baԴ)JB2'RܔLJТ.ҹR,d"mS"ؘJߝ0B*9ӥ`a mQќğHTH/-1dWI~ G: q}" F-`~Rb) *D*Eo-^*-ytĮ%Y-YCW 8H^fVn*^e*DR2 e%$(ڲelt?h4ӠR8YԚ)LʰOԫs<"H2eYnRMUb,ZZ0IV%2Vc޺|PC~ۯK-[]pROtv1/ b7˿m559D$B3=Q0.QFˈMv9)͕B ](m?*O3Ƣ.BZ80nR2˛|[v#oز&̧zuXXs`aԷ8Q~P2fZÁ0ۥAWs{Ȁ:.%GYqQVhE%P^\"9ʺA8O=:P9蹚5CŘ0+Wį;wWS;8h{½VONjjkrhDԂRG6Z*ٔ):+Ԋ&l;a!/%kyCWl<:D:<yURxc@6!)?S?nRA0 :M+C{@B9D0UtjDȻ)R)D٣wAtbQǻJCR֝wOnm C1qvx$GW>'YeEUx-+eo}"SGz|ooGl_lo5(cx*ق~,-fnt.g?zq>x6&enf6<(@{ J9sλ7e (Y)w MѽwFg nBطn1^-sօi4|9@U8m@[U'N|=% anLWR@1] pkb=?*?7yyft}sdTdDiJrYedB:#YAL) Y R_ M-Wd5SSay&f}_c ^bhpᐄjTK.˲|}rrylj?0k:<;\ƿ?+.PnZX}mu'grw;/e+^nGsx,?|ǞR\ JYQ9duUiX{-(HOO815.8I 9y ĩ4@N7Z&c97[))Mb0;Vvc0O6035Ci!acF)\R-`tWR+N G0^*n,%gڔZ0N(=jYvhoŹJ9 Ci!5Ka5|;K/ NCPz($h !tWR ӌ( K Q qi!57ՄJ>:]R9WR j GR*PJJ Ci!5ūF Z̄ZQ*hJE_4e BjA+'#JP R5̽R_mJAVOnjR vJ ڔZ ; LjRPJUJ>QOMQ"0|iXNRju(5(5c]UטG|e9E:~I6$LJvQ.+(Y*] L𓵻ܟR%M}>?\5/I7~q󑽻qP['Vnm1FaуQwI7Y2fm(Tug6 & 3V *F):=0fsROĿc2i-;f=]޽yc=~oSX#ORon6ɿkv3ů8/)+ޅfȔ cmbڂc6e$XDI"-qR+jI4VW*O^bA{W,ZQ4˒!sL`""rD8rȥpS2]'?ZhBaʿeXQu̡z^_[5EEF&vj2/W7'7{YYmC^Ɯ55!H)4ftQ)М+&QJIᑩyKBr;Re]nlҕR%#OR "l!NYu6KS*א$IL ȍ.) i. S:NG#l ; :)N;>֣lQjiR(嗟_?璏):58تzuL} @(ON>q& =r?sח8'*oaB˿)NԘc %l4w/'=►?j|%cI)75~6Y*;yޚ?| %*J櫯q:Kn}(0N1!p= ~An475,V悼N ݕDdB  ,1~U_lk[nO_}?sϗo_eU39r\]F[h+ϥ+AAda KPC2sA|o758#1nHKS%f}*JWW?_5lI|1H \KB/Zrr6@هG~N@ju&+;[)G?AEY+Γf L&K QJN1x}^|~ߒPs7+˭KKQk4ߕ2&^o= E3PKR5w)5>!Ơ<)|vy|_sޟpp3Q JݻGR=YL3 >xa gx5Zľw?nV{n:N;x>W1ҙw@3D AG _ʽ2Rz8:ѳn??}eÑ? uk/^|Ib~?!h'ݛ77; ~6'Έe< JCi!5?>v4eIa2R/=nB` El.Rk%Zh`3 +[u2N`ΐp-/DJIiKz@FK&sp_&|XR_Rvc8Q푬\dr)֎jSX44m=euՏQ,["`̲QE7J.l1I\1 Rt%g;=4QL0O,P^6& $5=hI9s zx'(*-;/;X'U\8׃ qH f^2Ζcl\Tr!ÿLm fm58.;=y1?S?e /3FJ"ODpjBJ8Ezx44"-DC%XT3c䆦X$,85N856a͙JI O~t.F1^C.}e"7)54@'0*]!SoaJd4Ӛg}8J$Z&tP) FC@p)f)cJe$,3"UR Nrp $^۞ !#Y`X0f94 TvA|[k;c4 A`<$FtFjz0AZ蘪lwg䕂tGź,J550Dy`` ^)EC6VJEk>%IUs500p*@d喤~@;& ZO2R 늙d7z"{"|SAC@{W3ەʹ,;hlRFwF)<1c;zX 'd A!di-1莐9 Y֝[qmִ󁍿Hx͚3hef]|%:޵E\KVYAnvZۨ Bucq.ϝC:DX;%0\\:ڌ]&㜲X}o,5.}ԑW1Axy9!!*!ɭ@J N I+DJ BД0R6a3I }0_࠮Iy.`17ݟ:BAg)d&T>dž&F pVL9,K38Y Rd I W"Q㫿@mʻgC9Z@3h|=nzܣwAtbQǻPF7vݢ;@3h|ػq9X@'u[Ƶ̻Ew4ԻgƘJ%[g ;m(bIj1[U,,ƑOܛ\ϝ[20{Vk-'.F?O8Xf:g7b~_#֧15&e.f6<(@{ J9sλ7yr}ldhe+A%Л{MW.f'/}y2g]fقi+q-0U>:~46q#鿢K-nT/ݗTKR`&EKNv*IļcWŒF4^__oզ;=mtbs[&lTjkt'v:,2X'pdD#CO-7$s!Ӆ0Onjo,;]L)4WnWqY +QzYeΔj äf]huYՇ8؜\RhvVtPdu]0)#EY96ךB4K"=!>lWQv2)WOhz9#D$+8!Dx:|_MB+mOP4(AѽqvaVڄ*X)l{dmA__b[*H UѾ{bb1,2 B)F1U9r[dJCBc|:0d"q>8@JlFd8S[lQWaTRה2̵&^<hZBwkZ0'sIXK&̮Υ}}wqWƮx K}HV0–tN >X=EHwiHGA J?NT)hO^2`D(vw&^3-+V3")m0u R>YdJH9O ΖjSKgjt?+B3^B-2j!(c`Hj|L'~yJt#&S~* }KY1(cTRM=Iڹ+ٰKFwU z}k=RԾj ac-視 vmB ` ![PJ-b4(ibdSn 9 ;"/dku*]7jS9R-yE0F=۵_W7꛶rsOQw)0Ls%Xv..9iHYWcp]JJ9彑@\༡t$ev2hFQۅHPCF/#iȶ2L-Fp5(napu!=ȸiD%*eֽEMY΁m.gd?Uyk^qܜL Tȇw }zn÷5XTi*H KIydGz,^8qN@z zJK5>XVJ<LMvt2-<듐f9c\|ym@3q(qѥTR@ qLa/@Իh(a4+G$7RBmZOLzx$hA8O־]>=zыmVD4XN9aЭK;u>8,e=ٺn0}׺n% Q/a< hV$U"K{'.CJ Yrì.=>j--['\{ uб}JY Ό0%]xEz9iX)Br [Bѻl )I?NP$Xkv݃t2I9˙dŷqzLwco~_:S+\e9!M*9?tWNK|6-~#Шp Ūs+FRTIϰ0}둝[xEg-;~b#Cɝ;¥$<%Lt ->Et'bmd N5p,\|zHn7܈Ze}0?5z?~_ؖO3 \lhUΑ\$ 5cCw$_uĒ %CQɗw(n s9jD|inךtuYnԻV'6}OwOj b)07ɎFF 1M(wmSЬRNȷ8L5#ʖڦ7mCFi;,dչ#R{]L 2nq-S![Im%4qStJ`Hڴ:qHfz/.^ CnQ|Knc_>իO]| mVܭ=ˋ]LXZӗ/Gszoּ] b]\׷O,?0kGԿE}jg-ʵK{Ow ?7l[ پ|mPh浕3j}E ˖u I7rs,Ty3FhcA edLdl;gKxNL>H>!B K<RgOIxFbnrgO~ \ #t7` |k1:U.WҢ^kԸ F{M<5Zz@na* Ԫ ]֍1Di=7BGsWѷZ77qPA$H#[o)&ڭ: XĻ<֖w~rR !Gw!9?_|}{zÛyGf~؎O"~>oUm\t១i//ph}ȑ{YK 7o#2l%Sż~*>%qϭw8EvxY fI,*s0͔,gզդ:{g"K%Trd< jjbPg4KTŋIqTaKV+ę'RxG/j BiXM|fI)Hsֲ{tfJjjÀtfI4r\̥/G}V_mZϋRJI.fOE)YJ=j\z,U:Jx jjItf)\Mai|nAz~6Vڞ{f4 R$8 Hcj%yY*eKr|,2 Ɉzi(WqDi,]X{OReX+Q(XLKV+Yj/!񹹖{{a;yO-ڰv${H)Smywd`sJ{:~QanZ(ЪGI~[c#^xmX`b;3i,;fz0Ymg>HtmǪ+ʕ ò'"Mw$)'(Kjċs7)D[=z <9YèJt9aGRsvʊ|9+%޹+ۂμψJμ&F :vJYܴ>D;Xk%~jSشY=V3kim%ֆ%FC Z_ΪV<;IhW p.arg ڐLb((O"P!MH޵uJqm-[LjkcHC{ \&^M# ;IHi&uKKHi$=%*CˍD(%$CG ~ZC(y*Fk-9FLH^X[CQ1WheWdm-uد 5TCaUKm*#i\"l!C [ŕcf\RpN@#vAvM@5s5mks3ڲ\ [(bdFEtV`j<\; ’)G Zͩ]դV뗯6F;y1J7T mN_OrVWO֮.~9F^xڄ =ƪ35.K3+Eni7lF.+\td!oDlRG0}wTl@A~wpou7ܽƗ/&mL#E]|wQ*qF2>HBۙkjN,J` -*Ё(4ZYEa(<|;jIOwD煠n*=ߠQ/$[ vMxP' }~h7;ׁ!i?H-Z3Ǔy#ѵ]ᶍ rAWALq>6CEZLX|ܡ>/\fLۜ H&K3+AN i6R@!= y@[( w> H't#C>h!pl%u.aD#ҙ TK\{ьLS+3Ϧ90 ' ++dfll٘f݌ C!c6H&~"cL+gZa"4N;Ij>jKR`>}EL/;dB hbo %k+Rp>NbayO*cc%$e!"b}r&|zlTjRpԄnJ>҇%5Q»Z(щ!S`-SusI\ǚٽ[wf~.2: AQjgAt"{db^vH.;}(r"nD&yǁGP;c(f@q݌4tP0|+[8>5iD|sX6H\r5<佫X"{@$cPc|U-q)hT5F$GP4BQy O57v?+vCm!h*.L~v?ݻwFPV KOo?Ŵ:~BU 2ΣvtL@d@ݰ4'2RD U#vh|U]7W 7o ]6Pv͈M(4i_\^jG D>q7JIcxqm4R[ FF?RC{瀼Đ~ /p>+'MDva!oDlJѯMk=Aa|7x^~B*~;Dz.,䍛hM)9wcFR11o4nCFZmylޱ-һua!oDk6YiecdlX1T~y@aB"[#^]ҬDH_,8cp~pneˏ]|7bp(usv}E9g~vkx_8?yet%,^tF }fu@M@K3^JmvnbH8SzˡB\w GB`|١i[F conۗerJIYh̀[֙gQD['ewY{n1:mQEi5e/1 ?-) 8GKl7kZ6_ ͳ@{r7'2 9Xטs_h3ӯOE b+10JFWɪO<M.}s.5#F܂@KxZxFKRզ~I:p)C[:A(2 y&bSHWm璉wK tRĻ ;Zͻw"n]X7$qֽ$~wK tRĻ EH2Ǒޭ y&ڦ"n⎞_y<>هF+l*wKuWΣ@:0 {_Kv)+>z-"705>?&3/]8ϩlTS^(]C# !Ț r6~־DE$RW1F!6; as{7W\5Q"?̡0)+LJ6rSYP:s`n 0VcSl^^]HzTW"(hpjQvU';U#'yR<s>I>U7CmGDYŻ]?n9rSW쟚AEz8m %R1 $JrV7B4橶i0ɯ!p}LJ.(f7s/aUؽ0?>u0`4~AN3JU(XNVe+%BX0ǝӖ̩sBe*F}K*X:`T%Eˏ$47omJHmVZfCJ{Hq3C#a3JX' JQh2CrԹN&#P)%&CAsyVd 7*/A)$-fqV9<%4κyVୟX-'WO˛3\.-+2g_=m朗=ٌt4hŧ7J+>\\&7o;#ÕK\y?+ٌlq>r ?S1!=3BkfM䦲_F69zly©v./ +!0wmEx_XٟPယ.G_ZC4];Ú.A^o֎a FweTq2#q~1nF) kITFo˵5%°Smu$.H%,Z\PYwMt\=xщKXDaLLt.hO1R!a뇨t:u.a2hܫsO;!KW?X;C;.0/a{% rd/a>-5]CcOn&%(Q$loc/k'%^ՌZMZ[HSmjdݴڰiht9!6 z?!QOttb}٘j1AY4B^9D0G&A#fX@'!mY) {f.!SR3cMc.>wث"=9T=:͚.aN {bWr#j5t/Q΀z|8ld?&BKX$c hDKhxiĂ㘦Xwca'xX#gw\]\\ow6&m?_؟ΚwgUޝUۻV:2&?;ڿ\'ǝ3`Ͼd#~,Xw6Z{Ȩ 2O4! v~>̈́zVc"۞uvm59ab&RR+)fl(=fi5YgJ5]PZ/dǭ^7F`Qzj(RNV@1&JYnZMPwQz(E\U.JMK+(=iC)[IդQz(5u] Ma\zuji=6JC).l$<.=mrRnVPSN?&:kAtݴ)T1@3U#k'9pD 1АZ!i߄=!l2o~ԫ1O:igȻ=Ai`5s3i2d?B3_Xi&-Z,B,0%eB)-FrpUǹW5??5zl@aL ,4g T2+u+ndur&c$:O|=}yۘς^TV#ԖS}c7ރ+1bcѯ7(g:t c|pQ62$!-̹`Zd2(T/Rs()‡v~zT<}Қ8фLҚFʚvfC5;nZ7D&-V5B~ʚQqq tt\IXوE~š(257vټS_kCc=7U U_<'pBkjfuSf*^Z4c8 ryvN6ŋ vh\뙠äVcDz>|yLhVV`2~ |BLk( ЏZ+yv>+ ý;!v\V#čmy; &A.֎Slw=Q$eE+bS^228=q}~-7k)̟k N| \pbÉa_ܯ {]<8xXsN=7Wft"QD*k;e҃x2t|py/A ['v4$W>J@ZEOc.kdQ:6+.H+4EXꢏjZijˁ}a'9,{ݱ^_~L/{l<T$#;' j-}wp>Iu+fG ;"o,_bi&sNJ%7h8'X^~HПiEhu"4&(KeF K;,{{ƒ͘A oӐHF.Hc&hN&54\sɴo+951![rcMrNnVUٵE^%}5W[ξ9{Қ#˟QmTspgȻ9V#MtӢ'!dT'FrLix8~չvJR?XlT،"KgoP>ޤ&馶Eps_~Ú@m. [/W:_hGdwnf!ꙋۏE"~xk⊙cE@|O!u ˄[_4M@Se2MMRR'"( S!6[I/طBOro#l0&~.$،#eKw<;5(͛DZ-H ?{M7~{S3'܉.:))zvE%.P_N=yŕEx<݀ٿy7 >Ob+u%#.Mi1V!r.)vV7&V7m ]JYnXI6Q].Ub(=fi5rTRy]N)' Qz(eRT5g(ej4QʥJtOǬ^7&kғF)zNBI={j f4Jɳǧf<{FsSF)?g^]q ǭ^7F-qi ?rQK 'cVVS]%zF)oUoQN$oF)*?r{O$P꥓Y]gHKQ9>%"S\yZP"dIyx^0$%RSxeJsLQp%Cq*Gs;|qk>잣;niŭC6zqCY<u}-d,m5hٱ#3Ӈߋu Gu^%O]Vh, gK"U,\b`sr ^aσ|ns%dv?JQ! bYJHDTB嬄f\0ɀQQ뫯=$Ey]Xeo7{o{ٵ>^$_͋2zh-a4S+i rp _dF6$ Y&rIMIHt f6 484\kO7iD{vCWp/#i6K`Ӵ<$ D2 Ԧ\2l"fc]I*e=^J Њ6A{5w^~ XIE)UG {L{UlȇcϿ|MRAC`TvC7iHlO:zHC-% JL#j2aLDEAivtфC4Gȩ@)9`}*zw~\h<'Ǘ'!S'ߩdi`CF x,\l./f;sELfLG_D˖lӔR8޵FhZA1wo+jЪA2֭/J0b:cWDǂN3[q0c]<r:6D<:Avl2+7y1#j!٪ 2\*Cܓwﱗ:ՐYCҶ 9X"$쌬\7!v6^(lV7 [5DQY1 b(i>RNAUӅl_i."X7D! Z]䭸͌n6 TWrQ[$-%s Tw[$%Za֓[d8䣹EvOf-Qy?g(Nt?"#WY^zqv'' rA:ƿ8k2s^>;I.t'.ʅcoxЃFVhu`ɸɄTU\ej}ncmBJ݁& $)$;Soe6!dg pfahAvkWr}odkSٲ[6B^8Ds0&`8~#%-9C-{dv!/1őoc*nUYk 1-&|莦e{@oۏ[Don>6q }KQmo.BK?AS7`na\7淦;;0pHK_>67~_{Q?.mcxvg:;low77؂}S}7ʏMS-})uˇmOdY=sMRAt[{[jk mC&@u2yo7ͧ/>\̫9B8 Xa|]cm\,n<_aՄ}=f?xmwƆ!nTAq%:,I& M8 qԠ_4(CqI_*,R e& ng3#O !He۳דAuAo[Eޒ3t UKteUTpuc1_x5dHu!*/}Q:.Ks8#7.9,Bh[! ]EYB<`>qFrcýIb/3O]Y'0楤*1qYл,=IOƉHN6F cXJҢteRŕjה` vLu#b*[4X[!0ym+.a+8]T/d`qމ 2{GXy XmJm)Tan*mj0X.+ŸZHڣ>(҇aUoyFu;W2ƚ/X;Ɗ*ٸ`SEEmB@y>p#G7΢fޘӽ1֚gG8@dl֡*)ȔV%hY`ͶD eg,7zcQBdORBvsOۚ7r7 x=rDE u#+ޛ8FȭM#t>j8F] ыA+? _y;'qD"eFRŸ)EB`r;ƋG3%-8!'hEJA8nNH{iCFY!? (Sh!GӴuXeBF=%PG^jȗf&͡ ktx!X^. q#=]3BE0bm<^ᐏ=={1 /dRD]W0}HaKuFo>s shR(LB ,G9 [o}ymymThOs' 6\)HaHDH!$MC?jHʒC8j1$qK/&ky 6OE Dx ]}۷Z$A {K7R˪!#̯" @ hG! :Hn V_㾝-Dnc C4S6k7lq䐙>bF9roH}=rpaEbm2n9dCHBwOi\y\yN־\yr%BC$Edh^ -64 ^=.NހutujmQ baLRP ;S ,i79_ͽo꿛i~~(!bu{l]6ܝΉyW%ǣΘ$[qNZGpj9J8=qS; (ã ͙HUE幭*ZmHPU@U@.!x4KcVFȫgËf\C UpcmEC`98d @[s#JUZcGةNܧx|*쮊"cK> t'9Y(+WQҚ Mֱ_^*JjK^EǹU{ +G~n Mw(۟8Oo|Vُ(նx?hc.kQ϶>aiowCępIC W opC̫f?jnxvf.zʐf/eXGv/كv}Mؗ(6Ρ),sV[MңLPJ&hyPwV xEEԺ8Zq ?6O޽+[*J b:+{}a2r7HXz Ԡ K% wMnQKIvό! }le[gJ' qUbRϷ?-^WT孭%UjE '@w܇=¸h!MƸS3Mds$M{{>њ XB6u[(eӨmWZ(mhA]8Ex=>'s]sfPyM*mUSyޣ2lYj+4|.zjoܴnwGѺi/>ۯmٞil\wFW{}gn^oCx\fmmZMtFO>47 n< bG3:8" } صƹn34O U3 [}:qj=E=4rhFѨxRB9TF,,`@6cU%úFgoEscꮅ?[~%+:oO Ϣsׯ0%:gݾlz2D чt6⨄`~wV,|cN$F*^ͮO wHb2fT&ܡ[q;ci9/^>ZyP>nRie/vtYdpwV 2tJ|B81~jJq2@>:3ff D.~os}4]O<:N7ruMn'Vcy+ֿ`vA trݖEafn"y)%ȲqkdcpLkTW1u=g݁#uw*Pʑ<[\Qz(b1(mi 签۷~R8–EWR8nfPsEE4F\x5@9ݶR6q(fUMJVҋF)D*N3Tnjr(#*RR8nWqwVҸU}ح8^JVFo<R82tV![Iu(dU9{ڻ4nqJV\/>~C"^j/|k ?|漱wƪ3 bG:'|u+S9s!9'TGKdxC]M/9߻@?J~v52y"q9n=2 lz~MڨCZWZD铥9.^A2n$zx&ɴh9+lOMwOf"Q2`ЮeIγYB5}JLvSYK)#OՋS*H-'sgq@`40kr) [-,͌$w[Ʋ_|a5^SϣNb/+CEi=# 56N W[yQdfgWto5828*Pto5=+J/ D;#f LgZ"40^d͉ז'E+dՒl[( cUdՏ<iC0K_!3qKb̛ҵVA}(y&ӆVỴSa8-/EX=jmͭ}0>"JiRTJ[PBPɐ,Zfnn4Q|nIgFL7 WMCSI?x'4`-8u0QقERR /8$ZI@k6޶PYYyTfE+:{QY9X"r&8vΎs^*̱A}J?_'8ʗIP5BЈNrɂZ2FEgNɕy;yTZ[|EL+KrװsgF'v82G[@кB!t [@xЯlXf˾E"w`)k(&J;`S4%󩜕fG&Sfɴ4 )TPRd12aǰ"CAJHX.Uڬ*`Nszİ<jpH1S th3%o Hd ͥMigGOd NS2ud'tZͥ(h.W6m<@fҐ{4 v?km)Zrw>#5ԗv@}M; Dާ1jr[: (*SM}N5C~kRmDmc*I"\}\liW#Mݦڧ\ݤxZ99In=.T*0+;;o}߱<1t#nWGSQnn>P>Nak\Vw7 R6O_ƫ \,Pj\ d!_)~nU8w+6mKTU"Sygm MMQrygFW9mیY V5w*݆/D۔ЁjvBy^ZyW c+|]Nw/<+ҷQGaeRбhIN#*pF*5*n Q`l I}9 ̕4Dh9+% F6($[qK7ny/"WY$7>Uq}sqdJ(n~_OsɁⵖ@;U>R#*81 *"g?h[jTJ&EckY2$ ~snn1fdLwU^{ֹ1שIK뢢>*^o\T(TdsluL҇F8tac? *^oY+qK964L ʨrBu\ҶG uEx(0Bڒx΅2E5O ^~8Fs9o}EqbkҰ ao+o9׃x+WcLYOm@gcHGX:: < ]2zNn`F= C+JzKbnbRvTşΆjװM#KKddQ͑>ÏA%xoeO>Yeoo. a+!R.H?ԸA X1TiO"e?7Wz9'/֧.d/+Nx'6(fEu@}4a3|N}Udx%5"2Q*o)瘐PFV"H C-E \ta0 ) )t9>Hc%=%,)ak9y ե5+w_U7~YwA VF)1VitЇ~ԡ>Ns"W鉺N+};'?EBz8ZΒsG˽{j /R;5TDfJf,Z.y+Ri)bĨE24|(we)QBi;ekmF/7(}Z]ˇ_ ew?U04ko75dd%Qȵz6LұRk)VJ0x A?QSDzŅ|GF/ WF"lqioW8'Ǘ)T(4' iT8F}J{JR|dh=Ơ7r~x=p(QQjS,%I\#Ǧ,KG>~ iɃFmu!U]@x$AٳJ(1z9D.LUfi\v 䱰v]P2|`W.>R.5W:~P@͛ko+>R]:ωq-}Ĭ$y]c&|C#b0R`W|[6Yu>kl-p{6ɵAp[iTKn߻7ѦwpT3o)) B#Ѝp0lBcIP>;_%'G3h >AEsAEk *ao+Cxv` Þux1_o#Nva {ck^ôh˽v:SW?9njU+yUԩ&z@wsa!_)N~nj3aFW9m|Wi"±[RL6|&˦9s׻ڽ3z21h"xJʻornCXnmJ* ZbsyvB+D+k)d#8V2Of P'?Nȩo,jߗBG#jG81볱<|`pF*5*n Qk1Ù@iFH-g~H@5nj^n9İ2&on33t#{K{VI$.$笆٧Ģ*yáv" yس${G +NDZ+N4`[!'%( &ԗ%Px@[Z2IG-V+jKNAD.ɨ$YB۾dRm!k&;3n)d {[s2tQ[*Pw]X*|Uo|٪ywMR%U4 ݩ % o{^2zbvU^hTG)[;R0Y;K-~>JKk&Tʐ,ZatoLAtJu\L#8&1:*`uu}pMM _}]p+>{s*B.&L{/B U.voVWBNF㷛 0/ p؈A@uAo~76Y6F]@G| ;wWk7y~(1}?=Y:$Cp 4a3O#K#H30>iVʑ1٠35DP]L;(AzdRk!!=3mdaAbV+PR)-% ueT%id.0C3{Rn UUKĨӑG#B6@NGy` MF%hEOTR=H@-Al&80$DH'؎h%DBhPUӥ((fM錦[PZf QIh!RͨX9bݏ,ME=4L$uTIy01Т*^l)@pu.+ 4up4e iFD3gy'R\sHʏQwCf@І$(Y2B$}dtwTA^+B%NT7N$hG|9)R TiJI Y)LKit?qǁwLqI%mꝴ6!#i䚚6l O2ޖֆӑ7GO^-HA&d6IustZ9g 3 .U|c( 6hύ*xJL, /1;. \e1dIEeԃLjav?|Ү +(}I\sbHVNmUuyȅSkP]Rծ&VP!r݅6Mك uS'5S`[ ^1kY}тl gT)e9jмa6JiczPF (pyf/ZE D5(+Eoya PIMhd*fq[HV(~ݫԞPqk5k; Ǐ*/pͯWv-Q7`a0~Ǐ\ܕ,_vGo"{k#sOJn6ρkͻZ$]ye)(hUݹDèwoۙ;+lnDm^TزRJPJ?I5!L[twns-`{a~gd88:k4ǯF*KɖwnlaƊl%أTbbp~j|'&XY`5lz)1*C*2ڪD+NX؁RXդWԾ.9E+@3cT sP F( :ڕ,R aT~@=eDp^cQ˞u{c1oE~<¯NkX3)مF#ޫ^GLEF2єeH,d)S9ݣae]C }4T&J (p 53V+.#@ Syz,*t*s6|}M|T%rWnAٍB] ~ZU}%iKr8xU8sp˨ eգhj&pӁݓ2PܤAEyWBi[šIiĤcBT5*f11vAW@!~wm"QSƢ::)7*-_iN7T3s~=&S ղL7e2TpC˓haH>4+تf~2SRjH5sNĆ{`]?\ofKB.v-(xS*`.RΆ3to!qa0yd arvE8(SJ¤aҬP" ɣ2C41R{znׂ痹sy\>=s`hP"&4^ _fwzb,L nw8n!oha7zS,ݯ߲!DYYg6m?+,ꥹP;65ɠk΃4Qfc8D<~6.g3Bv;y]˶iGB> .$TZyvKnnGZ(5[.i\PLْedg@=s2aI#XowtP(N"#Z$oV%6qNΦW"Fii/,.qw>y6ûye/ٯWoW_p(]upuQb9T(@D7ΟF͍a "j ~8 #^35AC.IU~mé-)u!ASY!0*3V wmlBl;r]`TuåO$rZ]}D AujI4J[$YRU${8XR™'8#PITID6ЭL]~*W\]lF_I-8-9%U ;O:X΍NVk$\C.-9.Ik1I<`ݺ6v!mo)Z}: ˱RF]F-Shtۢ^PcRн Mncey_ࢼylb{&JSFձzcv7lMizo~:&Ov3oK@na|4j˟c NF].|_֒;wn(5\8;]Tu6؛**|o]XsڛuX3O.n|hr8D~?W?ܟn#2`Ÿ́|ۙhb-:ӏ]c1uyxJ,'գF6 [,dslk5qk7h/M&4krӓ|"ENui,m :vJ6[&*[h'MM6=vJ6ngJև|"&SRFQӒZ Vhk]>V\S{P  ukA :%֗^o6QNez/qJ][wέ MYp/䣚!htDfa4tu*kOŨKvȥWt,=NgW+qt!fUԒ|1N*$D&N.AHGXd!@)E9Q+ FM{yLY@rTq#xk vY!Mz@#'*KE0^4ptEбy^!(AQ0\xS*,nR92I6)&h⢂/7*$hH D. Y3* fO<-sTI4 &g 5z}]q ^D ɼ!*hyfT$!e@F+7wE'x^x?e},^N/7c L Swey# MH$ѧg]))s4 YШQR!H.'6: {6LޝˀCNݜH2KKw~kk֗B&QI$K-IFKdT#fױ+D|$Sաh*@2hBA"epA*PJ{kqR^ J9p TJ9gd| 6R8AZ& [a ƑB!ۨr /8+`"he$BwA(5\o 79sxc)$l/]Mhzxu]B2jpnfEׁN(zRZ( AeehniԉD'L Aw-cfDq#)CSK iZ[U+%A+Z=YOH& ]]Bb$}$2#6xA`jh b޻O]*Gf m06 42x'1h㽽5 T Ac2\O|ESiU\p|8\CR6f"iʪWY#BtdT83PX՛NzdPcCv+<Ɍf6 <+QǢ\Z\]ĦcWa>-Yj/"Er(%dF=(iin$!]Xsd9oWNnXJP(}}rȇq&.IE1QLT93P?C`J};`W71 '꓾h;.. %}E }\]j+ϏFHitW}P}2:-5sH&C߷Dz!R.P!\h ?8HkvE"'LQc`sT+e=Q:~='D Ѷ{R~Z 9Ċ\;\dD`ɛbCMȨT.N;'j6;T s*,[+6 rWY$1*;AK)G$y2qEiPsbvꮸJ3OeܔX-8Qp:C`,ɂ slGk7u%SweY?B(,g復1\U!@d.Ɉ֔I҇+Xf9J ]w'L?|RVs5n"DXH3-ufh'J7Z36 5li69 S26qqq'ڊSD-g4+k5I_&xXዟusPL5(I5)yѽl7)!-roTg@,Dilw֦Z)"96dZzjc=x 搸}$ǦX!+\{_:8g*hrT8 U}3c>WK<-z|L[7>Kp˞qEI% uFRq宮 Pl8*}@MH(Y5]Ys9+~H82̉iNLFN(YmٴZ*lɊIZUćDЦef5NzΊMBԛTil191"r8~g'OZ Q[VT}ɄC ^(;Lz4.3$2x|+։D־#Gevб&;0GZ;ìPM+ͤl4EAaudLvgͬ\eX2Sǂu ˺X)Z'hL}֪RN7π߮pWxV#|[]dV8$۞k> e]V_>:>f&|_5پEHN@ 1Q&Xprd맗ŏ]"‚$fTS6׿5[~)ppluFBKQ{_tdۉ")֊2\0rlVCT 5$@Ud3;i$xesrhqotoHavgO 9^Q!4(e(Mo)^wZ4}9%䱵[Q+ PD( ~]$[IWo;"~@٢ėcj>%ONJ^Q۾j$U6d"GSw1<;YU:! B.՛>، ƾbAEcXa mSTl\Km:=Y4&#ƟcPѣE;n'<[sVo>`Cͨi]Dωm( JfCV>CçOC WȈ>D}65NzGE.|IxAj/ZԷfŵ $L0 Nϧkm yْtZI< (~() 5ƀNd;Slo˗m~kJI>bB{kUTs +N&}$ۃ[S@APYH#AxS'6{s6BFlo]g"*[W.V 1r'$Xl?U V^kvf$ts٠L2+3]H,g0t,ε^z.+1J`q%L5s@7L܇lP%ߛIud5VqxH[I@QY+9lAsuA{ Q_͛QY1zZl(^.M=9[n~{N <_jz$ XIjj^I **SI=fLꁭv}B[=K=Û!bBKˑvZH.gנT ;:߷uY&Ax!؄XezaT` J̦z@!LݎDo {$BN%17uAhȚG!z$Ld27m I6732V9qRu~7C@_[Z0[e /oۅ/QCJb>AbB,PUQSa`muC\-fN Y}XӞZPl^ Jk(FMY4vZCm%︂k$qJ?9c/KԦ~ Y~q$~\'٨XtUQ#xٮrU_ RG8,Zmԯڙ/ZzZj8-ᄱ;l:d,J//· ʰHalQdI!IVnB~|{ÌSGuf@ςW.۩eUbk XKF D=T|'VR[ ߍ 88Ͻ5{Ai2\oV~|BFH*]~!$+0Ua4 ־g˒iX*|wev0 8u B,s/ oЦ+[7IqRk DSMp?pb'`fUaj:zU%,Co@ѕ"81),CmVxt/h af/Qݨ)08r5QARD×;ElA QRo'C:]^׫/IugԎٽF|2~^qQ֋O%P 9IĤUh:mSLE߹(㹅|[!\{pANAҟeWTnKw[>JvNliwr?0#C~ʂr$(aB亽Cn֤i#sWR[|/.l y),קe5&x' A\؊Sm~"A7A[;LX'ᚳ;hXS@=G=j!ۖYE.9}|3h5U+xTs6ZmLa@́O W#|/q9s2`4G"(vF@W[\\dyUcm)KEt:E䚃q/.m(HslD%A:Fnwd |"&t04yшN-M8/w g!uv_ԋfY;[4UL_Wk ѺFzVgoWe@=޾Zo$$aH?CxTWg6I޾*IQQɃS.MCvz7 sOw"4:~2k9פ~tGnaf̢6^o,i*{r(CAHv[ Ԩ"D\9)wc8,z8Zg7>R C%gK*kT2ц=.~zqoI׏[dKnljQeaN!?k\$ᳱZrA?^tr2hh$8 ɰ\Lx,ؑHi6`;vTD5L$kL垫VTIg z69\58rG6{qa9 ~9Gp`d#+J\-큖FV_FF4RK Z" ^_yzw_g.?/[7z8=Y`xT __ KtMOov;j`Z;u\|qbd&l?⥀3p7.|Y/צKCtU\.NoAG$h\$wD"'CG$J eϷ?b C&K1UIe8>x;D_*%$zC267q Aybkۨ"n[?^-㢤k )~kR?gιfEjQccxǀd,>hS_tZugM翩",fjy}%w]-^|Y_UmR[.]ϒtyh-_^}][s6+Sz:9ҸON*ukWSR.mUt;VipFjnɑ*C6n6>RгktwӋW.f1/NMSS"I.&O7Mm)]|wpǾM U9V]iwp_fa;eL~ng7Owq*DL `ܮ.8_Zhr`ugtz'^UnI+|ڑ<h=mv 9"] X/~)K'!gf1V/))W]^ϿA'𠄖.n?'ߝǓǭǫW)~r ?0)J4#nTemR;'KQ; e)[ [qJO\FH\Ar'M+W\jYEQ#YoŠ8[wqrqLJ ͝gqzA@BBc%D>~q Ɂ@)stv4l4s̈h8rۭ_&?jqV Wt=(.W_u*/g7W?{[g/?[G[m l-B.4Z;-[P+* gy%{= &l4׭"1@Gm) d ]|P1OVU(QJk\SR@ȁ#\QmM~,0b| _HxCߕuFՁꉻ_moո[SVa9;?{D3wWtqjqnߜzK&~z{Nev7_j%-YW4NR/OՔKbjO&qx [T:74SH3%$;,oF톡I#.Jpjٸ8ڄu; [1E߭wp@42!lX{ToN&Czp6ݡCI'zR6}&|k/.{ sL'cXn=-eG&ԻDe&jgߟj/oh`bZjN iqS}jqv+qgм&ak/깺ܨUL>a'Q㪎%mdg7_ʎ!}_ [CEΦDfԇ}8Ћ}ڿ0\qحY(*,ԇ#Co@<5:F s#ny#H&#$?bvvPG!~x揷.ʖFT \.f:zGO'/?]\/c',WqyWfhJ8p muwjnսkq4B[ptd̰Su#% ^crr~Q8LW?zfqgF.!~>90ـ2' !)1X8ݹ+'P:n:^ HĐ&/Vu8c5{ BNb ^1z7THvd!DlJg &1x\ bL')m˄fyHօ|]jA8{}JM'+! iTs캌bGڤI! fp j.&5qQRY9_ʰTi"NAhZ4E )~;pJg91*XR@Fq-G[)7xjv<^1zR5cx؍B hj\=U Lbd,Z txN8ЙL/3tFp r5P9\D+[8dMe9*5F)- >ނ(\tƄfOBŁOQ}_6@(%=F7H BTKVٲpܗNr,wP&Tъ+fVjp@QF(ΣpG""%h%VX6Q@ +xyKWRj Qj+JZXA9>( U$*lQFW-{=Z{έHGB i2ĥ z~_TDRzЬ,XBqU%5aFEml*((dE-H1_y9P9<m` I߻}4i G6FL4,SSM3;ܕHK@pi!T!. BIZ|ͮcۯIbZegKm[-xq%V *nZ*giZ@4f|Cj(./ڳg0 790!1?|]X37,dw Ny-(|Fw, :Iz,䙛hM1z879A>#Ż F޼[Hօ.nMCa馇!ޕ;QJQ]vä8DPPJ Kk9 CTQK,ĖL{= ͺ%~B٩r_ԫ_kQ}u=?3ވ9V?bR,<stv~7mZ|CևWn>ץɗ#G}"y~U~ "/J;f Y)'m]_'cGѵ&Պ࿾RRC6R8MiNb?wzowC"CG)44&w$W)v/jw\4@(Iډʎs65N7m`A ,$f4'fL;=1#"U&j:W"ReM1#X-v-Vb.#B[.YHtp:l˒ \Yx! 53$ZtY%*=}^QyCELj!:LY?_Tޭ;z_ua997~$s4ӏ S"'UY\JbZxԲIVRJI¦JUTJ!?:QgDz ;V k6ۀF3B4IIG^Qm1h\9 ~'AF Cp]sLeDGA+r ]vTj>mj, c#-v8yu8jQ%v*Q#R&EFENT jxRu9[õ!GԂ+C מ'hZ>c?dm}2W'09q? rEWME wꦠ:ƬrV /-3Ԝҁ+>):qƪ1'q AGT8AWr3 C;0Ld˲Tޘ2%a)m5a :F=L+OF,$3e! C ZW\>JDɸS9DДҩjF eg=N->ֵŕDTD=srK9uQVVFdY)V1[ % ZiTpꁷmM z 4%c{M(#?!j> ))qMj&rbj4řkԥBJi(蓌6}M,dmPX !2(PP(bc[@ഷce< n"caS R1_LJ`JyywD32n3쟴:G_+=׸5`KuE?m&g*#i:.'U4p3DOtO9ZwчWn~F9L9FZi '!i0 yƂGJ _wJV$Bﲧa(SX%| )zû)`%лbc:HnF_b?_ݺgnY6%#M\bc:HnC".{n'"[MMz\APcF˸}1w+var/home/core/zuul-output/logs/kubelet.log0000644000000000000000006434477715136176335017735 0ustar rootrootJan 27 15:06:52 crc systemd[1]: Starting Kubernetes Kubelet... Jan 27 15:06:52 crc restorecon[4741]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Jan 27 15:06:52 crc restorecon[4741]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 27 15:06:52 crc restorecon[4741]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 27 15:06:52 crc restorecon[4741]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 27 15:06:52 crc restorecon[4741]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 27 15:06:52 crc restorecon[4741]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 27 15:06:52 crc restorecon[4741]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 27 15:06:52 crc restorecon[4741]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 27 15:06:52 crc restorecon[4741]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 27 15:06:52 crc restorecon[4741]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 27 15:06:52 crc restorecon[4741]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 27 15:06:52 crc restorecon[4741]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 27 15:06:52 crc restorecon[4741]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 27 15:06:52 crc restorecon[4741]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 27 15:06:52 crc restorecon[4741]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 27 15:06:52 crc restorecon[4741]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 27 15:06:52 crc restorecon[4741]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 27 15:06:52 crc restorecon[4741]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 27 15:06:52 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 27 15:06:52 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 27 15:06:52 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 27 15:06:52 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 27 15:06:52 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 27 15:06:53 crc restorecon[4741]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 27 15:06:53 crc restorecon[4741]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Jan 27 15:06:54 crc kubenswrapper[4772]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 27 15:06:54 crc kubenswrapper[4772]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Jan 27 15:06:54 crc kubenswrapper[4772]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 27 15:06:54 crc kubenswrapper[4772]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 27 15:06:54 crc kubenswrapper[4772]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jan 27 15:06:54 crc kubenswrapper[4772]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.430419 4772 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.434993 4772 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435018 4772 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435024 4772 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435032 4772 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435038 4772 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435044 4772 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435049 4772 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435055 4772 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435061 4772 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435066 4772 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435071 4772 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435078 4772 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435083 4772 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435089 4772 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435094 4772 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435099 4772 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435104 4772 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435110 4772 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435115 4772 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435120 4772 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435125 4772 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435130 4772 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435135 4772 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435141 4772 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435146 4772 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435151 4772 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435156 4772 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435161 4772 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435183 4772 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435189 4772 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435194 4772 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435199 4772 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435204 4772 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435212 4772 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435220 4772 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435226 4772 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435233 4772 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435239 4772 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435244 4772 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435250 4772 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435257 4772 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435262 4772 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435268 4772 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435273 4772 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435278 4772 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435285 4772 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435292 4772 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435300 4772 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435307 4772 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435314 4772 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435321 4772 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435327 4772 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435334 4772 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435341 4772 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435349 4772 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435358 4772 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435365 4772 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435377 4772 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435383 4772 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435388 4772 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435394 4772 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435401 4772 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435408 4772 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435413 4772 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435419 4772 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435425 4772 feature_gate.go:330] unrecognized feature gate: Example Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435430 4772 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435437 4772 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435442 4772 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435447 4772 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.435452 4772 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435591 4772 flags.go:64] FLAG: --address="0.0.0.0" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435604 4772 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435614 4772 flags.go:64] FLAG: --anonymous-auth="true" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435622 4772 flags.go:64] FLAG: --application-metrics-count-limit="100" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435630 4772 flags.go:64] FLAG: --authentication-token-webhook="false" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435636 4772 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435644 4772 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435652 4772 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435659 4772 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435665 4772 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435672 4772 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435682 4772 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435688 4772 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435694 4772 flags.go:64] FLAG: --cgroup-root="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435700 4772 flags.go:64] FLAG: --cgroups-per-qos="true" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435707 4772 flags.go:64] FLAG: --client-ca-file="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435714 4772 flags.go:64] FLAG: --cloud-config="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435720 4772 flags.go:64] FLAG: --cloud-provider="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435726 4772 flags.go:64] FLAG: --cluster-dns="[]" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435734 4772 flags.go:64] FLAG: --cluster-domain="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435740 4772 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435746 4772 flags.go:64] FLAG: --config-dir="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435752 4772 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435786 4772 flags.go:64] FLAG: --container-log-max-files="5" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435796 4772 flags.go:64] FLAG: --container-log-max-size="10Mi" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435803 4772 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435809 4772 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435815 4772 flags.go:64] FLAG: --containerd-namespace="k8s.io" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435821 4772 flags.go:64] FLAG: --contention-profiling="false" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435827 4772 flags.go:64] FLAG: --cpu-cfs-quota="true" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435833 4772 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435840 4772 flags.go:64] FLAG: --cpu-manager-policy="none" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435847 4772 flags.go:64] FLAG: --cpu-manager-policy-options="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435866 4772 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435872 4772 flags.go:64] FLAG: --enable-controller-attach-detach="true" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435878 4772 flags.go:64] FLAG: --enable-debugging-handlers="true" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435884 4772 flags.go:64] FLAG: --enable-load-reader="false" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435890 4772 flags.go:64] FLAG: --enable-server="true" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435896 4772 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435904 4772 flags.go:64] FLAG: --event-burst="100" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435911 4772 flags.go:64] FLAG: --event-qps="50" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435917 4772 flags.go:64] FLAG: --event-storage-age-limit="default=0" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435924 4772 flags.go:64] FLAG: --event-storage-event-limit="default=0" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435930 4772 flags.go:64] FLAG: --eviction-hard="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435937 4772 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435943 4772 flags.go:64] FLAG: --eviction-minimum-reclaim="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435949 4772 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435955 4772 flags.go:64] FLAG: --eviction-soft="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435961 4772 flags.go:64] FLAG: --eviction-soft-grace-period="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435967 4772 flags.go:64] FLAG: --exit-on-lock-contention="false" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435973 4772 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435979 4772 flags.go:64] FLAG: --experimental-mounter-path="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435986 4772 flags.go:64] FLAG: --fail-cgroupv1="false" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435992 4772 flags.go:64] FLAG: --fail-swap-on="true" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.435998 4772 flags.go:64] FLAG: --feature-gates="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436008 4772 flags.go:64] FLAG: --file-check-frequency="20s" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436015 4772 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436021 4772 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436027 4772 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436034 4772 flags.go:64] FLAG: --healthz-port="10248" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436040 4772 flags.go:64] FLAG: --help="false" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436046 4772 flags.go:64] FLAG: --hostname-override="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436052 4772 flags.go:64] FLAG: --housekeeping-interval="10s" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436058 4772 flags.go:64] FLAG: --http-check-frequency="20s" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436064 4772 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436070 4772 flags.go:64] FLAG: --image-credential-provider-config="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436076 4772 flags.go:64] FLAG: --image-gc-high-threshold="85" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436082 4772 flags.go:64] FLAG: --image-gc-low-threshold="80" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436089 4772 flags.go:64] FLAG: --image-service-endpoint="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436095 4772 flags.go:64] FLAG: --kernel-memcg-notification="false" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436102 4772 flags.go:64] FLAG: --kube-api-burst="100" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436108 4772 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436114 4772 flags.go:64] FLAG: --kube-api-qps="50" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436121 4772 flags.go:64] FLAG: --kube-reserved="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436127 4772 flags.go:64] FLAG: --kube-reserved-cgroup="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436132 4772 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436139 4772 flags.go:64] FLAG: --kubelet-cgroups="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436144 4772 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436151 4772 flags.go:64] FLAG: --lock-file="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436156 4772 flags.go:64] FLAG: --log-cadvisor-usage="false" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436163 4772 flags.go:64] FLAG: --log-flush-frequency="5s" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436188 4772 flags.go:64] FLAG: --log-json-info-buffer-size="0" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436197 4772 flags.go:64] FLAG: --log-json-split-stream="false" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436204 4772 flags.go:64] FLAG: --log-text-info-buffer-size="0" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436210 4772 flags.go:64] FLAG: --log-text-split-stream="false" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436216 4772 flags.go:64] FLAG: --logging-format="text" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436222 4772 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436229 4772 flags.go:64] FLAG: --make-iptables-util-chains="true" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436235 4772 flags.go:64] FLAG: --manifest-url="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436241 4772 flags.go:64] FLAG: --manifest-url-header="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436249 4772 flags.go:64] FLAG: --max-housekeeping-interval="15s" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436255 4772 flags.go:64] FLAG: --max-open-files="1000000" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436263 4772 flags.go:64] FLAG: --max-pods="110" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436269 4772 flags.go:64] FLAG: --maximum-dead-containers="-1" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436275 4772 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436281 4772 flags.go:64] FLAG: --memory-manager-policy="None" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436287 4772 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436293 4772 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436299 4772 flags.go:64] FLAG: --node-ip="192.168.126.11" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436305 4772 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436319 4772 flags.go:64] FLAG: --node-status-max-images="50" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436327 4772 flags.go:64] FLAG: --node-status-update-frequency="10s" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436333 4772 flags.go:64] FLAG: --oom-score-adj="-999" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436339 4772 flags.go:64] FLAG: --pod-cidr="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436347 4772 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436355 4772 flags.go:64] FLAG: --pod-manifest-path="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436361 4772 flags.go:64] FLAG: --pod-max-pids="-1" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436367 4772 flags.go:64] FLAG: --pods-per-core="0" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436374 4772 flags.go:64] FLAG: --port="10250" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436380 4772 flags.go:64] FLAG: --protect-kernel-defaults="false" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436386 4772 flags.go:64] FLAG: --provider-id="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436392 4772 flags.go:64] FLAG: --qos-reserved="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436398 4772 flags.go:64] FLAG: --read-only-port="10255" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436406 4772 flags.go:64] FLAG: --register-node="true" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436414 4772 flags.go:64] FLAG: --register-schedulable="true" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436422 4772 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436435 4772 flags.go:64] FLAG: --registry-burst="10" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436474 4772 flags.go:64] FLAG: --registry-qps="5" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436482 4772 flags.go:64] FLAG: --reserved-cpus="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436490 4772 flags.go:64] FLAG: --reserved-memory="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436500 4772 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436508 4772 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436515 4772 flags.go:64] FLAG: --rotate-certificates="false" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436521 4772 flags.go:64] FLAG: --rotate-server-certificates="false" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436527 4772 flags.go:64] FLAG: --runonce="false" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436533 4772 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436539 4772 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436546 4772 flags.go:64] FLAG: --seccomp-default="false" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436552 4772 flags.go:64] FLAG: --serialize-image-pulls="true" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436558 4772 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436564 4772 flags.go:64] FLAG: --storage-driver-db="cadvisor" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436571 4772 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436577 4772 flags.go:64] FLAG: --storage-driver-password="root" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436583 4772 flags.go:64] FLAG: --storage-driver-secure="false" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436590 4772 flags.go:64] FLAG: --storage-driver-table="stats" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436596 4772 flags.go:64] FLAG: --storage-driver-user="root" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436602 4772 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436608 4772 flags.go:64] FLAG: --sync-frequency="1m0s" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436614 4772 flags.go:64] FLAG: --system-cgroups="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436620 4772 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436632 4772 flags.go:64] FLAG: --system-reserved-cgroup="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436638 4772 flags.go:64] FLAG: --tls-cert-file="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436644 4772 flags.go:64] FLAG: --tls-cipher-suites="[]" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436652 4772 flags.go:64] FLAG: --tls-min-version="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436658 4772 flags.go:64] FLAG: --tls-private-key-file="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436671 4772 flags.go:64] FLAG: --topology-manager-policy="none" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436678 4772 flags.go:64] FLAG: --topology-manager-policy-options="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436684 4772 flags.go:64] FLAG: --topology-manager-scope="container" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436690 4772 flags.go:64] FLAG: --v="2" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436698 4772 flags.go:64] FLAG: --version="false" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436706 4772 flags.go:64] FLAG: --vmodule="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436713 4772 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.436720 4772 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436856 4772 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436863 4772 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436869 4772 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436874 4772 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436879 4772 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436885 4772 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436890 4772 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436895 4772 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436901 4772 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436906 4772 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436913 4772 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436919 4772 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436925 4772 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436930 4772 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436936 4772 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436941 4772 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436946 4772 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436951 4772 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436957 4772 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436962 4772 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436967 4772 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436972 4772 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436978 4772 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436983 4772 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436991 4772 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.436996 4772 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437002 4772 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437007 4772 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437014 4772 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437020 4772 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437026 4772 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437031 4772 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437037 4772 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437042 4772 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437048 4772 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437053 4772 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437058 4772 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437063 4772 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437069 4772 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437076 4772 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437082 4772 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437089 4772 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437096 4772 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437102 4772 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437109 4772 feature_gate.go:330] unrecognized feature gate: Example Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437115 4772 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437121 4772 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437127 4772 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437134 4772 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437141 4772 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437148 4772 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437155 4772 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437161 4772 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437192 4772 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437200 4772 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437206 4772 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437216 4772 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437223 4772 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437229 4772 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437240 4772 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437249 4772 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437256 4772 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437263 4772 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437270 4772 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437277 4772 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437285 4772 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437291 4772 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437297 4772 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437304 4772 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437313 4772 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.437321 4772 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.437341 4772 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.446432 4772 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.446477 4772 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446585 4772 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446602 4772 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446608 4772 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446614 4772 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446622 4772 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446630 4772 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446636 4772 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446642 4772 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446647 4772 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446653 4772 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446659 4772 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446665 4772 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446671 4772 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446676 4772 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446683 4772 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446688 4772 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446694 4772 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446699 4772 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446705 4772 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446710 4772 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446715 4772 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446720 4772 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446724 4772 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446729 4772 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446734 4772 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446740 4772 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446745 4772 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446750 4772 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446754 4772 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446759 4772 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446764 4772 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446769 4772 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446774 4772 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446780 4772 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446786 4772 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446791 4772 feature_gate.go:330] unrecognized feature gate: Example Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446796 4772 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446800 4772 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446805 4772 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446810 4772 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446815 4772 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446820 4772 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446825 4772 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446830 4772 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446835 4772 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446839 4772 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446844 4772 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446849 4772 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446854 4772 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446860 4772 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446868 4772 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446874 4772 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446879 4772 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446884 4772 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446889 4772 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446894 4772 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446899 4772 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446904 4772 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446908 4772 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446913 4772 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446918 4772 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446923 4772 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446928 4772 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446933 4772 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446938 4772 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446944 4772 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446950 4772 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446956 4772 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446962 4772 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446968 4772 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.446974 4772 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.446984 4772 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447156 4772 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447183 4772 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447189 4772 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447195 4772 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447200 4772 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447206 4772 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447211 4772 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447216 4772 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447221 4772 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447226 4772 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447232 4772 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447238 4772 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447243 4772 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447247 4772 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447253 4772 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447258 4772 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447263 4772 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447268 4772 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447273 4772 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447278 4772 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447283 4772 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447288 4772 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447292 4772 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447297 4772 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447302 4772 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447307 4772 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447312 4772 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447317 4772 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447321 4772 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447326 4772 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447331 4772 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447336 4772 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447341 4772 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447347 4772 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447352 4772 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447357 4772 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447362 4772 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447366 4772 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447373 4772 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447379 4772 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447385 4772 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447390 4772 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447397 4772 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447403 4772 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447410 4772 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447415 4772 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447420 4772 feature_gate.go:330] unrecognized feature gate: Example Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447425 4772 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447429 4772 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447435 4772 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447440 4772 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447446 4772 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447451 4772 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447457 4772 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447463 4772 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447469 4772 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447474 4772 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447479 4772 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447485 4772 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447491 4772 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447497 4772 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447502 4772 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447507 4772 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447513 4772 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447518 4772 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447523 4772 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447528 4772 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447533 4772 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447538 4772 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447543 4772 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.447548 4772 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.447555 4772 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.447739 4772 server.go:940] "Client rotation is on, will bootstrap in background" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.454537 4772 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.454643 4772 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.456730 4772 server.go:997] "Starting client certificate rotation" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.456764 4772 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.458711 4772 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-06 18:17:37.914092538 +0000 UTC Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.458929 4772 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.490416 4772 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 27 15:06:54 crc kubenswrapper[4772]: E0127 15:06:54.491952 4772 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.129.56.134:6443: connect: connection refused" logger="UnhandledError" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.494057 4772 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.512803 4772 log.go:25] "Validated CRI v1 runtime API" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.547156 4772 log.go:25] "Validated CRI v1 image API" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.548915 4772 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.554033 4772 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-01-27-15-01-40-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.554089 4772 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.575445 4772 manager.go:217] Machine: {Timestamp:2026-01-27 15:06:54.573479047 +0000 UTC m=+0.554088165 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:3933c4f3-43c9-48b4-998d-ee6c7e3cb9de BootID:3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:ef:3f:1d Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:ef:3f:1d Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:d9:12:63 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:88:df:b0 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:b0:26:c3 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:9d:1f:a3 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:99:7f:77 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:0a:7b:72:df:55:0a Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:ae:cc:75:46:58:1e Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.575747 4772 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.575910 4772 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.577891 4772 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.578066 4772 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.578106 4772 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.578306 4772 topology_manager.go:138] "Creating topology manager with none policy" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.578318 4772 container_manager_linux.go:303] "Creating device plugin manager" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.579282 4772 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.579315 4772 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.579520 4772 state_mem.go:36] "Initialized new in-memory state store" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.579611 4772 server.go:1245] "Using root directory" path="/var/lib/kubelet" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.584005 4772 kubelet.go:418] "Attempting to sync node with API server" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.584026 4772 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.584046 4772 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.584059 4772 kubelet.go:324] "Adding apiserver pod source" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.584070 4772 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.588119 4772 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.589064 4772 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.590959 4772 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.592359 4772 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.134:6443: connect: connection refused Jan 27 15:06:54 crc kubenswrapper[4772]: E0127 15:06:54.592462 4772 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.134:6443: connect: connection refused" logger="UnhandledError" Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.592439 4772 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.134:6443: connect: connection refused Jan 27 15:06:54 crc kubenswrapper[4772]: E0127 15:06:54.592518 4772 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.134:6443: connect: connection refused" logger="UnhandledError" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.592906 4772 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.592935 4772 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.592945 4772 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.592954 4772 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.592967 4772 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.592974 4772 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.592981 4772 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.593004 4772 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.593019 4772 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.593026 4772 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.593038 4772 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.593472 4772 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.595875 4772 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.596376 4772 server.go:1280] "Started kubelet" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.596526 4772 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.596990 4772 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.134:6443: connect: connection refused Jan 27 15:06:54 crc systemd[1]: Started Kubernetes Kubelet. Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.599572 4772 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.600652 4772 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.602114 4772 server.go:460] "Adding debug handlers to kubelet server" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.603485 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.603776 4772 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.604090 4772 volume_manager.go:287] "The desired_state_of_world populator starts" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.604197 4772 volume_manager.go:289] "Starting Kubelet Volume Manager" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.603908 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 16:01:57.657047466 +0000 UTC Jan 27 15:06:54 crc kubenswrapper[4772]: E0127 15:06:54.604072 4772 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.134:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.188e9ee612af890f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-27 15:06:54.596344079 +0000 UTC m=+0.576953177,LastTimestamp:2026-01-27 15:06:54.596344079 +0000 UTC m=+0.576953177,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.609846 4772 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.610251 4772 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.134:6443: connect: connection refused Jan 27 15:06:54 crc kubenswrapper[4772]: E0127 15:06:54.610307 4772 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.134:6443: connect: connection refused" logger="UnhandledError" Jan 27 15:06:54 crc kubenswrapper[4772]: E0127 15:06:54.604145 4772 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.610526 4772 factory.go:55] Registering systemd factory Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.610549 4772 factory.go:221] Registration of the systemd container factory successfully Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.610773 4772 factory.go:153] Registering CRI-O factory Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.610787 4772 factory.go:221] Registration of the crio container factory successfully Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.610903 4772 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.610929 4772 factory.go:103] Registering Raw factory Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.610944 4772 manager.go:1196] Started watching for new ooms in manager Jan 27 15:06:54 crc kubenswrapper[4772]: E0127 15:06:54.610966 4772 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.134:6443: connect: connection refused" interval="200ms" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.611947 4772 manager.go:319] Starting recovery of all containers Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.621719 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.621819 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.621834 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.621847 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.621865 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.621882 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.621897 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.621912 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.621925 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.621937 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.621975 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.622007 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.622031 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.622046 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.622068 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.622080 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.622114 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.622137 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.622162 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.622256 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.622288 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.622306 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.622353 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.622384 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.622402 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.622455 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.622504 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.625437 4772 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.625633 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.625747 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.625860 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.625962 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.626064 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.626152 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.626303 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.626404 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.626487 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.626569 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.626647 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.626726 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.626804 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.626990 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.627098 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.627222 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.627314 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.627419 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.627532 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.627613 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.627734 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.627872 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.627963 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.628051 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.628130 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.628244 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.628336 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.628417 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.628497 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.628605 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.628692 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.628774 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.628888 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.629017 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.629146 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.629352 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.629450 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.629544 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.629663 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.629806 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.629906 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.629984 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.630067 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.630221 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.630345 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.630438 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.630520 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.630626 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.630769 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.630562 4772 manager.go:324] Recovery completed Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.630891 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631023 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631044 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631057 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631069 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631081 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631094 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631105 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631117 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631131 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631143 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631153 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631179 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631194 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631205 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631216 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631227 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631239 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631252 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631264 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631278 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631289 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631301 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631314 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631327 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631413 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631451 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631475 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631497 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.631980 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632022 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632040 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632054 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632069 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632083 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632096 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632108 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632122 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632136 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632149 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632178 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632194 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632207 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632219 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632231 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632243 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632255 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632267 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632279 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632291 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632304 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632317 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632332 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632345 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632358 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632370 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632382 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632396 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632408 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632421 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632434 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632449 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632460 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632472 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632486 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632498 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632511 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632549 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632562 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632576 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632590 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632606 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632618 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632633 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632647 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632660 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632673 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632686 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632699 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632715 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632730 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632743 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632756 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632768 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632781 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632795 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632807 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632823 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632837 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632851 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632865 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632877 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632889 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632902 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632915 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632928 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632942 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632954 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632969 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632982 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.632995 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633008 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633021 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633038 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633050 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633063 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633075 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633087 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633098 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633110 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633122 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633134 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633146 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633160 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633189 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633202 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633214 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633227 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633239 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633250 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633261 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633275 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633287 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633299 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633311 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633329 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633342 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633354 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633365 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633377 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633390 4772 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633403 4772 reconstruct.go:97] "Volume reconstruction finished" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.633412 4772 reconciler.go:26] "Reconciler: start to sync state" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.647626 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.652039 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.652083 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.652095 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.652979 4772 cpu_manager.go:225] "Starting CPU manager" policy="none" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.653002 4772 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.653069 4772 state_mem.go:36] "Initialized new in-memory state store" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.659380 4772 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.661651 4772 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.661687 4772 status_manager.go:217] "Starting to sync pod status with apiserver" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.661716 4772 kubelet.go:2335] "Starting kubelet main sync loop" Jan 27 15:06:54 crc kubenswrapper[4772]: E0127 15:06:54.661843 4772 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 27 15:06:54 crc kubenswrapper[4772]: W0127 15:06:54.664463 4772 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.134:6443: connect: connection refused Jan 27 15:06:54 crc kubenswrapper[4772]: E0127 15:06:54.664554 4772 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.134:6443: connect: connection refused" logger="UnhandledError" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.667827 4772 policy_none.go:49] "None policy: Start" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.669080 4772 memory_manager.go:170] "Starting memorymanager" policy="None" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.669122 4772 state_mem.go:35] "Initializing new in-memory state store" Jan 27 15:06:54 crc kubenswrapper[4772]: E0127 15:06:54.711398 4772 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.721388 4772 manager.go:334] "Starting Device Plugin manager" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.721634 4772 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.721657 4772 server.go:79] "Starting device plugin registration server" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.722204 4772 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.722225 4772 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.722436 4772 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.722532 4772 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.722551 4772 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 27 15:06:54 crc kubenswrapper[4772]: E0127 15:06:54.729555 4772 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.762620 4772 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.762696 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.763713 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.763745 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.763757 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.763917 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.764180 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.764226 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.764466 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.764488 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.764499 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.764598 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.764758 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.764776 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.765137 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.765160 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.765180 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.765188 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.765194 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.765222 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.765232 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.765240 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.765197 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.765339 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.765511 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.765534 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.765933 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.765953 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.765960 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.766025 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.766102 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.766121 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.766133 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.766364 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.766383 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.766581 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.766613 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.766626 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.766774 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.766803 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.767294 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.767323 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.767331 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.767388 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.767403 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.767413 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:54 crc kubenswrapper[4772]: E0127 15:06:54.811695 4772 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.134:6443: connect: connection refused" interval="400ms" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.822318 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.823470 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.823513 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.823528 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.823585 4772 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 27 15:06:54 crc kubenswrapper[4772]: E0127 15:06:54.824099 4772 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.134:6443: connect: connection refused" node="crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.835956 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.835997 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.836023 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.836043 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.836063 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.836082 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.836106 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.836148 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.836197 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.836249 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.836284 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.836312 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.836343 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.836370 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.836396 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.937740 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.937808 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.937841 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.937885 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.937932 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.937976 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.938015 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.938017 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.938061 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.938022 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.938089 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.938147 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.938163 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.937955 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.938251 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.938152 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.938350 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.938384 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.938414 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.938424 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.938448 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.938482 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.938503 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.938547 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.938576 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.938579 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.938619 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.938676 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.938641 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: I0127 15:06:54.938832 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 27 15:06:54 crc kubenswrapper[4772]: E0127 15:06:54.951198 4772 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.134:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.188e9ee612af890f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-27 15:06:54.596344079 +0000 UTC m=+0.576953177,LastTimestamp:2026-01-27 15:06:54.596344079 +0000 UTC m=+0.576953177,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 27 15:06:55 crc kubenswrapper[4772]: I0127 15:06:55.025068 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:55 crc kubenswrapper[4772]: I0127 15:06:55.026548 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:55 crc kubenswrapper[4772]: I0127 15:06:55.026584 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:55 crc kubenswrapper[4772]: I0127 15:06:55.026596 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:55 crc kubenswrapper[4772]: I0127 15:06:55.026623 4772 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 27 15:06:55 crc kubenswrapper[4772]: E0127 15:06:55.027055 4772 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.134:6443: connect: connection refused" node="crc" Jan 27 15:06:55 crc kubenswrapper[4772]: I0127 15:06:55.100395 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 27 15:06:55 crc kubenswrapper[4772]: I0127 15:06:55.116439 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:06:55 crc kubenswrapper[4772]: I0127 15:06:55.123796 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 27 15:06:55 crc kubenswrapper[4772]: I0127 15:06:55.130364 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 27 15:06:55 crc kubenswrapper[4772]: I0127 15:06:55.131923 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 27 15:06:55 crc kubenswrapper[4772]: W0127 15:06:55.156253 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-b7446aa0404e33f408195fbb0a5d7000258cc2dcccf533b87b03def949c8451a WatchSource:0}: Error finding container b7446aa0404e33f408195fbb0a5d7000258cc2dcccf533b87b03def949c8451a: Status 404 returned error can't find the container with id b7446aa0404e33f408195fbb0a5d7000258cc2dcccf533b87b03def949c8451a Jan 27 15:06:55 crc kubenswrapper[4772]: W0127 15:06:55.157853 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-07214d993250107ab3c355de4cbedb841547f07ba78eb819ae7a8cf2f85d3254 WatchSource:0}: Error finding container 07214d993250107ab3c355de4cbedb841547f07ba78eb819ae7a8cf2f85d3254: Status 404 returned error can't find the container with id 07214d993250107ab3c355de4cbedb841547f07ba78eb819ae7a8cf2f85d3254 Jan 27 15:06:55 crc kubenswrapper[4772]: W0127 15:06:55.162636 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-fe47d7b7a8f551a26eb5563986df0792e2e7780fe682fc344ae48aad69d582e6 WatchSource:0}: Error finding container fe47d7b7a8f551a26eb5563986df0792e2e7780fe682fc344ae48aad69d582e6: Status 404 returned error can't find the container with id fe47d7b7a8f551a26eb5563986df0792e2e7780fe682fc344ae48aad69d582e6 Jan 27 15:06:55 crc kubenswrapper[4772]: W0127 15:06:55.163555 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-b64d7a630d80b6360b1f1a17d498dd9dfef664e1e946d3bb2d0127cea3dc9104 WatchSource:0}: Error finding container b64d7a630d80b6360b1f1a17d498dd9dfef664e1e946d3bb2d0127cea3dc9104: Status 404 returned error can't find the container with id b64d7a630d80b6360b1f1a17d498dd9dfef664e1e946d3bb2d0127cea3dc9104 Jan 27 15:06:55 crc kubenswrapper[4772]: W0127 15:06:55.165918 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-86d5acd4da3e8815a6f066738d8ae875fe76d45ca3c619c2319b08ea2150cd45 WatchSource:0}: Error finding container 86d5acd4da3e8815a6f066738d8ae875fe76d45ca3c619c2319b08ea2150cd45: Status 404 returned error can't find the container with id 86d5acd4da3e8815a6f066738d8ae875fe76d45ca3c619c2319b08ea2150cd45 Jan 27 15:06:55 crc kubenswrapper[4772]: E0127 15:06:55.212529 4772 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.134:6443: connect: connection refused" interval="800ms" Jan 27 15:06:55 crc kubenswrapper[4772]: I0127 15:06:55.427755 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:55 crc kubenswrapper[4772]: I0127 15:06:55.428922 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:55 crc kubenswrapper[4772]: I0127 15:06:55.428959 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:55 crc kubenswrapper[4772]: I0127 15:06:55.428971 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:55 crc kubenswrapper[4772]: I0127 15:06:55.429002 4772 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 27 15:06:55 crc kubenswrapper[4772]: E0127 15:06:55.429597 4772 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.134:6443: connect: connection refused" node="crc" Jan 27 15:06:55 crc kubenswrapper[4772]: W0127 15:06:55.557865 4772 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.134:6443: connect: connection refused Jan 27 15:06:55 crc kubenswrapper[4772]: E0127 15:06:55.557958 4772 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.134:6443: connect: connection refused" logger="UnhandledError" Jan 27 15:06:55 crc kubenswrapper[4772]: I0127 15:06:55.598001 4772 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.134:6443: connect: connection refused Jan 27 15:06:55 crc kubenswrapper[4772]: I0127 15:06:55.610133 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 03:10:21.574310172 +0000 UTC Jan 27 15:06:55 crc kubenswrapper[4772]: W0127 15:06:55.635797 4772 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.134:6443: connect: connection refused Jan 27 15:06:55 crc kubenswrapper[4772]: E0127 15:06:55.635892 4772 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.134:6443: connect: connection refused" logger="UnhandledError" Jan 27 15:06:55 crc kubenswrapper[4772]: I0127 15:06:55.670266 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"86d5acd4da3e8815a6f066738d8ae875fe76d45ca3c619c2319b08ea2150cd45"} Jan 27 15:06:55 crc kubenswrapper[4772]: I0127 15:06:55.671471 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fe47d7b7a8f551a26eb5563986df0792e2e7780fe682fc344ae48aad69d582e6"} Jan 27 15:06:55 crc kubenswrapper[4772]: I0127 15:06:55.672209 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"07214d993250107ab3c355de4cbedb841547f07ba78eb819ae7a8cf2f85d3254"} Jan 27 15:06:55 crc kubenswrapper[4772]: I0127 15:06:55.673077 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b7446aa0404e33f408195fbb0a5d7000258cc2dcccf533b87b03def949c8451a"} Jan 27 15:06:55 crc kubenswrapper[4772]: I0127 15:06:55.673758 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b64d7a630d80b6360b1f1a17d498dd9dfef664e1e946d3bb2d0127cea3dc9104"} Jan 27 15:06:55 crc kubenswrapper[4772]: W0127 15:06:55.852694 4772 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.134:6443: connect: connection refused Jan 27 15:06:55 crc kubenswrapper[4772]: E0127 15:06:55.852763 4772 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.134:6443: connect: connection refused" logger="UnhandledError" Jan 27 15:06:56 crc kubenswrapper[4772]: E0127 15:06:56.014754 4772 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.134:6443: connect: connection refused" interval="1.6s" Jan 27 15:06:56 crc kubenswrapper[4772]: W0127 15:06:56.183768 4772 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.134:6443: connect: connection refused Jan 27 15:06:56 crc kubenswrapper[4772]: E0127 15:06:56.183858 4772 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.134:6443: connect: connection refused" logger="UnhandledError" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.229714 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.231866 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.231907 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.231918 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.231943 4772 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 27 15:06:56 crc kubenswrapper[4772]: E0127 15:06:56.233104 4772 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.134:6443: connect: connection refused" node="crc" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.599030 4772 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.134:6443: connect: connection refused Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.611249 4772 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.611269 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 17:14:09.226128005 +0000 UTC Jan 27 15:06:56 crc kubenswrapper[4772]: E0127 15:06:56.612549 4772 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.129.56.134:6443: connect: connection refused" logger="UnhandledError" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.678293 4772 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="64098f88e61c989a4f2048d222906eacf8c0525f26e109913c1718c9dfb67d20" exitCode=0 Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.678389 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"64098f88e61c989a4f2048d222906eacf8c0525f26e109913c1718c9dfb67d20"} Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.678413 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.679765 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.679828 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.679846 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.680912 4772 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="8a8432f1dcc97ca9d30249542dfebc79f098859c0f7e04a637be764939fb6072" exitCode=0 Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.681021 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"8a8432f1dcc97ca9d30249542dfebc79f098859c0f7e04a637be764939fb6072"} Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.681180 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.682236 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.682277 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.682292 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.684229 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba"} Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.684268 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.684269 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328"} Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.684386 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285"} Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.684399 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c"} Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.685286 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.685321 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.685333 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.688404 4772 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5" exitCode=0 Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.688449 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5"} Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.688637 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.689806 4772 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9" exitCode=0 Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.689841 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9"} Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.689956 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.690064 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.690091 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.690105 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.691011 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.691032 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.691043 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.691765 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.692613 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.692645 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:56 crc kubenswrapper[4772]: I0127 15:06:56.692688 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.023302 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.064507 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 27 15:06:57 crc kubenswrapper[4772]: W0127 15:06:57.347344 4772 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.134:6443: connect: connection refused Jan 27 15:06:57 crc kubenswrapper[4772]: E0127 15:06:57.347431 4772 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.134:6443: connect: connection refused" logger="UnhandledError" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.482258 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.598434 4772 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.134:6443: connect: connection refused Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.611496 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 00:54:04.935516693 +0000 UTC Jan 27 15:06:57 crc kubenswrapper[4772]: E0127 15:06:57.615220 4772 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.134:6443: connect: connection refused" interval="3.2s" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.694980 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"109762c1c77a786a3c953ead1b65a17e9401f8ba205858ffc79f6e188d5005df"} Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.695094 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.696473 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.696511 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.696525 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.700804 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"21e07984f4ca6d9ea37d8213eab7f36c2a5342806beacdb07b50a15ffba13ded"} Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.700861 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a"} Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.700876 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d"} Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.700891 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59"} Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.700902 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b"} Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.700906 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.702469 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.702522 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.702540 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.704934 4772 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75" exitCode=0 Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.705037 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75"} Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.705053 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.706569 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.706602 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.706616 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.709681 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e8c8aeb62c273e8cb31f01742805bd40c4184821c4bbd94f319ff5f074265d75"} Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.709738 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3be141392fbb6ae8ce51085cfc186745bb5ac272899e3202aaed2b7093855c4f"} Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.709756 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f5b9360189745deea4c8950c4f3b73762c7cb098452c1f31c3df9ab99fc31ac7"} Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.709757 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.709817 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.711390 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.711427 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.711442 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.712262 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.712295 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.712312 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.735788 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.833819 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.834844 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.834876 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.834885 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:57 crc kubenswrapper[4772]: I0127 15:06:57.834904 4772 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 27 15:06:57 crc kubenswrapper[4772]: E0127 15:06:57.835282 4772 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.134:6443: connect: connection refused" node="crc" Jan 27 15:06:58 crc kubenswrapper[4772]: W0127 15:06:58.059626 4772 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.134:6443: connect: connection refused Jan 27 15:06:58 crc kubenswrapper[4772]: E0127 15:06:58.059713 4772 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.134:6443: connect: connection refused" logger="UnhandledError" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.263344 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.525256 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.525341 4772 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.525372 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": dial tcp 192.168.126.11:6443: connect: connection refused" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.598626 4772 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.134:6443: connect: connection refused Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.612446 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 17:53:02.965827067 +0000 UTC Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.716395 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.718558 4772 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="21e07984f4ca6d9ea37d8213eab7f36c2a5342806beacdb07b50a15ffba13ded" exitCode=255 Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.718642 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"21e07984f4ca6d9ea37d8213eab7f36c2a5342806beacdb07b50a15ffba13ded"} Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.718654 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.719596 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.719626 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.719637 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.720214 4772 scope.go:117] "RemoveContainer" containerID="21e07984f4ca6d9ea37d8213eab7f36c2a5342806beacdb07b50a15ffba13ded" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.721068 4772 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970" exitCode=0 Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.721154 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.721646 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.721980 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970"} Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.722020 4772 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.722064 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.722067 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.723039 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.723067 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.723079 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.723078 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.723113 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.723129 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.723376 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.723405 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.723438 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.723780 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.723802 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:58 crc kubenswrapper[4772]: I0127 15:06:58.723813 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:58 crc kubenswrapper[4772]: W0127 15:06:58.741557 4772 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.134:6443: connect: connection refused Jan 27 15:06:58 crc kubenswrapper[4772]: E0127 15:06:58.741632 4772 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.134:6443: connect: connection refused" logger="UnhandledError" Jan 27 15:06:58 crc kubenswrapper[4772]: W0127 15:06:58.831317 4772 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.134:6443: connect: connection refused Jan 27 15:06:58 crc kubenswrapper[4772]: E0127 15:06:58.831428 4772 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.134:6443: connect: connection refused" logger="UnhandledError" Jan 27 15:06:59 crc kubenswrapper[4772]: I0127 15:06:59.612710 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 17:11:10.198188402 +0000 UTC Jan 27 15:06:59 crc kubenswrapper[4772]: I0127 15:06:59.727980 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 27 15:06:59 crc kubenswrapper[4772]: I0127 15:06:59.730551 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67"} Jan 27 15:06:59 crc kubenswrapper[4772]: I0127 15:06:59.730598 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:59 crc kubenswrapper[4772]: I0127 15:06:59.732362 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:59 crc kubenswrapper[4772]: I0127 15:06:59.732413 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:59 crc kubenswrapper[4772]: I0127 15:06:59.732435 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:06:59 crc kubenswrapper[4772]: I0127 15:06:59.736015 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1"} Jan 27 15:06:59 crc kubenswrapper[4772]: I0127 15:06:59.736060 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c"} Jan 27 15:06:59 crc kubenswrapper[4772]: I0127 15:06:59.736079 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10"} Jan 27 15:06:59 crc kubenswrapper[4772]: I0127 15:06:59.736088 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:06:59 crc kubenswrapper[4772]: I0127 15:06:59.736090 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326"} Jan 27 15:06:59 crc kubenswrapper[4772]: I0127 15:06:59.737202 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:06:59 crc kubenswrapper[4772]: I0127 15:06:59.737231 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:06:59 crc kubenswrapper[4772]: I0127 15:06:59.737240 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:00 crc kubenswrapper[4772]: I0127 15:07:00.613629 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 03:03:39.705395721 +0000 UTC Jan 27 15:07:00 crc kubenswrapper[4772]: I0127 15:07:00.743488 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5"} Jan 27 15:07:00 crc kubenswrapper[4772]: I0127 15:07:00.743566 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:07:00 crc kubenswrapper[4772]: I0127 15:07:00.743601 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:07:00 crc kubenswrapper[4772]: I0127 15:07:00.743637 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:07:00 crc kubenswrapper[4772]: I0127 15:07:00.745497 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:00 crc kubenswrapper[4772]: I0127 15:07:00.745515 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:00 crc kubenswrapper[4772]: I0127 15:07:00.745548 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:00 crc kubenswrapper[4772]: I0127 15:07:00.745550 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:00 crc kubenswrapper[4772]: I0127 15:07:00.745569 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:00 crc kubenswrapper[4772]: I0127 15:07:00.745572 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:00 crc kubenswrapper[4772]: I0127 15:07:00.931669 4772 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 27 15:07:01 crc kubenswrapper[4772]: I0127 15:07:01.035445 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:07:01 crc kubenswrapper[4772]: I0127 15:07:01.036922 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:01 crc kubenswrapper[4772]: I0127 15:07:01.036953 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:01 crc kubenswrapper[4772]: I0127 15:07:01.036962 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:01 crc kubenswrapper[4772]: I0127 15:07:01.037006 4772 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 27 15:07:01 crc kubenswrapper[4772]: I0127 15:07:01.287129 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:07:01 crc kubenswrapper[4772]: I0127 15:07:01.615387 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 14:19:20.540339678 +0000 UTC Jan 27 15:07:01 crc kubenswrapper[4772]: I0127 15:07:01.745624 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:07:01 crc kubenswrapper[4772]: I0127 15:07:01.745624 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:07:01 crc kubenswrapper[4772]: I0127 15:07:01.746903 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:01 crc kubenswrapper[4772]: I0127 15:07:01.746920 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:01 crc kubenswrapper[4772]: I0127 15:07:01.746944 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:01 crc kubenswrapper[4772]: I0127 15:07:01.746950 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:01 crc kubenswrapper[4772]: I0127 15:07:01.746960 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:01 crc kubenswrapper[4772]: I0127 15:07:01.746963 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:02 crc kubenswrapper[4772]: I0127 15:07:02.616135 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 21:17:13.608597062 +0000 UTC Jan 27 15:07:02 crc kubenswrapper[4772]: I0127 15:07:02.672573 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Jan 27 15:07:02 crc kubenswrapper[4772]: I0127 15:07:02.747786 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:07:02 crc kubenswrapper[4772]: I0127 15:07:02.747893 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:07:02 crc kubenswrapper[4772]: I0127 15:07:02.748967 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:02 crc kubenswrapper[4772]: I0127 15:07:02.749085 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:02 crc kubenswrapper[4772]: I0127 15:07:02.749161 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:02 crc kubenswrapper[4772]: I0127 15:07:02.749340 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:02 crc kubenswrapper[4772]: I0127 15:07:02.749399 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:02 crc kubenswrapper[4772]: I0127 15:07:02.749414 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:03 crc kubenswrapper[4772]: I0127 15:07:03.524906 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 27 15:07:03 crc kubenswrapper[4772]: I0127 15:07:03.525049 4772 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 27 15:07:03 crc kubenswrapper[4772]: I0127 15:07:03.525088 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:07:03 crc kubenswrapper[4772]: I0127 15:07:03.526332 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:03 crc kubenswrapper[4772]: I0127 15:07:03.526395 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:03 crc kubenswrapper[4772]: I0127 15:07:03.526421 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:03 crc kubenswrapper[4772]: I0127 15:07:03.616722 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 16:11:23.891749624 +0000 UTC Jan 27 15:07:04 crc kubenswrapper[4772]: I0127 15:07:04.542203 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Jan 27 15:07:04 crc kubenswrapper[4772]: I0127 15:07:04.542596 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:07:04 crc kubenswrapper[4772]: I0127 15:07:04.544638 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:04 crc kubenswrapper[4772]: I0127 15:07:04.544677 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:04 crc kubenswrapper[4772]: I0127 15:07:04.544687 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:04 crc kubenswrapper[4772]: I0127 15:07:04.617272 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 08:37:21.852171994 +0000 UTC Jan 27 15:07:04 crc kubenswrapper[4772]: E0127 15:07:04.729714 4772 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 27 15:07:04 crc kubenswrapper[4772]: I0127 15:07:04.730858 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 27 15:07:04 crc kubenswrapper[4772]: I0127 15:07:04.731006 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:07:04 crc kubenswrapper[4772]: I0127 15:07:04.731989 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:04 crc kubenswrapper[4772]: I0127 15:07:04.732033 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:04 crc kubenswrapper[4772]: I0127 15:07:04.732048 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:05 crc kubenswrapper[4772]: I0127 15:07:05.617515 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 04:48:44.110438372 +0000 UTC Jan 27 15:07:06 crc kubenswrapper[4772]: I0127 15:07:06.525668 4772 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 27 15:07:06 crc kubenswrapper[4772]: I0127 15:07:06.525790 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 27 15:07:06 crc kubenswrapper[4772]: I0127 15:07:06.617963 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 09:01:26.846257702 +0000 UTC Jan 27 15:07:07 crc kubenswrapper[4772]: I0127 15:07:07.618722 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 12:03:43.652096101 +0000 UTC Jan 27 15:07:08 crc kubenswrapper[4772]: I0127 15:07:08.263914 4772 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 27 15:07:08 crc kubenswrapper[4772]: I0127 15:07:08.263994 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 27 15:07:08 crc kubenswrapper[4772]: I0127 15:07:08.619760 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 11:19:31.526392386 +0000 UTC Jan 27 15:07:09 crc kubenswrapper[4772]: I0127 15:07:09.459928 4772 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 27 15:07:09 crc kubenswrapper[4772]: I0127 15:07:09.459982 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 27 15:07:09 crc kubenswrapper[4772]: I0127 15:07:09.467624 4772 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 27 15:07:09 crc kubenswrapper[4772]: I0127 15:07:09.467695 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 27 15:07:09 crc kubenswrapper[4772]: I0127 15:07:09.620649 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 20:41:08.188648925 +0000 UTC Jan 27 15:07:10 crc kubenswrapper[4772]: I0127 15:07:10.621367 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 14:04:22.013766922 +0000 UTC Jan 27 15:07:11 crc kubenswrapper[4772]: I0127 15:07:11.622246 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 18:50:53.989294917 +0000 UTC Jan 27 15:07:12 crc kubenswrapper[4772]: I0127 15:07:12.623089 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 03:51:42.238454132 +0000 UTC Jan 27 15:07:13 crc kubenswrapper[4772]: I0127 15:07:13.531653 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:07:13 crc kubenswrapper[4772]: I0127 15:07:13.531810 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:07:13 crc kubenswrapper[4772]: I0127 15:07:13.533511 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:13 crc kubenswrapper[4772]: I0127 15:07:13.533559 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:13 crc kubenswrapper[4772]: I0127 15:07:13.533573 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:13 crc kubenswrapper[4772]: I0127 15:07:13.539030 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:07:13 crc kubenswrapper[4772]: I0127 15:07:13.623903 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 08:13:33.931890479 +0000 UTC Jan 27 15:07:13 crc kubenswrapper[4772]: I0127 15:07:13.779155 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:07:13 crc kubenswrapper[4772]: I0127 15:07:13.780812 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:13 crc kubenswrapper[4772]: I0127 15:07:13.780841 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:13 crc kubenswrapper[4772]: I0127 15:07:13.780850 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:14 crc kubenswrapper[4772]: E0127 15:07:14.440390 4772 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.442696 4772 trace.go:236] Trace[736418201]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Jan-2026 15:07:02.632) (total time: 11810ms): Jan 27 15:07:14 crc kubenswrapper[4772]: Trace[736418201]: ---"Objects listed" error: 11810ms (15:07:14.442) Jan 27 15:07:14 crc kubenswrapper[4772]: Trace[736418201]: [11.810231167s] [11.810231167s] END Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.442723 4772 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.443893 4772 trace.go:236] Trace[1971219875]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Jan-2026 15:07:03.607) (total time: 10836ms): Jan 27 15:07:14 crc kubenswrapper[4772]: Trace[1971219875]: ---"Objects listed" error: 10836ms (15:07:14.443) Jan 27 15:07:14 crc kubenswrapper[4772]: Trace[1971219875]: [10.836082889s] [10.836082889s] END Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.443914 4772 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.444382 4772 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.444458 4772 trace.go:236] Trace[261586842]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Jan-2026 15:07:04.320) (total time: 10124ms): Jan 27 15:07:14 crc kubenswrapper[4772]: Trace[261586842]: ---"Objects listed" error: 10124ms (15:07:14.444) Jan 27 15:07:14 crc kubenswrapper[4772]: Trace[261586842]: [10.124373696s] [10.124373696s] END Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.444472 4772 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.445257 4772 trace.go:236] Trace[1285288433]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Jan-2026 15:07:02.744) (total time: 11700ms): Jan 27 15:07:14 crc kubenswrapper[4772]: Trace[1285288433]: ---"Objects listed" error: 11700ms (15:07:14.445) Jan 27 15:07:14 crc kubenswrapper[4772]: Trace[1285288433]: [11.70081603s] [11.70081603s] END Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.445277 4772 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 27 15:07:14 crc kubenswrapper[4772]: E0127 15:07:14.446500 4772 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.466993 4772 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.482011 4772 csr.go:261] certificate signing request csr-2bhmf is approved, waiting to be issued Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.491646 4772 csr.go:257] certificate signing request csr-2bhmf is issued Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.495738 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.500489 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.501113 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.568231 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.579254 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.624843 4772 apiserver.go:52] "Watching apiserver" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.624852 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 20:28:14.708608036 +0000 UTC Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.628136 4772 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.628498 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.628799 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.628907 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.628929 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.628972 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.629097 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.629197 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:14 crc kubenswrapper[4772]: E0127 15:07:14.629207 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:07:14 crc kubenswrapper[4772]: E0127 15:07:14.629271 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:07:14 crc kubenswrapper[4772]: E0127 15:07:14.629478 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.632298 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.632447 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.633384 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.634114 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.634243 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.635638 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.635995 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.636054 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.640022 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.650218 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.666513 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.678118 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.693462 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.699999 4772 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:44474->192.168.126.11:17697: read: connection reset by peer" start-of-body= Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.700074 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:44474->192.168.126.11:17697: read: connection reset by peer" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.700362 4772 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.700389 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.709090 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.710646 4772 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.728901 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.744914 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745036 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745072 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745087 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745102 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745117 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745133 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745150 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745189 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745213 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745237 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745254 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745268 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745285 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745340 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745357 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745404 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745421 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745443 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745461 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745476 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745491 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745510 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745514 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745528 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745545 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745586 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745556 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745611 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745632 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745651 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745667 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745684 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745700 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745716 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745732 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745759 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745787 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745815 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745840 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745929 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745945 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745961 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745977 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745996 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746013 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746028 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746042 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746058 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746076 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746123 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746141 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746160 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746206 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746230 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746254 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746271 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746287 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746305 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746320 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746338 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746386 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746407 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746448 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746464 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746479 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746495 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746510 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746546 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746563 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746578 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746592 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746608 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746623 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746643 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746659 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746674 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746690 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746704 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746718 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746734 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746750 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746764 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746779 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746797 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746812 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746828 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746844 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746861 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746879 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746895 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746911 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746926 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746941 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746958 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746974 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746990 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747005 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747022 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745556 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745726 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745674 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745736 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745766 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745813 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.745953 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746021 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746140 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746227 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747160 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747194 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746225 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746320 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746443 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747223 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746497 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746644 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746751 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746801 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746964 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746976 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.746966 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747293 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747037 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747341 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747353 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747404 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747408 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747454 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747459 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747475 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747499 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747517 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747532 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747547 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747563 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747563 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747578 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747595 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747611 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747622 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747627 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747681 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747712 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747738 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747762 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747789 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747811 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747834 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747859 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747882 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747903 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747924 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747946 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747966 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.747985 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748005 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748028 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748048 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748069 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748091 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748095 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748117 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748141 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748184 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748209 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748233 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748256 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748278 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748299 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748321 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748342 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748362 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748384 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748406 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748430 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748453 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748478 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748502 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748525 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748552 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748574 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748596 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748622 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748642 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748663 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748686 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748693 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748709 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748734 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748759 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748781 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748802 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748823 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748846 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748870 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.748893 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.749373 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.749672 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.749704 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.749982 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.750041 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.750052 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.750536 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.750872 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.750907 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.750934 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.751157 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.751582 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.751591 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.751858 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.752083 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.752570 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.752647 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.752813 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.753174 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.753539 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.753656 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.753740 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.753804 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.754302 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.754647 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.754892 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.754943 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.755152 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.755524 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.755528 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.755673 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.755713 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.755996 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.756156 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.756249 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.753958 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.756398 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.756402 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.756710 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.756763 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.756749 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.757101 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.757338 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.757299 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.755850 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.757670 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.757702 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.757826 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.757828 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.757844 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.757839 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.757926 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.757957 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.757977 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.758002 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.758136 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.758383 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.758671 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.758708 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.760708 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.760734 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.760824 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.761726 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.762611 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.762627 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.762644 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.762660 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.762753 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.762782 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.764287 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.763160 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.764338 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.764406 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.764720 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.764760 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.758024 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.764920 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.764959 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.764988 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765015 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765042 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765066 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765093 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765120 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765145 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765191 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765216 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765241 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765264 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765289 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765313 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765335 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765445 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765472 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765497 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765522 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765545 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765568 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765592 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765615 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765637 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765663 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765736 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765769 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765796 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765825 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765853 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765876 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765900 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765926 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765953 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765977 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765999 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766023 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766045 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766070 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766154 4772 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766188 4772 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766201 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766214 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766227 4772 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766239 4772 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766259 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766272 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766285 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766298 4772 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766311 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766323 4772 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766335 4772 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766348 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766361 4772 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766374 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766386 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766398 4772 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766410 4772 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766423 4772 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766435 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766447 4772 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766460 4772 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766474 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766487 4772 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766499 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766511 4772 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766523 4772 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766534 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766547 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766559 4772 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766571 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766584 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766596 4772 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766608 4772 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766621 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766634 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766660 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766673 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766686 4772 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766699 4772 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766713 4772 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766724 4772 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766736 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766748 4772 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766761 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766773 4772 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766785 4772 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766796 4772 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766808 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766820 4772 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766832 4772 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766843 4772 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766856 4772 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766869 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766881 4772 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766892 4772 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766904 4772 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766916 4772 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766927 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766939 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766952 4772 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766963 4772 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766975 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766987 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.766999 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767011 4772 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767022 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767033 4772 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767045 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767057 4772 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767070 4772 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767082 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767095 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767108 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767120 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767132 4772 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767144 4772 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767156 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767220 4772 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767234 4772 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767247 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767259 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767271 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767282 4772 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767294 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767307 4772 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767318 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767330 4772 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767342 4772 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767389 4772 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767402 4772 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767413 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767424 4772 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767436 4772 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767447 4772 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767458 4772 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767470 4772 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767481 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767493 4772 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767504 4772 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767546 4772 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.767946 4772 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.764249 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.764641 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765075 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765483 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765739 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.765946 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.769016 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.769080 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.769047 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.769340 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.769376 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.769400 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.769546 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.769847 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.769907 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.770110 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.770396 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.770415 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.770700 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.770763 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.771103 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.771436 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.771586 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.771730 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.772390 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.772466 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.772713 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.772904 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.773437 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.773700 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.773897 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.774236 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.774510 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.774839 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.774915 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.775069 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.775603 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.776153 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.776618 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.776894 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.777374 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.777722 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.777966 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.778757 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.778967 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.779029 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 27 15:07:14 crc kubenswrapper[4772]: E0127 15:07:14.779153 4772 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 27 15:07:14 crc kubenswrapper[4772]: E0127 15:07:14.779221 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:15.279203022 +0000 UTC m=+21.259812210 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.779295 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.779448 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.779670 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.779709 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: E0127 15:07:14.779809 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:07:15.279795759 +0000 UTC m=+21.260404967 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.779917 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.779927 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.779934 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.779959 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.774281 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.780149 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.780223 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.780249 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.780369 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.780602 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.780743 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.780868 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.781081 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: E0127 15:07:14.781106 4772 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.781116 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: E0127 15:07:14.781194 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:15.281154317 +0000 UTC m=+21.261763515 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.781350 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.781467 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.781502 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.781621 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.781770 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.782011 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.782371 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.782432 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.782625 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.782810 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: E0127 15:07:14.785881 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 27 15:07:14 crc kubenswrapper[4772]: E0127 15:07:14.785899 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 27 15:07:14 crc kubenswrapper[4772]: E0127 15:07:14.785914 4772 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:14 crc kubenswrapper[4772]: E0127 15:07:14.785966 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:15.285950913 +0000 UTC m=+21.266560081 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.786209 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.774397 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.787431 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.787911 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.789238 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.789679 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.789734 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.789882 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.790423 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.790534 4772 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67" exitCode=255 Jan 27 15:07:14 crc kubenswrapper[4772]: E0127 15:07:14.790619 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 27 15:07:14 crc kubenswrapper[4772]: E0127 15:07:14.790661 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 27 15:07:14 crc kubenswrapper[4772]: E0127 15:07:14.790672 4772 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:14 crc kubenswrapper[4772]: E0127 15:07:14.790730 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:15.290697797 +0000 UTC m=+21.271306895 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.790756 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67"} Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.790824 4772 scope.go:117] "RemoveContainer" containerID="21e07984f4ca6d9ea37d8213eab7f36c2a5342806beacdb07b50a15ffba13ded" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.790884 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.791748 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.791951 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.792000 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.792123 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.793461 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.794954 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.796758 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.798027 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.799802 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.800106 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: E0127 15:07:14.802526 4772 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.806275 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.812317 4772 scope.go:117] "RemoveContainer" containerID="a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67" Jan 27 15:07:14 crc kubenswrapper[4772]: E0127 15:07:14.812733 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.812776 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.813049 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.813154 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.815422 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.816337 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.817554 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.820078 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.821635 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.826648 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.832814 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.837181 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.846633 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.858437 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.867897 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.867944 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868015 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868029 4772 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868040 4772 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868097 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868108 4772 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868117 4772 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868128 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868137 4772 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868146 4772 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868156 4772 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868181 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868193 4772 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868204 4772 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868215 4772 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868225 4772 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868235 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868245 4772 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868255 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868265 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868276 4772 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868287 4772 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868297 4772 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868306 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868316 4772 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868328 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868337 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868347 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868358 4772 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868368 4772 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868379 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868389 4772 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868399 4772 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868411 4772 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868422 4772 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868433 4772 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868444 4772 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868466 4772 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868478 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868488 4772 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868498 4772 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868508 4772 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868519 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868530 4772 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868540 4772 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868551 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868563 4772 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868573 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868594 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868607 4772 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868618 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868627 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868638 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868648 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868658 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868668 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868678 4772 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868690 4772 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868700 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868710 4772 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868719 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868728 4772 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868737 4772 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868747 4772 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868757 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868766 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868776 4772 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868785 4772 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868794 4772 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868804 4772 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868814 4772 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868823 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868832 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868842 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868850 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868861 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868870 4772 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868880 4772 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868890 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868901 4772 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868912 4772 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868922 4772 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868933 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868946 4772 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868958 4772 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868968 4772 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868978 4772 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.868989 4772 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.869002 4772 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.869014 4772 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.869024 4772 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.869035 4772 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.869045 4772 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.869093 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.869444 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.870322 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.881074 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.894375 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21e07984f4ca6d9ea37d8213eab7f36c2a5342806beacdb07b50a15ffba13ded\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"message\\\":\\\"W0127 15:06:57.758128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0127 15:06:57.758449 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769526417 cert, and key in /tmp/serving-cert-633789686/serving-signer.crt, /tmp/serving-cert-633789686/serving-signer.key\\\\nI0127 15:06:58.043108 1 observer_polling.go:159] Starting file observer\\\\nW0127 15:06:58.045578 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0127 15:06:58.045739 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:06:58.047315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-633789686/tls.crt::/tmp/serving-cert-633789686/tls.key\\\\\\\"\\\\nF0127 15:06:58.297465 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.906708 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.933670 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.943603 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.954611 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.961370 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 27 15:07:14 crc kubenswrapper[4772]: I0127 15:07:14.982019 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 27 15:07:14 crc kubenswrapper[4772]: W0127 15:07:14.989333 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-0c9c8240549313c1fb724b3845fa904773be10bf9695748d577f7630d7ee8265 WatchSource:0}: Error finding container 0c9c8240549313c1fb724b3845fa904773be10bf9695748d577f7630d7ee8265: Status 404 returned error can't find the container with id 0c9c8240549313c1fb724b3845fa904773be10bf9695748d577f7630d7ee8265 Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.376879 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:07:15 crc kubenswrapper[4772]: E0127 15:07:15.377071 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:07:16.377043392 +0000 UTC m=+22.357652490 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.377291 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.377317 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.377334 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.377565 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:15 crc kubenswrapper[4772]: E0127 15:07:15.377604 4772 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 27 15:07:15 crc kubenswrapper[4772]: E0127 15:07:15.377636 4772 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 27 15:07:15 crc kubenswrapper[4772]: E0127 15:07:15.377669 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:16.3776574 +0000 UTC m=+22.358266498 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 27 15:07:15 crc kubenswrapper[4772]: E0127 15:07:15.377750 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:16.377725342 +0000 UTC m=+22.358334510 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 27 15:07:15 crc kubenswrapper[4772]: E0127 15:07:15.377758 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 27 15:07:15 crc kubenswrapper[4772]: E0127 15:07:15.377795 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 27 15:07:15 crc kubenswrapper[4772]: E0127 15:07:15.377807 4772 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:15 crc kubenswrapper[4772]: E0127 15:07:15.377892 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:16.377877206 +0000 UTC m=+22.358486304 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:15 crc kubenswrapper[4772]: E0127 15:07:15.377954 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 27 15:07:15 crc kubenswrapper[4772]: E0127 15:07:15.377963 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 27 15:07:15 crc kubenswrapper[4772]: E0127 15:07:15.377970 4772 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:15 crc kubenswrapper[4772]: E0127 15:07:15.377992 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:16.377986039 +0000 UTC m=+22.358595127 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.379826 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.493130 4772 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-01-27 15:02:14 +0000 UTC, rotation deadline is 2026-11-07 11:43:22.531797809 +0000 UTC Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.493200 4772 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6812h36m7.038599849s for next certificate rotation Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.625834 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 22:28:43.659872933 +0000 UTC Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.662157 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:15 crc kubenswrapper[4772]: E0127 15:07:15.662289 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.794145 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449"} Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.794205 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"5a238964e704642f78de8ef46fa665accf609f3d0d53e2d04174c20478840530"} Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.795213 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.796864 4772 scope.go:117] "RemoveContainer" containerID="a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67" Jan 27 15:07:15 crc kubenswrapper[4772]: E0127 15:07:15.797000 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.797109 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"0c9c8240549313c1fb724b3845fa904773be10bf9695748d577f7630d7ee8265"} Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.798372 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842"} Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.798398 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015"} Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.798412 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6d2ab5bedc3e93a9e7981a953481fe73478caf4bfc9f40829609d914429ca3ec"} Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.813255 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21e07984f4ca6d9ea37d8213eab7f36c2a5342806beacdb07b50a15ffba13ded\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"message\\\":\\\"W0127 15:06:57.758128 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0127 15:06:57.758449 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769526417 cert, and key in /tmp/serving-cert-633789686/serving-signer.crt, /tmp/serving-cert-633789686/serving-signer.key\\\\nI0127 15:06:58.043108 1 observer_polling.go:159] Starting file observer\\\\nW0127 15:06:58.045578 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0127 15:06:58.045739 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:06:58.047315 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-633789686/tls.crt::/tmp/serving-cert-633789686/tls.key\\\\\\\"\\\\nF0127 15:06:58.297465 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.829087 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.842214 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.861646 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.876852 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.889224 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.903247 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.917849 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.934761 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.948299 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.960553 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.973959 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:15 crc kubenswrapper[4772]: I0127 15:07:15.996383 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.007873 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.035663 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.049688 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.087941 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.107371 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.314606 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-dtdj6"] Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.315217 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-q46tm"] Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.315413 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-dtdj6" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.315555 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-q46tm" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.318097 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.318308 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.318325 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.318434 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.318470 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.318573 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.318624 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.341302 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.365647 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.382613 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.386414 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.386524 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/fed65bae-f1c4-4c97-bb6d-d4144fe2532b-serviceca\") pod \"node-ca-q46tm\" (UID: \"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\") " pod="openshift-image-registry/node-ca-q46tm" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.386558 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:16 crc kubenswrapper[4772]: E0127 15:07:16.386645 4772 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 27 15:07:16 crc kubenswrapper[4772]: E0127 15:07:16.386667 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:07:18.386553619 +0000 UTC m=+24.367162727 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.386777 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:16 crc kubenswrapper[4772]: E0127 15:07:16.386922 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:18.386901669 +0000 UTC m=+24.367510767 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 27 15:07:16 crc kubenswrapper[4772]: E0127 15:07:16.386965 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 27 15:07:16 crc kubenswrapper[4772]: E0127 15:07:16.386985 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 27 15:07:16 crc kubenswrapper[4772]: E0127 15:07:16.386997 4772 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:16 crc kubenswrapper[4772]: E0127 15:07:16.387058 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:18.387048383 +0000 UTC m=+24.367657661 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.387121 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fed65bae-f1c4-4c97-bb6d-d4144fe2532b-host\") pod \"node-ca-q46tm\" (UID: \"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\") " pod="openshift-image-registry/node-ca-q46tm" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.387210 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.387247 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pswh\" (UniqueName: \"kubernetes.io/projected/95a893d4-4faa-40b2-b505-9698fe428ba8-kube-api-access-7pswh\") pod \"node-resolver-dtdj6\" (UID: \"95a893d4-4faa-40b2-b505-9698fe428ba8\") " pod="openshift-dns/node-resolver-dtdj6" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.387306 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:16 crc kubenswrapper[4772]: E0127 15:07:16.387360 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 27 15:07:16 crc kubenswrapper[4772]: E0127 15:07:16.387379 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.387379 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/95a893d4-4faa-40b2-b505-9698fe428ba8-hosts-file\") pod \"node-resolver-dtdj6\" (UID: \"95a893d4-4faa-40b2-b505-9698fe428ba8\") " pod="openshift-dns/node-resolver-dtdj6" Jan 27 15:07:16 crc kubenswrapper[4772]: E0127 15:07:16.387394 4772 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:16 crc kubenswrapper[4772]: E0127 15:07:16.387428 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:18.387419893 +0000 UTC m=+24.368028981 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.387444 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qln7r\" (UniqueName: \"kubernetes.io/projected/fed65bae-f1c4-4c97-bb6d-d4144fe2532b-kube-api-access-qln7r\") pod \"node-ca-q46tm\" (UID: \"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\") " pod="openshift-image-registry/node-ca-q46tm" Jan 27 15:07:16 crc kubenswrapper[4772]: E0127 15:07:16.387483 4772 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 27 15:07:16 crc kubenswrapper[4772]: E0127 15:07:16.387547 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:18.387540337 +0000 UTC m=+24.368149435 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.399669 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.416366 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.433461 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.447280 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.462345 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.473674 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.488804 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pswh\" (UniqueName: \"kubernetes.io/projected/95a893d4-4faa-40b2-b505-9698fe428ba8-kube-api-access-7pswh\") pod \"node-resolver-dtdj6\" (UID: \"95a893d4-4faa-40b2-b505-9698fe428ba8\") " pod="openshift-dns/node-resolver-dtdj6" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.488848 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/95a893d4-4faa-40b2-b505-9698fe428ba8-hosts-file\") pod \"node-resolver-dtdj6\" (UID: \"95a893d4-4faa-40b2-b505-9698fe428ba8\") " pod="openshift-dns/node-resolver-dtdj6" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.488872 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qln7r\" (UniqueName: \"kubernetes.io/projected/fed65bae-f1c4-4c97-bb6d-d4144fe2532b-kube-api-access-qln7r\") pod \"node-ca-q46tm\" (UID: \"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\") " pod="openshift-image-registry/node-ca-q46tm" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.488908 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/fed65bae-f1c4-4c97-bb6d-d4144fe2532b-serviceca\") pod \"node-ca-q46tm\" (UID: \"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\") " pod="openshift-image-registry/node-ca-q46tm" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.488939 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fed65bae-f1c4-4c97-bb6d-d4144fe2532b-host\") pod \"node-ca-q46tm\" (UID: \"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\") " pod="openshift-image-registry/node-ca-q46tm" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.488976 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/95a893d4-4faa-40b2-b505-9698fe428ba8-hosts-file\") pod \"node-resolver-dtdj6\" (UID: \"95a893d4-4faa-40b2-b505-9698fe428ba8\") " pod="openshift-dns/node-resolver-dtdj6" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.489012 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fed65bae-f1c4-4c97-bb6d-d4144fe2532b-host\") pod \"node-ca-q46tm\" (UID: \"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\") " pod="openshift-image-registry/node-ca-q46tm" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.490990 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/fed65bae-f1c4-4c97-bb6d-d4144fe2532b-serviceca\") pod \"node-ca-q46tm\" (UID: \"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\") " pod="openshift-image-registry/node-ca-q46tm" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.504408 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.515440 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qln7r\" (UniqueName: \"kubernetes.io/projected/fed65bae-f1c4-4c97-bb6d-d4144fe2532b-kube-api-access-qln7r\") pod \"node-ca-q46tm\" (UID: \"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\") " pod="openshift-image-registry/node-ca-q46tm" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.526661 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pswh\" (UniqueName: \"kubernetes.io/projected/95a893d4-4faa-40b2-b505-9698fe428ba8-kube-api-access-7pswh\") pod \"node-resolver-dtdj6\" (UID: \"95a893d4-4faa-40b2-b505-9698fe428ba8\") " pod="openshift-dns/node-resolver-dtdj6" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.533948 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.580595 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.597734 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.618195 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.626876 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 15:41:00.562975131 +0000 UTC Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.631018 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-dtdj6" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.638130 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-q46tm" Jan 27 15:07:16 crc kubenswrapper[4772]: W0127 15:07:16.645078 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95a893d4_4faa_40b2_b505_9698fe428ba8.slice/crio-1c0961d39afdfce06462f956256b15a99563f79b26cbab0f99cd2451b4fad029 WatchSource:0}: Error finding container 1c0961d39afdfce06462f956256b15a99563f79b26cbab0f99cd2451b4fad029: Status 404 returned error can't find the container with id 1c0961d39afdfce06462f956256b15a99563f79b26cbab0f99cd2451b4fad029 Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.664396 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.664986 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:16 crc kubenswrapper[4772]: E0127 15:07:16.665118 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:07:16 crc kubenswrapper[4772]: E0127 15:07:16.665234 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.669785 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.682542 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.683435 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.685519 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.686256 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.691898 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: W0127 15:07:16.692314 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfed65bae_f1c4_4c97_bb6d_d4144fe2532b.slice/crio-a3e708439b4fd068181bd4d2a3735a29f3f7f68a4cb602ed80fdec2f8dd0e6da WatchSource:0}: Error finding container a3e708439b4fd068181bd4d2a3735a29f3f7f68a4cb602ed80fdec2f8dd0e6da: Status 404 returned error can't find the container with id a3e708439b4fd068181bd4d2a3735a29f3f7f68a4cb602ed80fdec2f8dd0e6da Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.692904 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.693139 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.693719 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.695043 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.695883 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.697041 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.697747 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.699465 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.700118 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.700885 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.702071 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.705435 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.706511 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.713611 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.714887 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.715893 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.717325 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.718004 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.718539 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.720480 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.721022 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.725817 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.726768 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.727875 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.728707 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.729762 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.730355 4772 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.730534 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.731772 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.733096 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.733826 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.734606 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.736507 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.737732 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.738478 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.739904 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.740764 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.741889 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.742672 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.743932 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.745272 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.745831 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.746504 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.747659 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.749006 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.749730 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.750376 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.751651 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.752238 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.752589 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.754278 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.755108 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.766616 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.795075 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.801774 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-q46tm" event={"ID":"fed65bae-f1c4-4c97-bb6d-d4144fe2532b","Type":"ContainerStarted","Data":"a3e708439b4fd068181bd4d2a3735a29f3f7f68a4cb602ed80fdec2f8dd0e6da"} Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.804073 4772 scope.go:117] "RemoveContainer" containerID="a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.804195 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-dtdj6" event={"ID":"95a893d4-4faa-40b2-b505-9698fe428ba8","Type":"ContainerStarted","Data":"1c0961d39afdfce06462f956256b15a99563f79b26cbab0f99cd2451b4fad029"} Jan 27 15:07:16 crc kubenswrapper[4772]: E0127 15:07:16.804358 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Jan 27 15:07:16 crc kubenswrapper[4772]: I0127 15:07:16.810392 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:16Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.122756 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-4hwxn"] Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.123049 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-c7pdz"] Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.123277 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.125143 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.125408 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.125941 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.126147 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.126501 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.127132 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-x7jwx"] Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.127343 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.127908 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.133148 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.133357 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.133692 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.133728 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.133866 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.134071 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.136815 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.145727 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.162312 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.174707 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.186785 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.194816 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/67794a44-d793-4fd7-9e54-e40437f67c0b-proxy-tls\") pod \"machine-config-daemon-4hwxn\" (UID: \"67794a44-d793-4fd7-9e54-e40437f67c0b\") " pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.194886 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/67794a44-d793-4fd7-9e54-e40437f67c0b-mcd-auth-proxy-config\") pod \"machine-config-daemon-4hwxn\" (UID: \"67794a44-d793-4fd7-9e54-e40437f67c0b\") " pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.194915 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh6ph\" (UniqueName: \"kubernetes.io/projected/67794a44-d793-4fd7-9e54-e40437f67c0b-kube-api-access-lh6ph\") pod \"machine-config-daemon-4hwxn\" (UID: \"67794a44-d793-4fd7-9e54-e40437f67c0b\") " pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.195009 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/67794a44-d793-4fd7-9e54-e40437f67c0b-rootfs\") pod \"machine-config-daemon-4hwxn\" (UID: \"67794a44-d793-4fd7-9e54-e40437f67c0b\") " pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.200664 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.212121 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.224745 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.239011 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.252891 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.264457 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.273281 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.292321 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.295950 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-host-var-lib-kubelet\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.295987 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-hostroot\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.296003 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-multus-daemon-config\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.296024 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/67794a44-d793-4fd7-9e54-e40437f67c0b-rootfs\") pod \"machine-config-daemon-4hwxn\" (UID: \"67794a44-d793-4fd7-9e54-e40437f67c0b\") " pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.296040 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/67794a44-d793-4fd7-9e54-e40437f67c0b-mcd-auth-proxy-config\") pod \"machine-config-daemon-4hwxn\" (UID: \"67794a44-d793-4fd7-9e54-e40437f67c0b\") " pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.296054 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-cni-binary-copy\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.296068 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-host-run-netns\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.296082 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-host-var-lib-cni-bin\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.296096 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-host-run-multus-certs\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.296111 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-host-run-k8s-cni-cncf-io\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.296126 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1acef947-6310-4ac0-bc84-a06d91f84cb6-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-c7pdz\" (UID: \"1acef947-6310-4ac0-bc84-a06d91f84cb6\") " pod="openshift-multus/multus-additional-cni-plugins-c7pdz" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.296140 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-cnibin\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.296156 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d4sv\" (UniqueName: \"kubernetes.io/projected/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-kube-api-access-8d4sv\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.296185 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-multus-socket-dir-parent\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.296201 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh6ph\" (UniqueName: \"kubernetes.io/projected/67794a44-d793-4fd7-9e54-e40437f67c0b-kube-api-access-lh6ph\") pod \"machine-config-daemon-4hwxn\" (UID: \"67794a44-d793-4fd7-9e54-e40437f67c0b\") " pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.296220 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntsst\" (UniqueName: \"kubernetes.io/projected/1acef947-6310-4ac0-bc84-a06d91f84cb6-kube-api-access-ntsst\") pod \"multus-additional-cni-plugins-c7pdz\" (UID: \"1acef947-6310-4ac0-bc84-a06d91f84cb6\") " pod="openshift-multus/multus-additional-cni-plugins-c7pdz" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.296244 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1acef947-6310-4ac0-bc84-a06d91f84cb6-tuning-conf-dir\") pod \"multus-additional-cni-plugins-c7pdz\" (UID: \"1acef947-6310-4ac0-bc84-a06d91f84cb6\") " pod="openshift-multus/multus-additional-cni-plugins-c7pdz" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.296261 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/67794a44-d793-4fd7-9e54-e40437f67c0b-rootfs\") pod \"machine-config-daemon-4hwxn\" (UID: \"67794a44-d793-4fd7-9e54-e40437f67c0b\") " pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.296260 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-multus-conf-dir\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.296931 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-etc-kubernetes\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.297001 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-system-cni-dir\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.297019 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-multus-cni-dir\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.297019 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/67794a44-d793-4fd7-9e54-e40437f67c0b-mcd-auth-proxy-config\") pod \"machine-config-daemon-4hwxn\" (UID: \"67794a44-d793-4fd7-9e54-e40437f67c0b\") " pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.297057 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1acef947-6310-4ac0-bc84-a06d91f84cb6-cnibin\") pod \"multus-additional-cni-plugins-c7pdz\" (UID: \"1acef947-6310-4ac0-bc84-a06d91f84cb6\") " pod="openshift-multus/multus-additional-cni-plugins-c7pdz" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.297074 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1acef947-6310-4ac0-bc84-a06d91f84cb6-cni-binary-copy\") pod \"multus-additional-cni-plugins-c7pdz\" (UID: \"1acef947-6310-4ac0-bc84-a06d91f84cb6\") " pod="openshift-multus/multus-additional-cni-plugins-c7pdz" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.297114 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1acef947-6310-4ac0-bc84-a06d91f84cb6-system-cni-dir\") pod \"multus-additional-cni-plugins-c7pdz\" (UID: \"1acef947-6310-4ac0-bc84-a06d91f84cb6\") " pod="openshift-multus/multus-additional-cni-plugins-c7pdz" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.297129 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1acef947-6310-4ac0-bc84-a06d91f84cb6-os-release\") pod \"multus-additional-cni-plugins-c7pdz\" (UID: \"1acef947-6310-4ac0-bc84-a06d91f84cb6\") " pod="openshift-multus/multus-additional-cni-plugins-c7pdz" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.297150 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-host-var-lib-cni-multus\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.297188 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/67794a44-d793-4fd7-9e54-e40437f67c0b-proxy-tls\") pod \"machine-config-daemon-4hwxn\" (UID: \"67794a44-d793-4fd7-9e54-e40437f67c0b\") " pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.297259 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-os-release\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.302511 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/67794a44-d793-4fd7-9e54-e40437f67c0b-proxy-tls\") pod \"machine-config-daemon-4hwxn\" (UID: \"67794a44-d793-4fd7-9e54-e40437f67c0b\") " pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.320897 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.328340 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh6ph\" (UniqueName: \"kubernetes.io/projected/67794a44-d793-4fd7-9e54-e40437f67c0b-kube-api-access-lh6ph\") pod \"machine-config-daemon-4hwxn\" (UID: \"67794a44-d793-4fd7-9e54-e40437f67c0b\") " pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.334625 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.347960 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.362646 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.374475 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.390018 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398159 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-host-var-lib-cni-bin\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398221 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-host-run-multus-certs\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398246 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-host-run-k8s-cni-cncf-io\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398311 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1acef947-6310-4ac0-bc84-a06d91f84cb6-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-c7pdz\" (UID: \"1acef947-6310-4ac0-bc84-a06d91f84cb6\") " pod="openshift-multus/multus-additional-cni-plugins-c7pdz" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398338 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-cnibin\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398353 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-host-run-multus-certs\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398371 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d4sv\" (UniqueName: \"kubernetes.io/projected/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-kube-api-access-8d4sv\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398396 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntsst\" (UniqueName: \"kubernetes.io/projected/1acef947-6310-4ac0-bc84-a06d91f84cb6-kube-api-access-ntsst\") pod \"multus-additional-cni-plugins-c7pdz\" (UID: \"1acef947-6310-4ac0-bc84-a06d91f84cb6\") " pod="openshift-multus/multus-additional-cni-plugins-c7pdz" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398416 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-multus-socket-dir-parent\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398436 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1acef947-6310-4ac0-bc84-a06d91f84cb6-tuning-conf-dir\") pod \"multus-additional-cni-plugins-c7pdz\" (UID: \"1acef947-6310-4ac0-bc84-a06d91f84cb6\") " pod="openshift-multus/multus-additional-cni-plugins-c7pdz" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398455 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-multus-conf-dir\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398496 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-etc-kubernetes\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398513 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-multus-cni-dir\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398527 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1acef947-6310-4ac0-bc84-a06d91f84cb6-cnibin\") pod \"multus-additional-cni-plugins-c7pdz\" (UID: \"1acef947-6310-4ac0-bc84-a06d91f84cb6\") " pod="openshift-multus/multus-additional-cni-plugins-c7pdz" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398542 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1acef947-6310-4ac0-bc84-a06d91f84cb6-cni-binary-copy\") pod \"multus-additional-cni-plugins-c7pdz\" (UID: \"1acef947-6310-4ac0-bc84-a06d91f84cb6\") " pod="openshift-multus/multus-additional-cni-plugins-c7pdz" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398557 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-system-cni-dir\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398574 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1acef947-6310-4ac0-bc84-a06d91f84cb6-system-cni-dir\") pod \"multus-additional-cni-plugins-c7pdz\" (UID: \"1acef947-6310-4ac0-bc84-a06d91f84cb6\") " pod="openshift-multus/multus-additional-cni-plugins-c7pdz" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398588 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1acef947-6310-4ac0-bc84-a06d91f84cb6-os-release\") pod \"multus-additional-cni-plugins-c7pdz\" (UID: \"1acef947-6310-4ac0-bc84-a06d91f84cb6\") " pod="openshift-multus/multus-additional-cni-plugins-c7pdz" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398610 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-host-var-lib-cni-multus\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398632 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-os-release\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398670 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-hostroot\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398693 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-multus-daemon-config\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398717 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-host-var-lib-kubelet\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398735 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-cni-binary-copy\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398752 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-host-run-netns\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398811 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-host-run-netns\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398862 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-cnibin\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398856 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-host-var-lib-cni-bin\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398936 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1acef947-6310-4ac0-bc84-a06d91f84cb6-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-c7pdz\" (UID: \"1acef947-6310-4ac0-bc84-a06d91f84cb6\") " pod="openshift-multus/multus-additional-cni-plugins-c7pdz" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.398977 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-host-run-k8s-cni-cncf-io\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.399060 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-system-cni-dir\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.399134 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-hostroot\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.399204 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-os-release\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.399240 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-host-var-lib-kubelet\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.399486 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-multus-socket-dir-parent\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.399604 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1acef947-6310-4ac0-bc84-a06d91f84cb6-system-cni-dir\") pod \"multus-additional-cni-plugins-c7pdz\" (UID: \"1acef947-6310-4ac0-bc84-a06d91f84cb6\") " pod="openshift-multus/multus-additional-cni-plugins-c7pdz" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.399650 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1acef947-6310-4ac0-bc84-a06d91f84cb6-os-release\") pod \"multus-additional-cni-plugins-c7pdz\" (UID: \"1acef947-6310-4ac0-bc84-a06d91f84cb6\") " pod="openshift-multus/multus-additional-cni-plugins-c7pdz" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.399677 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-host-var-lib-cni-multus\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.399727 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-cni-binary-copy\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.399747 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-multus-daemon-config\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.399766 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-multus-conf-dir\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.399789 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-etc-kubernetes\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.399792 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1acef947-6310-4ac0-bc84-a06d91f84cb6-cnibin\") pod \"multus-additional-cni-plugins-c7pdz\" (UID: \"1acef947-6310-4ac0-bc84-a06d91f84cb6\") " pod="openshift-multus/multus-additional-cni-plugins-c7pdz" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.399789 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-multus-cni-dir\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.400126 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1acef947-6310-4ac0-bc84-a06d91f84cb6-cni-binary-copy\") pod \"multus-additional-cni-plugins-c7pdz\" (UID: \"1acef947-6310-4ac0-bc84-a06d91f84cb6\") " pod="openshift-multus/multus-additional-cni-plugins-c7pdz" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.400214 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1acef947-6310-4ac0-bc84-a06d91f84cb6-tuning-conf-dir\") pod \"multus-additional-cni-plugins-c7pdz\" (UID: \"1acef947-6310-4ac0-bc84-a06d91f84cb6\") " pod="openshift-multus/multus-additional-cni-plugins-c7pdz" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.400207 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.412312 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.414792 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d4sv\" (UniqueName: \"kubernetes.io/projected/87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8-kube-api-access-8d4sv\") pod \"multus-x7jwx\" (UID: \"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\") " pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.415862 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntsst\" (UniqueName: \"kubernetes.io/projected/1acef947-6310-4ac0-bc84-a06d91f84cb6-kube-api-access-ntsst\") pod \"multus-additional-cni-plugins-c7pdz\" (UID: \"1acef947-6310-4ac0-bc84-a06d91f84cb6\") " pod="openshift-multus/multus-additional-cni-plugins-c7pdz" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.424377 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.437245 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.438346 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.449114 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" Jan 27 15:07:17 crc kubenswrapper[4772]: W0127 15:07:17.450519 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67794a44_d793_4fd7_9e54_e40437f67c0b.slice/crio-fd0ddca1b8be2e12111ee0e50ac1b53dde3dd038c4cc631af4b78ec5dddd73cf WatchSource:0}: Error finding container fd0ddca1b8be2e12111ee0e50ac1b53dde3dd038c4cc631af4b78ec5dddd73cf: Status 404 returned error can't find the container with id fd0ddca1b8be2e12111ee0e50ac1b53dde3dd038c4cc631af4b78ec5dddd73cf Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.452316 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.460151 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-x7jwx" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.475002 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: W0127 15:07:17.475374 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87cb2a5b_099e_4a3b_a0bc_cba76a1a00a8.slice/crio-7cd942ff56dee37fd1e7c9671a18a0a4c9e67dd558fe1ef941783ed45ca45ea6 WatchSource:0}: Error finding container 7cd942ff56dee37fd1e7c9671a18a0a4c9e67dd558fe1ef941783ed45ca45ea6: Status 404 returned error can't find the container with id 7cd942ff56dee37fd1e7c9671a18a0a4c9e67dd558fe1ef941783ed45ca45ea6 Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.490653 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.505301 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.519791 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-n2khk"] Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.521742 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.523765 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.524279 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.524471 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.524890 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.525010 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.525131 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.526294 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.538175 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.588300 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.601515 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-run-ovn\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.601562 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-log-socket\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.601649 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-slash\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.601682 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-run-openvswitch\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.601698 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-cni-bin\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.601715 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/736264c8-cd18-479a-88ba-e1ec15dbfdae-ovn-node-metrics-cert\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.601732 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dt6g\" (UniqueName: \"kubernetes.io/projected/736264c8-cd18-479a-88ba-e1ec15dbfdae-kube-api-access-2dt6g\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.601753 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-kubelet\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.601772 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/736264c8-cd18-479a-88ba-e1ec15dbfdae-env-overrides\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.601791 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/736264c8-cd18-479a-88ba-e1ec15dbfdae-ovnkube-script-lib\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.601812 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-systemd-units\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.601829 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-run-systemd\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.601852 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-etc-openvswitch\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.601872 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-run-ovn-kubernetes\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.601899 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-run-netns\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.601965 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-node-log\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.602023 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-cni-netd\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.602058 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-var-lib-openvswitch\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.602083 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/736264c8-cd18-479a-88ba-e1ec15dbfdae-ovnkube-config\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.602133 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.621361 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.628072 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 01:35:13.127747721 +0000 UTC Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.637813 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.657955 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.662832 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:17 crc kubenswrapper[4772]: E0127 15:07:17.662969 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.673367 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.692970 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.702577 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.702620 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-run-ovn\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.702638 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-log-socket\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.702678 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-slash\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.702693 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-run-openvswitch\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.702710 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-cni-bin\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.702724 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/736264c8-cd18-479a-88ba-e1ec15dbfdae-ovn-node-metrics-cert\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.702740 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-kubelet\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.702753 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dt6g\" (UniqueName: \"kubernetes.io/projected/736264c8-cd18-479a-88ba-e1ec15dbfdae-kube-api-access-2dt6g\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.702773 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/736264c8-cd18-479a-88ba-e1ec15dbfdae-env-overrides\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.702787 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/736264c8-cd18-479a-88ba-e1ec15dbfdae-ovnkube-script-lib\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.702802 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-systemd-units\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.702817 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-run-systemd\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.702831 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-etc-openvswitch\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.702845 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-run-ovn-kubernetes\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.702868 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-run-netns\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.702882 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-node-log\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.702901 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-cni-netd\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.702916 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-var-lib-openvswitch\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.702930 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/736264c8-cd18-479a-88ba-e1ec15dbfdae-ovnkube-config\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.703559 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/736264c8-cd18-479a-88ba-e1ec15dbfdae-ovnkube-config\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.703608 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.703633 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-run-ovn\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.703655 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-log-socket\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.703675 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-slash\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.703696 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-run-openvswitch\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.703716 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-cni-bin\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.704454 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-run-systemd\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.704904 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-run-netns\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.704956 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-etc-openvswitch\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.704980 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-run-ovn-kubernetes\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.705003 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-systemd-units\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.705022 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/736264c8-cd18-479a-88ba-e1ec15dbfdae-env-overrides\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.705067 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-cni-netd\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.705092 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-node-log\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.705114 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-var-lib-openvswitch\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.705130 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-kubelet\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.705138 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/736264c8-cd18-479a-88ba-e1ec15dbfdae-ovnkube-script-lib\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.707834 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/736264c8-cd18-479a-88ba-e1ec15dbfdae-ovn-node-metrics-cert\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.721436 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.724431 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dt6g\" (UniqueName: \"kubernetes.io/projected/736264c8-cd18-479a-88ba-e1ec15dbfdae-kube-api-access-2dt6g\") pod \"ovnkube-node-n2khk\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.736206 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.754425 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.765659 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.781083 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.794714 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.807547 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2"} Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.809837 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x7jwx" event={"ID":"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8","Type":"ContainerStarted","Data":"ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33"} Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.809799 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.809883 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x7jwx" event={"ID":"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8","Type":"ContainerStarted","Data":"7cd942ff56dee37fd1e7c9671a18a0a4c9e67dd558fe1ef941783ed45ca45ea6"} Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.811258 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-q46tm" event={"ID":"fed65bae-f1c4-4c97-bb6d-d4144fe2532b","Type":"ContainerStarted","Data":"6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348"} Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.813335 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-dtdj6" event={"ID":"95a893d4-4faa-40b2-b505-9698fe428ba8","Type":"ContainerStarted","Data":"6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964"} Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.814861 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" event={"ID":"1acef947-6310-4ac0-bc84-a06d91f84cb6","Type":"ContainerStarted","Data":"af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538"} Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.814892 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" event={"ID":"1acef947-6310-4ac0-bc84-a06d91f84cb6","Type":"ContainerStarted","Data":"85e2f41d41be6c7cd6e458ef932bea5281182263cb6d8eb0efde7f5c3f5f3224"} Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.816735 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa"} Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.816791 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19"} Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.816805 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"fd0ddca1b8be2e12111ee0e50ac1b53dde3dd038c4cc631af4b78ec5dddd73cf"} Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.832316 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.845040 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.858944 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.872870 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.887784 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.898102 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.916643 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.921949 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: W0127 15:07:17.932352 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod736264c8_cd18_479a_88ba_e1ec15dbfdae.slice/crio-b6a209d8fc4e180971a6f92a0f3c7493472a2095b6c5303a9b0ce0f4e62056a9 WatchSource:0}: Error finding container b6a209d8fc4e180971a6f92a0f3c7493472a2095b6c5303a9b0ce0f4e62056a9: Status 404 returned error can't find the container with id b6a209d8fc4e180971a6f92a0f3c7493472a2095b6c5303a9b0ce0f4e62056a9 Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.942578 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:17 crc kubenswrapper[4772]: I0127 15:07:17.965436 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:17Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.003706 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:18Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.047293 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:18Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.089870 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:18Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.126804 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:18Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.164869 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:18Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.209873 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:18Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.245258 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:18Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.412123 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:07:18 crc kubenswrapper[4772]: E0127 15:07:18.412298 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:07:22.41227002 +0000 UTC m=+28.392879118 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.412384 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.412419 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.412453 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.412470 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:18 crc kubenswrapper[4772]: E0127 15:07:18.412578 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 27 15:07:18 crc kubenswrapper[4772]: E0127 15:07:18.412590 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 27 15:07:18 crc kubenswrapper[4772]: E0127 15:07:18.412600 4772 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:18 crc kubenswrapper[4772]: E0127 15:07:18.412601 4772 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 27 15:07:18 crc kubenswrapper[4772]: E0127 15:07:18.412635 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:22.41262848 +0000 UTC m=+28.393237568 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:18 crc kubenswrapper[4772]: E0127 15:07:18.412661 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 27 15:07:18 crc kubenswrapper[4772]: E0127 15:07:18.412684 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 27 15:07:18 crc kubenswrapper[4772]: E0127 15:07:18.412697 4772 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:18 crc kubenswrapper[4772]: E0127 15:07:18.412710 4772 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 27 15:07:18 crc kubenswrapper[4772]: E0127 15:07:18.412664 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:22.41264429 +0000 UTC m=+28.393253438 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 27 15:07:18 crc kubenswrapper[4772]: E0127 15:07:18.412892 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:22.412853486 +0000 UTC m=+28.393462634 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:18 crc kubenswrapper[4772]: E0127 15:07:18.412913 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:22.412902948 +0000 UTC m=+28.393512146 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.629063 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 00:22:47.813012344 +0000 UTC Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.662932 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.663014 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:18 crc kubenswrapper[4772]: E0127 15:07:18.663290 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:07:18 crc kubenswrapper[4772]: E0127 15:07:18.663417 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.821904 4772 generic.go:334] "Generic (PLEG): container finished" podID="1acef947-6310-4ac0-bc84-a06d91f84cb6" containerID="af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538" exitCode=0 Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.821971 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" event={"ID":"1acef947-6310-4ac0-bc84-a06d91f84cb6","Type":"ContainerDied","Data":"af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538"} Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.824758 4772 generic.go:334] "Generic (PLEG): container finished" podID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerID="c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d" exitCode=0 Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.825017 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerDied","Data":"c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d"} Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.825198 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerStarted","Data":"b6a209d8fc4e180971a6f92a0f3c7493472a2095b6c5303a9b0ce0f4e62056a9"} Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.842154 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:18Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.858039 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:18Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.873886 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:18Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.887348 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:18Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.900795 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:18Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.916412 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:18Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.929824 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:18Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.947824 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:18Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.961307 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:18Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.975297 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:18Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:18 crc kubenswrapper[4772]: I0127 15:07:18.991139 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:18Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.008449 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.027531 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.041042 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.057453 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.069902 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.081185 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.094802 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.108839 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.134432 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.149841 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.162215 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.176978 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.204076 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.244113 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.285556 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.328093 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.396086 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.418047 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.451389 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.629309 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 08:01:38.785341056 +0000 UTC Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.662865 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:19 crc kubenswrapper[4772]: E0127 15:07:19.662981 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.835793 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerStarted","Data":"0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b"} Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.836074 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerStarted","Data":"5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4"} Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.836186 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerStarted","Data":"45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5"} Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.836286 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerStarted","Data":"321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9"} Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.836375 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerStarted","Data":"3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854"} Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.836461 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerStarted","Data":"e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e"} Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.838027 4772 generic.go:334] "Generic (PLEG): container finished" podID="1acef947-6310-4ac0-bc84-a06d91f84cb6" containerID="e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1" exitCode=0 Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.838141 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" event={"ID":"1acef947-6310-4ac0-bc84-a06d91f84cb6","Type":"ContainerDied","Data":"e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1"} Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.856315 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.877711 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.892824 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.907140 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.931299 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.944387 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.956605 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.974483 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.987991 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:19 crc kubenswrapper[4772]: I0127 15:07:19.999218 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:19Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.009506 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:20Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.024675 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:20Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.040327 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:20Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.052455 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:20Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.070360 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:20Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.629641 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 00:29:12.132567393 +0000 UTC Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.662204 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.662237 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:20 crc kubenswrapper[4772]: E0127 15:07:20.662359 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:07:20 crc kubenswrapper[4772]: E0127 15:07:20.662573 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.843116 4772 generic.go:334] "Generic (PLEG): container finished" podID="1acef947-6310-4ac0-bc84-a06d91f84cb6" containerID="4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc" exitCode=0 Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.843194 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" event={"ID":"1acef947-6310-4ac0-bc84-a06d91f84cb6","Type":"ContainerDied","Data":"4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc"} Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.846571 4772 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.848429 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.848470 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.848484 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.848605 4772 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.855562 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:20Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.855597 4772 kubelet_node_status.go:115] "Node was previously registered" node="crc" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.855967 4772 kubelet_node_status.go:79] "Successfully registered node" node="crc" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.857003 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.857031 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.857041 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.857056 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.857065 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:20Z","lastTransitionTime":"2026-01-27T15:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.872913 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:20Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:20 crc kubenswrapper[4772]: E0127 15:07:20.878283 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:20Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.881964 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.881996 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.882007 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.882037 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.882051 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:20Z","lastTransitionTime":"2026-01-27T15:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.890189 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:20Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:20 crc kubenswrapper[4772]: E0127 15:07:20.899732 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:20Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.902506 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:20Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.904507 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.904553 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.904565 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.904583 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.904596 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:20Z","lastTransitionTime":"2026-01-27T15:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.916797 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:20Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:20 crc kubenswrapper[4772]: E0127 15:07:20.918641 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:20Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.924698 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.924941 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.925040 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.925125 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.925220 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:20Z","lastTransitionTime":"2026-01-27T15:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.932796 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:20Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:20 crc kubenswrapper[4772]: E0127 15:07:20.938117 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:20Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.942134 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.942185 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.942197 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.942213 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.942223 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:20Z","lastTransitionTime":"2026-01-27T15:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.943356 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:20Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:20 crc kubenswrapper[4772]: E0127 15:07:20.955615 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:20Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:20 crc kubenswrapper[4772]: E0127 15:07:20.955729 4772 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.957435 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.957482 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.957493 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.957510 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.957523 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:20Z","lastTransitionTime":"2026-01-27T15:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.965531 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:20Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.978887 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:20Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:20 crc kubenswrapper[4772]: I0127 15:07:20.990830 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:20Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.004047 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:21Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.016151 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:21Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.035041 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:21Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.049131 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:21Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.059226 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.059329 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.059386 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.059443 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.059498 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:21Z","lastTransitionTime":"2026-01-27T15:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.069894 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:21Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.162207 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.162247 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.162258 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.162275 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.162287 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:21Z","lastTransitionTime":"2026-01-27T15:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.264091 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.264370 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.264454 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.264521 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.264585 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:21Z","lastTransitionTime":"2026-01-27T15:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.367047 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.367328 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.367410 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.367492 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.367553 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:21Z","lastTransitionTime":"2026-01-27T15:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.469820 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.469854 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.469863 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.469878 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.469886 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:21Z","lastTransitionTime":"2026-01-27T15:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.572605 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.572664 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.572677 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.572707 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.572719 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:21Z","lastTransitionTime":"2026-01-27T15:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.630357 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 10:40:59.157086387 +0000 UTC Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.662905 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:21 crc kubenswrapper[4772]: E0127 15:07:21.663047 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.674848 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.675010 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.675037 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.675064 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.675080 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:21Z","lastTransitionTime":"2026-01-27T15:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.777019 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.777055 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.777068 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.777083 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.777093 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:21Z","lastTransitionTime":"2026-01-27T15:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.847762 4772 generic.go:334] "Generic (PLEG): container finished" podID="1acef947-6310-4ac0-bc84-a06d91f84cb6" containerID="3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd" exitCode=0 Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.847812 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" event={"ID":"1acef947-6310-4ac0-bc84-a06d91f84cb6","Type":"ContainerDied","Data":"3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd"} Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.851777 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerStarted","Data":"0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae"} Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.879844 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.879883 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.879891 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.879905 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.879915 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:21Z","lastTransitionTime":"2026-01-27T15:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.882695 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:21Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.895907 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:21Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.906850 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:21Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.921129 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:21Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.935460 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:21Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.948961 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:21Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.964577 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:21Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.980674 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:21Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.982047 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.982083 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.982292 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.982325 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.982362 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:21Z","lastTransitionTime":"2026-01-27T15:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:21 crc kubenswrapper[4772]: I0127 15:07:21.994631 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:21Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.010773 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:22Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.029888 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:22Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.046263 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:22Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.060580 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:22Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.075724 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:22Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.084670 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.084708 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.084718 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.084731 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.084741 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:22Z","lastTransitionTime":"2026-01-27T15:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.087859 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:22Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.186923 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.186958 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.186966 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.186979 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.186988 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:22Z","lastTransitionTime":"2026-01-27T15:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.289668 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.289716 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.289724 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.289740 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.289750 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:22Z","lastTransitionTime":"2026-01-27T15:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.392146 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.392215 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.392252 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.392269 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.392278 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:22Z","lastTransitionTime":"2026-01-27T15:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.449737 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.449851 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.449882 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.449916 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.449937 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:22 crc kubenswrapper[4772]: E0127 15:07:22.450051 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 27 15:07:22 crc kubenswrapper[4772]: E0127 15:07:22.450097 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 27 15:07:22 crc kubenswrapper[4772]: E0127 15:07:22.450105 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 27 15:07:22 crc kubenswrapper[4772]: E0127 15:07:22.450053 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:07:30.450025756 +0000 UTC m=+36.430634894 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:07:22 crc kubenswrapper[4772]: E0127 15:07:22.450110 4772 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:22 crc kubenswrapper[4772]: E0127 15:07:22.450190 4772 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 27 15:07:22 crc kubenswrapper[4772]: E0127 15:07:22.450127 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 27 15:07:22 crc kubenswrapper[4772]: E0127 15:07:22.450247 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:30.450219712 +0000 UTC m=+36.430828830 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:22 crc kubenswrapper[4772]: E0127 15:07:22.450281 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:30.450263363 +0000 UTC m=+36.430872561 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 27 15:07:22 crc kubenswrapper[4772]: E0127 15:07:22.450255 4772 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:22 crc kubenswrapper[4772]: E0127 15:07:22.450060 4772 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 27 15:07:22 crc kubenswrapper[4772]: E0127 15:07:22.450368 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:30.450356156 +0000 UTC m=+36.430965334 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:22 crc kubenswrapper[4772]: E0127 15:07:22.450394 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:30.450382146 +0000 UTC m=+36.430991324 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.494024 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.494056 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.494067 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.494081 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.494102 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:22Z","lastTransitionTime":"2026-01-27T15:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.596993 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.597056 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.597074 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.597099 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.597121 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:22Z","lastTransitionTime":"2026-01-27T15:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.630561 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 07:09:42.105757713 +0000 UTC Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.662367 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:22 crc kubenswrapper[4772]: E0127 15:07:22.662514 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.662595 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:22 crc kubenswrapper[4772]: E0127 15:07:22.662747 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.700035 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.700062 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.700071 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.700084 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.700092 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:22Z","lastTransitionTime":"2026-01-27T15:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.802777 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.802820 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.802829 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.802844 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.802853 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:22Z","lastTransitionTime":"2026-01-27T15:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.859694 4772 generic.go:334] "Generic (PLEG): container finished" podID="1acef947-6310-4ac0-bc84-a06d91f84cb6" containerID="c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57" exitCode=0 Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.859743 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" event={"ID":"1acef947-6310-4ac0-bc84-a06d91f84cb6","Type":"ContainerDied","Data":"c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57"} Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.873351 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:22Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.885760 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:22Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.904084 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:22Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.915718 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.915757 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.915771 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.915788 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.915800 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:22Z","lastTransitionTime":"2026-01-27T15:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.917531 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:22Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.935699 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:22Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:22 crc kubenswrapper[4772]: I0127 15:07:22.954094 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:22Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:22.964789 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:22Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:22.984861 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:22Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:22.997149 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:22Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.009514 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:23Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.018689 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.018724 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.018733 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.018748 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.018757 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:23Z","lastTransitionTime":"2026-01-27T15:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.023586 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:23Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.037469 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:23Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.056874 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:23Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.070128 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:23Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.080114 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:23Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.121945 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.121983 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.121996 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.122013 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.122026 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:23Z","lastTransitionTime":"2026-01-27T15:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.224477 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.224529 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.224538 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.224551 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.224561 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:23Z","lastTransitionTime":"2026-01-27T15:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.327462 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.327509 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.327522 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.327542 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.327557 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:23Z","lastTransitionTime":"2026-01-27T15:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.431495 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.431540 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.431555 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.431577 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.431594 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:23Z","lastTransitionTime":"2026-01-27T15:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.534156 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.534385 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.534468 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.534551 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.534627 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:23Z","lastTransitionTime":"2026-01-27T15:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.631016 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 18:36:14.169139746 +0000 UTC Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.636880 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.636921 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.636933 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.636949 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.636960 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:23Z","lastTransitionTime":"2026-01-27T15:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.662888 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:23 crc kubenswrapper[4772]: E0127 15:07:23.663022 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.739318 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.739357 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.739368 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.739384 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.739396 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:23Z","lastTransitionTime":"2026-01-27T15:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.841802 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.841868 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.841880 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.841897 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.841908 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:23Z","lastTransitionTime":"2026-01-27T15:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.870821 4772 generic.go:334] "Generic (PLEG): container finished" podID="1acef947-6310-4ac0-bc84-a06d91f84cb6" containerID="d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8" exitCode=0 Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.870866 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" event={"ID":"1acef947-6310-4ac0-bc84-a06d91f84cb6","Type":"ContainerDied","Data":"d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8"} Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.882958 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:23Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.905945 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:23Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.920256 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:23Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.933958 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:23Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.945047 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.945082 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.945090 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.945104 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.945114 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:23Z","lastTransitionTime":"2026-01-27T15:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.955556 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:23Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.973434 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:23Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.983713 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:23Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:23 crc kubenswrapper[4772]: I0127 15:07:23.996305 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:23Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.008225 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.021604 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.033025 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.048080 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.051366 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.051389 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.051398 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.051411 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.051419 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:24Z","lastTransitionTime":"2026-01-27T15:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.061912 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.074936 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.091587 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.153353 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.153383 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.153393 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.153409 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.153421 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:24Z","lastTransitionTime":"2026-01-27T15:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.256010 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.256051 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.256061 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.256080 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.256091 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:24Z","lastTransitionTime":"2026-01-27T15:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.358955 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.358999 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.359010 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.359027 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.359038 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:24Z","lastTransitionTime":"2026-01-27T15:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.456234 4772 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.461991 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.462033 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.462045 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.462061 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.462074 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:24Z","lastTransitionTime":"2026-01-27T15:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.564680 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.564722 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.564734 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.564750 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.564758 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:24Z","lastTransitionTime":"2026-01-27T15:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.632540 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 19:34:51.348688904 +0000 UTC Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.662905 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.663179 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:24 crc kubenswrapper[4772]: E0127 15:07:24.663365 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:07:24 crc kubenswrapper[4772]: E0127 15:07:24.663793 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.671044 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.671100 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.671112 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.671126 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.671137 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:24Z","lastTransitionTime":"2026-01-27T15:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.684910 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.701828 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.718474 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.732287 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.752209 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.764387 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.774006 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.774051 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.774061 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.774076 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.774086 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:24Z","lastTransitionTime":"2026-01-27T15:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.775066 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.791734 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.811574 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.826563 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.838673 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.853250 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.870528 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.879602 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.879648 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.879656 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.879673 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.879682 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:24Z","lastTransitionTime":"2026-01-27T15:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.882066 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.884249 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" event={"ID":"1acef947-6310-4ac0-bc84-a06d91f84cb6","Type":"ContainerStarted","Data":"790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490"} Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.889443 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerStarted","Data":"a1941175c2adff112ad9fafa1e24e1fec6c564305bb9ca8a437e2f0e9124dfec"} Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.889695 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.893948 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.907737 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.913232 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.920557 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.930365 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.938358 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.950602 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.961095 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.971999 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.983508 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.983552 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.983564 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.983582 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.983597 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:24Z","lastTransitionTime":"2026-01-27T15:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:24 crc kubenswrapper[4772]: I0127 15:07:24.989534 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1941175c2adff112ad9fafa1e24e1fec6c564305bb9ca8a437e2f0e9124dfec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.000068 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.011031 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.024511 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.036653 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.056013 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.069142 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.078391 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.085723 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.085775 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.085785 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.085800 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.085813 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:25Z","lastTransitionTime":"2026-01-27T15:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.096479 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.108714 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.116504 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.127153 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.139005 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.149723 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.160810 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.175323 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.188762 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.188800 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.188811 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.188827 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.188838 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:25Z","lastTransitionTime":"2026-01-27T15:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.189321 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.199815 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.217734 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1941175c2adff112ad9fafa1e24e1fec6c564305bb9ca8a437e2f0e9124dfec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.231767 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.245125 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.261521 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.274726 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.291629 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.291693 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.291710 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.291733 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.291752 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:25Z","lastTransitionTime":"2026-01-27T15:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.394217 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.394270 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.394282 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.394302 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.394315 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:25Z","lastTransitionTime":"2026-01-27T15:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.496290 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.496333 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.496344 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.496360 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.496368 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:25Z","lastTransitionTime":"2026-01-27T15:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.598832 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.598882 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.598898 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.598947 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.598961 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:25Z","lastTransitionTime":"2026-01-27T15:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.633423 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 23:39:04.687424573 +0000 UTC Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.662810 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:25 crc kubenswrapper[4772]: E0127 15:07:25.662998 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.701759 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.701798 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.701809 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.701825 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.701835 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:25Z","lastTransitionTime":"2026-01-27T15:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.804672 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.804996 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.805098 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.805228 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.805343 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:25Z","lastTransitionTime":"2026-01-27T15:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.892506 4772 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.893030 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.909397 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.909459 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.909471 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.909489 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.909499 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:25Z","lastTransitionTime":"2026-01-27T15:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.917964 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.931449 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.946727 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.959471 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:25 crc kubenswrapper[4772]: I0127 15:07:25.972313 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:25Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.005965 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:26Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.011600 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.011642 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.011654 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.011672 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.011686 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:26Z","lastTransitionTime":"2026-01-27T15:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.033157 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:26Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.052293 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:26Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.072484 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1941175c2adff112ad9fafa1e24e1fec6c564305bb9ca8a437e2f0e9124dfec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:26Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.085350 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:26Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.106030 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:26Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.114467 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.114514 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.114523 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.114538 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.114548 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:26Z","lastTransitionTime":"2026-01-27T15:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.121494 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:26Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.133968 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:26Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.153295 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:26Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.164736 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:26Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.173489 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:26Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.216912 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.216962 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.216975 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.216994 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.217007 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:26Z","lastTransitionTime":"2026-01-27T15:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.320112 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.320154 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.320198 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.320232 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.320266 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:26Z","lastTransitionTime":"2026-01-27T15:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.422311 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.422360 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.422372 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.422389 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.422401 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:26Z","lastTransitionTime":"2026-01-27T15:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.524675 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.524745 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.524757 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.524770 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.524796 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:26Z","lastTransitionTime":"2026-01-27T15:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.627773 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.627812 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.627823 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.627838 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.627848 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:26Z","lastTransitionTime":"2026-01-27T15:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.634183 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 20:05:07.996061563 +0000 UTC Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.662595 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.662657 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:26 crc kubenswrapper[4772]: E0127 15:07:26.662757 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:07:26 crc kubenswrapper[4772]: E0127 15:07:26.663061 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.730117 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.730156 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.730181 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.730198 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.730209 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:26Z","lastTransitionTime":"2026-01-27T15:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.832638 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.832679 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.832693 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.832709 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.832720 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:26Z","lastTransitionTime":"2026-01-27T15:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.895468 4772 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.985773 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.985822 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.985831 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.985845 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:26 crc kubenswrapper[4772]: I0127 15:07:26.985855 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:26Z","lastTransitionTime":"2026-01-27T15:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.088596 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.088627 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.088638 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.088652 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.088661 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:27Z","lastTransitionTime":"2026-01-27T15:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.191200 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.191248 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.191257 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.191270 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.191280 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:27Z","lastTransitionTime":"2026-01-27T15:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.295591 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.296687 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.296770 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.296839 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.296850 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:27Z","lastTransitionTime":"2026-01-27T15:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.399388 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.399426 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.399434 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.399448 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.399456 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:27Z","lastTransitionTime":"2026-01-27T15:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.502360 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.502424 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.502442 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.502465 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.502482 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:27Z","lastTransitionTime":"2026-01-27T15:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.604378 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.604403 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.604413 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.604428 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.604437 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:27Z","lastTransitionTime":"2026-01-27T15:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.634475 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 22:03:47.636543572 +0000 UTC Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.661978 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:27 crc kubenswrapper[4772]: E0127 15:07:27.662198 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.707706 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.707765 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.707780 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.707801 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.707814 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:27Z","lastTransitionTime":"2026-01-27T15:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.810763 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.810810 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.810821 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.810837 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.810849 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:27Z","lastTransitionTime":"2026-01-27T15:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.900496 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n2khk_736264c8-cd18-479a-88ba-e1ec15dbfdae/ovnkube-controller/0.log" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.903109 4772 generic.go:334] "Generic (PLEG): container finished" podID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerID="a1941175c2adff112ad9fafa1e24e1fec6c564305bb9ca8a437e2f0e9124dfec" exitCode=1 Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.903182 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerDied","Data":"a1941175c2adff112ad9fafa1e24e1fec6c564305bb9ca8a437e2f0e9124dfec"} Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.903721 4772 scope.go:117] "RemoveContainer" containerID="a1941175c2adff112ad9fafa1e24e1fec6c564305bb9ca8a437e2f0e9124dfec" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.912964 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.913011 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.913022 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.913038 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.913051 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:27Z","lastTransitionTime":"2026-01-27T15:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.919943 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:27Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.930504 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:27Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.943459 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:27Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.955865 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:27Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.967487 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:27Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:27 crc kubenswrapper[4772]: I0127 15:07:27.979781 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:27Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.001575 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1941175c2adff112ad9fafa1e24e1fec6c564305bb9ca8a437e2f0e9124dfec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1941175c2adff112ad9fafa1e24e1fec6c564305bb9ca8a437e2f0e9124dfec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:27Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:140\\\\nI0127 15:07:27.137663 6049 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0127 15:07:27.137944 6049 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0127 15:07:27.138014 6049 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0127 15:07:27.138393 6049 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0127 15:07:27.138457 6049 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0127 15:07:27.138485 6049 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0127 15:07:27.138538 6049 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0127 15:07:27.138563 6049 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0127 15:07:27.138584 6049 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0127 15:07:27.138489 6049 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0127 15:07:27.138607 6049 factory.go:656] Stopping watch factory\\\\nI0127 15:07:27.138647 6049 ovnkube.go:599] Stopped ovnkube\\\\nI0127 15:07:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:27Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.015751 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.015790 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.015802 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.015817 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.015851 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:28Z","lastTransitionTime":"2026-01-27T15:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.018367 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:28Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.040424 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:28Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.058212 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:28Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.075048 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:28Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.087782 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:28Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.097669 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:28Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.119231 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.119269 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.119281 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.119296 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.119307 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:28Z","lastTransitionTime":"2026-01-27T15:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.122248 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:28Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.133072 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:28Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.221820 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.221902 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.221913 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.221933 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.221946 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:28Z","lastTransitionTime":"2026-01-27T15:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.376128 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.376176 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.376186 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.376201 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.376212 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:28Z","lastTransitionTime":"2026-01-27T15:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.482120 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.482595 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.482611 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.482631 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.482646 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:28Z","lastTransitionTime":"2026-01-27T15:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.585246 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.585285 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.585296 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.585310 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.585320 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:28Z","lastTransitionTime":"2026-01-27T15:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.634955 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 08:13:17.435613329 +0000 UTC Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.662361 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:28 crc kubenswrapper[4772]: E0127 15:07:28.662475 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.662790 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:28 crc kubenswrapper[4772]: E0127 15:07:28.662863 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.687562 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.687597 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.687607 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.687620 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.687630 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:28Z","lastTransitionTime":"2026-01-27T15:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.790011 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.790092 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.790109 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.790126 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.790160 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:28Z","lastTransitionTime":"2026-01-27T15:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.892321 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.892363 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.892374 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.892388 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.892398 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:28Z","lastTransitionTime":"2026-01-27T15:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.908083 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n2khk_736264c8-cd18-479a-88ba-e1ec15dbfdae/ovnkube-controller/0.log" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.910646 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerStarted","Data":"44564ab0212c32423d179406999749f443cae1bd72ea8d12ff4411e23de77d97"} Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.910749 4772 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.927811 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:28Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.940311 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:28Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.951445 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:28Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.971760 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44564ab0212c32423d179406999749f443cae1bd72ea8d12ff4411e23de77d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1941175c2adff112ad9fafa1e24e1fec6c564305bb9ca8a437e2f0e9124dfec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:27Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:140\\\\nI0127 15:07:27.137663 6049 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0127 15:07:27.137944 6049 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0127 15:07:27.138014 6049 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0127 15:07:27.138393 6049 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0127 15:07:27.138457 6049 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0127 15:07:27.138485 6049 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0127 15:07:27.138538 6049 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0127 15:07:27.138563 6049 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0127 15:07:27.138584 6049 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0127 15:07:27.138489 6049 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0127 15:07:27.138607 6049 factory.go:656] Stopping watch factory\\\\nI0127 15:07:27.138647 6049 ovnkube.go:599] Stopped ovnkube\\\\nI0127 15:07:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:28Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.991507 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:28Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.995244 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.995317 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.995331 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.995397 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:28 crc kubenswrapper[4772]: I0127 15:07:28.995474 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:28Z","lastTransitionTime":"2026-01-27T15:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.008549 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.022616 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.037076 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.055815 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.073040 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx"] Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.073705 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.075588 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.075685 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.075789 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.081467 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3cc8fde5-4905-4fb1-b683-27ea4921b462-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-wkvpx\" (UID: \"3cc8fde5-4905-4fb1-b683-27ea4921b462\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.081560 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghx7x\" (UniqueName: \"kubernetes.io/projected/3cc8fde5-4905-4fb1-b683-27ea4921b462-kube-api-access-ghx7x\") pod \"ovnkube-control-plane-749d76644c-wkvpx\" (UID: \"3cc8fde5-4905-4fb1-b683-27ea4921b462\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.081599 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3cc8fde5-4905-4fb1-b683-27ea4921b462-env-overrides\") pod \"ovnkube-control-plane-749d76644c-wkvpx\" (UID: \"3cc8fde5-4905-4fb1-b683-27ea4921b462\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.081641 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3cc8fde5-4905-4fb1-b683-27ea4921b462-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-wkvpx\" (UID: \"3cc8fde5-4905-4fb1-b683-27ea4921b462\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.086023 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.097717 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.097748 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.097757 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.097773 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.097784 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:29Z","lastTransitionTime":"2026-01-27T15:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.099256 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.109567 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.118525 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.127855 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.140134 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.153226 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.162873 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.182878 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghx7x\" (UniqueName: \"kubernetes.io/projected/3cc8fde5-4905-4fb1-b683-27ea4921b462-kube-api-access-ghx7x\") pod \"ovnkube-control-plane-749d76644c-wkvpx\" (UID: \"3cc8fde5-4905-4fb1-b683-27ea4921b462\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.182933 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3cc8fde5-4905-4fb1-b683-27ea4921b462-env-overrides\") pod \"ovnkube-control-plane-749d76644c-wkvpx\" (UID: \"3cc8fde5-4905-4fb1-b683-27ea4921b462\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.182971 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3cc8fde5-4905-4fb1-b683-27ea4921b462-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-wkvpx\" (UID: \"3cc8fde5-4905-4fb1-b683-27ea4921b462\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.183043 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3cc8fde5-4905-4fb1-b683-27ea4921b462-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-wkvpx\" (UID: \"3cc8fde5-4905-4fb1-b683-27ea4921b462\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.183698 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3cc8fde5-4905-4fb1-b683-27ea4921b462-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-wkvpx\" (UID: \"3cc8fde5-4905-4fb1-b683-27ea4921b462\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.184457 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3cc8fde5-4905-4fb1-b683-27ea4921b462-env-overrides\") pod \"ovnkube-control-plane-749d76644c-wkvpx\" (UID: \"3cc8fde5-4905-4fb1-b683-27ea4921b462\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.186923 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44564ab0212c32423d179406999749f443cae1bd72ea8d12ff4411e23de77d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1941175c2adff112ad9fafa1e24e1fec6c564305bb9ca8a437e2f0e9124dfec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:27Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:140\\\\nI0127 15:07:27.137663 6049 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0127 15:07:27.137944 6049 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0127 15:07:27.138014 6049 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0127 15:07:27.138393 6049 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0127 15:07:27.138457 6049 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0127 15:07:27.138485 6049 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0127 15:07:27.138538 6049 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0127 15:07:27.138563 6049 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0127 15:07:27.138584 6049 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0127 15:07:27.138489 6049 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0127 15:07:27.138607 6049 factory.go:656] Stopping watch factory\\\\nI0127 15:07:27.138647 6049 ovnkube.go:599] Stopped ovnkube\\\\nI0127 15:07:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.190140 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3cc8fde5-4905-4fb1-b683-27ea4921b462-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-wkvpx\" (UID: \"3cc8fde5-4905-4fb1-b683-27ea4921b462\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.200353 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.200542 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.200599 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.200656 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.200724 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:29Z","lastTransitionTime":"2026-01-27T15:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.202284 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghx7x\" (UniqueName: \"kubernetes.io/projected/3cc8fde5-4905-4fb1-b683-27ea4921b462-kube-api-access-ghx7x\") pod \"ovnkube-control-plane-749d76644c-wkvpx\" (UID: \"3cc8fde5-4905-4fb1-b683-27ea4921b462\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.202937 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.215349 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.229555 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.241577 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.258450 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.269294 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.279861 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.291347 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.302219 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.304283 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.304409 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.304515 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.304617 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.304707 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:29Z","lastTransitionTime":"2026-01-27T15:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.312752 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.322477 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.333277 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8fde5-4905-4fb1-b683-27ea4921b462\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wkvpx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.385518 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.406953 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.407007 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.407023 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.407045 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.407063 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:29Z","lastTransitionTime":"2026-01-27T15:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.509395 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.509422 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.509431 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.509445 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.509455 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:29Z","lastTransitionTime":"2026-01-27T15:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.612033 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.612066 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.612076 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.612091 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.612100 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:29Z","lastTransitionTime":"2026-01-27T15:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.635955 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 00:18:35.166323642 +0000 UTC Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.662395 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:29 crc kubenswrapper[4772]: E0127 15:07:29.662516 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.714960 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.714999 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.715007 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.715021 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.715031 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:29Z","lastTransitionTime":"2026-01-27T15:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.817443 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.817487 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.817498 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.817514 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.817526 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:29Z","lastTransitionTime":"2026-01-27T15:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.915811 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n2khk_736264c8-cd18-479a-88ba-e1ec15dbfdae/ovnkube-controller/1.log" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.916577 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n2khk_736264c8-cd18-479a-88ba-e1ec15dbfdae/ovnkube-controller/0.log" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.918961 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.918997 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.919009 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.919025 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.919036 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:29Z","lastTransitionTime":"2026-01-27T15:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.919520 4772 generic.go:334] "Generic (PLEG): container finished" podID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerID="44564ab0212c32423d179406999749f443cae1bd72ea8d12ff4411e23de77d97" exitCode=1 Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.919580 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerDied","Data":"44564ab0212c32423d179406999749f443cae1bd72ea8d12ff4411e23de77d97"} Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.919608 4772 scope.go:117] "RemoveContainer" containerID="a1941175c2adff112ad9fafa1e24e1fec6c564305bb9ca8a437e2f0e9124dfec" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.920190 4772 scope.go:117] "RemoveContainer" containerID="44564ab0212c32423d179406999749f443cae1bd72ea8d12ff4411e23de77d97" Jan 27 15:07:29 crc kubenswrapper[4772]: E0127 15:07:29.920344 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-n2khk_openshift-ovn-kubernetes(736264c8-cd18-479a-88ba-e1ec15dbfdae)\"" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.923487 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" event={"ID":"3cc8fde5-4905-4fb1-b683-27ea4921b462","Type":"ContainerStarted","Data":"7b47fc0778ba4c5a1e12700735e6f9c52a7341b9eac61071607902a6ec8daf02"} Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.923540 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" event={"ID":"3cc8fde5-4905-4fb1-b683-27ea4921b462","Type":"ContainerStarted","Data":"1d0c05a9e18a9cfab5b84d6eca0e5dd9ecaea00074297e641ad4ddb1a294ea8e"} Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.932423 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.946353 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.956449 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.966134 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.977857 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8fde5-4905-4fb1-b683-27ea4921b462\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wkvpx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:29 crc kubenswrapper[4772]: I0127 15:07:29.998380 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:29Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.014750 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.021310 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.021338 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.021347 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.021360 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.021370 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:30Z","lastTransitionTime":"2026-01-27T15:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.032770 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.054705 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44564ab0212c32423d179406999749f443cae1bd72ea8d12ff4411e23de77d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1941175c2adff112ad9fafa1e24e1fec6c564305bb9ca8a437e2f0e9124dfec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:27Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:140\\\\nI0127 15:07:27.137663 6049 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0127 15:07:27.137944 6049 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0127 15:07:27.138014 6049 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0127 15:07:27.138393 6049 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0127 15:07:27.138457 6049 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0127 15:07:27.138485 6049 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0127 15:07:27.138538 6049 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0127 15:07:27.138563 6049 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0127 15:07:27.138584 6049 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0127 15:07:27.138489 6049 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0127 15:07:27.138607 6049 factory.go:656] Stopping watch factory\\\\nI0127 15:07:27.138647 6049 ovnkube.go:599] Stopped ovnkube\\\\nI0127 15:07:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44564ab0212c32423d179406999749f443cae1bd72ea8d12ff4411e23de77d97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:28Z\\\",\\\"message\\\":\\\"twork-check-target template LB for network=default: []services.LB{}\\\\nI0127 15:07:28.751281 6196 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.067880 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.080331 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.095101 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.107763 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.123384 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.123418 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.123429 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.123443 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.123454 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:30Z","lastTransitionTime":"2026-01-27T15:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.126572 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.140787 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.154818 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.225760 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.225801 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.225810 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.225825 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.225834 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:30Z","lastTransitionTime":"2026-01-27T15:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.328463 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.328504 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.328515 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.328529 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.328537 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:30Z","lastTransitionTime":"2026-01-27T15:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.431002 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.431345 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.431420 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.431529 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.431625 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:30Z","lastTransitionTime":"2026-01-27T15:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.493941 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.494051 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.494077 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.494107 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.494127 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:30 crc kubenswrapper[4772]: E0127 15:07:30.494246 4772 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 27 15:07:30 crc kubenswrapper[4772]: E0127 15:07:30.494294 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:46.494280857 +0000 UTC m=+52.474889955 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 27 15:07:30 crc kubenswrapper[4772]: E0127 15:07:30.494350 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:07:46.494301887 +0000 UTC m=+52.474910995 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:07:30 crc kubenswrapper[4772]: E0127 15:07:30.494594 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 27 15:07:30 crc kubenswrapper[4772]: E0127 15:07:30.494617 4772 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 27 15:07:30 crc kubenswrapper[4772]: E0127 15:07:30.494626 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 27 15:07:30 crc kubenswrapper[4772]: E0127 15:07:30.494651 4772 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:30 crc kubenswrapper[4772]: E0127 15:07:30.494663 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:46.494653517 +0000 UTC m=+52.475262825 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 27 15:07:30 crc kubenswrapper[4772]: E0127 15:07:30.494736 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:46.494723739 +0000 UTC m=+52.475333017 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:30 crc kubenswrapper[4772]: E0127 15:07:30.495048 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 27 15:07:30 crc kubenswrapper[4772]: E0127 15:07:30.495152 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 27 15:07:30 crc kubenswrapper[4772]: E0127 15:07:30.495273 4772 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:30 crc kubenswrapper[4772]: E0127 15:07:30.495518 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:46.495492361 +0000 UTC m=+52.476101469 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.534380 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.534443 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.534455 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.534478 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.534493 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:30Z","lastTransitionTime":"2026-01-27T15:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.546469 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-ql2vx"] Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.547007 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:30 crc kubenswrapper[4772]: E0127 15:07:30.547739 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.562397 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.576095 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.587796 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.594927 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8l57\" (UniqueName: \"kubernetes.io/projected/371016c8-5a23-427d-aa0a-0faa241d86a7-kube-api-access-v8l57\") pod \"network-metrics-daemon-ql2vx\" (UID: \"371016c8-5a23-427d-aa0a-0faa241d86a7\") " pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.594985 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs\") pod \"network-metrics-daemon-ql2vx\" (UID: \"371016c8-5a23-427d-aa0a-0faa241d86a7\") " pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.598772 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.619096 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.634623 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.636082 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 11:01:09.233295837 +0000 UTC Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.636534 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.636622 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.636636 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.636657 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.636671 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:30Z","lastTransitionTime":"2026-01-27T15:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.646248 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.660245 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.662355 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.662365 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:30 crc kubenswrapper[4772]: E0127 15:07:30.662547 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:07:30 crc kubenswrapper[4772]: E0127 15:07:30.662470 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.673568 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.684263 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.694997 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.695609 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8l57\" (UniqueName: \"kubernetes.io/projected/371016c8-5a23-427d-aa0a-0faa241d86a7-kube-api-access-v8l57\") pod \"network-metrics-daemon-ql2vx\" (UID: \"371016c8-5a23-427d-aa0a-0faa241d86a7\") " pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.695662 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs\") pod \"network-metrics-daemon-ql2vx\" (UID: \"371016c8-5a23-427d-aa0a-0faa241d86a7\") " pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:30 crc kubenswrapper[4772]: E0127 15:07:30.695802 4772 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 27 15:07:30 crc kubenswrapper[4772]: E0127 15:07:30.695876 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs podName:371016c8-5a23-427d-aa0a-0faa241d86a7 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:31.195856915 +0000 UTC m=+37.176466013 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs") pod "network-metrics-daemon-ql2vx" (UID: "371016c8-5a23-427d-aa0a-0faa241d86a7") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.704732 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8fde5-4905-4fb1-b683-27ea4921b462\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wkvpx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.714315 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ql2vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371016c8-5a23-427d-aa0a-0faa241d86a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ql2vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.714857 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8l57\" (UniqueName: \"kubernetes.io/projected/371016c8-5a23-427d-aa0a-0faa241d86a7-kube-api-access-v8l57\") pod \"network-metrics-daemon-ql2vx\" (UID: \"371016c8-5a23-427d-aa0a-0faa241d86a7\") " pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.726875 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.738766 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.738808 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.738820 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.738837 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.738853 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:30Z","lastTransitionTime":"2026-01-27T15:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.740600 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.749498 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.764324 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44564ab0212c32423d179406999749f443cae1bd72ea8d12ff4411e23de77d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1941175c2adff112ad9fafa1e24e1fec6c564305bb9ca8a437e2f0e9124dfec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:27Z\\\",\\\"message\\\":\\\"rmers/externalversions/factory.go:140\\\\nI0127 15:07:27.137663 6049 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0127 15:07:27.137944 6049 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0127 15:07:27.138014 6049 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0127 15:07:27.138393 6049 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0127 15:07:27.138457 6049 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0127 15:07:27.138485 6049 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0127 15:07:27.138538 6049 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0127 15:07:27.138563 6049 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0127 15:07:27.138584 6049 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0127 15:07:27.138489 6049 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0127 15:07:27.138607 6049 factory.go:656] Stopping watch factory\\\\nI0127 15:07:27.138647 6049 ovnkube.go:599] Stopped ovnkube\\\\nI0127 15:07:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44564ab0212c32423d179406999749f443cae1bd72ea8d12ff4411e23de77d97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:28Z\\\",\\\"message\\\":\\\"twork-check-target template LB for network=default: []services.LB{}\\\\nI0127 15:07:28.751281 6196 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.844516 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.844609 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.844645 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.844689 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.844714 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:30Z","lastTransitionTime":"2026-01-27T15:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.928863 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n2khk_736264c8-cd18-479a-88ba-e1ec15dbfdae/ovnkube-controller/1.log" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.932533 4772 scope.go:117] "RemoveContainer" containerID="44564ab0212c32423d179406999749f443cae1bd72ea8d12ff4411e23de77d97" Jan 27 15:07:30 crc kubenswrapper[4772]: E0127 15:07:30.932688 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-n2khk_openshift-ovn-kubernetes(736264c8-cd18-479a-88ba-e1ec15dbfdae)\"" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.935158 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" event={"ID":"3cc8fde5-4905-4fb1-b683-27ea4921b462","Type":"ContainerStarted","Data":"545dcc1be1335a1acf93f16d2e5b4a266dce5b7e736b7c56b80bbf56b3248ced"} Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.945068 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.947337 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.947372 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.947384 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.947400 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.947411 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:30Z","lastTransitionTime":"2026-01-27T15:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.956972 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:30 crc kubenswrapper[4772]: I0127 15:07:30.983962 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44564ab0212c32423d179406999749f443cae1bd72ea8d12ff4411e23de77d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44564ab0212c32423d179406999749f443cae1bd72ea8d12ff4411e23de77d97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:28Z\\\",\\\"message\\\":\\\"twork-check-target template LB for network=default: []services.LB{}\\\\nI0127 15:07:28.751281 6196 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-n2khk_openshift-ovn-kubernetes(736264c8-cd18-479a-88ba-e1ec15dbfdae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:30Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.008064 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.022241 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.036680 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.049428 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.049461 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.049469 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.049482 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.049491 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:31Z","lastTransitionTime":"2026-01-27T15:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.051850 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.065426 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.075114 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.093638 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.107243 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.119245 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.129667 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.140182 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8fde5-4905-4fb1-b683-27ea4921b462\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wkvpx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.151036 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ql2vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371016c8-5a23-427d-aa0a-0faa241d86a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ql2vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.151740 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.151766 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.151778 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.151796 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.151806 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:31Z","lastTransitionTime":"2026-01-27T15:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.163190 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.175740 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.194627 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44564ab0212c32423d179406999749f443cae1bd72ea8d12ff4411e23de77d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44564ab0212c32423d179406999749f443cae1bd72ea8d12ff4411e23de77d97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:28Z\\\",\\\"message\\\":\\\"twork-check-target template LB for network=default: []services.LB{}\\\\nI0127 15:07:28.751281 6196 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-n2khk_openshift-ovn-kubernetes(736264c8-cd18-479a-88ba-e1ec15dbfdae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.200242 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs\") pod \"network-metrics-daemon-ql2vx\" (UID: \"371016c8-5a23-427d-aa0a-0faa241d86a7\") " pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:31 crc kubenswrapper[4772]: E0127 15:07:31.200371 4772 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 27 15:07:31 crc kubenswrapper[4772]: E0127 15:07:31.200432 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs podName:371016c8-5a23-427d-aa0a-0faa241d86a7 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:32.200414192 +0000 UTC m=+38.181023290 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs") pod "network-metrics-daemon-ql2vx" (UID: "371016c8-5a23-427d-aa0a-0faa241d86a7") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.207985 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.219628 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.231330 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.245687 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.253915 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.253948 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.253958 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.253973 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.253985 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:31Z","lastTransitionTime":"2026-01-27T15:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.261034 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.278940 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.295378 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.316451 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.321450 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.321493 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.321503 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.321518 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.321526 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:31Z","lastTransitionTime":"2026-01-27T15:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:31 crc kubenswrapper[4772]: E0127 15:07:31.335007 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.335230 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.339192 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.339249 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.339262 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.339279 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.339292 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:31Z","lastTransitionTime":"2026-01-27T15:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.346540 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: E0127 15:07:31.351510 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.355198 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.355241 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.355256 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.355275 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.355288 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:31Z","lastTransitionTime":"2026-01-27T15:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.360469 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8fde5-4905-4fb1-b683-27ea4921b462\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b47fc0778ba4c5a1e12700735e6f9c52a7341b9eac61071607902a6ec8daf02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://545dcc1be1335a1acf93f16d2e5b4a266dce5b7e736b7c56b80bbf56b3248ced\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wkvpx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: E0127 15:07:31.366720 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.370685 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.370739 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.370751 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.370770 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.370783 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:31Z","lastTransitionTime":"2026-01-27T15:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.371245 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ql2vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371016c8-5a23-427d-aa0a-0faa241d86a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ql2vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: E0127 15:07:31.382727 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.384076 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.386757 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.386784 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.386793 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.386807 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.386816 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:31Z","lastTransitionTime":"2026-01-27T15:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.398952 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: E0127 15:07:31.400650 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: E0127 15:07:31.400824 4772 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.402580 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.402621 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.402630 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.402647 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.402659 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:31Z","lastTransitionTime":"2026-01-27T15:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.412685 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.423494 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.504716 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.504759 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.504769 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.504785 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.504796 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:31Z","lastTransitionTime":"2026-01-27T15:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.606733 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.606773 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.606782 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.606797 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.606806 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:31Z","lastTransitionTime":"2026-01-27T15:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.636208 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 06:54:38.358478323 +0000 UTC Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.662657 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:31 crc kubenswrapper[4772]: E0127 15:07:31.662829 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.663717 4772 scope.go:117] "RemoveContainer" containerID="a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.709407 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.709443 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.709454 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.709469 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.709480 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:31Z","lastTransitionTime":"2026-01-27T15:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.812410 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.812855 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.812874 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.812897 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.812915 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:31Z","lastTransitionTime":"2026-01-27T15:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.915337 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.915385 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.915396 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.915412 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.915424 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:31Z","lastTransitionTime":"2026-01-27T15:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.939446 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.941047 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08"} Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.954789 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.969307 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:31 crc kubenswrapper[4772]: I0127 15:07:31.983707 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.000353 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:31Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.012757 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:32Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.025314 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.025348 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.025360 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.025375 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.025386 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:32Z","lastTransitionTime":"2026-01-27T15:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.034415 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:32Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.057332 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:32Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.069769 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:32Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.085574 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:32Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.098710 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:32Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.108161 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8fde5-4905-4fb1-b683-27ea4921b462\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b47fc0778ba4c5a1e12700735e6f9c52a7341b9eac61071607902a6ec8daf02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://545dcc1be1335a1acf93f16d2e5b4a266dce5b7e736b7c56b80bbf56b3248ced\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wkvpx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:32Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.119977 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ql2vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371016c8-5a23-427d-aa0a-0faa241d86a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ql2vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:32Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.127557 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.127595 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.127603 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.127616 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.127626 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:32Z","lastTransitionTime":"2026-01-27T15:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.132582 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:32Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.144452 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:32Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.155298 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:32Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.173665 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44564ab0212c32423d179406999749f443cae1bd72ea8d12ff4411e23de77d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44564ab0212c32423d179406999749f443cae1bd72ea8d12ff4411e23de77d97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:28Z\\\",\\\"message\\\":\\\"twork-check-target template LB for network=default: []services.LB{}\\\\nI0127 15:07:28.751281 6196 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-n2khk_openshift-ovn-kubernetes(736264c8-cd18-479a-88ba-e1ec15dbfdae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:32Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.189838 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:32Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.208677 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs\") pod \"network-metrics-daemon-ql2vx\" (UID: \"371016c8-5a23-427d-aa0a-0faa241d86a7\") " pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:32 crc kubenswrapper[4772]: E0127 15:07:32.208848 4772 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 27 15:07:32 crc kubenswrapper[4772]: E0127 15:07:32.208913 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs podName:371016c8-5a23-427d-aa0a-0faa241d86a7 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:34.208895959 +0000 UTC m=+40.189505067 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs") pod "network-metrics-daemon-ql2vx" (UID: "371016c8-5a23-427d-aa0a-0faa241d86a7") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.229572 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.229608 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.229618 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.229634 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.229646 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:32Z","lastTransitionTime":"2026-01-27T15:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.331792 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.331831 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.331842 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.331858 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.331869 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:32Z","lastTransitionTime":"2026-01-27T15:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.434468 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.434525 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.434546 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.434573 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.434594 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:32Z","lastTransitionTime":"2026-01-27T15:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.536530 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.536568 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.536578 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.536593 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.536602 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:32Z","lastTransitionTime":"2026-01-27T15:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.636487 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 03:30:48.071867813 +0000 UTC Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.639403 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.639429 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.639438 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.639451 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.639461 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:32Z","lastTransitionTime":"2026-01-27T15:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.665290 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.665369 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:32 crc kubenswrapper[4772]: E0127 15:07:32.665464 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.665532 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:32 crc kubenswrapper[4772]: E0127 15:07:32.665679 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:07:32 crc kubenswrapper[4772]: E0127 15:07:32.666276 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.742494 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.742534 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.742543 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.742558 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.742569 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:32Z","lastTransitionTime":"2026-01-27T15:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.845883 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.845959 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.846226 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.846259 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.846283 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:32Z","lastTransitionTime":"2026-01-27T15:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.948756 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.948801 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.948811 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.948828 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:32 crc kubenswrapper[4772]: I0127 15:07:32.948841 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:32Z","lastTransitionTime":"2026-01-27T15:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.051092 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.051132 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.051147 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.051192 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.051209 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:33Z","lastTransitionTime":"2026-01-27T15:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.153612 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.153676 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.153699 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.153719 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.153733 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:33Z","lastTransitionTime":"2026-01-27T15:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.256438 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.256478 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.256487 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.256500 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.256509 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:33Z","lastTransitionTime":"2026-01-27T15:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.359536 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.359596 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.359625 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.359673 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.359696 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:33Z","lastTransitionTime":"2026-01-27T15:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.461919 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.461962 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.461970 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.461985 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.461996 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:33Z","lastTransitionTime":"2026-01-27T15:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.564253 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.564295 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.564307 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.564326 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.564338 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:33Z","lastTransitionTime":"2026-01-27T15:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.637443 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 19:59:45.302434912 +0000 UTC Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.662096 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:33 crc kubenswrapper[4772]: E0127 15:07:33.662257 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.667041 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.667106 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.667117 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.667129 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.667211 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:33Z","lastTransitionTime":"2026-01-27T15:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.769625 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.769664 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.769677 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.769694 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.769705 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:33Z","lastTransitionTime":"2026-01-27T15:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.872428 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.872461 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.872471 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.872484 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.872493 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:33Z","lastTransitionTime":"2026-01-27T15:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.975240 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.975284 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.975293 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.975308 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:33 crc kubenswrapper[4772]: I0127 15:07:33.975317 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:33Z","lastTransitionTime":"2026-01-27T15:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.079897 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.079950 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.079968 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.079992 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.080010 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:34Z","lastTransitionTime":"2026-01-27T15:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.181695 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.181731 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.181740 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.181752 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.181762 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:34Z","lastTransitionTime":"2026-01-27T15:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.230212 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs\") pod \"network-metrics-daemon-ql2vx\" (UID: \"371016c8-5a23-427d-aa0a-0faa241d86a7\") " pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:34 crc kubenswrapper[4772]: E0127 15:07:34.230373 4772 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 27 15:07:34 crc kubenswrapper[4772]: E0127 15:07:34.230428 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs podName:371016c8-5a23-427d-aa0a-0faa241d86a7 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:38.230413921 +0000 UTC m=+44.211023019 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs") pod "network-metrics-daemon-ql2vx" (UID: "371016c8-5a23-427d-aa0a-0faa241d86a7") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.284511 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.284545 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.284556 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.284570 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.284581 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:34Z","lastTransitionTime":"2026-01-27T15:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.387197 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.387239 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.387250 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.387267 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.387279 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:34Z","lastTransitionTime":"2026-01-27T15:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.489405 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.489453 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.489465 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.489484 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.489495 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:34Z","lastTransitionTime":"2026-01-27T15:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.591975 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.592009 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.592020 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.592035 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.592045 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:34Z","lastTransitionTime":"2026-01-27T15:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.637628 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 11:57:51.124484803 +0000 UTC Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.662916 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:34 crc kubenswrapper[4772]: E0127 15:07:34.663033 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.663325 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:34 crc kubenswrapper[4772]: E0127 15:07:34.663405 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.663454 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:34 crc kubenswrapper[4772]: E0127 15:07:34.663531 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.677286 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.689377 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.693576 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.693613 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.693624 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.693637 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.693645 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:34Z","lastTransitionTime":"2026-01-27T15:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.699573 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.717253 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44564ab0212c32423d179406999749f443cae1bd72ea8d12ff4411e23de77d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44564ab0212c32423d179406999749f443cae1bd72ea8d12ff4411e23de77d97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:28Z\\\",\\\"message\\\":\\\"twork-check-target template LB for network=default: []services.LB{}\\\\nI0127 15:07:28.751281 6196 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-n2khk_openshift-ovn-kubernetes(736264c8-cd18-479a-88ba-e1ec15dbfdae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.731828 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.746575 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.759676 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.771016 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.788469 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.796773 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.796808 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.796819 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.796835 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.796846 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:34Z","lastTransitionTime":"2026-01-27T15:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.802287 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.813927 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.829584 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.849545 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.867797 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.889868 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.900769 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.900804 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.900814 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.900829 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.900840 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:34Z","lastTransitionTime":"2026-01-27T15:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.905400 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8fde5-4905-4fb1-b683-27ea4921b462\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b47fc0778ba4c5a1e12700735e6f9c52a7341b9eac61071607902a6ec8daf02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://545dcc1be1335a1acf93f16d2e5b4a266dce5b7e736b7c56b80bbf56b3248ced\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wkvpx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:34 crc kubenswrapper[4772]: I0127 15:07:34.917606 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ql2vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371016c8-5a23-427d-aa0a-0faa241d86a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ql2vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.003698 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.003736 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.003748 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.003766 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.003779 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:35Z","lastTransitionTime":"2026-01-27T15:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.138317 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.138780 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.138871 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.138945 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.139106 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:35Z","lastTransitionTime":"2026-01-27T15:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.241882 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.241933 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.241946 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.241964 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.241977 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:35Z","lastTransitionTime":"2026-01-27T15:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.344742 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.344789 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.344798 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.344941 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.344964 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:35Z","lastTransitionTime":"2026-01-27T15:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.447392 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.447450 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.447466 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.447484 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.447498 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:35Z","lastTransitionTime":"2026-01-27T15:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.549808 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.549842 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.549858 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.549875 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.549885 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:35Z","lastTransitionTime":"2026-01-27T15:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.638452 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 21:15:00.125298241 +0000 UTC Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.651470 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.651503 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.651514 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.651531 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.651543 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:35Z","lastTransitionTime":"2026-01-27T15:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.662759 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:35 crc kubenswrapper[4772]: E0127 15:07:35.662902 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.754368 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.754400 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.754409 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.754422 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.754430 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:35Z","lastTransitionTime":"2026-01-27T15:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.856307 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.856355 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.856369 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.856385 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.856394 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:35Z","lastTransitionTime":"2026-01-27T15:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.958480 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.958539 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.958547 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.958562 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:35 crc kubenswrapper[4772]: I0127 15:07:35.958573 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:35Z","lastTransitionTime":"2026-01-27T15:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.060595 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.060631 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.060645 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.060661 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.060674 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:36Z","lastTransitionTime":"2026-01-27T15:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.162667 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.162701 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.162711 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.162726 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.162736 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:36Z","lastTransitionTime":"2026-01-27T15:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.265975 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.266058 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.266081 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.266106 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.266123 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:36Z","lastTransitionTime":"2026-01-27T15:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.369413 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.369455 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.369465 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.369482 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.369492 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:36Z","lastTransitionTime":"2026-01-27T15:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.472897 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.472969 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.472990 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.473016 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.473037 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:36Z","lastTransitionTime":"2026-01-27T15:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.575531 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.576323 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.576384 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.576516 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.576583 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:36Z","lastTransitionTime":"2026-01-27T15:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.639325 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 02:23:52.257149762 +0000 UTC Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.662257 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.662273 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.662478 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:36 crc kubenswrapper[4772]: E0127 15:07:36.662560 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:07:36 crc kubenswrapper[4772]: E0127 15:07:36.662653 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:07:36 crc kubenswrapper[4772]: E0127 15:07:36.662754 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.680210 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.680269 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.680280 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.680297 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.680309 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:36Z","lastTransitionTime":"2026-01-27T15:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.782747 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.782784 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.782797 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.782811 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.782822 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:36Z","lastTransitionTime":"2026-01-27T15:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.885569 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.885612 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.885630 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.885648 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.885659 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:36Z","lastTransitionTime":"2026-01-27T15:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.988006 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.988047 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.988062 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.988081 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:36 crc kubenswrapper[4772]: I0127 15:07:36.988094 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:36Z","lastTransitionTime":"2026-01-27T15:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.091620 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.091689 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.091714 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.091745 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.091769 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:37Z","lastTransitionTime":"2026-01-27T15:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.195087 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.195149 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.195371 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.195393 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.195409 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:37Z","lastTransitionTime":"2026-01-27T15:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.297854 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.297895 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.297904 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.297917 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.297927 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:37Z","lastTransitionTime":"2026-01-27T15:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.401221 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.401286 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.401310 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.401335 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.401350 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:37Z","lastTransitionTime":"2026-01-27T15:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.504999 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.505042 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.505054 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.505081 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.505093 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:37Z","lastTransitionTime":"2026-01-27T15:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.607508 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.607564 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.607577 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.607597 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.607611 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:37Z","lastTransitionTime":"2026-01-27T15:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.639840 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 18:47:17.14797465 +0000 UTC Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.662522 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:37 crc kubenswrapper[4772]: E0127 15:07:37.662655 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.709695 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.709751 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.709761 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.709778 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.709791 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:37Z","lastTransitionTime":"2026-01-27T15:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.813010 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.813063 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.813073 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.813086 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.813095 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:37Z","lastTransitionTime":"2026-01-27T15:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.916210 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.916259 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.916271 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.916287 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:37 crc kubenswrapper[4772]: I0127 15:07:37.916298 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:37Z","lastTransitionTime":"2026-01-27T15:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.018944 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.019302 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.019313 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.019328 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.019340 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:38Z","lastTransitionTime":"2026-01-27T15:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.121927 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.121993 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.122012 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.122040 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.122058 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:38Z","lastTransitionTime":"2026-01-27T15:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.223971 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.224035 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.224052 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.224076 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.224095 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:38Z","lastTransitionTime":"2026-01-27T15:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.263378 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.282118 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs\") pod \"network-metrics-daemon-ql2vx\" (UID: \"371016c8-5a23-427d-aa0a-0faa241d86a7\") " pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:38 crc kubenswrapper[4772]: E0127 15:07:38.282399 4772 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 27 15:07:38 crc kubenswrapper[4772]: E0127 15:07:38.282482 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs podName:371016c8-5a23-427d-aa0a-0faa241d86a7 nodeName:}" failed. No retries permitted until 2026-01-27 15:07:46.28246009 +0000 UTC m=+52.263069228 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs") pod "network-metrics-daemon-ql2vx" (UID: "371016c8-5a23-427d-aa0a-0faa241d86a7") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.327151 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.327305 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.327323 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.327351 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.327369 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:38Z","lastTransitionTime":"2026-01-27T15:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.430607 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.430653 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.430663 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.430678 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.430689 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:38Z","lastTransitionTime":"2026-01-27T15:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.534029 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.534079 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.534090 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.534130 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.534144 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:38Z","lastTransitionTime":"2026-01-27T15:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.637039 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.637096 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.637114 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.637138 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.637156 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:38Z","lastTransitionTime":"2026-01-27T15:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.640594 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 21:21:05.69207103 +0000 UTC Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.661916 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.661921 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:38 crc kubenswrapper[4772]: E0127 15:07:38.662086 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.662110 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:38 crc kubenswrapper[4772]: E0127 15:07:38.662195 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:07:38 crc kubenswrapper[4772]: E0127 15:07:38.662283 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.740316 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.740393 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.740413 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.740755 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.741033 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:38Z","lastTransitionTime":"2026-01-27T15:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.844451 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.844570 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.844591 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.844618 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.844635 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:38Z","lastTransitionTime":"2026-01-27T15:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.948297 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.948381 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.948393 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.948411 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:38 crc kubenswrapper[4772]: I0127 15:07:38.948425 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:38Z","lastTransitionTime":"2026-01-27T15:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.050881 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.050943 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.050960 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.050976 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.050988 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:39Z","lastTransitionTime":"2026-01-27T15:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.152860 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.152914 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.152930 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.152962 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.152980 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:39Z","lastTransitionTime":"2026-01-27T15:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.255872 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.255927 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.255939 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.255962 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.256009 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:39Z","lastTransitionTime":"2026-01-27T15:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.362757 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.362848 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.362889 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.362923 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.362947 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:39Z","lastTransitionTime":"2026-01-27T15:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.466303 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.466351 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.466367 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.466387 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.466402 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:39Z","lastTransitionTime":"2026-01-27T15:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.569249 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.569296 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.569310 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.569332 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.569347 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:39Z","lastTransitionTime":"2026-01-27T15:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.641354 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 10:31:19.011289632 +0000 UTC Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.662853 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:39 crc kubenswrapper[4772]: E0127 15:07:39.663065 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.671988 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.672022 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.672031 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.672043 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.672052 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:39Z","lastTransitionTime":"2026-01-27T15:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.773882 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.773930 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.773947 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.773964 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.773976 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:39Z","lastTransitionTime":"2026-01-27T15:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.876267 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.876302 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.876311 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.876325 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.876334 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:39Z","lastTransitionTime":"2026-01-27T15:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.978532 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.978629 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.978646 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.978668 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:39 crc kubenswrapper[4772]: I0127 15:07:39.978684 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:39Z","lastTransitionTime":"2026-01-27T15:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.081490 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.081536 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.081549 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.081568 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.081582 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:40Z","lastTransitionTime":"2026-01-27T15:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.183894 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.183928 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.183939 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.183952 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.183961 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:40Z","lastTransitionTime":"2026-01-27T15:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.286505 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.286759 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.286837 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.286899 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.286955 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:40Z","lastTransitionTime":"2026-01-27T15:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.389654 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.389689 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.389697 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.389710 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.389718 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:40Z","lastTransitionTime":"2026-01-27T15:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.492203 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.492443 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.492512 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.492580 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.492636 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:40Z","lastTransitionTime":"2026-01-27T15:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.594896 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.594927 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.594935 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.594947 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.594956 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:40Z","lastTransitionTime":"2026-01-27T15:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.641470 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 13:43:14.017470977 +0000 UTC Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.662801 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.662831 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:40 crc kubenswrapper[4772]: E0127 15:07:40.662928 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.663076 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:40 crc kubenswrapper[4772]: E0127 15:07:40.663125 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:07:40 crc kubenswrapper[4772]: E0127 15:07:40.663456 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.696909 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.696944 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.696954 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.696969 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.696979 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:40Z","lastTransitionTime":"2026-01-27T15:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.799070 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.799340 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.799513 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.799608 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.799689 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:40Z","lastTransitionTime":"2026-01-27T15:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.902060 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.902358 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.902510 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.902672 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:40 crc kubenswrapper[4772]: I0127 15:07:40.902793 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:40Z","lastTransitionTime":"2026-01-27T15:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.006528 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.006569 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.006581 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.006597 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.006607 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:41Z","lastTransitionTime":"2026-01-27T15:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.109445 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.109495 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.109507 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.109525 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.109536 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:41Z","lastTransitionTime":"2026-01-27T15:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.211717 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.211760 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.211769 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.211783 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.211792 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:41Z","lastTransitionTime":"2026-01-27T15:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.314392 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.314496 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.314513 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.314530 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.314542 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:41Z","lastTransitionTime":"2026-01-27T15:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.417460 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.417498 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.417509 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.417527 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.417540 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:41Z","lastTransitionTime":"2026-01-27T15:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.423074 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.423101 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.423114 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.423128 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.423138 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:41Z","lastTransitionTime":"2026-01-27T15:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:41 crc kubenswrapper[4772]: E0127 15:07:41.434406 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:41Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.438064 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.438099 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.438112 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.438130 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.438140 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:41Z","lastTransitionTime":"2026-01-27T15:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:41 crc kubenswrapper[4772]: E0127 15:07:41.450119 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:41Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.453085 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.453120 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.453128 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.453141 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.453149 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:41Z","lastTransitionTime":"2026-01-27T15:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:41 crc kubenswrapper[4772]: E0127 15:07:41.463659 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:41Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.467085 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.467120 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.467131 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.467145 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.467183 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:41Z","lastTransitionTime":"2026-01-27T15:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:41 crc kubenswrapper[4772]: E0127 15:07:41.476845 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:41Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.479437 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.479468 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.479481 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.479499 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.479511 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:41Z","lastTransitionTime":"2026-01-27T15:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:41 crc kubenswrapper[4772]: E0127 15:07:41.490079 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:41Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:41 crc kubenswrapper[4772]: E0127 15:07:41.490372 4772 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.519557 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.519595 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.519606 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.519622 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.519633 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:41Z","lastTransitionTime":"2026-01-27T15:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.622401 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.622644 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.622770 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.622867 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.622955 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:41Z","lastTransitionTime":"2026-01-27T15:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.641760 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 11:56:29.40995986 +0000 UTC Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.662224 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:41 crc kubenswrapper[4772]: E0127 15:07:41.662399 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.663120 4772 scope.go:117] "RemoveContainer" containerID="44564ab0212c32423d179406999749f443cae1bd72ea8d12ff4411e23de77d97" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.726695 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.726746 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.726759 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.726777 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.726790 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:41Z","lastTransitionTime":"2026-01-27T15:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.829128 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.829211 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.829229 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.829255 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.829275 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:41Z","lastTransitionTime":"2026-01-27T15:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.931729 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.931794 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.931812 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.931834 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.931853 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:41Z","lastTransitionTime":"2026-01-27T15:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.974877 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n2khk_736264c8-cd18-479a-88ba-e1ec15dbfdae/ovnkube-controller/1.log" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.978093 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerStarted","Data":"c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7"} Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.978281 4772 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 27 15:07:41 crc kubenswrapper[4772]: I0127 15:07:41.991227 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:41Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.003023 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:42Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.042262 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.042316 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.042327 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.042342 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.042353 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:42Z","lastTransitionTime":"2026-01-27T15:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.046133 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8fde5-4905-4fb1-b683-27ea4921b462\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b47fc0778ba4c5a1e12700735e6f9c52a7341b9eac61071607902a6ec8daf02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://545dcc1be1335a1acf93f16d2e5b4a266dce5b7e736b7c56b80bbf56b3248ced\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wkvpx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:42Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.061988 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ql2vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371016c8-5a23-427d-aa0a-0faa241d86a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ql2vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:42Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.080900 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:42Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.093687 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:42Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.106434 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:42Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.124404 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:42Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.142244 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44564ab0212c32423d179406999749f443cae1bd72ea8d12ff4411e23de77d97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:28Z\\\",\\\"message\\\":\\\"twork-check-target template LB for network=default: []services.LB{}\\\\nI0127 15:07:28.751281 6196 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:42Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.144111 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.144229 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.144241 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.144257 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.144268 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:42Z","lastTransitionTime":"2026-01-27T15:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.154874 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:42Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.166760 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:42Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.183453 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:42Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.197588 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:42Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.210852 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:42Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.220008 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:42Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.240478 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:42Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.246256 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.246302 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.246312 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.246330 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.246340 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:42Z","lastTransitionTime":"2026-01-27T15:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.258120 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:42Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.348452 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.348491 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.348502 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.348519 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.348530 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:42Z","lastTransitionTime":"2026-01-27T15:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.450921 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.450977 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.450989 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.451003 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.451013 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:42Z","lastTransitionTime":"2026-01-27T15:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.552813 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.552847 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.552858 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.552872 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.552885 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:42Z","lastTransitionTime":"2026-01-27T15:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.641882 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 02:01:46.178645859 +0000 UTC Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.654830 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.654858 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.654866 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.654880 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.654890 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:42Z","lastTransitionTime":"2026-01-27T15:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.662410 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.662495 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:42 crc kubenswrapper[4772]: E0127 15:07:42.662532 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.662410 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:42 crc kubenswrapper[4772]: E0127 15:07:42.662635 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:07:42 crc kubenswrapper[4772]: E0127 15:07:42.662719 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.756858 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.756915 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.756937 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.756965 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.756986 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:42Z","lastTransitionTime":"2026-01-27T15:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.859370 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.859425 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.859436 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.859454 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.859467 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:42Z","lastTransitionTime":"2026-01-27T15:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.961789 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.961831 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.961842 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.961869 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.961884 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:42Z","lastTransitionTime":"2026-01-27T15:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.981665 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n2khk_736264c8-cd18-479a-88ba-e1ec15dbfdae/ovnkube-controller/2.log" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.982328 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n2khk_736264c8-cd18-479a-88ba-e1ec15dbfdae/ovnkube-controller/1.log" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.984632 4772 generic.go:334] "Generic (PLEG): container finished" podID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerID="c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7" exitCode=1 Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.984675 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerDied","Data":"c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7"} Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.984723 4772 scope.go:117] "RemoveContainer" containerID="44564ab0212c32423d179406999749f443cae1bd72ea8d12ff4411e23de77d97" Jan 27 15:07:42 crc kubenswrapper[4772]: I0127 15:07:42.985377 4772 scope.go:117] "RemoveContainer" containerID="c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7" Jan 27 15:07:42 crc kubenswrapper[4772]: E0127 15:07:42.985512 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-n2khk_openshift-ovn-kubernetes(736264c8-cd18-479a-88ba-e1ec15dbfdae)\"" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.004767 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:43Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.023381 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:43Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.036753 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:43Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.049680 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:43Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.059574 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:43Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.063861 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.063907 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.063920 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.063939 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.063954 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:43Z","lastTransitionTime":"2026-01-27T15:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.078603 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:43Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.090213 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:43Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.100785 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:43Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.132572 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:43Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.164031 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8fde5-4905-4fb1-b683-27ea4921b462\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b47fc0778ba4c5a1e12700735e6f9c52a7341b9eac61071607902a6ec8daf02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://545dcc1be1335a1acf93f16d2e5b4a266dce5b7e736b7c56b80bbf56b3248ced\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wkvpx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:43Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.165692 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.165730 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.165740 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.165755 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.165765 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:43Z","lastTransitionTime":"2026-01-27T15:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.174757 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ql2vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371016c8-5a23-427d-aa0a-0faa241d86a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ql2vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:43Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.187850 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:43Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.199274 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:43Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.210275 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:43Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.222420 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:43Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.247991 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44564ab0212c32423d179406999749f443cae1bd72ea8d12ff4411e23de77d97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:28Z\\\",\\\"message\\\":\\\"twork-check-target template LB for network=default: []services.LB{}\\\\nI0127 15:07:28.751281 6196 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:42Z\\\",\\\"message\\\":\\\"g/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:1936, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0127 15:07:42.483073 6419 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:43Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.267915 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.267975 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.267984 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.267999 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.268027 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:43Z","lastTransitionTime":"2026-01-27T15:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.269979 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:43Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.370288 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.370330 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.370341 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.370358 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.370370 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:43Z","lastTransitionTime":"2026-01-27T15:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.474019 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.474114 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.474139 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.474233 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.474263 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:43Z","lastTransitionTime":"2026-01-27T15:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.576576 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.576629 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.576646 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.576668 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.576684 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:43Z","lastTransitionTime":"2026-01-27T15:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.642440 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 18:11:04.781888721 +0000 UTC Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.662453 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:43 crc kubenswrapper[4772]: E0127 15:07:43.662608 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.678577 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.678637 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.678654 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.678678 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.678696 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:43Z","lastTransitionTime":"2026-01-27T15:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.782087 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.782141 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.782157 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.782200 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.782219 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:43Z","lastTransitionTime":"2026-01-27T15:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.884579 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.884655 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.884667 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.884684 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.884697 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:43Z","lastTransitionTime":"2026-01-27T15:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.987239 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.987283 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.987294 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.987311 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.987322 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:43Z","lastTransitionTime":"2026-01-27T15:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:43 crc kubenswrapper[4772]: I0127 15:07:43.989453 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n2khk_736264c8-cd18-479a-88ba-e1ec15dbfdae/ovnkube-controller/2.log" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.090497 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.090587 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.090613 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.090640 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.090662 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:44Z","lastTransitionTime":"2026-01-27T15:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.193178 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.193228 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.193263 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.193283 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.193293 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:44Z","lastTransitionTime":"2026-01-27T15:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.295518 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.295569 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.295585 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.295606 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.295618 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:44Z","lastTransitionTime":"2026-01-27T15:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.398236 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.398275 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.398283 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.398297 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.398306 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:44Z","lastTransitionTime":"2026-01-27T15:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.500352 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.500386 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.500394 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.500408 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.500425 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:44Z","lastTransitionTime":"2026-01-27T15:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.603556 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.603605 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.603622 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.603639 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.603650 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:44Z","lastTransitionTime":"2026-01-27T15:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.643314 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 09:56:31.689912996 +0000 UTC Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.662671 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:44 crc kubenswrapper[4772]: E0127 15:07:44.662799 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.663309 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.663368 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:44 crc kubenswrapper[4772]: E0127 15:07:44.663385 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:07:44 crc kubenswrapper[4772]: E0127 15:07:44.663510 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.674988 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:44Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.684906 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:44Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.693090 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:44Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.703052 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8fde5-4905-4fb1-b683-27ea4921b462\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b47fc0778ba4c5a1e12700735e6f9c52a7341b9eac61071607902a6ec8daf02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://545dcc1be1335a1acf93f16d2e5b4a266dce5b7e736b7c56b80bbf56b3248ced\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wkvpx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:44Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.705971 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.705997 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.706005 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.706018 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.706026 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:44Z","lastTransitionTime":"2026-01-27T15:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.713325 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ql2vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371016c8-5a23-427d-aa0a-0faa241d86a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ql2vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:44Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.724902 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:44Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.735549 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:44Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.748096 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:44Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.764140 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44564ab0212c32423d179406999749f443cae1bd72ea8d12ff4411e23de77d97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:28Z\\\",\\\"message\\\":\\\"twork-check-target template LB for network=default: []services.LB{}\\\\nI0127 15:07:28.751281 6196 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:42Z\\\",\\\"message\\\":\\\"g/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:1936, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0127 15:07:42.483073 6419 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:44Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.775732 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:44Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.787732 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:44Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.801571 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:44Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.807746 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.807782 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.807793 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.807807 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.807818 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:44Z","lastTransitionTime":"2026-01-27T15:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.815858 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:44Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.827215 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:44Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.838893 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:44Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.848772 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:44Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.866145 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:44Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.909671 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.909715 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.909730 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.909745 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:44 crc kubenswrapper[4772]: I0127 15:07:44.909757 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:44Z","lastTransitionTime":"2026-01-27T15:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.012447 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.012496 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.012505 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.012521 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.012532 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:45Z","lastTransitionTime":"2026-01-27T15:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.115212 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.115265 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.115281 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.115302 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.115320 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:45Z","lastTransitionTime":"2026-01-27T15:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.218465 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.218514 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.218529 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.218550 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.218563 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:45Z","lastTransitionTime":"2026-01-27T15:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.321231 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.321278 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.321288 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.321311 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.321333 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:45Z","lastTransitionTime":"2026-01-27T15:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.424685 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.424759 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.424775 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.424800 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.424820 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:45Z","lastTransitionTime":"2026-01-27T15:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.527107 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.527139 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.527148 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.527160 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.527216 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:45Z","lastTransitionTime":"2026-01-27T15:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.629334 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.629384 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.629395 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.629413 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.629430 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:45Z","lastTransitionTime":"2026-01-27T15:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.644462 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 09:20:26.808390846 +0000 UTC Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.661951 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:45 crc kubenswrapper[4772]: E0127 15:07:45.662087 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.732216 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.732260 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.732271 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.732289 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.732302 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:45Z","lastTransitionTime":"2026-01-27T15:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.835694 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.835765 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.835776 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.835794 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.835805 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:45Z","lastTransitionTime":"2026-01-27T15:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.937979 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.938047 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.938063 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.938088 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:45 crc kubenswrapper[4772]: I0127 15:07:45.938105 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:45Z","lastTransitionTime":"2026-01-27T15:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.040333 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.040367 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.040378 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.040395 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.040405 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:46Z","lastTransitionTime":"2026-01-27T15:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.142691 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.142729 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.142742 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.142759 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.142772 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:46Z","lastTransitionTime":"2026-01-27T15:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.245768 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.245805 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.245813 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.245827 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.245855 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:46Z","lastTransitionTime":"2026-01-27T15:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.290410 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs\") pod \"network-metrics-daemon-ql2vx\" (UID: \"371016c8-5a23-427d-aa0a-0faa241d86a7\") " pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:46 crc kubenswrapper[4772]: E0127 15:07:46.290606 4772 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 27 15:07:46 crc kubenswrapper[4772]: E0127 15:07:46.290692 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs podName:371016c8-5a23-427d-aa0a-0faa241d86a7 nodeName:}" failed. No retries permitted until 2026-01-27 15:08:02.290667502 +0000 UTC m=+68.271276660 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs") pod "network-metrics-daemon-ql2vx" (UID: "371016c8-5a23-427d-aa0a-0faa241d86a7") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.348558 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.348609 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.348634 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.348657 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.348673 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:46Z","lastTransitionTime":"2026-01-27T15:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.456440 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.456484 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.456495 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.456510 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.456520 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:46Z","lastTransitionTime":"2026-01-27T15:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.561105 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.561153 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.561182 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.561200 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.561214 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:46Z","lastTransitionTime":"2026-01-27T15:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.592985 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.593126 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:46 crc kubenswrapper[4772]: E0127 15:07:46.593157 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:08:18.593135207 +0000 UTC m=+84.573744305 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.593214 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:46 crc kubenswrapper[4772]: E0127 15:07:46.593258 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 27 15:07:46 crc kubenswrapper[4772]: E0127 15:07:46.593274 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 27 15:07:46 crc kubenswrapper[4772]: E0127 15:07:46.593286 4772 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:46 crc kubenswrapper[4772]: E0127 15:07:46.593303 4772 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 27 15:07:46 crc kubenswrapper[4772]: E0127 15:07:46.593321 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-27 15:08:18.593311902 +0000 UTC m=+84.573921000 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:46 crc kubenswrapper[4772]: E0127 15:07:46.593337 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-27 15:08:18.593328042 +0000 UTC m=+84.573937150 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.593260 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.593367 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:46 crc kubenswrapper[4772]: E0127 15:07:46.593444 4772 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 27 15:07:46 crc kubenswrapper[4772]: E0127 15:07:46.593466 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 27 15:07:46 crc kubenswrapper[4772]: E0127 15:07:46.593489 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 27 15:07:46 crc kubenswrapper[4772]: E0127 15:07:46.593498 4772 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:46 crc kubenswrapper[4772]: E0127 15:07:46.593530 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-27 15:08:18.593509547 +0000 UTC m=+84.574118645 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 27 15:07:46 crc kubenswrapper[4772]: E0127 15:07:46.593548 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-27 15:08:18.593542078 +0000 UTC m=+84.574151176 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.645258 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 07:27:43.545897704 +0000 UTC Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.662607 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.662692 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:46 crc kubenswrapper[4772]: E0127 15:07:46.662710 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.662772 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:46 crc kubenswrapper[4772]: E0127 15:07:46.662876 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:07:46 crc kubenswrapper[4772]: E0127 15:07:46.662995 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.664203 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.664234 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.664245 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.664261 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.664272 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:46Z","lastTransitionTime":"2026-01-27T15:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.767064 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.767103 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.767115 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.767133 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.767145 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:46Z","lastTransitionTime":"2026-01-27T15:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.831857 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.833102 4772 scope.go:117] "RemoveContainer" containerID="c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7" Jan 27 15:07:46 crc kubenswrapper[4772]: E0127 15:07:46.833482 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-n2khk_openshift-ovn-kubernetes(736264c8-cd18-479a-88ba-e1ec15dbfdae)\"" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.856491 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:46Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.870706 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.870763 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.870781 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.870808 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.870831 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:46Z","lastTransitionTime":"2026-01-27T15:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.873282 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:46Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.892590 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:46Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.916810 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:42Z\\\",\\\"message\\\":\\\"g/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:1936, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0127 15:07:42.483073 6419 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-n2khk_openshift-ovn-kubernetes(736264c8-cd18-479a-88ba-e1ec15dbfdae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:46Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.933517 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:46Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.949013 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:46Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.967445 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:46Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.973042 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.973093 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.973109 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.973136 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.973154 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:46Z","lastTransitionTime":"2026-01-27T15:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:46 crc kubenswrapper[4772]: I0127 15:07:46.983272 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:46Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.007867 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.023910 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.037230 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.049930 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.064015 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.075555 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.075595 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.075606 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.075625 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.075638 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:47Z","lastTransitionTime":"2026-01-27T15:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.075638 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.088748 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.101341 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8fde5-4905-4fb1-b683-27ea4921b462\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b47fc0778ba4c5a1e12700735e6f9c52a7341b9eac61071607902a6ec8daf02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://545dcc1be1335a1acf93f16d2e5b4a266dce5b7e736b7c56b80bbf56b3248ced\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wkvpx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.112367 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ql2vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371016c8-5a23-427d-aa0a-0faa241d86a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ql2vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.178301 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.178402 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.178429 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.178468 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.178498 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:47Z","lastTransitionTime":"2026-01-27T15:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.281979 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.282043 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.283806 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.283833 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.283844 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:47Z","lastTransitionTime":"2026-01-27T15:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.386359 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.386412 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.386425 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.386442 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.386454 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:47Z","lastTransitionTime":"2026-01-27T15:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.488434 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.488472 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.488483 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.488499 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.488511 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:47Z","lastTransitionTime":"2026-01-27T15:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.591649 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.591720 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.591745 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.591777 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.591800 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:47Z","lastTransitionTime":"2026-01-27T15:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.645660 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 14:22:27.576648238 +0000 UTC Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.662016 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:47 crc kubenswrapper[4772]: E0127 15:07:47.662144 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.694092 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.694193 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.694219 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.694244 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.694262 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:47Z","lastTransitionTime":"2026-01-27T15:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.741324 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.752264 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.754250 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.775158 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:42Z\\\",\\\"message\\\":\\\"g/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:1936, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0127 15:07:42.483073 6419 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-n2khk_openshift-ovn-kubernetes(736264c8-cd18-479a-88ba-e1ec15dbfdae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.790349 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.796592 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.796625 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.796635 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.796650 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.796660 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:47Z","lastTransitionTime":"2026-01-27T15:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.802050 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.816783 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.830022 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.843774 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.854956 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.873774 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.886530 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.896116 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.898892 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.898926 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.898939 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.898954 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.898966 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:47Z","lastTransitionTime":"2026-01-27T15:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.906413 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.920404 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8fde5-4905-4fb1-b683-27ea4921b462\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b47fc0778ba4c5a1e12700735e6f9c52a7341b9eac61071607902a6ec8daf02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://545dcc1be1335a1acf93f16d2e5b4a266dce5b7e736b7c56b80bbf56b3248ced\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wkvpx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.936661 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ql2vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371016c8-5a23-427d-aa0a-0faa241d86a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ql2vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.950081 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.962611 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:47 crc kubenswrapper[4772]: I0127 15:07:47.974198 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:47Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.001463 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.001498 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.001508 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.001523 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.001533 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:48Z","lastTransitionTime":"2026-01-27T15:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.103526 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.103567 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.103576 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.103591 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.103600 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:48Z","lastTransitionTime":"2026-01-27T15:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.206039 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.206098 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.206117 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.206141 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.206225 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:48Z","lastTransitionTime":"2026-01-27T15:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.267284 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.280249 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:48Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.290820 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:48Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.307048 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:42Z\\\",\\\"message\\\":\\\"g/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:1936, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0127 15:07:42.483073 6419 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-n2khk_openshift-ovn-kubernetes(736264c8-cd18-479a-88ba-e1ec15dbfdae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:48Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.308260 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.308294 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.308308 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.308324 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.308336 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:48Z","lastTransitionTime":"2026-01-27T15:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.321342 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:48Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.333367 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6cfdf02-101c-4f18-9ebe-16002352afce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5b9360189745deea4c8950c4f3b73762c7cb098452c1f31c3df9ab99fc31ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3be141392fbb6ae8ce51085cfc186745bb5ac272899e3202aaed2b7093855c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8c8aeb62c273e8cb31f01742805bd40c4184821c4bbd94f319ff5f074265d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64098f88e61c989a4f2048d222906eacf8c0525f26e109913c1718c9dfb67d20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64098f88e61c989a4f2048d222906eacf8c0525f26e109913c1718c9dfb67d20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:48Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.345081 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:48Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.358700 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:48Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.371007 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:48Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.383222 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:48Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.393832 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:48Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.410444 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.410491 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.410501 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.410519 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.410532 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:48Z","lastTransitionTime":"2026-01-27T15:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.415997 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:48Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.430155 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:48Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.444367 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:48Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.457008 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:48Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.468734 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8fde5-4905-4fb1-b683-27ea4921b462\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b47fc0778ba4c5a1e12700735e6f9c52a7341b9eac61071607902a6ec8daf02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://545dcc1be1335a1acf93f16d2e5b4a266dce5b7e736b7c56b80bbf56b3248ced\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wkvpx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:48Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.481052 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ql2vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371016c8-5a23-427d-aa0a-0faa241d86a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ql2vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:48Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.492653 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:48Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.502708 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:48Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.512636 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.512693 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.512708 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.512728 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.512740 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:48Z","lastTransitionTime":"2026-01-27T15:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.614955 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.615002 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.615010 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.615025 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.615034 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:48Z","lastTransitionTime":"2026-01-27T15:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.646230 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 09:19:41.122006038 +0000 UTC Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.663076 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.663140 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.663097 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:48 crc kubenswrapper[4772]: E0127 15:07:48.663275 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:07:48 crc kubenswrapper[4772]: E0127 15:07:48.663370 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:07:48 crc kubenswrapper[4772]: E0127 15:07:48.663457 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.718483 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.718524 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.718533 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.718549 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.718560 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:48Z","lastTransitionTime":"2026-01-27T15:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.821327 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.821375 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.821391 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.821476 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.821497 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:48Z","lastTransitionTime":"2026-01-27T15:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.924014 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.924066 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.924075 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.924093 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:48 crc kubenswrapper[4772]: I0127 15:07:48.924107 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:48Z","lastTransitionTime":"2026-01-27T15:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.027212 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.027255 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.027266 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.027284 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.027295 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:49Z","lastTransitionTime":"2026-01-27T15:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.130292 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.130365 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.130386 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.130414 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.130434 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:49Z","lastTransitionTime":"2026-01-27T15:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.234335 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.234671 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.234691 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.234713 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.234728 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:49Z","lastTransitionTime":"2026-01-27T15:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.337407 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.337442 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.337450 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.337464 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.337472 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:49Z","lastTransitionTime":"2026-01-27T15:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.439653 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.439707 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.439716 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.439732 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.439745 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:49Z","lastTransitionTime":"2026-01-27T15:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.543010 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.543074 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.543097 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.543127 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.543148 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:49Z","lastTransitionTime":"2026-01-27T15:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.646132 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.646198 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.646210 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.646226 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.646237 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:49Z","lastTransitionTime":"2026-01-27T15:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.646358 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 10:56:58.117468312 +0000 UTC Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.662757 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:49 crc kubenswrapper[4772]: E0127 15:07:49.662903 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.748829 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.748878 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.748903 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.748925 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.748940 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:49Z","lastTransitionTime":"2026-01-27T15:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.851651 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.851698 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.851711 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.851726 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.851738 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:49Z","lastTransitionTime":"2026-01-27T15:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.953698 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.953729 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.953738 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.953752 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:49 crc kubenswrapper[4772]: I0127 15:07:49.953760 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:49Z","lastTransitionTime":"2026-01-27T15:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.056285 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.056364 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.056380 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.056399 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.056411 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:50Z","lastTransitionTime":"2026-01-27T15:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.159066 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.159110 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.159122 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.159140 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.159154 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:50Z","lastTransitionTime":"2026-01-27T15:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.262327 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.262357 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.262365 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.262378 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.262387 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:50Z","lastTransitionTime":"2026-01-27T15:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.365380 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.365434 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.365451 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.365483 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.365507 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:50Z","lastTransitionTime":"2026-01-27T15:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.468127 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.468164 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.468218 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.468234 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.468246 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:50Z","lastTransitionTime":"2026-01-27T15:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.575568 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.575633 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.575645 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.575677 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.575698 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:50Z","lastTransitionTime":"2026-01-27T15:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.647458 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 18:15:57.156439563 +0000 UTC Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.663535 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.663622 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.663734 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:50 crc kubenswrapper[4772]: E0127 15:07:50.663762 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:07:50 crc kubenswrapper[4772]: E0127 15:07:50.663933 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:07:50 crc kubenswrapper[4772]: E0127 15:07:50.663991 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.678826 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.678886 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.678897 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.678919 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.678931 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:50Z","lastTransitionTime":"2026-01-27T15:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.782090 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.782148 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.782160 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.782227 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.782241 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:50Z","lastTransitionTime":"2026-01-27T15:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.884634 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.884682 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.884694 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.884713 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.884726 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:50Z","lastTransitionTime":"2026-01-27T15:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.987229 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.987274 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.987285 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.987300 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:50 crc kubenswrapper[4772]: I0127 15:07:50.987309 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:50Z","lastTransitionTime":"2026-01-27T15:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.090021 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.090070 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.090092 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.090110 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.090123 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:51Z","lastTransitionTime":"2026-01-27T15:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.193683 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.193731 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.193742 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.193758 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.193769 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:51Z","lastTransitionTime":"2026-01-27T15:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.296319 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.296393 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.296442 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.296467 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.296482 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:51Z","lastTransitionTime":"2026-01-27T15:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.399110 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.399183 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.399194 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.399210 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.399220 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:51Z","lastTransitionTime":"2026-01-27T15:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.502851 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.502901 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.502916 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.502939 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.502957 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:51Z","lastTransitionTime":"2026-01-27T15:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.605142 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.605224 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.605236 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.605257 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.605271 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:51Z","lastTransitionTime":"2026-01-27T15:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.610057 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.610104 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.610128 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.610147 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.610160 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:51Z","lastTransitionTime":"2026-01-27T15:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:51 crc kubenswrapper[4772]: E0127 15:07:51.625143 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:51Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.630620 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.630658 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.630669 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.630686 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.630697 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:51Z","lastTransitionTime":"2026-01-27T15:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:51 crc kubenswrapper[4772]: E0127 15:07:51.644408 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:51Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.647846 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 21:49:09.111859564 +0000 UTC Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.648561 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.648597 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.648606 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.648622 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.648630 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:51Z","lastTransitionTime":"2026-01-27T15:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:51 crc kubenswrapper[4772]: E0127 15:07:51.659323 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:51Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.661905 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:51 crc kubenswrapper[4772]: E0127 15:07:51.662005 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.663766 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.663792 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.663800 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.663811 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.663821 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:51Z","lastTransitionTime":"2026-01-27T15:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:51 crc kubenswrapper[4772]: E0127 15:07:51.674509 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:51Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.677704 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.677753 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.677765 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.677782 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.677795 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:51Z","lastTransitionTime":"2026-01-27T15:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:51 crc kubenswrapper[4772]: E0127 15:07:51.690000 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:51Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:51 crc kubenswrapper[4772]: E0127 15:07:51.690111 4772 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.708302 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.708333 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.708344 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.708361 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.708371 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:51Z","lastTransitionTime":"2026-01-27T15:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.810295 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.810346 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.810356 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.810378 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.810388 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:51Z","lastTransitionTime":"2026-01-27T15:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.912767 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.912833 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.912850 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.912875 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:51 crc kubenswrapper[4772]: I0127 15:07:51.912892 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:51Z","lastTransitionTime":"2026-01-27T15:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.014320 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.014356 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.014368 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.014381 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.014390 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:52Z","lastTransitionTime":"2026-01-27T15:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.116533 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.116601 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.116620 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.116650 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.116668 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:52Z","lastTransitionTime":"2026-01-27T15:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.219671 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.219728 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.219746 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.219769 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.219786 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:52Z","lastTransitionTime":"2026-01-27T15:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.323241 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.323304 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.323329 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.323358 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.323384 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:52Z","lastTransitionTime":"2026-01-27T15:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.426234 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.426295 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.426313 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.426340 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.426359 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:52Z","lastTransitionTime":"2026-01-27T15:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.529988 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.530029 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.530042 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.530061 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.530074 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:52Z","lastTransitionTime":"2026-01-27T15:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.632984 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.633034 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.633050 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.633071 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.633086 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:52Z","lastTransitionTime":"2026-01-27T15:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.648382 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 13:18:54.752362309 +0000 UTC Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.662805 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:52 crc kubenswrapper[4772]: E0127 15:07:52.663339 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.662817 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.663829 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:52 crc kubenswrapper[4772]: E0127 15:07:52.664108 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:07:52 crc kubenswrapper[4772]: E0127 15:07:52.667759 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.735994 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.736047 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.736058 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.736076 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.736089 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:52Z","lastTransitionTime":"2026-01-27T15:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.838604 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.838639 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.838647 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.838662 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.838671 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:52Z","lastTransitionTime":"2026-01-27T15:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.941980 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.942021 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.942033 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.942047 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:52 crc kubenswrapper[4772]: I0127 15:07:52.942057 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:52Z","lastTransitionTime":"2026-01-27T15:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.044457 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.044509 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.044540 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.044564 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.044581 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:53Z","lastTransitionTime":"2026-01-27T15:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.147431 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.147464 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.147476 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.147491 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.147502 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:53Z","lastTransitionTime":"2026-01-27T15:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.250408 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.250461 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.250472 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.250489 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.250499 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:53Z","lastTransitionTime":"2026-01-27T15:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.352637 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.352700 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.352708 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.352721 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.352731 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:53Z","lastTransitionTime":"2026-01-27T15:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.456245 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.456312 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.456332 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.456356 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.456375 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:53Z","lastTransitionTime":"2026-01-27T15:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.559341 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.559388 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.559399 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.559419 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.559434 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:53Z","lastTransitionTime":"2026-01-27T15:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.648831 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 04:00:08.54394988 +0000 UTC Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.661977 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.662038 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.662051 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.662062 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.662072 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:53Z","lastTransitionTime":"2026-01-27T15:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.662083 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:53 crc kubenswrapper[4772]: E0127 15:07:53.662277 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.764789 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.764831 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.764842 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.764861 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.764874 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:53Z","lastTransitionTime":"2026-01-27T15:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.867796 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.867840 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.867856 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.867871 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.867881 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:53Z","lastTransitionTime":"2026-01-27T15:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.971693 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.971766 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.971790 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.971822 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:53 crc kubenswrapper[4772]: I0127 15:07:53.971861 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:53Z","lastTransitionTime":"2026-01-27T15:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.074366 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.074398 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.074406 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.074420 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.074428 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:54Z","lastTransitionTime":"2026-01-27T15:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.176963 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.177008 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.177020 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.177036 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.177050 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:54Z","lastTransitionTime":"2026-01-27T15:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.279841 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.279894 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.279909 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.279928 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.279941 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:54Z","lastTransitionTime":"2026-01-27T15:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.382137 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.382233 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.382249 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.382272 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.382288 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:54Z","lastTransitionTime":"2026-01-27T15:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.484865 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.484924 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.484932 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.484947 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.484956 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:54Z","lastTransitionTime":"2026-01-27T15:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.586863 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.586935 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.586959 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.586988 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.587009 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:54Z","lastTransitionTime":"2026-01-27T15:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.649021 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 20:29:46.546595802 +0000 UTC Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.661986 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.662005 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:54 crc kubenswrapper[4772]: E0127 15:07:54.662107 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.662176 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:54 crc kubenswrapper[4772]: E0127 15:07:54.662261 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:07:54 crc kubenswrapper[4772]: E0127 15:07:54.662370 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.676310 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:54Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.686650 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:54Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.689270 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.689290 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.689300 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.689314 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.689323 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:54Z","lastTransitionTime":"2026-01-27T15:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.697503 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8fde5-4905-4fb1-b683-27ea4921b462\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b47fc0778ba4c5a1e12700735e6f9c52a7341b9eac61071607902a6ec8daf02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://545dcc1be1335a1acf93f16d2e5b4a266dce5b7e736b7c56b80bbf56b3248ced\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wkvpx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:54Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.707093 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ql2vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371016c8-5a23-427d-aa0a-0faa241d86a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ql2vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:54Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.719140 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:54Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.731376 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:54Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.744540 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:54Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.756079 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:54Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.771260 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:42Z\\\",\\\"message\\\":\\\"g/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:1936, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0127 15:07:42.483073 6419 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-n2khk_openshift-ovn-kubernetes(736264c8-cd18-479a-88ba-e1ec15dbfdae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:54Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.783220 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:54Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.791257 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.791321 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.791336 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.791353 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.792210 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:54Z","lastTransitionTime":"2026-01-27T15:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.796707 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6cfdf02-101c-4f18-9ebe-16002352afce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5b9360189745deea4c8950c4f3b73762c7cb098452c1f31c3df9ab99fc31ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3be141392fbb6ae8ce51085cfc186745bb5ac272899e3202aaed2b7093855c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8c8aeb62c273e8cb31f01742805bd40c4184821c4bbd94f319ff5f074265d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64098f88e61c989a4f2048d222906eacf8c0525f26e109913c1718c9dfb67d20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64098f88e61c989a4f2048d222906eacf8c0525f26e109913c1718c9dfb67d20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:54Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.806628 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:54Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.817836 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:54Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.831451 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:54Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.850031 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:54Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.860839 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:54Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.881367 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:54Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.894137 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.894193 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.894203 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.894220 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.894232 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:54Z","lastTransitionTime":"2026-01-27T15:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.894398 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:07:54Z is after 2025-08-24T17:21:41Z" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.997035 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.997073 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.997083 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.997100 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:54 crc kubenswrapper[4772]: I0127 15:07:54.997112 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:54Z","lastTransitionTime":"2026-01-27T15:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.099061 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.099102 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.099117 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.099140 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.099153 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:55Z","lastTransitionTime":"2026-01-27T15:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.201592 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.201625 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.201635 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.201650 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.201662 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:55Z","lastTransitionTime":"2026-01-27T15:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.305067 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.305123 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.305139 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.305222 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.305246 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:55Z","lastTransitionTime":"2026-01-27T15:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.407805 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.407858 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.407874 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.407895 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.407908 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:55Z","lastTransitionTime":"2026-01-27T15:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.510456 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.510516 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.510538 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.510568 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.510591 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:55Z","lastTransitionTime":"2026-01-27T15:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.614269 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.614710 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.614879 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.615038 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.615222 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:55Z","lastTransitionTime":"2026-01-27T15:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.649498 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 02:51:52.967908726 +0000 UTC Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.662590 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:55 crc kubenswrapper[4772]: E0127 15:07:55.662840 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.717896 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.717933 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.717944 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.717959 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.717972 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:55Z","lastTransitionTime":"2026-01-27T15:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.820365 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.820434 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.820458 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.820487 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.820511 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:55Z","lastTransitionTime":"2026-01-27T15:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.923599 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.923643 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.923653 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.923669 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:55 crc kubenswrapper[4772]: I0127 15:07:55.923682 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:55Z","lastTransitionTime":"2026-01-27T15:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.026499 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.026867 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.026982 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.027056 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.027128 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:56Z","lastTransitionTime":"2026-01-27T15:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.130729 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.130766 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.130777 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.130793 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.130803 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:56Z","lastTransitionTime":"2026-01-27T15:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.233703 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.233995 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.234068 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.234146 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.234243 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:56Z","lastTransitionTime":"2026-01-27T15:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.337452 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.337493 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.337501 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.337521 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.337531 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:56Z","lastTransitionTime":"2026-01-27T15:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.439945 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.439984 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.440001 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.440023 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.440033 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:56Z","lastTransitionTime":"2026-01-27T15:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.543431 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.543492 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.543504 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.543528 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.543543 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:56Z","lastTransitionTime":"2026-01-27T15:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.646465 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.646925 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.646936 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.646953 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.646964 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:56Z","lastTransitionTime":"2026-01-27T15:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.649656 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 10:41:36.127466939 +0000 UTC Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.661948 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.662043 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.661970 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:56 crc kubenswrapper[4772]: E0127 15:07:56.662150 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:07:56 crc kubenswrapper[4772]: E0127 15:07:56.662272 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:07:56 crc kubenswrapper[4772]: E0127 15:07:56.662372 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.749809 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.749887 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.749906 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.749935 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.749957 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:56Z","lastTransitionTime":"2026-01-27T15:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.853086 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.853145 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.853187 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.853214 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.853237 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:56Z","lastTransitionTime":"2026-01-27T15:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.956316 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.956354 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.956365 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.956382 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:56 crc kubenswrapper[4772]: I0127 15:07:56.956393 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:56Z","lastTransitionTime":"2026-01-27T15:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.099486 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.099572 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.099594 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.099611 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.099625 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:57Z","lastTransitionTime":"2026-01-27T15:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.202484 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.202509 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.202517 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.202530 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.202539 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:57Z","lastTransitionTime":"2026-01-27T15:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.305094 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.305158 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.305193 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.305207 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.305217 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:57Z","lastTransitionTime":"2026-01-27T15:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.407033 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.407065 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.407074 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.407087 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.407097 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:57Z","lastTransitionTime":"2026-01-27T15:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.509958 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.510050 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.510062 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.510077 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.510088 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:57Z","lastTransitionTime":"2026-01-27T15:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.613455 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.613526 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.613540 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.613567 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.613583 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:57Z","lastTransitionTime":"2026-01-27T15:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.649897 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 18:40:31.847490531 +0000 UTC Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.662287 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:57 crc kubenswrapper[4772]: E0127 15:07:57.662406 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.716137 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.716208 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.716220 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.716236 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.716247 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:57Z","lastTransitionTime":"2026-01-27T15:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.819747 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.819817 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.819842 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.819871 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.819891 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:57Z","lastTransitionTime":"2026-01-27T15:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.922941 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.922969 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.922979 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.922993 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:57 crc kubenswrapper[4772]: I0127 15:07:57.923003 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:57Z","lastTransitionTime":"2026-01-27T15:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.025345 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.025374 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.025385 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.025407 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.025419 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:58Z","lastTransitionTime":"2026-01-27T15:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.127251 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.127293 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.127309 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.127329 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.127346 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:58Z","lastTransitionTime":"2026-01-27T15:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.228913 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.228953 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.228964 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.228983 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.228994 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:58Z","lastTransitionTime":"2026-01-27T15:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.330935 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.331253 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.331353 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.331452 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.331538 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:58Z","lastTransitionTime":"2026-01-27T15:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.434619 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.434654 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.434664 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.434678 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.434688 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:58Z","lastTransitionTime":"2026-01-27T15:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.537702 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.537797 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.537817 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.537841 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.537859 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:58Z","lastTransitionTime":"2026-01-27T15:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.640727 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.640794 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.640813 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.640832 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.640846 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:58Z","lastTransitionTime":"2026-01-27T15:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.650970 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 18:33:36.513932384 +0000 UTC Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.662405 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.662447 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.662480 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:07:58 crc kubenswrapper[4772]: E0127 15:07:58.662539 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:07:58 crc kubenswrapper[4772]: E0127 15:07:58.662602 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:07:58 crc kubenswrapper[4772]: E0127 15:07:58.662712 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.743397 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.743808 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.743969 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.744130 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.744323 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:58Z","lastTransitionTime":"2026-01-27T15:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.847105 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.847379 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.847447 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.847529 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.847604 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:58Z","lastTransitionTime":"2026-01-27T15:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.950121 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.950189 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.950206 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.950227 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:58 crc kubenswrapper[4772]: I0127 15:07:58.950244 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:58Z","lastTransitionTime":"2026-01-27T15:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.051967 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.051991 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.051999 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.052010 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.052019 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:59Z","lastTransitionTime":"2026-01-27T15:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.155341 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.155377 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.155388 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.155403 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.155413 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:59Z","lastTransitionTime":"2026-01-27T15:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.257437 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.257469 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.257478 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.257491 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.257501 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:59Z","lastTransitionTime":"2026-01-27T15:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.359750 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.359787 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.359797 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.359811 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.359822 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:59Z","lastTransitionTime":"2026-01-27T15:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.462460 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.462502 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.462513 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.462529 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.462542 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:59Z","lastTransitionTime":"2026-01-27T15:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.565134 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.565227 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.565244 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.565274 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.565289 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:59Z","lastTransitionTime":"2026-01-27T15:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.651932 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 15:12:28.388235496 +0000 UTC Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.662346 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:07:59 crc kubenswrapper[4772]: E0127 15:07:59.662468 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.667232 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.667266 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.667275 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.667289 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.667299 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:59Z","lastTransitionTime":"2026-01-27T15:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.769235 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.769296 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.769308 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.769325 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.769337 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:59Z","lastTransitionTime":"2026-01-27T15:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.871667 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.871716 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.871730 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.871748 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.871762 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:59Z","lastTransitionTime":"2026-01-27T15:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.973798 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.973836 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.973862 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.973879 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:07:59 crc kubenswrapper[4772]: I0127 15:07:59.973889 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:07:59Z","lastTransitionTime":"2026-01-27T15:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.076699 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.076749 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.076757 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.076771 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.076779 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:00Z","lastTransitionTime":"2026-01-27T15:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.179779 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.179828 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.179838 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.179854 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.179863 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:00Z","lastTransitionTime":"2026-01-27T15:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.282573 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.282612 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.282622 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.282640 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.282650 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:00Z","lastTransitionTime":"2026-01-27T15:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.384934 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.384977 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.384988 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.385004 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.385015 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:00Z","lastTransitionTime":"2026-01-27T15:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.488132 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.488192 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.488201 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.488217 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.488229 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:00Z","lastTransitionTime":"2026-01-27T15:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.590615 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.590654 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.590668 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.590687 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.590701 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:00Z","lastTransitionTime":"2026-01-27T15:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.652806 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 00:34:46.230937063 +0000 UTC Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.664782 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:00 crc kubenswrapper[4772]: E0127 15:08:00.664898 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.665094 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:00 crc kubenswrapper[4772]: E0127 15:08:00.665157 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.665929 4772 scope.go:117] "RemoveContainer" containerID="c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7" Jan 27 15:08:00 crc kubenswrapper[4772]: E0127 15:08:00.666074 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-n2khk_openshift-ovn-kubernetes(736264c8-cd18-479a-88ba-e1ec15dbfdae)\"" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.666244 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:00 crc kubenswrapper[4772]: E0127 15:08:00.666316 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.693728 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.693767 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.693778 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.693794 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.693804 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:00Z","lastTransitionTime":"2026-01-27T15:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.799104 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.799141 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.799149 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.799183 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.799194 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:00Z","lastTransitionTime":"2026-01-27T15:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.901922 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.901950 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.901959 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.901974 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:00 crc kubenswrapper[4772]: I0127 15:08:00.901985 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:00Z","lastTransitionTime":"2026-01-27T15:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.004730 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.004772 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.004781 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.004801 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.004813 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:01Z","lastTransitionTime":"2026-01-27T15:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.107687 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.107736 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.107748 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.107768 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.107781 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:01Z","lastTransitionTime":"2026-01-27T15:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.209792 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.209854 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.209868 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.209885 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.209896 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:01Z","lastTransitionTime":"2026-01-27T15:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.312043 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.312087 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.312099 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.312116 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.312126 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:01Z","lastTransitionTime":"2026-01-27T15:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.414696 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.414731 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.414743 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.414758 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.414770 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:01Z","lastTransitionTime":"2026-01-27T15:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.517606 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.517648 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.517661 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.517678 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.517688 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:01Z","lastTransitionTime":"2026-01-27T15:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.620692 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.620721 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.620731 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.620745 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.620755 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:01Z","lastTransitionTime":"2026-01-27T15:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.653142 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 01:57:17.807469065 +0000 UTC Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.662787 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:01 crc kubenswrapper[4772]: E0127 15:08:01.662886 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.722729 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.722764 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.722774 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.722792 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.722802 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:01Z","lastTransitionTime":"2026-01-27T15:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.825527 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.825559 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.825570 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.825586 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.825599 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:01Z","lastTransitionTime":"2026-01-27T15:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.929504 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.929546 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.929557 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.929574 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.929584 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:01Z","lastTransitionTime":"2026-01-27T15:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.942377 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.942429 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.942444 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.942462 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.942474 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:01Z","lastTransitionTime":"2026-01-27T15:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:01 crc kubenswrapper[4772]: E0127 15:08:01.954952 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:01Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.958931 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.958971 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.958982 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.958998 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.959009 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:01Z","lastTransitionTime":"2026-01-27T15:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:01 crc kubenswrapper[4772]: E0127 15:08:01.969228 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:01Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.973404 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.973446 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.973459 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.973483 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.973495 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:01Z","lastTransitionTime":"2026-01-27T15:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:01 crc kubenswrapper[4772]: E0127 15:08:01.984507 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:01Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.988371 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.988415 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.988425 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.988444 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:01 crc kubenswrapper[4772]: I0127 15:08:01.988461 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:01Z","lastTransitionTime":"2026-01-27T15:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:01 crc kubenswrapper[4772]: E0127 15:08:01.999449 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:01Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.002285 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.002310 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.002320 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.002333 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.002341 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:02Z","lastTransitionTime":"2026-01-27T15:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:02 crc kubenswrapper[4772]: E0127 15:08:02.013640 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:02Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:02 crc kubenswrapper[4772]: E0127 15:08:02.013743 4772 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.032284 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.032314 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.032322 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.032334 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.032343 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:02Z","lastTransitionTime":"2026-01-27T15:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.135068 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.135102 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.135110 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.135123 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.135132 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:02Z","lastTransitionTime":"2026-01-27T15:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.237273 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.237311 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.237323 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.237339 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.237351 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:02Z","lastTransitionTime":"2026-01-27T15:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.339261 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.339302 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.339315 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.339332 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.339343 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:02Z","lastTransitionTime":"2026-01-27T15:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.361678 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs\") pod \"network-metrics-daemon-ql2vx\" (UID: \"371016c8-5a23-427d-aa0a-0faa241d86a7\") " pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:02 crc kubenswrapper[4772]: E0127 15:08:02.361799 4772 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 27 15:08:02 crc kubenswrapper[4772]: E0127 15:08:02.361840 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs podName:371016c8-5a23-427d-aa0a-0faa241d86a7 nodeName:}" failed. No retries permitted until 2026-01-27 15:08:34.361825703 +0000 UTC m=+100.342434801 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs") pod "network-metrics-daemon-ql2vx" (UID: "371016c8-5a23-427d-aa0a-0faa241d86a7") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.441947 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.441994 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.442011 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.442030 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.442039 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:02Z","lastTransitionTime":"2026-01-27T15:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.544687 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.544729 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.544744 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.544765 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.544777 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:02Z","lastTransitionTime":"2026-01-27T15:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.646341 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.646397 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.646407 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.646423 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.646434 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:02Z","lastTransitionTime":"2026-01-27T15:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.654186 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 00:21:49.964793684 +0000 UTC Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.662544 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.662574 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:02 crc kubenswrapper[4772]: E0127 15:08:02.662687 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.662734 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:02 crc kubenswrapper[4772]: E0127 15:08:02.662858 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:02 crc kubenswrapper[4772]: E0127 15:08:02.662963 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.748727 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.748772 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.748782 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.748796 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.748807 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:02Z","lastTransitionTime":"2026-01-27T15:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.851025 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.851062 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.851070 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.851093 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.851103 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:02Z","lastTransitionTime":"2026-01-27T15:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.958376 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.958421 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.958454 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.958474 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:02 crc kubenswrapper[4772]: I0127 15:08:02.958484 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:02Z","lastTransitionTime":"2026-01-27T15:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.069214 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.069272 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.069283 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.069301 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.069312 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:03Z","lastTransitionTime":"2026-01-27T15:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.171366 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.171405 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.171414 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.171431 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.171440 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:03Z","lastTransitionTime":"2026-01-27T15:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.273501 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.273537 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.273546 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.273560 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.273569 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:03Z","lastTransitionTime":"2026-01-27T15:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.375488 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.375566 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.375578 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.375596 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.375606 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:03Z","lastTransitionTime":"2026-01-27T15:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.477615 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.477677 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.477686 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.477736 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.477748 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:03Z","lastTransitionTime":"2026-01-27T15:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.580392 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.580439 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.580453 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.580471 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.580484 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:03Z","lastTransitionTime":"2026-01-27T15:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.654466 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 06:51:56.369202357 +0000 UTC Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.662801 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:03 crc kubenswrapper[4772]: E0127 15:08:03.662902 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.682583 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.682608 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.682616 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.682627 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.682636 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:03Z","lastTransitionTime":"2026-01-27T15:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.785403 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.785499 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.785530 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.785563 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.785590 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:03Z","lastTransitionTime":"2026-01-27T15:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.888000 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.888047 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.888060 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.888078 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.888091 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:03Z","lastTransitionTime":"2026-01-27T15:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.990855 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.990890 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.990899 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.990913 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:03 crc kubenswrapper[4772]: I0127 15:08:03.990923 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:03Z","lastTransitionTime":"2026-01-27T15:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.093439 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.093471 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.093478 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.093493 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.093502 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:04Z","lastTransitionTime":"2026-01-27T15:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.195918 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.195960 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.195972 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.195988 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.195999 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:04Z","lastTransitionTime":"2026-01-27T15:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.298292 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.298330 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.298341 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.298356 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.298368 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:04Z","lastTransitionTime":"2026-01-27T15:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.400757 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.400814 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.400829 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.400846 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.400857 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:04Z","lastTransitionTime":"2026-01-27T15:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.533219 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.533259 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.533270 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.533288 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.533299 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:04Z","lastTransitionTime":"2026-01-27T15:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.635853 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.635893 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.635902 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.635916 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.635926 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:04Z","lastTransitionTime":"2026-01-27T15:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.655092 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 23:00:43.81977253 +0000 UTC Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.662438 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.662512 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.662438 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:04 crc kubenswrapper[4772]: E0127 15:08:04.662615 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:04 crc kubenswrapper[4772]: E0127 15:08:04.662769 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:04 crc kubenswrapper[4772]: E0127 15:08:04.662875 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.683505 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:04Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.696558 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:04Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.704751 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:04Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.713209 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ql2vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371016c8-5a23-427d-aa0a-0faa241d86a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ql2vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:04Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.722898 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:04Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.732812 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:04Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.741806 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.741834 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.741844 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.741858 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.741869 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:04Z","lastTransitionTime":"2026-01-27T15:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.744881 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:04Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.754232 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:04Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.762597 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8fde5-4905-4fb1-b683-27ea4921b462\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b47fc0778ba4c5a1e12700735e6f9c52a7341b9eac61071607902a6ec8daf02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://545dcc1be1335a1acf93f16d2e5b4a266dce5b7e736b7c56b80bbf56b3248ced\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wkvpx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:04Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.773975 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:04Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.783180 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6cfdf02-101c-4f18-9ebe-16002352afce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5b9360189745deea4c8950c4f3b73762c7cb098452c1f31c3df9ab99fc31ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3be141392fbb6ae8ce51085cfc186745bb5ac272899e3202aaed2b7093855c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8c8aeb62c273e8cb31f01742805bd40c4184821c4bbd94f319ff5f074265d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64098f88e61c989a4f2048d222906eacf8c0525f26e109913c1718c9dfb67d20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64098f88e61c989a4f2048d222906eacf8c0525f26e109913c1718c9dfb67d20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:04Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.793290 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:04Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.802242 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:04Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.824957 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:42Z\\\",\\\"message\\\":\\\"g/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:1936, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0127 15:07:42.483073 6419 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-n2khk_openshift-ovn-kubernetes(736264c8-cd18-479a-88ba-e1ec15dbfdae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:04Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.835188 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:04Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.843614 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.843667 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.843680 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.843694 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.843704 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:04Z","lastTransitionTime":"2026-01-27T15:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.846633 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:04Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.873009 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:04Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.891748 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:04Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.946494 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.946541 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.946564 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.946584 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:04 crc kubenswrapper[4772]: I0127 15:08:04.946598 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:04Z","lastTransitionTime":"2026-01-27T15:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.048787 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.048821 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.048830 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.048842 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.048850 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:05Z","lastTransitionTime":"2026-01-27T15:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.051864 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7jwx_87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8/kube-multus/0.log" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.051905 4772 generic.go:334] "Generic (PLEG): container finished" podID="87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8" containerID="ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33" exitCode=1 Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.051932 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x7jwx" event={"ID":"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8","Type":"ContainerDied","Data":"ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33"} Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.052318 4772 scope.go:117] "RemoveContainer" containerID="ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.066858 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:05Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.079181 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:05Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.095926 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:42Z\\\",\\\"message\\\":\\\"g/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:1936, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0127 15:07:42.483073 6419 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-n2khk_openshift-ovn-kubernetes(736264c8-cd18-479a-88ba-e1ec15dbfdae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:05Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.109369 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:05Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.119964 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6cfdf02-101c-4f18-9ebe-16002352afce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5b9360189745deea4c8950c4f3b73762c7cb098452c1f31c3df9ab99fc31ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3be141392fbb6ae8ce51085cfc186745bb5ac272899e3202aaed2b7093855c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8c8aeb62c273e8cb31f01742805bd40c4184821c4bbd94f319ff5f074265d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64098f88e61c989a4f2048d222906eacf8c0525f26e109913c1718c9dfb67d20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64098f88e61c989a4f2048d222906eacf8c0525f26e109913c1718c9dfb67d20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:05Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.131737 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:05Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.145087 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:05Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.150356 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.150393 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.150406 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.150444 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.150456 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:05Z","lastTransitionTime":"2026-01-27T15:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.158601 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:08:04Z\\\",\\\"message\\\":\\\"2026-01-27T15:07:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_85dd4958-953b-469d-b48e-3b998cab8570\\\\n2026-01-27T15:07:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_85dd4958-953b-469d-b48e-3b998cab8570 to /host/opt/cni/bin/\\\\n2026-01-27T15:07:19Z [verbose] multus-daemon started\\\\n2026-01-27T15:07:19Z [verbose] Readiness Indicator file check\\\\n2026-01-27T15:08:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:05Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.171741 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:05Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.180361 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:05Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.196105 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:05Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.206633 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:05Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.215549 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:05Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.224356 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:05Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.233382 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8fde5-4905-4fb1-b683-27ea4921b462\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b47fc0778ba4c5a1e12700735e6f9c52a7341b9eac61071607902a6ec8daf02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://545dcc1be1335a1acf93f16d2e5b4a266dce5b7e736b7c56b80bbf56b3248ced\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wkvpx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:05Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.243713 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ql2vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371016c8-5a23-427d-aa0a-0faa241d86a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ql2vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:05Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.252302 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.252335 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.252344 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.252358 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.252367 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:05Z","lastTransitionTime":"2026-01-27T15:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.264277 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:05Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.274628 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:05Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.354309 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.354355 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.354367 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.354382 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.354392 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:05Z","lastTransitionTime":"2026-01-27T15:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.456507 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.456560 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.456576 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.456596 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.456613 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:05Z","lastTransitionTime":"2026-01-27T15:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.558607 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.558644 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.558655 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.558670 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.558679 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:05Z","lastTransitionTime":"2026-01-27T15:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.655764 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 18:52:15.672424013 +0000 UTC Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.660778 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.660845 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.660870 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.660907 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.660957 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:05Z","lastTransitionTime":"2026-01-27T15:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.661983 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:05 crc kubenswrapper[4772]: E0127 15:08:05.662106 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.763765 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.763812 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.763824 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.763842 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.763853 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:05Z","lastTransitionTime":"2026-01-27T15:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.865706 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.865755 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.865766 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.865781 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.865791 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:05Z","lastTransitionTime":"2026-01-27T15:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.968933 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.968972 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.968986 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.969003 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:05 crc kubenswrapper[4772]: I0127 15:08:05.969015 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:05Z","lastTransitionTime":"2026-01-27T15:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.056988 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7jwx_87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8/kube-multus/0.log" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.057038 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x7jwx" event={"ID":"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8","Type":"ContainerStarted","Data":"9f72b451fa77f3fce2c251de546110ab49c7c9e0122759f6ef29a32fde422356"} Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.070188 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:06Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.071203 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.071258 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.071275 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.071297 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.071314 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:06Z","lastTransitionTime":"2026-01-27T15:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.087291 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:42Z\\\",\\\"message\\\":\\\"g/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:1936, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0127 15:07:42.483073 6419 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-n2khk_openshift-ovn-kubernetes(736264c8-cd18-479a-88ba-e1ec15dbfdae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:06Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.100636 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:06Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.111898 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6cfdf02-101c-4f18-9ebe-16002352afce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5b9360189745deea4c8950c4f3b73762c7cb098452c1f31c3df9ab99fc31ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3be141392fbb6ae8ce51085cfc186745bb5ac272899e3202aaed2b7093855c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8c8aeb62c273e8cb31f01742805bd40c4184821c4bbd94f319ff5f074265d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64098f88e61c989a4f2048d222906eacf8c0525f26e109913c1718c9dfb67d20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64098f88e61c989a4f2048d222906eacf8c0525f26e109913c1718c9dfb67d20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:06Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.124082 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:06Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.138550 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:06Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.151107 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f72b451fa77f3fce2c251de546110ab49c7c9e0122759f6ef29a32fde422356\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:08:04Z\\\",\\\"message\\\":\\\"2026-01-27T15:07:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_85dd4958-953b-469d-b48e-3b998cab8570\\\\n2026-01-27T15:07:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_85dd4958-953b-469d-b48e-3b998cab8570 to /host/opt/cni/bin/\\\\n2026-01-27T15:07:19Z [verbose] multus-daemon started\\\\n2026-01-27T15:07:19Z [verbose] Readiness Indicator file check\\\\n2026-01-27T15:08:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:06Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.161817 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:06Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.171231 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:06Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.174069 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.174096 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.174107 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.174121 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.174132 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:06Z","lastTransitionTime":"2026-01-27T15:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.189914 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:06Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.219861 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:06Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.238813 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:06Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.255693 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:06Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.268668 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8fde5-4905-4fb1-b683-27ea4921b462\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b47fc0778ba4c5a1e12700735e6f9c52a7341b9eac61071607902a6ec8daf02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://545dcc1be1335a1acf93f16d2e5b4a266dce5b7e736b7c56b80bbf56b3248ced\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wkvpx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:06Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.276496 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.276560 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.276575 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.276594 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.276618 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:06Z","lastTransitionTime":"2026-01-27T15:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.281364 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ql2vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371016c8-5a23-427d-aa0a-0faa241d86a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ql2vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:06Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.295959 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:06Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.307569 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:06Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.319493 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:06Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.386215 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.386253 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.386263 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.386279 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.386291 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:06Z","lastTransitionTime":"2026-01-27T15:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.488372 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.488401 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.488409 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.488439 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.488448 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:06Z","lastTransitionTime":"2026-01-27T15:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.590942 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.590981 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.590993 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.591009 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.591020 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:06Z","lastTransitionTime":"2026-01-27T15:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.656281 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 00:39:23.673343576 +0000 UTC Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.662664 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.662666 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:06 crc kubenswrapper[4772]: E0127 15:08:06.662775 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:06 crc kubenswrapper[4772]: E0127 15:08:06.662833 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.662669 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:06 crc kubenswrapper[4772]: E0127 15:08:06.662916 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.693522 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.693576 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.693589 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.693607 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.693628 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:06Z","lastTransitionTime":"2026-01-27T15:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.796502 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.796559 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.796567 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.796585 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.796595 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:06Z","lastTransitionTime":"2026-01-27T15:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.899200 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.899244 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.899253 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.899269 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:06 crc kubenswrapper[4772]: I0127 15:08:06.899279 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:06Z","lastTransitionTime":"2026-01-27T15:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.002430 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.002477 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.002487 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.002506 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.002522 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:07Z","lastTransitionTime":"2026-01-27T15:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.104499 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.104544 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.104557 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.104575 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.104587 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:07Z","lastTransitionTime":"2026-01-27T15:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.207990 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.208059 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.208072 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.208109 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.208122 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:07Z","lastTransitionTime":"2026-01-27T15:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.310642 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.310686 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.310696 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.310710 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.310718 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:07Z","lastTransitionTime":"2026-01-27T15:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.413273 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.413332 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.413343 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.413358 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.413367 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:07Z","lastTransitionTime":"2026-01-27T15:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.516109 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.516158 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.516194 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.516213 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.516225 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:07Z","lastTransitionTime":"2026-01-27T15:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.619091 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.619130 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.619139 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.619155 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.619180 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:07Z","lastTransitionTime":"2026-01-27T15:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.657435 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 14:12:23.172922165 +0000 UTC Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.662730 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:07 crc kubenswrapper[4772]: E0127 15:08:07.662839 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.721211 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.721243 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.721251 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.721263 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.721272 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:07Z","lastTransitionTime":"2026-01-27T15:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.823657 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.823711 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.823720 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.823736 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.823746 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:07Z","lastTransitionTime":"2026-01-27T15:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.925889 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.925925 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.925934 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.925950 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:07 crc kubenswrapper[4772]: I0127 15:08:07.925959 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:07Z","lastTransitionTime":"2026-01-27T15:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.027942 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.027994 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.028004 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.028018 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.028028 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:08Z","lastTransitionTime":"2026-01-27T15:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.130377 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.130415 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.130429 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.130446 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.130458 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:08Z","lastTransitionTime":"2026-01-27T15:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.231969 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.231999 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.232006 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.232019 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.232027 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:08Z","lastTransitionTime":"2026-01-27T15:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.334152 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.334226 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.334241 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.334260 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.334275 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:08Z","lastTransitionTime":"2026-01-27T15:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.436847 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.436895 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.436906 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.436959 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.436970 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:08Z","lastTransitionTime":"2026-01-27T15:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.539084 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.539129 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.539160 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.539194 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.539205 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:08Z","lastTransitionTime":"2026-01-27T15:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.641924 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.641966 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.641980 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.641998 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.642009 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:08Z","lastTransitionTime":"2026-01-27T15:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.658489 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 09:57:10.470768523 +0000 UTC Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.662803 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.662909 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:08 crc kubenswrapper[4772]: E0127 15:08:08.663036 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.663094 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:08 crc kubenswrapper[4772]: E0127 15:08:08.663105 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:08 crc kubenswrapper[4772]: E0127 15:08:08.663331 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.744885 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.744928 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.744938 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.744955 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.744965 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:08Z","lastTransitionTime":"2026-01-27T15:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.847697 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.847745 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.847757 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.847774 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.847786 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:08Z","lastTransitionTime":"2026-01-27T15:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.949893 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.949934 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.949947 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.949965 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:08 crc kubenswrapper[4772]: I0127 15:08:08.949976 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:08Z","lastTransitionTime":"2026-01-27T15:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.052271 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.052316 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.052326 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.052340 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.052349 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:09Z","lastTransitionTime":"2026-01-27T15:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.154571 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.154603 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.154613 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.154627 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.154637 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:09Z","lastTransitionTime":"2026-01-27T15:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.257239 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.257281 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.257292 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.257307 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.257316 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:09Z","lastTransitionTime":"2026-01-27T15:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.359471 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.359536 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.359551 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.359568 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.359579 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:09Z","lastTransitionTime":"2026-01-27T15:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.462354 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.462394 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.462403 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.462417 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.462427 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:09Z","lastTransitionTime":"2026-01-27T15:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.565612 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.565678 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.565696 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.565726 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.565748 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:09Z","lastTransitionTime":"2026-01-27T15:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.658750 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 23:32:29.990523342 +0000 UTC Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.661857 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:09 crc kubenswrapper[4772]: E0127 15:08:09.661978 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.668441 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.668475 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.668486 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.668500 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.668508 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:09Z","lastTransitionTime":"2026-01-27T15:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.772022 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.772080 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.772091 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.772107 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.772118 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:09Z","lastTransitionTime":"2026-01-27T15:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.874928 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.875247 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.875259 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.875277 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.875287 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:09Z","lastTransitionTime":"2026-01-27T15:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.977559 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.977593 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.977601 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.977618 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:09 crc kubenswrapper[4772]: I0127 15:08:09.977626 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:09Z","lastTransitionTime":"2026-01-27T15:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.080127 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.080153 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.080187 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.080201 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.080209 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:10Z","lastTransitionTime":"2026-01-27T15:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.183250 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.183311 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.183341 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.183371 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.183389 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:10Z","lastTransitionTime":"2026-01-27T15:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.290409 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.290453 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.290464 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.290480 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.290491 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:10Z","lastTransitionTime":"2026-01-27T15:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.393364 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.393411 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.393423 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.393439 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.393453 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:10Z","lastTransitionTime":"2026-01-27T15:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.495945 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.496002 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.496015 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.496033 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.496045 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:10Z","lastTransitionTime":"2026-01-27T15:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.598445 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.598497 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.598512 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.598531 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.598542 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:10Z","lastTransitionTime":"2026-01-27T15:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.659359 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 05:33:31.392985993 +0000 UTC Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.662667 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.662787 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:10 crc kubenswrapper[4772]: E0127 15:08:10.662795 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:10 crc kubenswrapper[4772]: E0127 15:08:10.662868 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.662667 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:10 crc kubenswrapper[4772]: E0127 15:08:10.662939 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.700696 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.700733 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.700741 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.700755 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.700764 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:10Z","lastTransitionTime":"2026-01-27T15:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.802587 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.802619 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.802628 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.802641 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.802650 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:10Z","lastTransitionTime":"2026-01-27T15:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.904953 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.905003 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.905014 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.905029 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:10 crc kubenswrapper[4772]: I0127 15:08:10.905039 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:10Z","lastTransitionTime":"2026-01-27T15:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.008502 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.008561 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.008579 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.008602 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.008619 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:11Z","lastTransitionTime":"2026-01-27T15:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.111546 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.111619 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.111642 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.111670 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.111691 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:11Z","lastTransitionTime":"2026-01-27T15:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.214550 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.214596 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.214619 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.214639 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.214652 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:11Z","lastTransitionTime":"2026-01-27T15:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.318121 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.318216 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.318238 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.318261 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.318273 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:11Z","lastTransitionTime":"2026-01-27T15:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.421468 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.421511 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.421522 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.421540 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.421552 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:11Z","lastTransitionTime":"2026-01-27T15:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.524193 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.524237 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.524247 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.524262 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.524272 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:11Z","lastTransitionTime":"2026-01-27T15:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.627603 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.627982 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.628133 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.628195 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.628215 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:11Z","lastTransitionTime":"2026-01-27T15:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.659899 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 03:32:48.024226201 +0000 UTC Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.662218 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:11 crc kubenswrapper[4772]: E0127 15:08:11.662330 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.730404 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.730444 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.730452 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.730464 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.730473 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:11Z","lastTransitionTime":"2026-01-27T15:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.833607 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.833650 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.833661 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.833675 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.833685 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:11Z","lastTransitionTime":"2026-01-27T15:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.936596 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.936664 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.936676 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.936695 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:11 crc kubenswrapper[4772]: I0127 15:08:11.936708 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:11Z","lastTransitionTime":"2026-01-27T15:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.039630 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.039671 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.039685 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.039705 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.039720 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:12Z","lastTransitionTime":"2026-01-27T15:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.062205 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.062252 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.062266 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.062282 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.062291 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:12Z","lastTransitionTime":"2026-01-27T15:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:12 crc kubenswrapper[4772]: E0127 15:08:12.073118 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:12Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.077854 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.077938 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.077962 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.078017 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.078039 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:12Z","lastTransitionTime":"2026-01-27T15:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:12 crc kubenswrapper[4772]: E0127 15:08:12.097286 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:12Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.099957 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.099983 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.099999 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.100020 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.100031 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:12Z","lastTransitionTime":"2026-01-27T15:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:12 crc kubenswrapper[4772]: E0127 15:08:12.110096 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:12Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.113903 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.113931 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.113956 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.113969 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.113977 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:12Z","lastTransitionTime":"2026-01-27T15:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:12 crc kubenswrapper[4772]: E0127 15:08:12.125601 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:12Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.128452 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.128486 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.128495 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.128507 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.128516 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:12Z","lastTransitionTime":"2026-01-27T15:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:12 crc kubenswrapper[4772]: E0127 15:08:12.139943 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:12Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:12 crc kubenswrapper[4772]: E0127 15:08:12.140109 4772 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.141983 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.142237 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.142263 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.142286 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.142308 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:12Z","lastTransitionTime":"2026-01-27T15:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.247407 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.247467 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.247484 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.247505 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.247899 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:12Z","lastTransitionTime":"2026-01-27T15:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.350757 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.350784 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.350795 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.350807 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.350815 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:12Z","lastTransitionTime":"2026-01-27T15:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.454108 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.454136 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.454145 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.454185 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.454199 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:12Z","lastTransitionTime":"2026-01-27T15:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.556526 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.556564 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.556573 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.556587 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.556596 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:12Z","lastTransitionTime":"2026-01-27T15:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.659978 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.660035 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.660047 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.660007 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 05:14:26.187134993 +0000 UTC Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.660064 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.660080 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:12Z","lastTransitionTime":"2026-01-27T15:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.662454 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.662493 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.662509 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:12 crc kubenswrapper[4772]: E0127 15:08:12.662676 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:12 crc kubenswrapper[4772]: E0127 15:08:12.662794 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:12 crc kubenswrapper[4772]: E0127 15:08:12.662855 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.663404 4772 scope.go:117] "RemoveContainer" containerID="c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.763513 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.763564 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.763576 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.763599 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.763613 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:12Z","lastTransitionTime":"2026-01-27T15:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.866285 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.866323 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.866354 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.866370 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.866381 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:12Z","lastTransitionTime":"2026-01-27T15:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.968553 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.968592 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.968602 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.968616 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:12 crc kubenswrapper[4772]: I0127 15:08:12.968625 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:12Z","lastTransitionTime":"2026-01-27T15:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.072999 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.073029 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.073038 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.073051 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.073060 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:13Z","lastTransitionTime":"2026-01-27T15:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.079149 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n2khk_736264c8-cd18-479a-88ba-e1ec15dbfdae/ovnkube-controller/2.log" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.082907 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerStarted","Data":"8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a"} Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.083364 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.099739 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:13Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.117351 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f72b451fa77f3fce2c251de546110ab49c7c9e0122759f6ef29a32fde422356\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:08:04Z\\\",\\\"message\\\":\\\"2026-01-27T15:07:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_85dd4958-953b-469d-b48e-3b998cab8570\\\\n2026-01-27T15:07:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_85dd4958-953b-469d-b48e-3b998cab8570 to /host/opt/cni/bin/\\\\n2026-01-27T15:07:19Z [verbose] multus-daemon started\\\\n2026-01-27T15:07:19Z [verbose] Readiness Indicator file check\\\\n2026-01-27T15:08:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:13Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.128086 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:13Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.145739 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:13Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.164511 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:13Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.174829 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.174854 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.174863 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.174876 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.174885 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:13Z","lastTransitionTime":"2026-01-27T15:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.183820 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:13Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.193470 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:13Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.202225 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:13Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.211649 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8fde5-4905-4fb1-b683-27ea4921b462\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b47fc0778ba4c5a1e12700735e6f9c52a7341b9eac61071607902a6ec8daf02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://545dcc1be1335a1acf93f16d2e5b4a266dce5b7e736b7c56b80bbf56b3248ced\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wkvpx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:13Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.222916 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ql2vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371016c8-5a23-427d-aa0a-0faa241d86a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ql2vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:13Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.234760 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:13Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.245036 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:13Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.258608 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:13Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.268620 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:13Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.276668 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.276740 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.276752 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.276768 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.276800 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:13Z","lastTransitionTime":"2026-01-27T15:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.288524 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:42Z\\\",\\\"message\\\":\\\"g/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:1936, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0127 15:07:42.483073 6419 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:13Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.301267 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:13Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.312077 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6cfdf02-101c-4f18-9ebe-16002352afce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5b9360189745deea4c8950c4f3b73762c7cb098452c1f31c3df9ab99fc31ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3be141392fbb6ae8ce51085cfc186745bb5ac272899e3202aaed2b7093855c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8c8aeb62c273e8cb31f01742805bd40c4184821c4bbd94f319ff5f074265d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64098f88e61c989a4f2048d222906eacf8c0525f26e109913c1718c9dfb67d20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64098f88e61c989a4f2048d222906eacf8c0525f26e109913c1718c9dfb67d20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:13Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.326107 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:13Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.379709 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.379747 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.379758 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.379775 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.379786 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:13Z","lastTransitionTime":"2026-01-27T15:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.484989 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.485047 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.485058 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.485076 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.485091 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:13Z","lastTransitionTime":"2026-01-27T15:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.587644 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.587700 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.587716 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.587739 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.587753 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:13Z","lastTransitionTime":"2026-01-27T15:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.660109 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 07:10:11.341317044 +0000 UTC Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.662390 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:13 crc kubenswrapper[4772]: E0127 15:08:13.662522 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.690869 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.690942 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.690962 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.690984 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.690999 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:13Z","lastTransitionTime":"2026-01-27T15:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.793991 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.794049 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.794060 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.794079 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.794090 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:13Z","lastTransitionTime":"2026-01-27T15:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.896788 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.896822 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.896830 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.896843 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.896859 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:13Z","lastTransitionTime":"2026-01-27T15:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.999236 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.999273 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.999283 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.999299 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:13 crc kubenswrapper[4772]: I0127 15:08:13.999311 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:13Z","lastTransitionTime":"2026-01-27T15:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.087008 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n2khk_736264c8-cd18-479a-88ba-e1ec15dbfdae/ovnkube-controller/3.log" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.087580 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n2khk_736264c8-cd18-479a-88ba-e1ec15dbfdae/ovnkube-controller/2.log" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.089869 4772 generic.go:334] "Generic (PLEG): container finished" podID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerID="8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a" exitCode=1 Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.089903 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerDied","Data":"8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a"} Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.089936 4772 scope.go:117] "RemoveContainer" containerID="c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.090621 4772 scope.go:117] "RemoveContainer" containerID="8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a" Jan 27 15:08:14 crc kubenswrapper[4772]: E0127 15:08:14.090801 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-n2khk_openshift-ovn-kubernetes(736264c8-cd18-479a-88ba-e1ec15dbfdae)\"" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.102739 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.102804 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.102816 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.102835 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.102845 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:14Z","lastTransitionTime":"2026-01-27T15:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.112590 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.126585 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.139370 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.150807 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8fde5-4905-4fb1-b683-27ea4921b462\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b47fc0778ba4c5a1e12700735e6f9c52a7341b9eac61071607902a6ec8daf02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://545dcc1be1335a1acf93f16d2e5b4a266dce5b7e736b7c56b80bbf56b3248ced\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wkvpx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.161216 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ql2vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371016c8-5a23-427d-aa0a-0faa241d86a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ql2vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.178722 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.190056 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6cfdf02-101c-4f18-9ebe-16002352afce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5b9360189745deea4c8950c4f3b73762c7cb098452c1f31c3df9ab99fc31ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3be141392fbb6ae8ce51085cfc186745bb5ac272899e3202aaed2b7093855c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8c8aeb62c273e8cb31f01742805bd40c4184821c4bbd94f319ff5f074265d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64098f88e61c989a4f2048d222906eacf8c0525f26e109913c1718c9dfb67d20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64098f88e61c989a4f2048d222906eacf8c0525f26e109913c1718c9dfb67d20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.201708 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.205048 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.205084 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.205093 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.205108 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.205120 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:14Z","lastTransitionTime":"2026-01-27T15:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.213831 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.230368 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:42Z\\\",\\\"message\\\":\\\"g/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:1936, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0127 15:07:42.483073 6419 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:08:13Z\\\",\\\"message\\\":\\\"-secret-name:openshift-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc006e9cc6b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: openshift-controller-manager-operator,},ClusterIP:10.217.5.58,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.58],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF0127 15:08:13.535689 6829 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.242711 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.254760 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.269921 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.281784 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f72b451fa77f3fce2c251de546110ab49c7c9e0122759f6ef29a32fde422356\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:08:04Z\\\",\\\"message\\\":\\\"2026-01-27T15:07:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_85dd4958-953b-469d-b48e-3b998cab8570\\\\n2026-01-27T15:07:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_85dd4958-953b-469d-b48e-3b998cab8570 to /host/opt/cni/bin/\\\\n2026-01-27T15:07:19Z [verbose] multus-daemon started\\\\n2026-01-27T15:07:19Z [verbose] Readiness Indicator file check\\\\n2026-01-27T15:08:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.293758 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.307274 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.307304 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.307312 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.307324 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.307335 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:14Z","lastTransitionTime":"2026-01-27T15:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.309699 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.319797 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.338547 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.409478 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.409507 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.409515 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.409527 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.409535 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:14Z","lastTransitionTime":"2026-01-27T15:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.512628 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.512685 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.512703 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.512726 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.512744 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:14Z","lastTransitionTime":"2026-01-27T15:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.614671 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.614711 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.614720 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.614734 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.614744 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:14Z","lastTransitionTime":"2026-01-27T15:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.661225 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 18:12:51.564234114 +0000 UTC Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.662498 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:14 crc kubenswrapper[4772]: E0127 15:08:14.662661 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.662679 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.662745 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:14 crc kubenswrapper[4772]: E0127 15:08:14.662852 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:14 crc kubenswrapper[4772]: E0127 15:08:14.662924 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.677131 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.687616 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.702798 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8fde5-4905-4fb1-b683-27ea4921b462\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b47fc0778ba4c5a1e12700735e6f9c52a7341b9eac61071607902a6ec8daf02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://545dcc1be1335a1acf93f16d2e5b4a266dce5b7e736b7c56b80bbf56b3248ced\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wkvpx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.717281 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.717279 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ql2vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371016c8-5a23-427d-aa0a-0faa241d86a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ql2vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.717333 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.717358 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.717373 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.717382 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:14Z","lastTransitionTime":"2026-01-27T15:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.738116 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.752406 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.767900 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.781475 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.805246 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c77c5fad0f4c478526015a7e2dcc6a18a4a586ead55b8bb16ff40d61ca66f4a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:07:42Z\\\",\\\"message\\\":\\\"g/owner\\\\\\\":\\\\\\\"openshift-ingress/router-internal-default\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.176\\\\\\\", Port:1936, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0127 15:07:42.483073 6419 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:08:13Z\\\",\\\"message\\\":\\\"-secret-name:openshift-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc006e9cc6b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: openshift-controller-manager-operator,},ClusterIP:10.217.5.58,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.58],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF0127 15:08:13.535689 6829 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.819884 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.819922 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.819936 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.819953 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.819964 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:14Z","lastTransitionTime":"2026-01-27T15:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.820047 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.831281 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6cfdf02-101c-4f18-9ebe-16002352afce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5b9360189745deea4c8950c4f3b73762c7cb098452c1f31c3df9ab99fc31ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3be141392fbb6ae8ce51085cfc186745bb5ac272899e3202aaed2b7093855c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8c8aeb62c273e8cb31f01742805bd40c4184821c4bbd94f319ff5f074265d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64098f88e61c989a4f2048d222906eacf8c0525f26e109913c1718c9dfb67d20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64098f88e61c989a4f2048d222906eacf8c0525f26e109913c1718c9dfb67d20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.846062 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.862448 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.877232 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f72b451fa77f3fce2c251de546110ab49c7c9e0122759f6ef29a32fde422356\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:08:04Z\\\",\\\"message\\\":\\\"2026-01-27T15:07:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_85dd4958-953b-469d-b48e-3b998cab8570\\\\n2026-01-27T15:07:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_85dd4958-953b-469d-b48e-3b998cab8570 to /host/opt/cni/bin/\\\\n2026-01-27T15:07:19Z [verbose] multus-daemon started\\\\n2026-01-27T15:07:19Z [verbose] Readiness Indicator file check\\\\n2026-01-27T15:08:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.891241 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.900985 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.920013 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.922249 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.922291 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.922300 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.922316 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.922326 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:14Z","lastTransitionTime":"2026-01-27T15:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:14 crc kubenswrapper[4772]: I0127 15:08:14.932660 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:14Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.023795 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.023835 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.023846 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.023863 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.023875 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:15Z","lastTransitionTime":"2026-01-27T15:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.094235 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n2khk_736264c8-cd18-479a-88ba-e1ec15dbfdae/ovnkube-controller/3.log" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.097153 4772 scope.go:117] "RemoveContainer" containerID="8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a" Jan 27 15:08:15 crc kubenswrapper[4772]: E0127 15:08:15.097440 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-n2khk_openshift-ovn-kubernetes(736264c8-cd18-479a-88ba-e1ec15dbfdae)\"" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.108034 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.118100 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.126500 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.126572 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.126595 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.126626 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.126649 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:15Z","lastTransitionTime":"2026-01-27T15:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.141920 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.154728 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f72b451fa77f3fce2c251de546110ab49c7c9e0122759f6ef29a32fde422356\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:08:04Z\\\",\\\"message\\\":\\\"2026-01-27T15:07:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_85dd4958-953b-469d-b48e-3b998cab8570\\\\n2026-01-27T15:07:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_85dd4958-953b-469d-b48e-3b998cab8570 to /host/opt/cni/bin/\\\\n2026-01-27T15:07:19Z [verbose] multus-daemon started\\\\n2026-01-27T15:07:19Z [verbose] Readiness Indicator file check\\\\n2026-01-27T15:08:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.172804 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.186871 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.195265 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.204892 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.216193 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.226740 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.233690 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.233734 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.233747 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.233762 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.233774 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:15Z","lastTransitionTime":"2026-01-27T15:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.242498 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.253517 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8fde5-4905-4fb1-b683-27ea4921b462\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b47fc0778ba4c5a1e12700735e6f9c52a7341b9eac61071607902a6ec8daf02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://545dcc1be1335a1acf93f16d2e5b4a266dce5b7e736b7c56b80bbf56b3248ced\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wkvpx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.261390 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ql2vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371016c8-5a23-427d-aa0a-0faa241d86a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ql2vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.271593 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.281141 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6cfdf02-101c-4f18-9ebe-16002352afce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5b9360189745deea4c8950c4f3b73762c7cb098452c1f31c3df9ab99fc31ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3be141392fbb6ae8ce51085cfc186745bb5ac272899e3202aaed2b7093855c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8c8aeb62c273e8cb31f01742805bd40c4184821c4bbd94f319ff5f074265d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64098f88e61c989a4f2048d222906eacf8c0525f26e109913c1718c9dfb67d20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64098f88e61c989a4f2048d222906eacf8c0525f26e109913c1718c9dfb67d20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.292416 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.301807 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.317630 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:08:13Z\\\",\\\"message\\\":\\\"-secret-name:openshift-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc006e9cc6b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: openshift-controller-manager-operator,},ClusterIP:10.217.5.58,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.58],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF0127 15:08:13.535689 6829 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:08:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-n2khk_openshift-ovn-kubernetes(736264c8-cd18-479a-88ba-e1ec15dbfdae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:15Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.336142 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.336195 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.336214 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.336230 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.336334 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:15Z","lastTransitionTime":"2026-01-27T15:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.439264 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.439318 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.439342 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.439364 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.439379 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:15Z","lastTransitionTime":"2026-01-27T15:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.542088 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.542131 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.542143 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.542160 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.542193 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:15Z","lastTransitionTime":"2026-01-27T15:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.645141 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.645198 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.645207 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.645221 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.645229 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:15Z","lastTransitionTime":"2026-01-27T15:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.662344 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 08:47:23.925426665 +0000 UTC Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.662415 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:15 crc kubenswrapper[4772]: E0127 15:08:15.662548 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.748643 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.748682 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.748690 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.748703 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.748713 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:15Z","lastTransitionTime":"2026-01-27T15:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.851649 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.851684 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.851692 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.851705 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.851714 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:15Z","lastTransitionTime":"2026-01-27T15:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.953931 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.953962 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.953970 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.953983 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:15 crc kubenswrapper[4772]: I0127 15:08:15.953992 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:15Z","lastTransitionTime":"2026-01-27T15:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.056376 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.056402 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.056410 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.056428 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.056446 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:16Z","lastTransitionTime":"2026-01-27T15:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.159490 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.159580 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.159616 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.159648 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.159672 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:16Z","lastTransitionTime":"2026-01-27T15:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.262991 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.263030 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.263073 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.263089 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.263098 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:16Z","lastTransitionTime":"2026-01-27T15:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.365582 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.365623 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.365632 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.365646 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.365656 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:16Z","lastTransitionTime":"2026-01-27T15:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.467993 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.468053 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.468067 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.468084 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.468095 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:16Z","lastTransitionTime":"2026-01-27T15:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.571637 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.571681 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.571691 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.571708 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.571721 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:16Z","lastTransitionTime":"2026-01-27T15:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.662853 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.662867 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 08:56:51.946547702 +0000 UTC Jan 27 15:08:16 crc kubenswrapper[4772]: E0127 15:08:16.662985 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.662853 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.663058 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:16 crc kubenswrapper[4772]: E0127 15:08:16.663127 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:16 crc kubenswrapper[4772]: E0127 15:08:16.663224 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.674003 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.674037 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.674048 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.674063 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.674077 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:16Z","lastTransitionTime":"2026-01-27T15:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.776409 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.776469 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.776486 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.776512 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.776529 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:16Z","lastTransitionTime":"2026-01-27T15:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.879453 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.879525 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.879537 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.879556 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.879569 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:16Z","lastTransitionTime":"2026-01-27T15:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.981980 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.982027 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.982039 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.982054 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:16 crc kubenswrapper[4772]: I0127 15:08:16.982063 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:16Z","lastTransitionTime":"2026-01-27T15:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.084227 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.084568 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.084578 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.084592 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.084601 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:17Z","lastTransitionTime":"2026-01-27T15:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.186441 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.186484 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.186494 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.186510 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.186521 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:17Z","lastTransitionTime":"2026-01-27T15:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.288910 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.288985 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.289003 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.289028 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.289047 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:17Z","lastTransitionTime":"2026-01-27T15:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.392508 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.392548 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.392566 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.392581 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.392591 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:17Z","lastTransitionTime":"2026-01-27T15:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.496198 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.496239 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.496252 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.496274 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.496292 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:17Z","lastTransitionTime":"2026-01-27T15:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.598429 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.598792 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.598911 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.599053 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.599210 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:17Z","lastTransitionTime":"2026-01-27T15:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.662628 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:17 crc kubenswrapper[4772]: E0127 15:08:17.662780 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.662967 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 18:02:34.461479057 +0000 UTC Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.701472 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.701544 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.701567 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.701593 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.701613 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:17Z","lastTransitionTime":"2026-01-27T15:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.804149 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.804206 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.804219 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.804235 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.804247 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:17Z","lastTransitionTime":"2026-01-27T15:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.907476 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.907526 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.907543 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.907564 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:17 crc kubenswrapper[4772]: I0127 15:08:17.907577 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:17Z","lastTransitionTime":"2026-01-27T15:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.009750 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.009783 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.009793 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.009805 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.009813 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:18Z","lastTransitionTime":"2026-01-27T15:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.112697 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.112732 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.112744 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.112762 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.112774 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:18Z","lastTransitionTime":"2026-01-27T15:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.215424 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.215471 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.215486 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.215503 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.215515 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:18Z","lastTransitionTime":"2026-01-27T15:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.318647 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.318720 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.318729 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.318746 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.318755 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:18Z","lastTransitionTime":"2026-01-27T15:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.421355 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.421401 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.421418 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.421441 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.421456 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:18Z","lastTransitionTime":"2026-01-27T15:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.524289 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.524337 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.524348 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.524366 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.524378 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:18Z","lastTransitionTime":"2026-01-27T15:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.628061 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.628130 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.628159 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.628232 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.628257 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:18Z","lastTransitionTime":"2026-01-27T15:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.644508 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.644631 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.644657 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.644691 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:18 crc kubenswrapper[4772]: E0127 15:08:18.644732 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:22.644692111 +0000 UTC m=+148.625301209 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:08:18 crc kubenswrapper[4772]: E0127 15:08:18.644752 4772 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.644807 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:18 crc kubenswrapper[4772]: E0127 15:08:18.644931 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 27 15:08:18 crc kubenswrapper[4772]: E0127 15:08:18.644971 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 27 15:08:18 crc kubenswrapper[4772]: E0127 15:08:18.644991 4772 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:08:18 crc kubenswrapper[4772]: E0127 15:08:18.644931 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 27 15:08:18 crc kubenswrapper[4772]: E0127 15:08:18.644967 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-27 15:09:22.644916478 +0000 UTC m=+148.625525776 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 27 15:08:18 crc kubenswrapper[4772]: E0127 15:08:18.645079 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-27 15:09:22.645052192 +0000 UTC m=+148.625661400 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:08:18 crc kubenswrapper[4772]: E0127 15:08:18.645086 4772 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 27 15:08:18 crc kubenswrapper[4772]: E0127 15:08:18.645114 4772 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:08:18 crc kubenswrapper[4772]: E0127 15:08:18.644811 4772 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 27 15:08:18 crc kubenswrapper[4772]: E0127 15:08:18.645199 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-27 15:09:22.645179195 +0000 UTC m=+148.625788293 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 27 15:08:18 crc kubenswrapper[4772]: E0127 15:08:18.645234 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-27 15:09:22.645226127 +0000 UTC m=+148.625835225 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.662723 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.662760 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:18 crc kubenswrapper[4772]: E0127 15:08:18.662879 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.662942 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.663063 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 15:36:33.756985932 +0000 UTC Jan 27 15:08:18 crc kubenswrapper[4772]: E0127 15:08:18.663131 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:18 crc kubenswrapper[4772]: E0127 15:08:18.663232 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.731281 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.731340 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.731353 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.731370 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.731385 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:18Z","lastTransitionTime":"2026-01-27T15:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.834189 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.834229 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.834238 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.834252 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.834263 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:18Z","lastTransitionTime":"2026-01-27T15:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.936697 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.936735 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.936746 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.936763 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:18 crc kubenswrapper[4772]: I0127 15:08:18.936772 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:18Z","lastTransitionTime":"2026-01-27T15:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.039492 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.039529 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.039540 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.039557 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.039567 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:19Z","lastTransitionTime":"2026-01-27T15:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.142059 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.142100 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.142108 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.142125 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.142135 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:19Z","lastTransitionTime":"2026-01-27T15:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.245485 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.245546 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.245569 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.245591 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.245602 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:19Z","lastTransitionTime":"2026-01-27T15:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.347876 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.347927 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.347938 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.347953 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.347962 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:19Z","lastTransitionTime":"2026-01-27T15:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.450656 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.450693 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.450703 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.450717 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.450726 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:19Z","lastTransitionTime":"2026-01-27T15:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.553469 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.553556 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.553572 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.553589 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.553624 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:19Z","lastTransitionTime":"2026-01-27T15:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.656369 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.656426 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.656442 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.656466 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.656482 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:19Z","lastTransitionTime":"2026-01-27T15:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.662659 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:19 crc kubenswrapper[4772]: E0127 15:08:19.662836 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.663147 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 17:47:22.8914139 +0000 UTC Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.758328 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.758389 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.758406 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.758428 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.758446 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:19Z","lastTransitionTime":"2026-01-27T15:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.862030 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.862077 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.862089 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.862106 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.862117 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:19Z","lastTransitionTime":"2026-01-27T15:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.964886 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.964939 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.964955 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.964972 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:19 crc kubenswrapper[4772]: I0127 15:08:19.964984 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:19Z","lastTransitionTime":"2026-01-27T15:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.067688 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.067733 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.067742 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.067758 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.067768 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:20Z","lastTransitionTime":"2026-01-27T15:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.170292 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.170367 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.170388 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.170416 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.170437 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:20Z","lastTransitionTime":"2026-01-27T15:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.273246 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.273294 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.273308 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.273338 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.273350 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:20Z","lastTransitionTime":"2026-01-27T15:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.376547 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.376609 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.376626 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.376651 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.376668 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:20Z","lastTransitionTime":"2026-01-27T15:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.479676 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.479721 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.479738 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.479756 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.479766 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:20Z","lastTransitionTime":"2026-01-27T15:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.582969 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.583015 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.583026 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.583046 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.583064 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:20Z","lastTransitionTime":"2026-01-27T15:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.662351 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.662475 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.662483 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:20 crc kubenswrapper[4772]: E0127 15:08:20.662608 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:20 crc kubenswrapper[4772]: E0127 15:08:20.662709 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:20 crc kubenswrapper[4772]: E0127 15:08:20.662843 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.663518 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 02:58:51.448823276 +0000 UTC Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.685748 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.685792 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.685800 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.685815 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.685825 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:20Z","lastTransitionTime":"2026-01-27T15:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.787965 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.788006 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.788014 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.788027 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.788036 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:20Z","lastTransitionTime":"2026-01-27T15:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.891380 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.891434 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.891450 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.891471 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.891485 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:20Z","lastTransitionTime":"2026-01-27T15:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.995517 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.995562 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.995574 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.995591 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:20 crc kubenswrapper[4772]: I0127 15:08:20.995606 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:20Z","lastTransitionTime":"2026-01-27T15:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.098953 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.098990 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.099017 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.099036 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.099069 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:21Z","lastTransitionTime":"2026-01-27T15:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.201532 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.201595 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.201610 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.201627 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.201639 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:21Z","lastTransitionTime":"2026-01-27T15:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.305588 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.305656 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.305692 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.305734 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.305758 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:21Z","lastTransitionTime":"2026-01-27T15:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.408297 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.408349 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.408357 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.408373 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.408383 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:21Z","lastTransitionTime":"2026-01-27T15:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.511696 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.511754 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.511771 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.511796 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.511813 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:21Z","lastTransitionTime":"2026-01-27T15:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.614921 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.614967 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.614976 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.614994 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.615006 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:21Z","lastTransitionTime":"2026-01-27T15:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.662696 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:21 crc kubenswrapper[4772]: E0127 15:08:21.663515 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.663653 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 00:12:42.761807844 +0000 UTC Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.716925 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.716969 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.716981 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.716997 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.717007 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:21Z","lastTransitionTime":"2026-01-27T15:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.820625 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.820683 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.820696 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.820715 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.820730 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:21Z","lastTransitionTime":"2026-01-27T15:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.923364 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.923412 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.923425 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.923446 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:21 crc kubenswrapper[4772]: I0127 15:08:21.923473 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:21Z","lastTransitionTime":"2026-01-27T15:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.026277 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.026399 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.026418 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.026444 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.026464 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:22Z","lastTransitionTime":"2026-01-27T15:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.128640 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.128681 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.128693 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.128709 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.128721 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:22Z","lastTransitionTime":"2026-01-27T15:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.231467 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.231536 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.231554 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.231579 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.231595 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:22Z","lastTransitionTime":"2026-01-27T15:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.334591 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.334643 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.334658 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.334679 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.334693 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:22Z","lastTransitionTime":"2026-01-27T15:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.438138 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.438268 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.438300 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.438322 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.438339 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:22Z","lastTransitionTime":"2026-01-27T15:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.449746 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.449783 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.449791 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.449809 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.449819 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:22Z","lastTransitionTime":"2026-01-27T15:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:22 crc kubenswrapper[4772]: E0127 15:08:22.464150 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:22Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.470683 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.470747 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.470766 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.470790 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.470808 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:22Z","lastTransitionTime":"2026-01-27T15:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:22 crc kubenswrapper[4772]: E0127 15:08:22.493022 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:22Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.498230 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.498311 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.498336 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.498370 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.498393 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:22Z","lastTransitionTime":"2026-01-27T15:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:22 crc kubenswrapper[4772]: E0127 15:08:22.520436 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:22Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.526017 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.526088 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.526112 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.526143 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.526222 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:22Z","lastTransitionTime":"2026-01-27T15:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:22 crc kubenswrapper[4772]: E0127 15:08:22.540004 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:22Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.544155 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.544281 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.544298 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.544320 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.544336 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:22Z","lastTransitionTime":"2026-01-27T15:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:22 crc kubenswrapper[4772]: E0127 15:08:22.562919 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:22Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:22 crc kubenswrapper[4772]: E0127 15:08:22.563134 4772 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.565243 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.565308 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.565328 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.565350 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.565366 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:22Z","lastTransitionTime":"2026-01-27T15:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.662360 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.662467 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:22 crc kubenswrapper[4772]: E0127 15:08:22.662607 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.662636 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:22 crc kubenswrapper[4772]: E0127 15:08:22.662713 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:22 crc kubenswrapper[4772]: E0127 15:08:22.662842 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.664853 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 08:58:03.60725289 +0000 UTC Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.668640 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.668690 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.668709 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.668729 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.668744 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:22Z","lastTransitionTime":"2026-01-27T15:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.771709 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.771785 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.771806 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.771833 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.771854 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:22Z","lastTransitionTime":"2026-01-27T15:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.874612 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.874650 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.874659 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.874677 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.874688 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:22Z","lastTransitionTime":"2026-01-27T15:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.977158 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.977224 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.977235 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.977250 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:22 crc kubenswrapper[4772]: I0127 15:08:22.977261 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:22Z","lastTransitionTime":"2026-01-27T15:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.080153 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.080230 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.080239 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.080252 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.080282 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:23Z","lastTransitionTime":"2026-01-27T15:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.183617 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.183653 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.183664 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.183699 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.183709 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:23Z","lastTransitionTime":"2026-01-27T15:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.285965 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.285999 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.286007 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.286021 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.286030 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:23Z","lastTransitionTime":"2026-01-27T15:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.388646 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.388742 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.388768 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.388798 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.388819 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:23Z","lastTransitionTime":"2026-01-27T15:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.491253 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.491287 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.491298 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.491315 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.491327 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:23Z","lastTransitionTime":"2026-01-27T15:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.593577 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.593603 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.593611 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.593623 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.593631 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:23Z","lastTransitionTime":"2026-01-27T15:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.662977 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:23 crc kubenswrapper[4772]: E0127 15:08:23.663223 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.665186 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 13:33:29.670525141 +0000 UTC Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.695981 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.696004 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.696012 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.696024 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.696033 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:23Z","lastTransitionTime":"2026-01-27T15:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.798775 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.798832 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.798849 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.798872 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.798889 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:23Z","lastTransitionTime":"2026-01-27T15:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.902368 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.902409 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.902419 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.902434 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:23 crc kubenswrapper[4772]: I0127 15:08:23.902444 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:23Z","lastTransitionTime":"2026-01-27T15:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.003985 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.004023 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.004039 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.004055 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.004065 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:24Z","lastTransitionTime":"2026-01-27T15:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.106925 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.106976 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.106988 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.107007 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.107021 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:24Z","lastTransitionTime":"2026-01-27T15:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.208915 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.208982 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.208999 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.209028 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.209087 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:24Z","lastTransitionTime":"2026-01-27T15:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.311437 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.311487 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.311499 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.311516 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.311529 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:24Z","lastTransitionTime":"2026-01-27T15:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.413854 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.413889 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.413901 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.413917 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.413928 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:24Z","lastTransitionTime":"2026-01-27T15:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.516193 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.516236 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.516248 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.516265 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.516277 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:24Z","lastTransitionTime":"2026-01-27T15:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.618722 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.618767 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.618783 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.618798 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.618810 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:24Z","lastTransitionTime":"2026-01-27T15:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.662524 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.662565 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.662529 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:24 crc kubenswrapper[4772]: E0127 15:08:24.662674 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:24 crc kubenswrapper[4772]: E0127 15:08:24.662790 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:24 crc kubenswrapper[4772]: E0127 15:08:24.662884 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.665555 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 21:24:58.031957623 +0000 UTC Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.677716 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f72b451fa77f3fce2c251de546110ab49c7c9e0122759f6ef29a32fde422356\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:08:04Z\\\",\\\"message\\\":\\\"2026-01-27T15:07:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_85dd4958-953b-469d-b48e-3b998cab8570\\\\n2026-01-27T15:07:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_85dd4958-953b-469d-b48e-3b998cab8570 to /host/opt/cni/bin/\\\\n2026-01-27T15:07:19Z [verbose] multus-daemon started\\\\n2026-01-27T15:07:19Z [verbose] Readiness Indicator file check\\\\n2026-01-27T15:08:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.690356 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.702493 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.715630 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.720851 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.720896 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.720907 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.720929 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.720941 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:24Z","lastTransitionTime":"2026-01-27T15:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.743526 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.760966 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.774387 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.787088 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8fde5-4905-4fb1-b683-27ea4921b462\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b47fc0778ba4c5a1e12700735e6f9c52a7341b9eac61071607902a6ec8daf02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://545dcc1be1335a1acf93f16d2e5b4a266dce5b7e736b7c56b80bbf56b3248ced\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wkvpx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.800352 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ql2vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371016c8-5a23-427d-aa0a-0faa241d86a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ql2vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.816627 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.822925 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.822965 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.822976 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.822993 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.823004 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:24Z","lastTransitionTime":"2026-01-27T15:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.829561 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.840874 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.854630 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.876361 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:08:13Z\\\",\\\"message\\\":\\\"-secret-name:openshift-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc006e9cc6b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: openshift-controller-manager-operator,},ClusterIP:10.217.5.58,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.58],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF0127 15:08:13.535689 6829 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:08:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-n2khk_openshift-ovn-kubernetes(736264c8-cd18-479a-88ba-e1ec15dbfdae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.892104 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.902443 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6cfdf02-101c-4f18-9ebe-16002352afce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5b9360189745deea4c8950c4f3b73762c7cb098452c1f31c3df9ab99fc31ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3be141392fbb6ae8ce51085cfc186745bb5ac272899e3202aaed2b7093855c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8c8aeb62c273e8cb31f01742805bd40c4184821c4bbd94f319ff5f074265d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64098f88e61c989a4f2048d222906eacf8c0525f26e109913c1718c9dfb67d20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64098f88e61c989a4f2048d222906eacf8c0525f26e109913c1718c9dfb67d20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.913490 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.923802 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:24Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.925088 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.925138 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.925150 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.925191 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:24 crc kubenswrapper[4772]: I0127 15:08:24.925204 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:24Z","lastTransitionTime":"2026-01-27T15:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.027715 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.027768 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.027778 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.027794 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.027804 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:25Z","lastTransitionTime":"2026-01-27T15:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.129843 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.129908 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.129919 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.129933 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.129945 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:25Z","lastTransitionTime":"2026-01-27T15:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.232808 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.232896 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.232914 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.232939 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.232958 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:25Z","lastTransitionTime":"2026-01-27T15:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.335428 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.335469 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.335481 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.335499 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.335511 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:25Z","lastTransitionTime":"2026-01-27T15:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.438570 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.438704 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.438717 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.438736 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.438751 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:25Z","lastTransitionTime":"2026-01-27T15:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.541144 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.541196 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.541204 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.541219 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.541229 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:25Z","lastTransitionTime":"2026-01-27T15:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.643237 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.643302 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.643311 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.643324 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.643333 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:25Z","lastTransitionTime":"2026-01-27T15:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.662120 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:25 crc kubenswrapper[4772]: E0127 15:08:25.662369 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.663188 4772 scope.go:117] "RemoveContainer" containerID="8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a" Jan 27 15:08:25 crc kubenswrapper[4772]: E0127 15:08:25.663337 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-n2khk_openshift-ovn-kubernetes(736264c8-cd18-479a-88ba-e1ec15dbfdae)\"" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.666560 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 02:55:52.778190418 +0000 UTC Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.672967 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.745664 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.745720 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.745770 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.745789 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.745799 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:25Z","lastTransitionTime":"2026-01-27T15:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.847676 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.847712 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.847720 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.847735 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.847745 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:25Z","lastTransitionTime":"2026-01-27T15:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.950215 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.950252 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.950262 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.950279 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:25 crc kubenswrapper[4772]: I0127 15:08:25.950289 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:25Z","lastTransitionTime":"2026-01-27T15:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.052408 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.052441 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.052449 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.052463 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.052472 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:26Z","lastTransitionTime":"2026-01-27T15:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.155534 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.155575 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.155586 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.155602 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.155612 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:26Z","lastTransitionTime":"2026-01-27T15:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.258317 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.258381 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.258393 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.258407 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.258418 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:26Z","lastTransitionTime":"2026-01-27T15:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.361547 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.361616 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.361633 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.361658 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.361674 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:26Z","lastTransitionTime":"2026-01-27T15:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.464107 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.464156 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.464182 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.464200 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.464212 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:26Z","lastTransitionTime":"2026-01-27T15:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.566578 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.566618 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.566628 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.566642 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.566652 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:26Z","lastTransitionTime":"2026-01-27T15:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.662280 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.662359 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:26 crc kubenswrapper[4772]: E0127 15:08:26.662464 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.662307 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:26 crc kubenswrapper[4772]: E0127 15:08:26.662560 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:26 crc kubenswrapper[4772]: E0127 15:08:26.662629 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.667003 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 17:41:01.673331451 +0000 UTC Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.668954 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.669006 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.669027 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.669132 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.669257 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:26Z","lastTransitionTime":"2026-01-27T15:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.771890 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.771933 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.771945 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.771962 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.771976 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:26Z","lastTransitionTime":"2026-01-27T15:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.875022 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.875064 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.875075 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.875093 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.875107 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:26Z","lastTransitionTime":"2026-01-27T15:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.977533 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.977571 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.977580 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.977594 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:26 crc kubenswrapper[4772]: I0127 15:08:26.977603 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:26Z","lastTransitionTime":"2026-01-27T15:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.079529 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.079567 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.079578 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.079595 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.079607 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:27Z","lastTransitionTime":"2026-01-27T15:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.181937 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.181982 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.181993 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.182011 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.182023 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:27Z","lastTransitionTime":"2026-01-27T15:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.285479 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.285526 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.285539 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.285560 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.285571 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:27Z","lastTransitionTime":"2026-01-27T15:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.388729 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.388780 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.388798 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.388819 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.388838 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:27Z","lastTransitionTime":"2026-01-27T15:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.491368 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.491401 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.491410 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.491423 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.491433 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:27Z","lastTransitionTime":"2026-01-27T15:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.594374 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.594439 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.594463 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.594497 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.594520 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:27Z","lastTransitionTime":"2026-01-27T15:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.662387 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:27 crc kubenswrapper[4772]: E0127 15:08:27.662574 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.667761 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 06:25:56.007323477 +0000 UTC Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.697632 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.697686 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.697704 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.697729 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.697747 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:27Z","lastTransitionTime":"2026-01-27T15:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.800574 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.800644 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.800668 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.800698 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.800720 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:27Z","lastTransitionTime":"2026-01-27T15:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.904084 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.904153 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.904215 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.904244 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:27 crc kubenswrapper[4772]: I0127 15:08:27.904265 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:27Z","lastTransitionTime":"2026-01-27T15:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.007767 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.007843 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.007866 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.007899 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.007923 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:28Z","lastTransitionTime":"2026-01-27T15:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.111094 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.111151 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.111162 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.111239 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.111256 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:28Z","lastTransitionTime":"2026-01-27T15:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.213979 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.214031 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.214042 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.214060 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.214071 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:28Z","lastTransitionTime":"2026-01-27T15:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.317550 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.317632 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.317656 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.317686 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.317711 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:28Z","lastTransitionTime":"2026-01-27T15:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.420808 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.420889 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.420907 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.420936 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.420955 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:28Z","lastTransitionTime":"2026-01-27T15:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.525339 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.525419 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.525437 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.525465 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.525483 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:28Z","lastTransitionTime":"2026-01-27T15:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.628995 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.629064 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.629084 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.629117 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.629136 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:28Z","lastTransitionTime":"2026-01-27T15:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.663065 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.663280 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.663373 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:28 crc kubenswrapper[4772]: E0127 15:08:28.663498 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:28 crc kubenswrapper[4772]: E0127 15:08:28.663576 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:28 crc kubenswrapper[4772]: E0127 15:08:28.663742 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.668607 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 10:23:54.613707222 +0000 UTC Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.731532 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.731571 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.731584 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.731603 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.731617 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:28Z","lastTransitionTime":"2026-01-27T15:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.834754 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.834798 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.834810 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.834828 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.834840 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:28Z","lastTransitionTime":"2026-01-27T15:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.936994 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.937044 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.937052 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.937065 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:28 crc kubenswrapper[4772]: I0127 15:08:28.937074 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:28Z","lastTransitionTime":"2026-01-27T15:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.040084 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.040157 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.040226 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.040262 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.040287 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:29Z","lastTransitionTime":"2026-01-27T15:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.141855 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.141896 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.141907 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.141922 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.141932 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:29Z","lastTransitionTime":"2026-01-27T15:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.244103 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.244150 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.244159 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.244276 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.244295 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:29Z","lastTransitionTime":"2026-01-27T15:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.346805 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.346847 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.346858 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.346876 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.346888 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:29Z","lastTransitionTime":"2026-01-27T15:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.449532 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.449575 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.449591 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.449605 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.449616 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:29Z","lastTransitionTime":"2026-01-27T15:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.552473 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.552547 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.552568 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.552597 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.552620 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:29Z","lastTransitionTime":"2026-01-27T15:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.654973 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.655017 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.655026 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.655040 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.655049 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:29Z","lastTransitionTime":"2026-01-27T15:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.662435 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:29 crc kubenswrapper[4772]: E0127 15:08:29.662591 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.668875 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 01:04:24.947734451 +0000 UTC Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.757259 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.757310 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.757333 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.757355 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.757370 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:29Z","lastTransitionTime":"2026-01-27T15:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.859850 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.859890 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.859916 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.859943 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.859959 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:29Z","lastTransitionTime":"2026-01-27T15:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.963259 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.963294 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.963303 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.963318 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:29 crc kubenswrapper[4772]: I0127 15:08:29.963326 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:29Z","lastTransitionTime":"2026-01-27T15:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.065879 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.065917 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.065934 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.065951 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.065962 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:30Z","lastTransitionTime":"2026-01-27T15:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.167887 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.167925 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.167937 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.167953 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.167966 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:30Z","lastTransitionTime":"2026-01-27T15:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.271740 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.271822 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.271847 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.271878 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.271902 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:30Z","lastTransitionTime":"2026-01-27T15:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.375346 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.375427 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.375445 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.375469 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.375486 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:30Z","lastTransitionTime":"2026-01-27T15:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.478345 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.478395 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.478417 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.478445 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.478464 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:30Z","lastTransitionTime":"2026-01-27T15:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.580888 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.581239 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.581252 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.581269 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.581281 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:30Z","lastTransitionTime":"2026-01-27T15:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.662782 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.662840 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.662782 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:30 crc kubenswrapper[4772]: E0127 15:08:30.662917 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:30 crc kubenswrapper[4772]: E0127 15:08:30.663192 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:30 crc kubenswrapper[4772]: E0127 15:08:30.663238 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.669150 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 11:12:21.075251844 +0000 UTC Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.684074 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.684113 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.684126 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.684143 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.684156 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:30Z","lastTransitionTime":"2026-01-27T15:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.786890 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.786985 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.786995 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.787008 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.787018 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:30Z","lastTransitionTime":"2026-01-27T15:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.889808 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.889858 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.889874 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.889895 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.889908 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:30Z","lastTransitionTime":"2026-01-27T15:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.991918 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.991965 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.991975 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.991990 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:30 crc kubenswrapper[4772]: I0127 15:08:30.992000 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:30Z","lastTransitionTime":"2026-01-27T15:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.095866 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.095921 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.095929 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.095945 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.095955 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:31Z","lastTransitionTime":"2026-01-27T15:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.198790 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.198849 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.198868 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.198893 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.198910 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:31Z","lastTransitionTime":"2026-01-27T15:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.301249 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.301278 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.301286 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.301299 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.301308 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:31Z","lastTransitionTime":"2026-01-27T15:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.403884 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.403937 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.403953 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.403973 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.403991 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:31Z","lastTransitionTime":"2026-01-27T15:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.506228 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.506325 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.506344 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.506360 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.506371 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:31Z","lastTransitionTime":"2026-01-27T15:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.608956 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.609047 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.609067 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.609092 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.609134 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:31Z","lastTransitionTime":"2026-01-27T15:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.662852 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:31 crc kubenswrapper[4772]: E0127 15:08:31.663003 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.670268 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 14:04:13.722393681 +0000 UTC Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.711420 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.711473 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.711488 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.711508 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.711519 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:31Z","lastTransitionTime":"2026-01-27T15:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.813578 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.813641 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.813663 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.813690 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.813711 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:31Z","lastTransitionTime":"2026-01-27T15:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.916215 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.916263 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.916278 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.916295 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:31 crc kubenswrapper[4772]: I0127 15:08:31.916307 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:31Z","lastTransitionTime":"2026-01-27T15:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.019069 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.019111 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.019124 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.019144 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.019155 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:32Z","lastTransitionTime":"2026-01-27T15:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.121629 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.121714 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.121728 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.121750 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.121764 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:32Z","lastTransitionTime":"2026-01-27T15:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.223795 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.223865 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.223888 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.223919 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.223942 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:32Z","lastTransitionTime":"2026-01-27T15:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.327076 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.327123 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.327134 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.327155 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.327192 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:32Z","lastTransitionTime":"2026-01-27T15:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.430991 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.431356 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.431442 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.431473 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.431488 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:32Z","lastTransitionTime":"2026-01-27T15:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.533938 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.534006 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.534025 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.534047 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.534067 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:32Z","lastTransitionTime":"2026-01-27T15:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.637002 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.637139 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.637157 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.637190 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.637202 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:32Z","lastTransitionTime":"2026-01-27T15:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.662543 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:32 crc kubenswrapper[4772]: E0127 15:08:32.662777 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.662869 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.662574 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:32 crc kubenswrapper[4772]: E0127 15:08:32.663028 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:32 crc kubenswrapper[4772]: E0127 15:08:32.663146 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.670458 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 17:34:18.814230479 +0000 UTC Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.740238 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.740277 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.740288 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.740303 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.740314 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:32Z","lastTransitionTime":"2026-01-27T15:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.843576 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.843662 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.843686 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.843716 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.843738 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:32Z","lastTransitionTime":"2026-01-27T15:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.946243 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.946318 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.946341 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.946371 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.946395 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:32Z","lastTransitionTime":"2026-01-27T15:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.957768 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.957817 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.957828 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.957846 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.957858 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:32Z","lastTransitionTime":"2026-01-27T15:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:32 crc kubenswrapper[4772]: E0127 15:08:32.976944 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:32Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.981252 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.981282 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.981292 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.981306 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.981315 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:32Z","lastTransitionTime":"2026-01-27T15:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:32 crc kubenswrapper[4772]: E0127 15:08:32.991985 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:32Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.995032 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.995067 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.995077 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.995092 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:32 crc kubenswrapper[4772]: I0127 15:08:32.995101 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:32Z","lastTransitionTime":"2026-01-27T15:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:33 crc kubenswrapper[4772]: E0127 15:08:33.005134 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:33Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.007749 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.007779 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.007788 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.007800 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.007810 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:33Z","lastTransitionTime":"2026-01-27T15:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:33 crc kubenswrapper[4772]: E0127 15:08:33.018752 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:33Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.021231 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.021268 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.021279 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.021290 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.021299 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:33Z","lastTransitionTime":"2026-01-27T15:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:33 crc kubenswrapper[4772]: E0127 15:08:33.033204 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:33Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:33 crc kubenswrapper[4772]: E0127 15:08:33.033313 4772 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.048675 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.048715 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.048725 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.048738 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.048746 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:33Z","lastTransitionTime":"2026-01-27T15:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.150830 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.150865 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.150877 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.150891 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.150902 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:33Z","lastTransitionTime":"2026-01-27T15:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.253754 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.253808 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.253820 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.253839 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.253853 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:33Z","lastTransitionTime":"2026-01-27T15:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.356620 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.356683 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.356697 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.356716 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.356732 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:33Z","lastTransitionTime":"2026-01-27T15:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.459968 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.460032 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.460048 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.460068 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.460080 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:33Z","lastTransitionTime":"2026-01-27T15:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.562124 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.562230 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.562248 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.562273 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.562289 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:33Z","lastTransitionTime":"2026-01-27T15:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.662562 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:33 crc kubenswrapper[4772]: E0127 15:08:33.662766 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.664589 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.664650 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.664660 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.664694 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.664706 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:33Z","lastTransitionTime":"2026-01-27T15:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.671116 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 09:21:29.382700586 +0000 UTC Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.766992 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.767019 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.767027 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.767040 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.767049 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:33Z","lastTransitionTime":"2026-01-27T15:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.870594 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.870649 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.870666 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.870687 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.870704 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:33Z","lastTransitionTime":"2026-01-27T15:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.973727 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.973766 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.973776 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.973792 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:33 crc kubenswrapper[4772]: I0127 15:08:33.973802 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:33Z","lastTransitionTime":"2026-01-27T15:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.077094 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.077147 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.077157 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.077190 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.077201 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:34Z","lastTransitionTime":"2026-01-27T15:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.180458 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.180512 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.180528 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.180546 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.180560 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:34Z","lastTransitionTime":"2026-01-27T15:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.283897 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.283947 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.283961 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.283984 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.284001 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:34Z","lastTransitionTime":"2026-01-27T15:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.387399 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.387442 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.387455 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.387473 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.387485 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:34Z","lastTransitionTime":"2026-01-27T15:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.416022 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs\") pod \"network-metrics-daemon-ql2vx\" (UID: \"371016c8-5a23-427d-aa0a-0faa241d86a7\") " pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:34 crc kubenswrapper[4772]: E0127 15:08:34.416245 4772 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 27 15:08:34 crc kubenswrapper[4772]: E0127 15:08:34.416340 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs podName:371016c8-5a23-427d-aa0a-0faa241d86a7 nodeName:}" failed. No retries permitted until 2026-01-27 15:09:38.416318497 +0000 UTC m=+164.396927645 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs") pod "network-metrics-daemon-ql2vx" (UID: "371016c8-5a23-427d-aa0a-0faa241d86a7") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.490059 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.490106 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.490117 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.490135 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.490147 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:34Z","lastTransitionTime":"2026-01-27T15:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.592537 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.592563 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.592571 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.592583 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.592591 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:34Z","lastTransitionTime":"2026-01-27T15:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.662756 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.662887 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:34 crc kubenswrapper[4772]: E0127 15:08:34.662931 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.662993 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:34 crc kubenswrapper[4772]: E0127 15:08:34.663090 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:34 crc kubenswrapper[4772]: E0127 15:08:34.663162 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.671577 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 09:02:18.202682362 +0000 UTC Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.681239 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a2b20172f77b5d595f6543c954936ade12fd0cf0625b1abe17cb400adfd8842\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ab1226587c192dfc7094511fb2d0ce13cd3e47e84a683ec1a3a175f2496c015\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.693364 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67794a44-d793-4fd7-9e54-e40437f67c0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80a5eaeeb793907d34b34a1bf5727c3da1dd01beb45fd8ebdc224b650b9f9aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lh6ph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4hwxn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.695231 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.695429 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.695447 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.695465 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.695476 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:34Z","lastTransitionTime":"2026-01-27T15:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.713158 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"736264c8-cd18-479a-88ba-e1ec15dbfdae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:08:13Z\\\",\\\"message\\\":\\\"-secret-name:openshift-controller-manager-operator-serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc006e9cc6b \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: openshift-controller-manager-operator,},ClusterIP:10.217.5.58,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.58],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF0127 15:08:13.535689 6829 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:08:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-n2khk_openshift-ovn-kubernetes(736264c8-cd18-479a-88ba-e1ec15dbfdae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2dt6g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-n2khk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.728315 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11f71341-1cdc-430d-8d90-a87af2a493f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0127 15:07:09.417986 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0127 15:07:09.418905 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2455853299/tls.crt::/tmp/serving-cert-2455853299/tls.key\\\\\\\"\\\\nI0127 15:07:14.676378 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0127 15:07:14.679920 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0127 15:07:14.679946 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0127 15:07:14.679972 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0127 15:07:14.679980 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0127 15:07:14.686726 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0127 15:07:14.686753 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686793 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0127 15:07:14.686798 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0127 15:07:14.686801 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0127 15:07:14.686805 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0127 15:07:14.686807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0127 15:07:14.686800 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0127 15:07:14.690409 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.740394 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6cfdf02-101c-4f18-9ebe-16002352afce\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5b9360189745deea4c8950c4f3b73762c7cb098452c1f31c3df9ab99fc31ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3be141392fbb6ae8ce51085cfc186745bb5ac272899e3202aaed2b7093855c4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8c8aeb62c273e8cb31f01742805bd40c4184821c4bbd94f319ff5f074265d75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64098f88e61c989a4f2048d222906eacf8c0525f26e109913c1718c9dfb67d20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64098f88e61c989a4f2048d222906eacf8c0525f26e109913c1718c9dfb67d20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.755308 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.769134 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acef947-6310-4ac0-bc84-a06d91f84cb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://790f07e4c1ef52bf6e541034bdd5cc70277cdd5522fd74919677e8dc97f13490\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af9580aa20d3a39dcbe87c6eb9b7f294c101a80c6a360abe4caf6e47270bc538\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e668395335c1e28bff6af3da4358538cbc78d8a8837cc1bf1fa58053d0a792d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4981ab36202d45fb2e3b5474b83bb97a70ee85761ed4595969d3629a9a14c7cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e7feb8db10ed98640d39d316ec7bee08cf060c2e5056803a4a7b6afe5bc9dbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c31746847c1db9cc19a3f43cedd3a346b1991a96cca59a07c709e8a57f546e57\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9ee51cff497be766cad5cd7ead7b3e51e5e98af36d4af1a9d0da91a55a079d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:07:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntsst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c7pdz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.781604 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x7jwx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f72b451fa77f3fce2c251de546110ab49c7c9e0122759f6ef29a32fde422356\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-27T15:08:04Z\\\",\\\"message\\\":\\\"2026-01-27T15:07:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_85dd4958-953b-469d-b48e-3b998cab8570\\\\n2026-01-27T15:07:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_85dd4958-953b-469d-b48e-3b998cab8570 to /host/opt/cni/bin/\\\\n2026-01-27T15:07:19Z [verbose] multus-daemon started\\\\n2026-01-27T15:07:19Z [verbose] Readiness Indicator file check\\\\n2026-01-27T15:08:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d4sv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x7jwx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.793028 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7947125a-23ff-4bc4-9f9a-743173e3bf96\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ec44d988753e2d56359eaeae0782085a35526439972fa9fdef4dca6a95a285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cf43337dd65773891148fae7e97baf3bad08bcc42d35ccbcb396924a5e1328\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a044713f62e865404e3a08eb2b72e000eb4418bd86be24564fc1dff9c3fb8ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.798801 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.798853 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.798861 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.798879 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.798890 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:34Z","lastTransitionTime":"2026-01-27T15:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.802036 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e737111-6155-41bb-8a02-7dd880eade01\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109762c1c77a786a3c953ead1b65a17e9401f8ba205858ffc79f6e188d5005df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a8432f1dcc97ca9d30249542dfebc79f098859c0f7e04a637be764939fb6072\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a8432f1dcc97ca9d30249542dfebc79f098859c0f7e04a637be764939fb6072\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.810551 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dtdj6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95a893d4-4faa-40b2-b505-9698fe428ba8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e03e42cc2bd2d62e397d8138f1bfb5d4f3ef8ca22faec9ae48f6bcf5b22d964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7pswh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dtdj6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.827243 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9eb558c7-0e97-42c7-96e9-8b170ac2a3df\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6faa9eac1866fffe234666ebc0ccebaa65ed897a10df3c2f4c60af170a24ff10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f731c47cd6cbadce3d6caf107e6f7b47ccf52a192f27d5c88455d94ed51e724c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58ba8d2a22f341cf2a156befb63aabda598800197b206ac09ceae36296b428d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af25a6897fe69ff57615b72ae31abd93d49a9dfd960f98842eaf8db1b327fe5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c88a3bea343764365f32d13ed70b783845cce76318d2378c7bbcc2005e97326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b19c9dbf8cf2f44ddd32a80e600c191dfb1efb1ca81580a520c4f8eb767332a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f477eb4e1b649a5aa1f5e85f03691700b168b968e263e772a6c54f298e10b75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:56Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cddf104d24d1462376c31b95bf9f29f354d4d7c5bb60f0b3391b6f4d692b7970\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-27T15:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-27T15:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:06:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.838890 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b1528b753084926155afff8e9a49f8be08e24697c358fc5079cd2ef0e88449\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.848763 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4877bb2bd2a893b384e8153ff7b20a81a4640200989e85efd4c654bb9ecf0c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.857137 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46tm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed65bae-f1c4-4c97-bb6d-d4144fe2532b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a94e333403eaaf5f1c05153d18d284dda3a2cde1d727e5652613049041fe348\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qln7r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46tm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.867609 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cc8fde5-4905-4fb1-b683-27ea4921b462\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b47fc0778ba4c5a1e12700735e6f9c52a7341b9eac61071607902a6ec8daf02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://545dcc1be1335a1acf93f16d2e5b4a266dce5b7e736b7c56b80bbf56b3248ced\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-27T15:07:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghx7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-wkvpx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.876750 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-ql2vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"371016c8-5a23-427d-aa0a-0faa241d86a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8l57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-27T15:07:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-ql2vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.886691 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.896348 4772 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-27T15:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:34Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.900837 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.900867 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.900880 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.900896 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:34 crc kubenswrapper[4772]: I0127 15:08:34.900908 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:34Z","lastTransitionTime":"2026-01-27T15:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.002803 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.002849 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.002862 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.002879 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.002892 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:35Z","lastTransitionTime":"2026-01-27T15:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.105576 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.105618 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.105627 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.105643 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.105654 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:35Z","lastTransitionTime":"2026-01-27T15:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.207848 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.207897 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.207908 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.207926 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.207937 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:35Z","lastTransitionTime":"2026-01-27T15:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.310000 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.310033 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.310043 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.310057 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.310067 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:35Z","lastTransitionTime":"2026-01-27T15:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.413313 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.413396 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.413428 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.413467 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.413489 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:35Z","lastTransitionTime":"2026-01-27T15:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.516310 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.516351 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.516362 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.516378 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.516390 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:35Z","lastTransitionTime":"2026-01-27T15:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.617902 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.617929 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.617937 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.617951 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.617960 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:35Z","lastTransitionTime":"2026-01-27T15:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.662896 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:35 crc kubenswrapper[4772]: E0127 15:08:35.663049 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.672098 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 21:09:33.025518333 +0000 UTC Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.720012 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.720091 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.720117 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.720145 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.720163 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:35Z","lastTransitionTime":"2026-01-27T15:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.822329 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.822361 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.822371 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.822385 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.822395 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:35Z","lastTransitionTime":"2026-01-27T15:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.925535 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.925641 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.925687 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.925707 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:35 crc kubenswrapper[4772]: I0127 15:08:35.925721 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:35Z","lastTransitionTime":"2026-01-27T15:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.028618 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.028658 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.028667 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.028683 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.028692 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:36Z","lastTransitionTime":"2026-01-27T15:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.131301 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.131411 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.131447 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.131476 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.131497 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:36Z","lastTransitionTime":"2026-01-27T15:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.234565 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.234609 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.234620 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.234635 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.234644 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:36Z","lastTransitionTime":"2026-01-27T15:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.336731 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.336793 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.336812 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.336840 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.336858 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:36Z","lastTransitionTime":"2026-01-27T15:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.440009 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.440063 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.440076 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.440094 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.440107 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:36Z","lastTransitionTime":"2026-01-27T15:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.542655 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.542706 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.542729 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.542750 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.542765 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:36Z","lastTransitionTime":"2026-01-27T15:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.644656 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.644714 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.644731 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.644756 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.644771 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:36Z","lastTransitionTime":"2026-01-27T15:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.662519 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.662544 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.662676 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:36 crc kubenswrapper[4772]: E0127 15:08:36.662817 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:36 crc kubenswrapper[4772]: E0127 15:08:36.662882 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:36 crc kubenswrapper[4772]: E0127 15:08:36.662961 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.672680 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 12:06:33.848817085 +0000 UTC Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.746986 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.747025 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.747036 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.747051 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.747062 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:36Z","lastTransitionTime":"2026-01-27T15:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.849363 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.849399 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.849409 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.849424 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.849447 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:36Z","lastTransitionTime":"2026-01-27T15:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.951892 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.951948 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.951960 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.951977 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:36 crc kubenswrapper[4772]: I0127 15:08:36.951990 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:36Z","lastTransitionTime":"2026-01-27T15:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.053849 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.053888 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.053899 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.053916 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.053926 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:37Z","lastTransitionTime":"2026-01-27T15:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.155546 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.155602 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.155616 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.155633 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.155646 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:37Z","lastTransitionTime":"2026-01-27T15:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.257243 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.257282 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.257294 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.257311 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.257324 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:37Z","lastTransitionTime":"2026-01-27T15:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.359734 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.359784 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.359800 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.359815 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.359826 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:37Z","lastTransitionTime":"2026-01-27T15:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.461772 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.461806 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.461814 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.461827 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.461835 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:37Z","lastTransitionTime":"2026-01-27T15:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.564466 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.564542 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.564555 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.564572 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.564581 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:37Z","lastTransitionTime":"2026-01-27T15:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.662926 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:37 crc kubenswrapper[4772]: E0127 15:08:37.663095 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.666968 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.667051 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.667071 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.667096 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.667115 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:37Z","lastTransitionTime":"2026-01-27T15:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.672832 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 18:57:41.81792312 +0000 UTC Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.770220 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.770284 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.770299 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.770322 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.770336 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:37Z","lastTransitionTime":"2026-01-27T15:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.872638 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.872687 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.872698 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.872724 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.872741 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:37Z","lastTransitionTime":"2026-01-27T15:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.974754 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.974789 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.974796 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.974810 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:37 crc kubenswrapper[4772]: I0127 15:08:37.974821 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:37Z","lastTransitionTime":"2026-01-27T15:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.076662 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.076705 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.076714 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.076729 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.076738 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:38Z","lastTransitionTime":"2026-01-27T15:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.180037 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.180077 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.180086 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.180098 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.180108 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:38Z","lastTransitionTime":"2026-01-27T15:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.282402 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.282445 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.282493 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.282514 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.282525 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:38Z","lastTransitionTime":"2026-01-27T15:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.385026 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.385101 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.385118 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.385143 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.385158 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:38Z","lastTransitionTime":"2026-01-27T15:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.487720 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.487767 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.487788 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.487810 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.487825 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:38Z","lastTransitionTime":"2026-01-27T15:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.591492 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.591539 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.591549 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.591568 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.591578 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:38Z","lastTransitionTime":"2026-01-27T15:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.662035 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.662205 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:38 crc kubenswrapper[4772]: E0127 15:08:38.662295 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.662355 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:38 crc kubenswrapper[4772]: E0127 15:08:38.662523 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:38 crc kubenswrapper[4772]: E0127 15:08:38.662643 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.673094 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 07:18:27.231225909 +0000 UTC Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.694164 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.694237 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.694250 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.694271 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.694326 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:38Z","lastTransitionTime":"2026-01-27T15:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.798101 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.798138 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.798148 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.798186 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.798204 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:38Z","lastTransitionTime":"2026-01-27T15:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.901053 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.901102 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.901111 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.901126 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:38 crc kubenswrapper[4772]: I0127 15:08:38.901135 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:38Z","lastTransitionTime":"2026-01-27T15:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.003462 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.003503 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.003519 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.003542 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.003559 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:39Z","lastTransitionTime":"2026-01-27T15:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.106494 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.106534 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.106544 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.106560 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.106569 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:39Z","lastTransitionTime":"2026-01-27T15:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.209206 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.209293 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.209310 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.209335 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.209351 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:39Z","lastTransitionTime":"2026-01-27T15:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.311703 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.311732 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.311740 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.311753 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.311762 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:39Z","lastTransitionTime":"2026-01-27T15:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.414641 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.414693 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.414707 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.414726 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.414739 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:39Z","lastTransitionTime":"2026-01-27T15:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.517328 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.517368 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.517379 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.517397 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.517409 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:39Z","lastTransitionTime":"2026-01-27T15:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.620849 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.620914 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.620933 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.620956 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.620968 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:39Z","lastTransitionTime":"2026-01-27T15:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.662323 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:39 crc kubenswrapper[4772]: E0127 15:08:39.662964 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.663481 4772 scope.go:117] "RemoveContainer" containerID="8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a" Jan 27 15:08:39 crc kubenswrapper[4772]: E0127 15:08:39.663753 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-n2khk_openshift-ovn-kubernetes(736264c8-cd18-479a-88ba-e1ec15dbfdae)\"" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.673274 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 08:03:33.417504168 +0000 UTC Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.724476 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.724558 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.724573 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.724597 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.724613 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:39Z","lastTransitionTime":"2026-01-27T15:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.827579 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.827642 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.827654 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.827675 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.827691 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:39Z","lastTransitionTime":"2026-01-27T15:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.930382 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.930451 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.930466 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.930489 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:39 crc kubenswrapper[4772]: I0127 15:08:39.930507 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:39Z","lastTransitionTime":"2026-01-27T15:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.032819 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.032932 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.032943 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.032956 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.032965 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:40Z","lastTransitionTime":"2026-01-27T15:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.136798 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.136835 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.136843 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.136857 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.136866 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:40Z","lastTransitionTime":"2026-01-27T15:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.239402 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.239446 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.239459 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.239474 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.239482 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:40Z","lastTransitionTime":"2026-01-27T15:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.342300 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.342365 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.342388 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.342422 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.342442 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:40Z","lastTransitionTime":"2026-01-27T15:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.444278 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.444326 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.444341 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.444360 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.444372 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:40Z","lastTransitionTime":"2026-01-27T15:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.546975 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.547033 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.547043 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.547057 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.547066 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:40Z","lastTransitionTime":"2026-01-27T15:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.649859 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.649904 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.649923 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.649941 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.649952 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:40Z","lastTransitionTime":"2026-01-27T15:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.662335 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:40 crc kubenswrapper[4772]: E0127 15:08:40.662476 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.662513 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.662571 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:40 crc kubenswrapper[4772]: E0127 15:08:40.662623 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:40 crc kubenswrapper[4772]: E0127 15:08:40.662659 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.673857 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 00:21:21.841021375 +0000 UTC Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.751987 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.752024 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.752032 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.752046 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.752055 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:40Z","lastTransitionTime":"2026-01-27T15:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.854219 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.854254 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.854263 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.854276 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.854285 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:40Z","lastTransitionTime":"2026-01-27T15:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.956629 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.956689 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.956708 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.956733 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:40 crc kubenswrapper[4772]: I0127 15:08:40.956750 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:40Z","lastTransitionTime":"2026-01-27T15:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.059258 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.059306 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.059316 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.059331 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.059340 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:41Z","lastTransitionTime":"2026-01-27T15:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.162811 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.162849 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.162859 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.162874 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.162884 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:41Z","lastTransitionTime":"2026-01-27T15:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.264955 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.264995 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.265008 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.265025 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.265037 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:41Z","lastTransitionTime":"2026-01-27T15:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.367618 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.367682 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.367712 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.367739 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.367757 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:41Z","lastTransitionTime":"2026-01-27T15:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.470897 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.470935 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.470943 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.470956 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.470964 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:41Z","lastTransitionTime":"2026-01-27T15:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.573330 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.573363 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.573371 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.573384 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.573393 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:41Z","lastTransitionTime":"2026-01-27T15:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.662360 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:41 crc kubenswrapper[4772]: E0127 15:08:41.662497 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.674467 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 02:50:15.293223886 +0000 UTC Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.676047 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.676079 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.676090 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.676107 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.676116 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:41Z","lastTransitionTime":"2026-01-27T15:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.779636 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.779698 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.779716 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.779745 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.779762 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:41Z","lastTransitionTime":"2026-01-27T15:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.883260 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.883337 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.883356 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.883381 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.883399 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:41Z","lastTransitionTime":"2026-01-27T15:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.986147 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.986245 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.986263 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.986291 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:41 crc kubenswrapper[4772]: I0127 15:08:41.986309 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:41Z","lastTransitionTime":"2026-01-27T15:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.088389 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.088435 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.088446 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.088466 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.088479 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:42Z","lastTransitionTime":"2026-01-27T15:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.190925 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.191006 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.191031 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.191062 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.191087 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:42Z","lastTransitionTime":"2026-01-27T15:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.294884 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.294927 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.294936 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.294953 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.294964 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:42Z","lastTransitionTime":"2026-01-27T15:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.397413 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.397459 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.397469 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.397487 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.397499 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:42Z","lastTransitionTime":"2026-01-27T15:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.499666 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.499716 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.499745 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.499763 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.499774 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:42Z","lastTransitionTime":"2026-01-27T15:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.602455 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.602498 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.602511 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.602530 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.602546 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:42Z","lastTransitionTime":"2026-01-27T15:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.662709 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.662863 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:42 crc kubenswrapper[4772]: E0127 15:08:42.663033 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.663061 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:42 crc kubenswrapper[4772]: E0127 15:08:42.663589 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:42 crc kubenswrapper[4772]: E0127 15:08:42.663681 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.674650 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 11:58:18.536286688 +0000 UTC Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.705370 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.705420 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.705433 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.705449 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.705461 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:42Z","lastTransitionTime":"2026-01-27T15:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.808581 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.808643 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.808654 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.808669 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.808679 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:42Z","lastTransitionTime":"2026-01-27T15:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.911967 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.912068 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.912091 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.912113 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:42 crc kubenswrapper[4772]: I0127 15:08:42.912127 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:42Z","lastTransitionTime":"2026-01-27T15:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.015519 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.015558 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.015567 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.015582 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.015591 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:43Z","lastTransitionTime":"2026-01-27T15:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.108866 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.108906 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.108918 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.108938 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.108955 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:43Z","lastTransitionTime":"2026-01-27T15:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:43 crc kubenswrapper[4772]: E0127 15:08:43.122486 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:43Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.129684 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.129725 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.129735 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.129758 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.129772 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:43Z","lastTransitionTime":"2026-01-27T15:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:43 crc kubenswrapper[4772]: E0127 15:08:43.142909 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:43Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.146929 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.146964 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.146975 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.147000 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.147011 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:43Z","lastTransitionTime":"2026-01-27T15:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:43 crc kubenswrapper[4772]: E0127 15:08:43.161569 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:43Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.165930 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.165977 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.165989 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.166007 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.166020 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:43Z","lastTransitionTime":"2026-01-27T15:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:43 crc kubenswrapper[4772]: E0127 15:08:43.178019 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:43Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.181381 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.181418 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.181426 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.181441 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.181450 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:43Z","lastTransitionTime":"2026-01-27T15:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:43 crc kubenswrapper[4772]: E0127 15:08:43.191904 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-27T15:08:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3e5a8897-2f3c-4c79-87ef-bcf3ebf59cdb\\\",\\\"systemUUID\\\":\\\"3933c4f3-43c9-48b4-998d-ee6c7e3cb9de\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-27T15:08:43Z is after 2025-08-24T17:21:41Z" Jan 27 15:08:43 crc kubenswrapper[4772]: E0127 15:08:43.192019 4772 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.193288 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.193323 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.193335 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.193350 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.193361 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:43Z","lastTransitionTime":"2026-01-27T15:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.295917 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.295962 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.295973 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.295991 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.296002 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:43Z","lastTransitionTime":"2026-01-27T15:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.398454 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.398510 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.398522 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.398542 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.398553 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:43Z","lastTransitionTime":"2026-01-27T15:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.501532 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.501574 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.501585 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.501599 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.501608 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:43Z","lastTransitionTime":"2026-01-27T15:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.604244 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.604319 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.604346 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.604365 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.604382 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:43Z","lastTransitionTime":"2026-01-27T15:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.661896 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:43 crc kubenswrapper[4772]: E0127 15:08:43.662217 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.675212 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 10:30:16.891761801 +0000 UTC Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.706903 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.706941 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.706952 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.706968 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.706979 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:43Z","lastTransitionTime":"2026-01-27T15:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.809466 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.809532 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.809549 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.809567 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.809579 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:43Z","lastTransitionTime":"2026-01-27T15:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.912226 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.912274 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.912287 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.912305 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:43 crc kubenswrapper[4772]: I0127 15:08:43.912318 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:43Z","lastTransitionTime":"2026-01-27T15:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.015335 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.015381 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.015391 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.015404 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.015414 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:44Z","lastTransitionTime":"2026-01-27T15:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.117455 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.117735 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.117814 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.117885 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.117950 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:44Z","lastTransitionTime":"2026-01-27T15:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.220228 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.220261 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.220288 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.220314 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.220326 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:44Z","lastTransitionTime":"2026-01-27T15:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.323340 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.323395 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.323412 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.323431 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.323442 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:44Z","lastTransitionTime":"2026-01-27T15:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.426054 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.426102 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.426115 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.426132 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.426144 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:44Z","lastTransitionTime":"2026-01-27T15:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.529379 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.529433 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.529445 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.529462 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.529471 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:44Z","lastTransitionTime":"2026-01-27T15:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.633351 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.633413 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.633425 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.633448 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.633461 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:44Z","lastTransitionTime":"2026-01-27T15:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.662715 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.662830 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.662908 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:44 crc kubenswrapper[4772]: E0127 15:08:44.663139 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:44 crc kubenswrapper[4772]: E0127 15:08:44.663230 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:44 crc kubenswrapper[4772]: E0127 15:08:44.663291 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.676062 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 08:14:04.428354454 +0000 UTC Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.703729 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-dtdj6" podStartSLOduration=88.703678825 podStartE2EDuration="1m28.703678825s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:08:44.703432318 +0000 UTC m=+110.684041446" watchObservedRunningTime="2026-01-27 15:08:44.703678825 +0000 UTC m=+110.684287923" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.736401 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.736510 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.736525 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.736550 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.736564 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:44Z","lastTransitionTime":"2026-01-27T15:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.739089 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=90.739065049 podStartE2EDuration="1m30.739065049s" podCreationTimestamp="2026-01-27 15:07:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:08:44.735674421 +0000 UTC m=+110.716283519" watchObservedRunningTime="2026-01-27 15:08:44.739065049 +0000 UTC m=+110.719674167" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.824393 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-q46tm" podStartSLOduration=88.824372436 podStartE2EDuration="1m28.824372436s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:08:44.810629889 +0000 UTC m=+110.791239007" watchObservedRunningTime="2026-01-27 15:08:44.824372436 +0000 UTC m=+110.804981534" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.834038 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-wkvpx" podStartSLOduration=87.834017065 podStartE2EDuration="1m27.834017065s" podCreationTimestamp="2026-01-27 15:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:08:44.824079088 +0000 UTC m=+110.804688196" watchObservedRunningTime="2026-01-27 15:08:44.834017065 +0000 UTC m=+110.814626163" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.838985 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.839264 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.839348 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.839440 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.839521 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:44Z","lastTransitionTime":"2026-01-27T15:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.861296 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=57.861279224 podStartE2EDuration="57.861279224s" podCreationTimestamp="2026-01-27 15:07:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:08:44.860983105 +0000 UTC m=+110.841592203" watchObservedRunningTime="2026-01-27 15:08:44.861279224 +0000 UTC m=+110.841888322" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.887269 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podStartSLOduration=88.887250985 podStartE2EDuration="1m28.887250985s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:08:44.886636117 +0000 UTC m=+110.867245225" watchObservedRunningTime="2026-01-27 15:08:44.887250985 +0000 UTC m=+110.867860093" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.922610 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=90.922580067 podStartE2EDuration="1m30.922580067s" podCreationTimestamp="2026-01-27 15:07:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:08:44.921960849 +0000 UTC m=+110.902569947" watchObservedRunningTime="2026-01-27 15:08:44.922580067 +0000 UTC m=+110.903189165" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.934072 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=19.934049168 podStartE2EDuration="19.934049168s" podCreationTimestamp="2026-01-27 15:08:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:08:44.932767431 +0000 UTC m=+110.913376529" watchObservedRunningTime="2026-01-27 15:08:44.934049168 +0000 UTC m=+110.914658266" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.942678 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.942728 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.942738 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.942759 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.942772 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:44Z","lastTransitionTime":"2026-01-27T15:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:44 crc kubenswrapper[4772]: I0127 15:08:44.980956 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-c7pdz" podStartSLOduration=88.980932184 podStartE2EDuration="1m28.980932184s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:08:44.979952356 +0000 UTC m=+110.960561454" watchObservedRunningTime="2026-01-27 15:08:44.980932184 +0000 UTC m=+110.961541282" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.026103 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-x7jwx" podStartSLOduration=89.02608288 podStartE2EDuration="1m29.02608288s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:08:45.009323996 +0000 UTC m=+110.989933094" watchObservedRunningTime="2026-01-27 15:08:45.02608288 +0000 UTC m=+111.006691978" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.045188 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.045240 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.045250 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.045268 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.045279 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:45Z","lastTransitionTime":"2026-01-27T15:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.147771 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.147813 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.147825 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.147843 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.147854 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:45Z","lastTransitionTime":"2026-01-27T15:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.251688 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.251740 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.251751 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.251774 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.251785 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:45Z","lastTransitionTime":"2026-01-27T15:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.356097 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.356132 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.356140 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.356154 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.356179 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:45Z","lastTransitionTime":"2026-01-27T15:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.459841 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.459912 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.459933 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.459958 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.459977 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:45Z","lastTransitionTime":"2026-01-27T15:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.563725 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.563793 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.563809 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.563836 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.563850 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:45Z","lastTransitionTime":"2026-01-27T15:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.662494 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:45 crc kubenswrapper[4772]: E0127 15:08:45.663230 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.666711 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.666767 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.666783 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.666801 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.666815 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:45Z","lastTransitionTime":"2026-01-27T15:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.676841 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 16:28:44.801174951 +0000 UTC Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.769761 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.769801 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.769810 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.769827 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.769836 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:45Z","lastTransitionTime":"2026-01-27T15:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.872801 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.872862 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.872880 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.872900 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.872915 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:45Z","lastTransitionTime":"2026-01-27T15:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.976057 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.976099 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.976108 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.976125 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:45 crc kubenswrapper[4772]: I0127 15:08:45.976136 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:45Z","lastTransitionTime":"2026-01-27T15:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.079111 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.079337 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.079377 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.079408 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.079430 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:46Z","lastTransitionTime":"2026-01-27T15:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.181439 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.181470 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.181480 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.181496 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.181507 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:46Z","lastTransitionTime":"2026-01-27T15:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.284012 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.284046 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.284053 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.284068 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.284078 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:46Z","lastTransitionTime":"2026-01-27T15:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.386336 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.386386 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.386399 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.386418 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.386429 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:46Z","lastTransitionTime":"2026-01-27T15:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.490089 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.490159 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.490190 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.490211 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.490223 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:46Z","lastTransitionTime":"2026-01-27T15:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.592336 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.592369 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.592378 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.592393 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.592403 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:46Z","lastTransitionTime":"2026-01-27T15:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.662887 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.662961 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.662889 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:46 crc kubenswrapper[4772]: E0127 15:08:46.663015 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:46 crc kubenswrapper[4772]: E0127 15:08:46.663214 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:46 crc kubenswrapper[4772]: E0127 15:08:46.663335 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.678044 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 09:43:43.180326489 +0000 UTC Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.694996 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.695052 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.695060 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.695073 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.695083 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:46Z","lastTransitionTime":"2026-01-27T15:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.797530 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.797575 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.797583 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.797597 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.797606 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:46Z","lastTransitionTime":"2026-01-27T15:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.899664 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.899700 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.899709 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.899724 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:46 crc kubenswrapper[4772]: I0127 15:08:46.899734 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:46Z","lastTransitionTime":"2026-01-27T15:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.001319 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.001364 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.001376 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.001395 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.001407 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:47Z","lastTransitionTime":"2026-01-27T15:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.104034 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.104080 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.104091 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.104108 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.104120 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:47Z","lastTransitionTime":"2026-01-27T15:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.206555 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.206588 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.206596 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.206608 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.206617 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:47Z","lastTransitionTime":"2026-01-27T15:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.309333 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.309365 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.309375 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.309405 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.309419 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:47Z","lastTransitionTime":"2026-01-27T15:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.412314 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.412346 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.412355 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.412369 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.412381 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:47Z","lastTransitionTime":"2026-01-27T15:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.515221 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.515259 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.515268 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.515283 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.515292 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:47Z","lastTransitionTime":"2026-01-27T15:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.618159 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.618217 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.618227 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.618240 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.618249 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:47Z","lastTransitionTime":"2026-01-27T15:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.662252 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:47 crc kubenswrapper[4772]: E0127 15:08:47.662388 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.679143 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 15:32:06.965387102 +0000 UTC Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.720107 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.720153 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.720163 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.720211 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.720224 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:47Z","lastTransitionTime":"2026-01-27T15:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.822884 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.822922 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.822939 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.822954 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.822966 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:47Z","lastTransitionTime":"2026-01-27T15:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.925105 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.925136 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.925144 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.925158 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:47 crc kubenswrapper[4772]: I0127 15:08:47.925199 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:47Z","lastTransitionTime":"2026-01-27T15:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.027308 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.027338 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.027348 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.027365 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.027376 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:48Z","lastTransitionTime":"2026-01-27T15:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.484751 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.484787 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.484797 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.484812 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.484821 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:48Z","lastTransitionTime":"2026-01-27T15:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.586677 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.586720 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.586731 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.586746 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.586757 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:48Z","lastTransitionTime":"2026-01-27T15:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.662312 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.662353 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:48 crc kubenswrapper[4772]: E0127 15:08:48.662643 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.662729 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:48 crc kubenswrapper[4772]: E0127 15:08:48.662794 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:48 crc kubenswrapper[4772]: E0127 15:08:48.662924 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.679447 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 06:17:02.770680884 +0000 UTC Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.689827 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.689863 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.689871 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.689886 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.689899 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:48Z","lastTransitionTime":"2026-01-27T15:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.791966 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.792003 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.792014 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.792027 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.792036 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:48Z","lastTransitionTime":"2026-01-27T15:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.894488 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.894548 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.894560 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.894577 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.894590 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:48Z","lastTransitionTime":"2026-01-27T15:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.997344 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.997386 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.997396 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.997410 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:48 crc kubenswrapper[4772]: I0127 15:08:48.997420 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:48Z","lastTransitionTime":"2026-01-27T15:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.099594 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.099634 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.099644 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.099659 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.099670 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:49Z","lastTransitionTime":"2026-01-27T15:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.202529 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.203254 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.203283 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.203306 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.203319 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:49Z","lastTransitionTime":"2026-01-27T15:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.307075 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.307132 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.307148 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.307189 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.307206 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:49Z","lastTransitionTime":"2026-01-27T15:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.410499 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.410566 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.410597 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.410624 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.410648 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:49Z","lastTransitionTime":"2026-01-27T15:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.513396 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.513446 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.513459 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.513475 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.513485 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:49Z","lastTransitionTime":"2026-01-27T15:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.616088 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.616208 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.616258 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.616281 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.616297 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:49Z","lastTransitionTime":"2026-01-27T15:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.662775 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:49 crc kubenswrapper[4772]: E0127 15:08:49.662955 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.680229 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 17:09:09.640463772 +0000 UTC Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.718963 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.719005 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.719028 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.719048 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.719061 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:49Z","lastTransitionTime":"2026-01-27T15:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.822307 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.822359 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.822375 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.822400 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.822418 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:49Z","lastTransitionTime":"2026-01-27T15:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.925958 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.926007 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.926028 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.926055 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:49 crc kubenswrapper[4772]: I0127 15:08:49.926067 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:49Z","lastTransitionTime":"2026-01-27T15:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.029156 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.029263 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.029288 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.029319 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.029340 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:50Z","lastTransitionTime":"2026-01-27T15:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.132640 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.132721 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.132737 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.132763 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.132778 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:50Z","lastTransitionTime":"2026-01-27T15:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.235298 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.235385 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.235400 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.235419 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.235433 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:50Z","lastTransitionTime":"2026-01-27T15:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.338768 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.338835 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.338853 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.338880 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.338893 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:50Z","lastTransitionTime":"2026-01-27T15:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.443655 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.443729 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.443744 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.443766 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.443783 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:50Z","lastTransitionTime":"2026-01-27T15:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.546888 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.546968 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.546990 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.547030 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.547070 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:50Z","lastTransitionTime":"2026-01-27T15:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.649227 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.649267 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.649280 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.649297 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.649310 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:50Z","lastTransitionTime":"2026-01-27T15:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.662495 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:50 crc kubenswrapper[4772]: E0127 15:08:50.662656 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.662887 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:50 crc kubenswrapper[4772]: E0127 15:08:50.662974 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.663292 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:50 crc kubenswrapper[4772]: E0127 15:08:50.663366 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.681285 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 11:11:19.139125138 +0000 UTC Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.751461 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.751493 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.751501 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.751515 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.751525 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:50Z","lastTransitionTime":"2026-01-27T15:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.853534 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.853571 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.853580 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.853596 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.853607 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:50Z","lastTransitionTime":"2026-01-27T15:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.956895 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.956944 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.956956 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.956974 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:50 crc kubenswrapper[4772]: I0127 15:08:50.956986 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:50Z","lastTransitionTime":"2026-01-27T15:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.059347 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.059407 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.059428 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.059450 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.059465 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:51Z","lastTransitionTime":"2026-01-27T15:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.162634 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.162678 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.162689 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.162707 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.162720 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:51Z","lastTransitionTime":"2026-01-27T15:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.265112 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.265144 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.265186 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.265204 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.265214 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:51Z","lastTransitionTime":"2026-01-27T15:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.369492 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.369569 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.369589 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.369613 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.369631 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:51Z","lastTransitionTime":"2026-01-27T15:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.472908 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.472946 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.472955 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.472973 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.472985 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:51Z","lastTransitionTime":"2026-01-27T15:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.499121 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7jwx_87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8/kube-multus/1.log" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.499781 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7jwx_87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8/kube-multus/0.log" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.499912 4772 generic.go:334] "Generic (PLEG): container finished" podID="87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8" containerID="9f72b451fa77f3fce2c251de546110ab49c7c9e0122759f6ef29a32fde422356" exitCode=1 Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.500039 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x7jwx" event={"ID":"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8","Type":"ContainerDied","Data":"9f72b451fa77f3fce2c251de546110ab49c7c9e0122759f6ef29a32fde422356"} Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.500148 4772 scope.go:117] "RemoveContainer" containerID="ba06c066217d03c059fbd555552d87574ea4ec17f72937330155f4bfbc4e3a33" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.500841 4772 scope.go:117] "RemoveContainer" containerID="9f72b451fa77f3fce2c251de546110ab49c7c9e0122759f6ef29a32fde422356" Jan 27 15:08:51 crc kubenswrapper[4772]: E0127 15:08:51.501118 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-x7jwx_openshift-multus(87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8)\"" pod="openshift-multus/multus-x7jwx" podUID="87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.528355 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=97.528308138 podStartE2EDuration="1m37.528308138s" podCreationTimestamp="2026-01-27 15:07:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:08:45.026014918 +0000 UTC m=+111.006624016" watchObservedRunningTime="2026-01-27 15:08:51.528308138 +0000 UTC m=+117.508917236" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.583591 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.583633 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.583642 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.583660 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.583669 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:51Z","lastTransitionTime":"2026-01-27T15:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.662884 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:51 crc kubenswrapper[4772]: E0127 15:08:51.663085 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.682552 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 14:03:21.587268973 +0000 UTC Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.687452 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.687507 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.687527 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.687548 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.687559 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:51Z","lastTransitionTime":"2026-01-27T15:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.790476 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.790544 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.790560 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.790585 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.790602 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:51Z","lastTransitionTime":"2026-01-27T15:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.893482 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.893531 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.893543 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.893559 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.893576 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:51Z","lastTransitionTime":"2026-01-27T15:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.995791 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.995838 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.995852 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.995868 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:51 crc kubenswrapper[4772]: I0127 15:08:51.995879 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:51Z","lastTransitionTime":"2026-01-27T15:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.099535 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.099614 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.099631 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.099656 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.099676 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:52Z","lastTransitionTime":"2026-01-27T15:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.202988 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.203039 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.203051 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.203070 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.203083 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:52Z","lastTransitionTime":"2026-01-27T15:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.305566 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.305654 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.305667 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.305683 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.305695 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:52Z","lastTransitionTime":"2026-01-27T15:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.408068 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.408113 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.408124 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.408139 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.408150 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:52Z","lastTransitionTime":"2026-01-27T15:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.504730 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7jwx_87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8/kube-multus/1.log" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.510275 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.510315 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.510329 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.510344 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.510355 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:52Z","lastTransitionTime":"2026-01-27T15:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.613121 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.613155 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.613163 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.613193 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.613201 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:52Z","lastTransitionTime":"2026-01-27T15:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.662897 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.662910 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.662926 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:52 crc kubenswrapper[4772]: E0127 15:08:52.663079 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:52 crc kubenswrapper[4772]: E0127 15:08:52.663305 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:52 crc kubenswrapper[4772]: E0127 15:08:52.663466 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.685136 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 17:57:49.888874258 +0000 UTC Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.716591 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.716829 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.716917 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.717040 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.717139 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:52Z","lastTransitionTime":"2026-01-27T15:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.819951 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.820003 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.820014 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.820032 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.820043 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:52Z","lastTransitionTime":"2026-01-27T15:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.922657 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.922695 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.922704 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.922718 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:52 crc kubenswrapper[4772]: I0127 15:08:52.922727 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:52Z","lastTransitionTime":"2026-01-27T15:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.025235 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.025273 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.025283 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.025297 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.025308 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:53Z","lastTransitionTime":"2026-01-27T15:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.128406 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.128446 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.128454 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.128468 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.128479 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:53Z","lastTransitionTime":"2026-01-27T15:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.231133 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.231189 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.231199 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.231213 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.231224 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:53Z","lastTransitionTime":"2026-01-27T15:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.233434 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.233477 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.233489 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.233508 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.233519 4772 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-27T15:08:53Z","lastTransitionTime":"2026-01-27T15:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.276892 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-jv7kr"] Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.277285 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jv7kr" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.280545 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.280918 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.281050 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.281192 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.324561 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/765e5a84-2a7d-43fa-a02c-e3800d9b6fd8-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-jv7kr\" (UID: \"765e5a84-2a7d-43fa-a02c-e3800d9b6fd8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jv7kr" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.324830 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/765e5a84-2a7d-43fa-a02c-e3800d9b6fd8-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-jv7kr\" (UID: \"765e5a84-2a7d-43fa-a02c-e3800d9b6fd8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jv7kr" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.324975 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/765e5a84-2a7d-43fa-a02c-e3800d9b6fd8-service-ca\") pod \"cluster-version-operator-5c965bbfc6-jv7kr\" (UID: \"765e5a84-2a7d-43fa-a02c-e3800d9b6fd8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jv7kr" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.325073 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/765e5a84-2a7d-43fa-a02c-e3800d9b6fd8-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-jv7kr\" (UID: \"765e5a84-2a7d-43fa-a02c-e3800d9b6fd8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jv7kr" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.325192 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/765e5a84-2a7d-43fa-a02c-e3800d9b6fd8-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-jv7kr\" (UID: \"765e5a84-2a7d-43fa-a02c-e3800d9b6fd8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jv7kr" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.426123 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/765e5a84-2a7d-43fa-a02c-e3800d9b6fd8-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-jv7kr\" (UID: \"765e5a84-2a7d-43fa-a02c-e3800d9b6fd8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jv7kr" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.426982 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/765e5a84-2a7d-43fa-a02c-e3800d9b6fd8-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-jv7kr\" (UID: \"765e5a84-2a7d-43fa-a02c-e3800d9b6fd8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jv7kr" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.427019 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/765e5a84-2a7d-43fa-a02c-e3800d9b6fd8-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-jv7kr\" (UID: \"765e5a84-2a7d-43fa-a02c-e3800d9b6fd8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jv7kr" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.427042 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/765e5a84-2a7d-43fa-a02c-e3800d9b6fd8-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-jv7kr\" (UID: \"765e5a84-2a7d-43fa-a02c-e3800d9b6fd8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jv7kr" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.427116 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/765e5a84-2a7d-43fa-a02c-e3800d9b6fd8-service-ca\") pod \"cluster-version-operator-5c965bbfc6-jv7kr\" (UID: \"765e5a84-2a7d-43fa-a02c-e3800d9b6fd8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jv7kr" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.427260 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/765e5a84-2a7d-43fa-a02c-e3800d9b6fd8-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-jv7kr\" (UID: \"765e5a84-2a7d-43fa-a02c-e3800d9b6fd8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jv7kr" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.427268 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/765e5a84-2a7d-43fa-a02c-e3800d9b6fd8-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-jv7kr\" (UID: \"765e5a84-2a7d-43fa-a02c-e3800d9b6fd8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jv7kr" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.428743 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/765e5a84-2a7d-43fa-a02c-e3800d9b6fd8-service-ca\") pod \"cluster-version-operator-5c965bbfc6-jv7kr\" (UID: \"765e5a84-2a7d-43fa-a02c-e3800d9b6fd8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jv7kr" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.432256 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/765e5a84-2a7d-43fa-a02c-e3800d9b6fd8-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-jv7kr\" (UID: \"765e5a84-2a7d-43fa-a02c-e3800d9b6fd8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jv7kr" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.444491 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/765e5a84-2a7d-43fa-a02c-e3800d9b6fd8-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-jv7kr\" (UID: \"765e5a84-2a7d-43fa-a02c-e3800d9b6fd8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jv7kr" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.596103 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jv7kr" Jan 27 15:08:53 crc kubenswrapper[4772]: W0127 15:08:53.617752 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod765e5a84_2a7d_43fa_a02c_e3800d9b6fd8.slice/crio-c223f5a24ed190a2600efb87a34c7e266fac233ace6d0338754e7f0d86e41ffa WatchSource:0}: Error finding container c223f5a24ed190a2600efb87a34c7e266fac233ace6d0338754e7f0d86e41ffa: Status 404 returned error can't find the container with id c223f5a24ed190a2600efb87a34c7e266fac233ace6d0338754e7f0d86e41ffa Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.661837 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:53 crc kubenswrapper[4772]: E0127 15:08:53.662057 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.662944 4772 scope.go:117] "RemoveContainer" containerID="8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a" Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.685631 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 04:55:54.576859208 +0000 UTC Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.685679 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Jan 27 15:08:53 crc kubenswrapper[4772]: I0127 15:08:53.693733 4772 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 27 15:08:54 crc kubenswrapper[4772]: I0127 15:08:54.514027 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n2khk_736264c8-cd18-479a-88ba-e1ec15dbfdae/ovnkube-controller/3.log" Jan 27 15:08:54 crc kubenswrapper[4772]: I0127 15:08:54.517180 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerStarted","Data":"7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823"} Jan 27 15:08:54 crc kubenswrapper[4772]: I0127 15:08:54.517604 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:08:54 crc kubenswrapper[4772]: I0127 15:08:54.519012 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jv7kr" event={"ID":"765e5a84-2a7d-43fa-a02c-e3800d9b6fd8","Type":"ContainerStarted","Data":"2d1bb5111d302a6e71b243f0bcee9a98b6fc9a0bebc10bd5befbef42ffcdad72"} Jan 27 15:08:54 crc kubenswrapper[4772]: I0127 15:08:54.519053 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jv7kr" event={"ID":"765e5a84-2a7d-43fa-a02c-e3800d9b6fd8","Type":"ContainerStarted","Data":"c223f5a24ed190a2600efb87a34c7e266fac233ace6d0338754e7f0d86e41ffa"} Jan 27 15:08:54 crc kubenswrapper[4772]: I0127 15:08:54.542838 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" podStartSLOduration=98.542822159 podStartE2EDuration="1m38.542822159s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:08:54.54214716 +0000 UTC m=+120.522756268" watchObservedRunningTime="2026-01-27 15:08:54.542822159 +0000 UTC m=+120.523431257" Jan 27 15:08:54 crc kubenswrapper[4772]: I0127 15:08:54.575267 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-jv7kr" podStartSLOduration=98.575247587 podStartE2EDuration="1m38.575247587s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:08:54.555581178 +0000 UTC m=+120.536190286" watchObservedRunningTime="2026-01-27 15:08:54.575247587 +0000 UTC m=+120.555856685" Jan 27 15:08:54 crc kubenswrapper[4772]: I0127 15:08:54.575569 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-ql2vx"] Jan 27 15:08:54 crc kubenswrapper[4772]: I0127 15:08:54.575667 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:54 crc kubenswrapper[4772]: E0127 15:08:54.575746 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:54 crc kubenswrapper[4772]: E0127 15:08:54.662738 4772 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Jan 27 15:08:54 crc kubenswrapper[4772]: I0127 15:08:54.662737 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:54 crc kubenswrapper[4772]: I0127 15:08:54.662782 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:54 crc kubenswrapper[4772]: E0127 15:08:54.663658 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:54 crc kubenswrapper[4772]: E0127 15:08:54.663732 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:54 crc kubenswrapper[4772]: E0127 15:08:54.750500 4772 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 27 15:08:55 crc kubenswrapper[4772]: I0127 15:08:55.662969 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:55 crc kubenswrapper[4772]: E0127 15:08:55.663118 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:55 crc kubenswrapper[4772]: I0127 15:08:55.662964 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:55 crc kubenswrapper[4772]: E0127 15:08:55.663629 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:56 crc kubenswrapper[4772]: I0127 15:08:56.662433 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:56 crc kubenswrapper[4772]: E0127 15:08:56.662581 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:56 crc kubenswrapper[4772]: I0127 15:08:56.662636 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:56 crc kubenswrapper[4772]: E0127 15:08:56.662836 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:57 crc kubenswrapper[4772]: I0127 15:08:57.662451 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:57 crc kubenswrapper[4772]: I0127 15:08:57.662628 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:57 crc kubenswrapper[4772]: E0127 15:08:57.662743 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:57 crc kubenswrapper[4772]: E0127 15:08:57.662923 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:58 crc kubenswrapper[4772]: I0127 15:08:58.662898 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:08:58 crc kubenswrapper[4772]: I0127 15:08:58.663002 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:08:58 crc kubenswrapper[4772]: E0127 15:08:58.663111 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:08:58 crc kubenswrapper[4772]: E0127 15:08:58.663295 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:08:59 crc kubenswrapper[4772]: I0127 15:08:59.662589 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:08:59 crc kubenswrapper[4772]: E0127 15:08:59.662730 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:08:59 crc kubenswrapper[4772]: I0127 15:08:59.662589 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:08:59 crc kubenswrapper[4772]: E0127 15:08:59.662828 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:08:59 crc kubenswrapper[4772]: E0127 15:08:59.751954 4772 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 27 15:09:00 crc kubenswrapper[4772]: I0127 15:09:00.662063 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:09:00 crc kubenswrapper[4772]: E0127 15:09:00.662347 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:09:00 crc kubenswrapper[4772]: I0127 15:09:00.662539 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:09:00 crc kubenswrapper[4772]: E0127 15:09:00.662773 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:09:01 crc kubenswrapper[4772]: I0127 15:09:01.662626 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:09:01 crc kubenswrapper[4772]: I0127 15:09:01.662653 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:09:01 crc kubenswrapper[4772]: E0127 15:09:01.662795 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:09:01 crc kubenswrapper[4772]: E0127 15:09:01.662922 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:09:02 crc kubenswrapper[4772]: I0127 15:09:02.662884 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:09:02 crc kubenswrapper[4772]: E0127 15:09:02.663029 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:09:02 crc kubenswrapper[4772]: I0127 15:09:02.663916 4772 scope.go:117] "RemoveContainer" containerID="9f72b451fa77f3fce2c251de546110ab49c7c9e0122759f6ef29a32fde422356" Jan 27 15:09:02 crc kubenswrapper[4772]: I0127 15:09:02.664582 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:09:02 crc kubenswrapper[4772]: E0127 15:09:02.664871 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:09:03 crc kubenswrapper[4772]: I0127 15:09:03.547579 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7jwx_87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8/kube-multus/1.log" Jan 27 15:09:03 crc kubenswrapper[4772]: I0127 15:09:03.547894 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x7jwx" event={"ID":"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8","Type":"ContainerStarted","Data":"a5fee45d3fc79618abfe1fb780f6741fbf20558f07d7edf5c931f442a9c1c7dd"} Jan 27 15:09:03 crc kubenswrapper[4772]: I0127 15:09:03.662631 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:09:03 crc kubenswrapper[4772]: I0127 15:09:03.662657 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:09:03 crc kubenswrapper[4772]: E0127 15:09:03.662817 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 27 15:09:03 crc kubenswrapper[4772]: E0127 15:09:03.662961 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-ql2vx" podUID="371016c8-5a23-427d-aa0a-0faa241d86a7" Jan 27 15:09:04 crc kubenswrapper[4772]: I0127 15:09:04.662781 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:09:04 crc kubenswrapper[4772]: E0127 15:09:04.664208 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 27 15:09:04 crc kubenswrapper[4772]: I0127 15:09:04.664367 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:09:04 crc kubenswrapper[4772]: E0127 15:09:04.664549 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 27 15:09:05 crc kubenswrapper[4772]: I0127 15:09:05.662287 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:09:05 crc kubenswrapper[4772]: I0127 15:09:05.662391 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:09:05 crc kubenswrapper[4772]: I0127 15:09:05.665366 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 27 15:09:05 crc kubenswrapper[4772]: I0127 15:09:05.665554 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 27 15:09:05 crc kubenswrapper[4772]: I0127 15:09:05.665896 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 27 15:09:05 crc kubenswrapper[4772]: I0127 15:09:05.668029 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 27 15:09:06 crc kubenswrapper[4772]: I0127 15:09:06.662886 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:09:06 crc kubenswrapper[4772]: I0127 15:09:06.663010 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:09:06 crc kubenswrapper[4772]: I0127 15:09:06.665317 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 27 15:09:06 crc kubenswrapper[4772]: I0127 15:09:06.665502 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.823085 4772 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.881790 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6pclx"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.882505 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.882813 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.883564 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.885435 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2h2z8"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.886035 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.887758 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-mfh29"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.888402 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-mfh29" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.889140 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.889671 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.889695 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.890334 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.899110 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.899990 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.901498 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.903256 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.903648 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.904013 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.905718 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.913067 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.913387 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.913631 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.913773 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.913914 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.913937 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.914080 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.914160 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.914342 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.914433 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.914507 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.914555 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.916067 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.916517 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.916838 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-bck4j"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.917315 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zq27x"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.917546 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g667l"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.917825 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bck4j" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.917917 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g667l" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.918224 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zq27x" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.918481 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.919340 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.921022 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.921190 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.921289 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.921317 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.921376 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.921379 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.921553 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.921574 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.921624 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.921713 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-7qfrl"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.922195 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-npths"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.922475 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-npths" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.921317 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.922776 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.921744 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.923726 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.924231 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-klfsg"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.924619 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfsg" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.925525 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.926109 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.926254 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.926397 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.926506 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.926642 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.926798 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.926842 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.927012 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.927107 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.927212 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.927304 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.927365 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4v228"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.927524 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.927692 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.927728 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-tgmck"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.928104 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-tgmck" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.928434 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4v228" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.928646 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-vswtw"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.928943 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-vswtw" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.929993 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6pclx"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.945869 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.945919 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-mfh29"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.945929 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g667l"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.946850 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.947444 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.947467 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.947586 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.947752 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.947920 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.948031 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.948059 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.948131 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.948206 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.948293 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.948392 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.948436 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.948299 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.948398 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.948502 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.948670 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.948741 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.948766 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.950469 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/34e7a553-e424-472e-a143-76e7e08e57aa-encryption-config\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.950508 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d1aba7eb-5916-4023-90f2-10152ad89b63-audit-policies\") pod \"apiserver-7bbb656c7d-wk7gd\" (UID: \"d1aba7eb-5916-4023-90f2-10152ad89b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.950532 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/34e7a553-e424-472e-a143-76e7e08e57aa-node-pullsecrets\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.950556 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-config\") pod \"controller-manager-879f6c89f-6pclx\" (UID: \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.950579 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/625f7e2d-0e3f-4c2c-8f49-b09fc3638536-images\") pod \"machine-api-operator-5694c8668f-mfh29\" (UID: \"625f7e2d-0e3f-4c2c-8f49-b09fc3638536\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mfh29" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.950598 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-client-ca\") pod \"controller-manager-879f6c89f-6pclx\" (UID: \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.950620 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/34e7a553-e424-472e-a143-76e7e08e57aa-audit-dir\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.950642 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d1aba7eb-5916-4023-90f2-10152ad89b63-etcd-client\") pod \"apiserver-7bbb656c7d-wk7gd\" (UID: \"d1aba7eb-5916-4023-90f2-10152ad89b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.950662 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d1aba7eb-5916-4023-90f2-10152ad89b63-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wk7gd\" (UID: \"d1aba7eb-5916-4023-90f2-10152ad89b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.950682 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkjwx\" (UniqueName: \"kubernetes.io/projected/d1aba7eb-5916-4023-90f2-10152ad89b63-kube-api-access-pkjwx\") pod \"apiserver-7bbb656c7d-wk7gd\" (UID: \"d1aba7eb-5916-4023-90f2-10152ad89b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.950704 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-serving-cert\") pod \"controller-manager-879f6c89f-6pclx\" (UID: \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.950723 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1aba7eb-5916-4023-90f2-10152ad89b63-serving-cert\") pod \"apiserver-7bbb656c7d-wk7gd\" (UID: \"d1aba7eb-5916-4023-90f2-10152ad89b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.950746 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8d519648-7eaa-49bb-9a09-bd91d09d98c0-client-ca\") pod \"route-controller-manager-6576b87f9c-r9glz\" (UID: \"8d519648-7eaa-49bb-9a09-bd91d09d98c0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.950768 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/34e7a553-e424-472e-a143-76e7e08e57aa-audit\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.950807 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d1aba7eb-5916-4023-90f2-10152ad89b63-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wk7gd\" (UID: \"d1aba7eb-5916-4023-90f2-10152ad89b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.950827 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d1aba7eb-5916-4023-90f2-10152ad89b63-encryption-config\") pod \"apiserver-7bbb656c7d-wk7gd\" (UID: \"d1aba7eb-5916-4023-90f2-10152ad89b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.950850 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl5hf\" (UniqueName: \"kubernetes.io/projected/34e7a553-e424-472e-a143-76e7e08e57aa-kube-api-access-tl5hf\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.950874 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m24gk\" (UniqueName: \"kubernetes.io/projected/8d519648-7eaa-49bb-9a09-bd91d09d98c0-kube-api-access-m24gk\") pod \"route-controller-manager-6576b87f9c-r9glz\" (UID: \"8d519648-7eaa-49bb-9a09-bd91d09d98c0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.950905 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34e7a553-e424-472e-a143-76e7e08e57aa-serving-cert\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.950924 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d1aba7eb-5916-4023-90f2-10152ad89b63-audit-dir\") pod \"apiserver-7bbb656c7d-wk7gd\" (UID: \"d1aba7eb-5916-4023-90f2-10152ad89b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.950947 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/34e7a553-e424-472e-a143-76e7e08e57aa-etcd-serving-ca\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.950972 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/34e7a553-e424-472e-a143-76e7e08e57aa-etcd-client\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.950998 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/625f7e2d-0e3f-4c2c-8f49-b09fc3638536-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-mfh29\" (UID: \"625f7e2d-0e3f-4c2c-8f49-b09fc3638536\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mfh29" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.951019 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29qn6\" (UniqueName: \"kubernetes.io/projected/625f7e2d-0e3f-4c2c-8f49-b09fc3638536-kube-api-access-29qn6\") pod \"machine-api-operator-5694c8668f-mfh29\" (UID: \"625f7e2d-0e3f-4c2c-8f49-b09fc3638536\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mfh29" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.951040 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-6pclx\" (UID: \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.951065 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c7gp\" (UniqueName: \"kubernetes.io/projected/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-kube-api-access-8c7gp\") pod \"controller-manager-879f6c89f-6pclx\" (UID: \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.951086 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34e7a553-e424-472e-a143-76e7e08e57aa-config\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.951119 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/625f7e2d-0e3f-4c2c-8f49-b09fc3638536-config\") pod \"machine-api-operator-5694c8668f-mfh29\" (UID: \"625f7e2d-0e3f-4c2c-8f49-b09fc3638536\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mfh29" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.951139 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d519648-7eaa-49bb-9a09-bd91d09d98c0-serving-cert\") pod \"route-controller-manager-6576b87f9c-r9glz\" (UID: \"8d519648-7eaa-49bb-9a09-bd91d09d98c0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.951159 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/34e7a553-e424-472e-a143-76e7e08e57aa-image-import-ca\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.951290 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34e7a553-e424-472e-a143-76e7e08e57aa-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.951312 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d519648-7eaa-49bb-9a09-bd91d09d98c0-config\") pod \"route-controller-manager-6576b87f9c-r9glz\" (UID: \"8d519648-7eaa-49bb-9a09-bd91d09d98c0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.952816 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.956417 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-7qfrl"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.957726 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-npths"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.958114 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2h2z8"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.959039 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-crlcr"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.974863 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.989312 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fgw98"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.989899 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.989881 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.989907 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.991915 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.992097 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.992311 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.992347 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.992409 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.992460 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.992642 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-bck4j"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.995773 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.995760 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-jplbk"] Jan 27 15:09:13 crc kubenswrapper[4772]: I0127 15:09:13.998935 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-w4j4h"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.000463 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.000699 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.000869 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.003551 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4j4h" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.004417 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-jplbk" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.004823 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.004911 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.005076 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.005051 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.005085 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.005345 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.005452 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.005514 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.005632 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.007343 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.007364 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-bhgv8"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.008293 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-vswtw"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.008394 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-bhgv8" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.011942 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-7k7sg"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.012652 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4v228"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.012734 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-7k7sg" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.014139 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzkn2"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.014896 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzkn2" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.021537 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.021759 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.025006 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cv2z7"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.025709 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-pwmhd"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.026142 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-pwmhd" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.026452 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cv2z7" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.027085 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.027290 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.028006 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.028712 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.035828 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c6n8m"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.036254 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.036364 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c6n8m" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.037707 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.051983 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34e7a553-e424-472e-a143-76e7e08e57aa-serving-cert\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052024 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/34e7a553-e424-472e-a143-76e7e08e57aa-etcd-serving-ca\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052054 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d1aba7eb-5916-4023-90f2-10152ad89b63-audit-dir\") pod \"apiserver-7bbb656c7d-wk7gd\" (UID: \"d1aba7eb-5916-4023-90f2-10152ad89b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052075 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/34e7a553-e424-472e-a143-76e7e08e57aa-etcd-client\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052100 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/625f7e2d-0e3f-4c2c-8f49-b09fc3638536-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-mfh29\" (UID: \"625f7e2d-0e3f-4c2c-8f49-b09fc3638536\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mfh29" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052126 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29qn6\" (UniqueName: \"kubernetes.io/projected/625f7e2d-0e3f-4c2c-8f49-b09fc3638536-kube-api-access-29qn6\") pod \"machine-api-operator-5694c8668f-mfh29\" (UID: \"625f7e2d-0e3f-4c2c-8f49-b09fc3638536\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mfh29" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052148 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-6pclx\" (UID: \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052190 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c7gp\" (UniqueName: \"kubernetes.io/projected/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-kube-api-access-8c7gp\") pod \"controller-manager-879f6c89f-6pclx\" (UID: \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052215 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34e7a553-e424-472e-a143-76e7e08e57aa-config\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052250 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/625f7e2d-0e3f-4c2c-8f49-b09fc3638536-config\") pod \"machine-api-operator-5694c8668f-mfh29\" (UID: \"625f7e2d-0e3f-4c2c-8f49-b09fc3638536\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mfh29" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052273 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d519648-7eaa-49bb-9a09-bd91d09d98c0-serving-cert\") pod \"route-controller-manager-6576b87f9c-r9glz\" (UID: \"8d519648-7eaa-49bb-9a09-bd91d09d98c0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052293 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/34e7a553-e424-472e-a143-76e7e08e57aa-image-import-ca\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052316 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34e7a553-e424-472e-a143-76e7e08e57aa-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052339 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d519648-7eaa-49bb-9a09-bd91d09d98c0-config\") pod \"route-controller-manager-6576b87f9c-r9glz\" (UID: \"8d519648-7eaa-49bb-9a09-bd91d09d98c0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052366 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39dd090e-b988-4c36-88f0-c0cb28a23e8b-config\") pod \"authentication-operator-69f744f599-tgmck\" (UID: \"39dd090e-b988-4c36-88f0-c0cb28a23e8b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tgmck" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052399 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/34e7a553-e424-472e-a143-76e7e08e57aa-encryption-config\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052424 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/39dd090e-b988-4c36-88f0-c0cb28a23e8b-service-ca-bundle\") pod \"authentication-operator-69f744f599-tgmck\" (UID: \"39dd090e-b988-4c36-88f0-c0cb28a23e8b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tgmck" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052448 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d1aba7eb-5916-4023-90f2-10152ad89b63-audit-policies\") pod \"apiserver-7bbb656c7d-wk7gd\" (UID: \"d1aba7eb-5916-4023-90f2-10152ad89b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052471 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/34e7a553-e424-472e-a143-76e7e08e57aa-node-pullsecrets\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052493 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qztp\" (UniqueName: \"kubernetes.io/projected/39dd090e-b988-4c36-88f0-c0cb28a23e8b-kube-api-access-5qztp\") pod \"authentication-operator-69f744f599-tgmck\" (UID: \"39dd090e-b988-4c36-88f0-c0cb28a23e8b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tgmck" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052518 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-config\") pod \"controller-manager-879f6c89f-6pclx\" (UID: \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052540 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/625f7e2d-0e3f-4c2c-8f49-b09fc3638536-images\") pod \"machine-api-operator-5694c8668f-mfh29\" (UID: \"625f7e2d-0e3f-4c2c-8f49-b09fc3638536\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mfh29" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052560 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-client-ca\") pod \"controller-manager-879f6c89f-6pclx\" (UID: \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052581 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/34e7a553-e424-472e-a143-76e7e08e57aa-audit-dir\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052604 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d1aba7eb-5916-4023-90f2-10152ad89b63-etcd-client\") pod \"apiserver-7bbb656c7d-wk7gd\" (UID: \"d1aba7eb-5916-4023-90f2-10152ad89b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052625 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d1aba7eb-5916-4023-90f2-10152ad89b63-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wk7gd\" (UID: \"d1aba7eb-5916-4023-90f2-10152ad89b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052648 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/39dd090e-b988-4c36-88f0-c0cb28a23e8b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-tgmck\" (UID: \"39dd090e-b988-4c36-88f0-c0cb28a23e8b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tgmck" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052670 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkjwx\" (UniqueName: \"kubernetes.io/projected/d1aba7eb-5916-4023-90f2-10152ad89b63-kube-api-access-pkjwx\") pod \"apiserver-7bbb656c7d-wk7gd\" (UID: \"d1aba7eb-5916-4023-90f2-10152ad89b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052696 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-serving-cert\") pod \"controller-manager-879f6c89f-6pclx\" (UID: \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052717 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1aba7eb-5916-4023-90f2-10152ad89b63-serving-cert\") pod \"apiserver-7bbb656c7d-wk7gd\" (UID: \"d1aba7eb-5916-4023-90f2-10152ad89b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052740 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39dd090e-b988-4c36-88f0-c0cb28a23e8b-serving-cert\") pod \"authentication-operator-69f744f599-tgmck\" (UID: \"39dd090e-b988-4c36-88f0-c0cb28a23e8b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tgmck" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052764 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8d519648-7eaa-49bb-9a09-bd91d09d98c0-client-ca\") pod \"route-controller-manager-6576b87f9c-r9glz\" (UID: \"8d519648-7eaa-49bb-9a09-bd91d09d98c0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052786 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/34e7a553-e424-472e-a143-76e7e08e57aa-audit\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052824 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d1aba7eb-5916-4023-90f2-10152ad89b63-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wk7gd\" (UID: \"d1aba7eb-5916-4023-90f2-10152ad89b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052845 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d1aba7eb-5916-4023-90f2-10152ad89b63-encryption-config\") pod \"apiserver-7bbb656c7d-wk7gd\" (UID: \"d1aba7eb-5916-4023-90f2-10152ad89b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052867 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl5hf\" (UniqueName: \"kubernetes.io/projected/34e7a553-e424-472e-a143-76e7e08e57aa-kube-api-access-tl5hf\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.052889 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m24gk\" (UniqueName: \"kubernetes.io/projected/8d519648-7eaa-49bb-9a09-bd91d09d98c0-kube-api-access-m24gk\") pod \"route-controller-manager-6576b87f9c-r9glz\" (UID: \"8d519648-7eaa-49bb-9a09-bd91d09d98c0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.053489 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d1aba7eb-5916-4023-90f2-10152ad89b63-audit-policies\") pod \"apiserver-7bbb656c7d-wk7gd\" (UID: \"d1aba7eb-5916-4023-90f2-10152ad89b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.054208 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/34e7a553-e424-472e-a143-76e7e08e57aa-node-pullsecrets\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.063507 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1aba7eb-5916-4023-90f2-10152ad89b63-serving-cert\") pod \"apiserver-7bbb656c7d-wk7gd\" (UID: \"d1aba7eb-5916-4023-90f2-10152ad89b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.064496 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8d519648-7eaa-49bb-9a09-bd91d09d98c0-client-ca\") pod \"route-controller-manager-6576b87f9c-r9glz\" (UID: \"8d519648-7eaa-49bb-9a09-bd91d09d98c0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.065053 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/34e7a553-e424-472e-a143-76e7e08e57aa-audit\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.065470 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34e7a553-e424-472e-a143-76e7e08e57aa-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.065700 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d1aba7eb-5916-4023-90f2-10152ad89b63-audit-dir\") pod \"apiserver-7bbb656c7d-wk7gd\" (UID: \"d1aba7eb-5916-4023-90f2-10152ad89b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.066674 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d1aba7eb-5916-4023-90f2-10152ad89b63-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wk7gd\" (UID: \"d1aba7eb-5916-4023-90f2-10152ad89b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.066800 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d1aba7eb-5916-4023-90f2-10152ad89b63-etcd-client\") pod \"apiserver-7bbb656c7d-wk7gd\" (UID: \"d1aba7eb-5916-4023-90f2-10152ad89b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.067069 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d1aba7eb-5916-4023-90f2-10152ad89b63-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wk7gd\" (UID: \"d1aba7eb-5916-4023-90f2-10152ad89b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.067136 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/34e7a553-e424-472e-a143-76e7e08e57aa-etcd-serving-ca\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.067637 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d519648-7eaa-49bb-9a09-bd91d09d98c0-config\") pod \"route-controller-manager-6576b87f9c-r9glz\" (UID: \"8d519648-7eaa-49bb-9a09-bd91d09d98c0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.067710 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-tgmck"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.067750 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fbwkz"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.068071 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d519648-7eaa-49bb-9a09-bd91d09d98c0-serving-cert\") pod \"route-controller-manager-6576b87f9c-r9glz\" (UID: \"8d519648-7eaa-49bb-9a09-bd91d09d98c0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.081329 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/34e7a553-e424-472e-a143-76e7e08e57aa-image-import-ca\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.086341 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d1aba7eb-5916-4023-90f2-10152ad89b63-encryption-config\") pod \"apiserver-7bbb656c7d-wk7gd\" (UID: \"d1aba7eb-5916-4023-90f2-10152ad89b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.086765 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34e7a553-e424-472e-a143-76e7e08e57aa-config\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.087545 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/625f7e2d-0e3f-4c2c-8f49-b09fc3638536-images\") pod \"machine-api-operator-5694c8668f-mfh29\" (UID: \"625f7e2d-0e3f-4c2c-8f49-b09fc3638536\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mfh29" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.090956 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/625f7e2d-0e3f-4c2c-8f49-b09fc3638536-config\") pod \"machine-api-operator-5694c8668f-mfh29\" (UID: \"625f7e2d-0e3f-4c2c-8f49-b09fc3638536\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mfh29" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.091124 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/34e7a553-e424-472e-a143-76e7e08e57aa-etcd-client\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.095031 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-serving-cert\") pod \"controller-manager-879f6c89f-6pclx\" (UID: \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.095276 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/34e7a553-e424-472e-a143-76e7e08e57aa-audit-dir\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.095663 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-jplbk"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.095697 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zq27x"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.095857 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fbwkz" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.096594 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-client-ca\") pod \"controller-manager-879f6c89f-6pclx\" (UID: \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.098220 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-config\") pod \"controller-manager-879f6c89f-6pclx\" (UID: \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.098713 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34e7a553-e424-472e-a143-76e7e08e57aa-serving-cert\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.100242 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/34e7a553-e424-472e-a143-76e7e08e57aa-encryption-config\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.100278 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f9fqs"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.100620 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/625f7e2d-0e3f-4c2c-8f49-b09fc3638536-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-mfh29\" (UID: \"625f7e2d-0e3f-4c2c-8f49-b09fc3638536\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mfh29" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.102014 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-6pclx\" (UID: \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.112336 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.112537 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.112917 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.113075 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.117334 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-w4j4h"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.117361 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-bhgv8"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.117371 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.117449 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f9fqs" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.121917 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.123190 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.123213 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fgw98"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.124368 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.124529 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-djmb4"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.124970 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-djmb4" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.125024 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4lj2h"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.125623 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.126191 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492100-r2zj6"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.126769 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492100-r2zj6" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.127299 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-fw6bh"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.127667 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-fw6bh" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.128772 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.131441 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-xqtff"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.132565 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xqtff" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.134515 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jdcpn"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.135063 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jdcpn" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.139420 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zkmjj"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.141392 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zkmjj" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.143086 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cv2z7"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.144643 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.148884 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5vmlj"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.149494 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5vmlj" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.150884 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9wv77"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.151501 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-9wv77" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.152267 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-24dmv"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.153263 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-24dmv" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.153443 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/39dd090e-b988-4c36-88f0-c0cb28a23e8b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-tgmck\" (UID: \"39dd090e-b988-4c36-88f0-c0cb28a23e8b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tgmck" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.153493 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39dd090e-b988-4c36-88f0-c0cb28a23e8b-serving-cert\") pod \"authentication-operator-69f744f599-tgmck\" (UID: \"39dd090e-b988-4c36-88f0-c0cb28a23e8b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tgmck" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.153577 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39dd090e-b988-4c36-88f0-c0cb28a23e8b-config\") pod \"authentication-operator-69f744f599-tgmck\" (UID: \"39dd090e-b988-4c36-88f0-c0cb28a23e8b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tgmck" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.153594 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/39dd090e-b988-4c36-88f0-c0cb28a23e8b-service-ca-bundle\") pod \"authentication-operator-69f744f599-tgmck\" (UID: \"39dd090e-b988-4c36-88f0-c0cb28a23e8b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tgmck" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.153615 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qztp\" (UniqueName: \"kubernetes.io/projected/39dd090e-b988-4c36-88f0-c0cb28a23e8b-kube-api-access-5qztp\") pod \"authentication-operator-69f744f599-tgmck\" (UID: \"39dd090e-b988-4c36-88f0-c0cb28a23e8b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tgmck" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.154512 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/39dd090e-b988-4c36-88f0-c0cb28a23e8b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-tgmck\" (UID: \"39dd090e-b988-4c36-88f0-c0cb28a23e8b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tgmck" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.154576 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/39dd090e-b988-4c36-88f0-c0cb28a23e8b-service-ca-bundle\") pod \"authentication-operator-69f744f599-tgmck\" (UID: \"39dd090e-b988-4c36-88f0-c0cb28a23e8b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tgmck" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.154661 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39dd090e-b988-4c36-88f0-c0cb28a23e8b-config\") pod \"authentication-operator-69f744f599-tgmck\" (UID: \"39dd090e-b988-4c36-88f0-c0cb28a23e8b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tgmck" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.156949 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzkn2"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.157446 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39dd090e-b988-4c36-88f0-c0cb28a23e8b-serving-cert\") pod \"authentication-operator-69f744f599-tgmck\" (UID: \"39dd090e-b988-4c36-88f0-c0cb28a23e8b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tgmck" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.158327 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-njx6w"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.159071 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-njx6w" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.160467 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cf6v7"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.161146 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cf6v7" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.161867 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mnltb"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.162548 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mnltb" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.167344 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.167787 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-v7l4k"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.169196 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v7l4k" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.171583 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-crlcr"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.173709 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-pwmhd"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.174816 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-djmb4"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.175474 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5vmlj"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.178967 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-24dmv"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.180955 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jdcpn"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.184154 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.187285 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492100-r2zj6"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.189498 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9wv77"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.192854 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4lj2h"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.194399 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-v7l4k"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.195942 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c6n8m"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.197067 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f9fqs"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.198314 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-58dhr"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.205094 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fbwkz"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.205236 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-58dhr" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.208125 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.208430 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-xqtff"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.211811 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-fw6bh"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.213113 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zkmjj"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.214304 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cf6v7"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.215637 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-njx6w"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.216798 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mnltb"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.217782 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-hbbxh"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.218837 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.218974 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-hbbxh"] Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.224816 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.244584 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.264885 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.286489 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.307594 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.325187 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.346441 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.365673 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.386535 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.405211 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.425878 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.445200 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.485806 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.505810 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.527783 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.546161 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.567092 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.585673 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.607813 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.626750 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.645673 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.665841 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.685976 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.704784 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.745310 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.765041 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.784857 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.806020 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.825332 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.868469 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m24gk\" (UniqueName: \"kubernetes.io/projected/8d519648-7eaa-49bb-9a09-bd91d09d98c0-kube-api-access-m24gk\") pod \"route-controller-manager-6576b87f9c-r9glz\" (UID: \"8d519648-7eaa-49bb-9a09-bd91d09d98c0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.928839 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.937659 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkjwx\" (UniqueName: \"kubernetes.io/projected/d1aba7eb-5916-4023-90f2-10152ad89b63-kube-api-access-pkjwx\") pod \"apiserver-7bbb656c7d-wk7gd\" (UID: \"d1aba7eb-5916-4023-90f2-10152ad89b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.946076 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.946978 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c7gp\" (UniqueName: \"kubernetes.io/projected/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-kube-api-access-8c7gp\") pod \"controller-manager-879f6c89f-6pclx\" (UID: \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.948184 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl5hf\" (UniqueName: \"kubernetes.io/projected/34e7a553-e424-472e-a143-76e7e08e57aa-kube-api-access-tl5hf\") pod \"apiserver-76f77b778f-2h2z8\" (UID: \"34e7a553-e424-472e-a143-76e7e08e57aa\") " pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.948924 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29qn6\" (UniqueName: \"kubernetes.io/projected/625f7e2d-0e3f-4c2c-8f49-b09fc3638536-kube-api-access-29qn6\") pod \"machine-api-operator-5694c8668f-mfh29\" (UID: \"625f7e2d-0e3f-4c2c-8f49-b09fc3638536\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mfh29" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.965619 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 27 15:09:14 crc kubenswrapper[4772]: I0127 15:09:14.985911 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.005781 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.025674 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.045614 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.064900 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.085396 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.102202 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz"] Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.105053 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.126020 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.130489 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.143661 4772 request.go:700] Waited for 1.018420389s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca-operator/secrets?fieldSelector=metadata.name%3Dserving-cert&limit=500&resourceVersion=0 Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.145888 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.151087 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.155038 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.165814 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.196215 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.200577 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-mfh29" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.205643 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.228578 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.249925 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.275045 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.287550 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.293704 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6pclx"] Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.304870 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.326131 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.333580 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2h2z8"] Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.345378 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.364643 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.380018 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd"] Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.386608 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.406266 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.415838 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-mfh29"] Jan 27 15:09:15 crc kubenswrapper[4772]: W0127 15:09:15.422829 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod625f7e2d_0e3f_4c2c_8f49_b09fc3638536.slice/crio-810797dbb78d0461ed7aa6192665d20e18e1d6246178b7721049b2d5b19a90b6 WatchSource:0}: Error finding container 810797dbb78d0461ed7aa6192665d20e18e1d6246178b7721049b2d5b19a90b6: Status 404 returned error can't find the container with id 810797dbb78d0461ed7aa6192665d20e18e1d6246178b7721049b2d5b19a90b6 Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.424444 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.445666 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.465482 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.485296 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.505561 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.526235 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.544779 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.565085 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.585476 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.588274 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-mfh29" event={"ID":"625f7e2d-0e3f-4c2c-8f49-b09fc3638536","Type":"ContainerStarted","Data":"b5d5b46006fddbc4deae6b1d3905bab70e2b1c8b23766630fb99d618a421958a"} Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.588352 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-mfh29" event={"ID":"625f7e2d-0e3f-4c2c-8f49-b09fc3638536","Type":"ContainerStarted","Data":"810797dbb78d0461ed7aa6192665d20e18e1d6246178b7721049b2d5b19a90b6"} Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.589295 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" event={"ID":"3dfd9a91-e760-4c80-96e6-ca6525aa86b8","Type":"ContainerStarted","Data":"bb01029e32299fb52d56d061afc654aa573880622860ea04caf54ad26b9a84eb"} Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.589336 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" event={"ID":"3dfd9a91-e760-4c80-96e6-ca6525aa86b8","Type":"ContainerStarted","Data":"380eb12e1295d0270de3d27b76c2692262e194cd8678145268c2765050e2b23e"} Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.589631 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.591973 4772 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-6pclx container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.592020 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" podUID="3dfd9a91-e760-4c80-96e6-ca6525aa86b8" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.593801 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" event={"ID":"34e7a553-e424-472e-a143-76e7e08e57aa","Type":"ContainerStarted","Data":"10dc65c23141e7590f682165f4eacee477055ad9f3dde4720a2c88ce24d8486f"} Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.596424 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" event={"ID":"d1aba7eb-5916-4023-90f2-10152ad89b63","Type":"ContainerStarted","Data":"07119c87526ca057f6f02dc0420297e1a13ea1238c1309af265353e4d0396523"} Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.598560 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" event={"ID":"8d519648-7eaa-49bb-9a09-bd91d09d98c0","Type":"ContainerStarted","Data":"3506ad5b58f850018e8ca4a14f82aae1d0b2f9ec52328d0669f3f49efb696d0f"} Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.598611 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" event={"ID":"8d519648-7eaa-49bb-9a09-bd91d09d98c0","Type":"ContainerStarted","Data":"326c33fb529962a602f8dfd5dbe7dcbd0ebb132fe4709244f27812007b261a68"} Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.598956 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.601052 4772 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-r9glz container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.601096 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" podUID="8d519648-7eaa-49bb-9a09-bd91d09d98c0" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.605331 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.625243 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.647498 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.665653 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.685324 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.704946 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.724997 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.747132 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.764980 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.786436 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.824752 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qztp\" (UniqueName: \"kubernetes.io/projected/39dd090e-b988-4c36-88f0-c0cb28a23e8b-kube-api-access-5qztp\") pod \"authentication-operator-69f744f599-tgmck\" (UID: \"39dd090e-b988-4c36-88f0-c0cb28a23e8b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tgmck" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.825747 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.845841 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.866226 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.885651 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.905257 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.925419 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.946606 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.956621 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-tgmck" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.965515 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 27 15:09:15 crc kubenswrapper[4772]: I0127 15:09:15.986953 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.006091 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.027434 4772 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.050753 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.065242 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.168261 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-tgmck"] Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.172995 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/877de785-bc18-4c1c-970a-1e6533539467-trusted-ca\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.173072 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/06cdc094-b372-4016-bc5e-4c15a28e032e-audit-policies\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.173124 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.173158 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f9ab8468-b920-41e9-a5f2-1af70f1b5ffd-bound-sa-token\") pod \"ingress-operator-5b745b69d9-w4j4h\" (UID: \"f9ab8468-b920-41e9-a5f2-1af70f1b5ffd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4j4h" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.173228 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd05c164-83f7-4ebe-bbe8-9db6707741c5-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-mzkn2\" (UID: \"dd05c164-83f7-4ebe-bbe8-9db6707741c5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzkn2" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.173254 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.173300 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5f6v\" (UniqueName: \"kubernetes.io/projected/06cdc094-b372-4016-bc5e-4c15a28e032e-kube-api-access-l5f6v\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.173335 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.173384 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-oauth-serving-cert\") pod \"console-f9d7485db-7qfrl\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.173405 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/192d0f8f-10f9-43e2-a24a-2019aae0db44-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zq27x\" (UID: \"192d0f8f-10f9-43e2-a24a-2019aae0db44\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zq27x" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.173447 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4225ddc-bdcd-4158-811b-113234d0c3d0-service-ca-bundle\") pod \"router-default-5444994796-7k7sg\" (UID: \"c4225ddc-bdcd-4158-811b-113234d0c3d0\") " pod="openshift-ingress/router-default-5444994796-7k7sg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.173470 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c4225ddc-bdcd-4158-811b-113234d0c3d0-metrics-certs\") pod \"router-default-5444994796-7k7sg\" (UID: \"c4225ddc-bdcd-4158-811b-113234d0c3d0\") " pod="openshift-ingress/router-default-5444994796-7k7sg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.173491 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/192d0f8f-10f9-43e2-a24a-2019aae0db44-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zq27x\" (UID: \"192d0f8f-10f9-43e2-a24a-2019aae0db44\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zq27x" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.173534 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.173652 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.173851 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5z5pj\" (UniqueName: \"kubernetes.io/projected/c4225ddc-bdcd-4158-811b-113234d0c3d0-kube-api-access-5z5pj\") pod \"router-default-5444994796-7k7sg\" (UID: \"c4225ddc-bdcd-4158-811b-113234d0c3d0\") " pod="openshift-ingress/router-default-5444994796-7k7sg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.173880 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/ee07e18c-9f40-41c3-b2fb-05fd325976e4-available-featuregates\") pod \"openshift-config-operator-7777fb866f-bck4j\" (UID: \"ee07e18c-9f40-41c3-b2fb-05fd325976e4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bck4j" Jan 27 15:09:16 crc kubenswrapper[4772]: E0127 15:09:16.174021 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:16.673950135 +0000 UTC m=+142.654559343 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.177535 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mswpx\" (UniqueName: \"kubernetes.io/projected/2c9eef7e-3996-45b6-ab7b-50d319dc1117-kube-api-access-mswpx\") pod \"cluster-image-registry-operator-dc59b4c8b-4v228\" (UID: \"2c9eef7e-3996-45b6-ab7b-50d319dc1117\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4v228" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.177606 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3150273d-63f7-4908-bcc5-2403e123d1e7-config-volume\") pod \"dns-default-jplbk\" (UID: \"3150273d-63f7-4908-bcc5-2403e123d1e7\") " pod="openshift-dns/dns-default-jplbk" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.177927 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg5jv\" (UniqueName: \"kubernetes.io/projected/3150273d-63f7-4908-bcc5-2403e123d1e7-kube-api-access-dg5jv\") pod \"dns-default-jplbk\" (UID: \"3150273d-63f7-4908-bcc5-2403e123d1e7\") " pod="openshift-dns/dns-default-jplbk" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.178046 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j667k\" (UniqueName: \"kubernetes.io/projected/5276546c-f731-4bd0-bb93-b5cd19b0992c-kube-api-access-j667k\") pod \"console-operator-58897d9998-npths\" (UID: \"5276546c-f731-4bd0-bb93-b5cd19b0992c\") " pod="openshift-console-operator/console-operator-58897d9998-npths" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.178088 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.178705 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmgkc\" (UniqueName: \"kubernetes.io/projected/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-kube-api-access-vmgkc\") pod \"console-f9d7485db-7qfrl\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.178767 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2c9eef7e-3996-45b6-ab7b-50d319dc1117-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-4v228\" (UID: \"2c9eef7e-3996-45b6-ab7b-50d319dc1117\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4v228" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.178891 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/877de785-bc18-4c1c-970a-1e6533539467-registry-certificates\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.178949 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c4225ddc-bdcd-4158-811b-113234d0c3d0-stats-auth\") pod \"router-default-5444994796-7k7sg\" (UID: \"c4225ddc-bdcd-4158-811b-113234d0c3d0\") " pod="openshift-ingress/router-default-5444994796-7k7sg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.179006 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2zfh\" (UniqueName: \"kubernetes.io/projected/ce4b4e2e-496b-4334-8736-db4f25473731-kube-api-access-k2zfh\") pod \"dns-operator-744455d44c-bhgv8\" (UID: \"ce4b4e2e-496b-4334-8736-db4f25473731\") " pod="openshift-dns-operator/dns-operator-744455d44c-bhgv8" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.180065 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5276546c-f731-4bd0-bb93-b5cd19b0992c-trusted-ca\") pod \"console-operator-58897d9998-npths\" (UID: \"5276546c-f731-4bd0-bb93-b5cd19b0992c\") " pod="openshift-console-operator/console-operator-58897d9998-npths" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.180469 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.180861 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-console-oauth-config\") pod \"console-f9d7485db-7qfrl\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.180890 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8800c113-7b51-4554-8e52-c1d0df1a08be-machine-approver-tls\") pod \"machine-approver-56656f9798-klfsg\" (UID: \"8800c113-7b51-4554-8e52-c1d0df1a08be\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfsg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.180917 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-console-serving-cert\") pod \"console-f9d7485db-7qfrl\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.180945 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-service-ca\") pod \"console-f9d7485db-7qfrl\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.180976 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/06cdc094-b372-4016-bc5e-4c15a28e032e-audit-dir\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181059 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfglr\" (UniqueName: \"kubernetes.io/projected/f9ab8468-b920-41e9-a5f2-1af70f1b5ffd-kube-api-access-cfglr\") pod \"ingress-operator-5b745b69d9-w4j4h\" (UID: \"f9ab8468-b920-41e9-a5f2-1af70f1b5ffd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4j4h" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181090 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/877de785-bc18-4c1c-970a-1e6533539467-registry-tls\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181114 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181145 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee07e18c-9f40-41c3-b2fb-05fd325976e4-serving-cert\") pod \"openshift-config-operator-7777fb866f-bck4j\" (UID: \"ee07e18c-9f40-41c3-b2fb-05fd325976e4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bck4j" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181213 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ce4b4e2e-496b-4334-8736-db4f25473731-metrics-tls\") pod \"dns-operator-744455d44c-bhgv8\" (UID: \"ce4b4e2e-496b-4334-8736-db4f25473731\") " pod="openshift-dns-operator/dns-operator-744455d44c-bhgv8" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181259 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5276546c-f731-4bd0-bb93-b5cd19b0992c-serving-cert\") pod \"console-operator-58897d9998-npths\" (UID: \"5276546c-f731-4bd0-bb93-b5cd19b0992c\") " pod="openshift-console-operator/console-operator-58897d9998-npths" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181288 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwmzq\" (UniqueName: \"kubernetes.io/projected/17bdd07d-f7e5-47f8-b730-724d5cc8e3d2-kube-api-access-cwmzq\") pod \"downloads-7954f5f757-vswtw\" (UID: \"17bdd07d-f7e5-47f8-b730-724d5cc8e3d2\") " pod="openshift-console/downloads-7954f5f757-vswtw" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181317 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f9ab8468-b920-41e9-a5f2-1af70f1b5ffd-trusted-ca\") pod \"ingress-operator-5b745b69d9-w4j4h\" (UID: \"f9ab8468-b920-41e9-a5f2-1af70f1b5ffd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4j4h" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181378 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/877de785-bc18-4c1c-970a-1e6533539467-installation-pull-secrets\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181406 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3150273d-63f7-4908-bcc5-2403e123d1e7-metrics-tls\") pod \"dns-default-jplbk\" (UID: \"3150273d-63f7-4908-bcc5-2403e123d1e7\") " pod="openshift-dns/dns-default-jplbk" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181432 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2c9eef7e-3996-45b6-ab7b-50d319dc1117-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-4v228\" (UID: \"2c9eef7e-3996-45b6-ab7b-50d319dc1117\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4v228" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181484 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s88q\" (UniqueName: \"kubernetes.io/projected/192d0f8f-10f9-43e2-a24a-2019aae0db44-kube-api-access-7s88q\") pod \"openshift-apiserver-operator-796bbdcf4f-zq27x\" (UID: \"192d0f8f-10f9-43e2-a24a-2019aae0db44\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zq27x" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181515 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smmpv\" (UniqueName: \"kubernetes.io/projected/ee07e18c-9f40-41c3-b2fb-05fd325976e4-kube-api-access-smmpv\") pod \"openshift-config-operator-7777fb866f-bck4j\" (UID: \"ee07e18c-9f40-41c3-b2fb-05fd325976e4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bck4j" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181560 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181590 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dd05c164-83f7-4ebe-bbe8-9db6707741c5-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-mzkn2\" (UID: \"dd05c164-83f7-4ebe-bbe8-9db6707741c5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzkn2" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181618 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181654 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h66t\" (UniqueName: \"kubernetes.io/projected/877de785-bc18-4c1c-970a-1e6533539467-kube-api-access-5h66t\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181679 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8800c113-7b51-4554-8e52-c1d0df1a08be-auth-proxy-config\") pod \"machine-approver-56656f9798-klfsg\" (UID: \"8800c113-7b51-4554-8e52-c1d0df1a08be\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfsg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181709 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/06755863-1a8b-4f4d-a304-03bfd45725ec-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-g667l\" (UID: \"06755863-1a8b-4f4d-a304-03bfd45725ec\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g667l" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181736 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2c9eef7e-3996-45b6-ab7b-50d319dc1117-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-4v228\" (UID: \"2c9eef7e-3996-45b6-ab7b-50d319dc1117\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4v228" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181779 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5276546c-f731-4bd0-bb93-b5cd19b0992c-config\") pod \"console-operator-58897d9998-npths\" (UID: \"5276546c-f731-4bd0-bb93-b5cd19b0992c\") " pod="openshift-console-operator/console-operator-58897d9998-npths" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181807 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-trusted-ca-bundle\") pod \"console-f9d7485db-7qfrl\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181849 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/877de785-bc18-4c1c-970a-1e6533539467-ca-trust-extracted\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181877 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lbs2\" (UniqueName: \"kubernetes.io/projected/8800c113-7b51-4554-8e52-c1d0df1a08be-kube-api-access-4lbs2\") pod \"machine-approver-56656f9798-klfsg\" (UID: \"8800c113-7b51-4554-8e52-c1d0df1a08be\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfsg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181905 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twqnj\" (UniqueName: \"kubernetes.io/projected/06755863-1a8b-4f4d-a304-03bfd45725ec-kube-api-access-twqnj\") pod \"cluster-samples-operator-665b6dd947-g667l\" (UID: \"06755863-1a8b-4f4d-a304-03bfd45725ec\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g667l" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181931 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd05c164-83f7-4ebe-bbe8-9db6707741c5-config\") pod \"kube-apiserver-operator-766d6c64bb-mzkn2\" (UID: \"dd05c164-83f7-4ebe-bbe8-9db6707741c5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzkn2" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.181956 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.182023 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c4225ddc-bdcd-4158-811b-113234d0c3d0-default-certificate\") pod \"router-default-5444994796-7k7sg\" (UID: \"c4225ddc-bdcd-4158-811b-113234d0c3d0\") " pod="openshift-ingress/router-default-5444994796-7k7sg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.182046 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.182096 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8800c113-7b51-4554-8e52-c1d0df1a08be-config\") pod \"machine-approver-56656f9798-klfsg\" (UID: \"8800c113-7b51-4554-8e52-c1d0df1a08be\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfsg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.182122 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-console-config\") pod \"console-f9d7485db-7qfrl\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.182325 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/877de785-bc18-4c1c-970a-1e6533539467-bound-sa-token\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.182365 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f9ab8468-b920-41e9-a5f2-1af70f1b5ffd-metrics-tls\") pod \"ingress-operator-5b745b69d9-w4j4h\" (UID: \"f9ab8468-b920-41e9-a5f2-1af70f1b5ffd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4j4h" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.283452 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:16 crc kubenswrapper[4772]: E0127 15:09:16.283471 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:16.783450997 +0000 UTC m=+142.764060095 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.283991 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-console-serving-cert\") pod \"console-f9d7485db-7qfrl\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284019 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-service-ca\") pod \"console-f9d7485db-7qfrl\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284041 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dbd05529-6d54-416a-8df0-5973ee3179b6-serving-cert\") pod \"service-ca-operator-777779d784-djmb4\" (UID: \"dbd05529-6d54-416a-8df0-5973ee3179b6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-djmb4" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284065 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfglr\" (UniqueName: \"kubernetes.io/projected/f9ab8468-b920-41e9-a5f2-1af70f1b5ffd-kube-api-access-cfglr\") pod \"ingress-operator-5b745b69d9-w4j4h\" (UID: \"f9ab8468-b920-41e9-a5f2-1af70f1b5ffd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4j4h" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284082 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/794cdbb9-3392-465a-8a0a-a78a465aee2b-plugins-dir\") pod \"csi-hostpathplugin-hbbxh\" (UID: \"794cdbb9-3392-465a-8a0a-a78a465aee2b\") " pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284098 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8vjd\" (UniqueName: \"kubernetes.io/projected/55440e9e-5d99-4244-8b5c-55e2d270313b-kube-api-access-s8vjd\") pod \"machine-config-operator-74547568cd-njx6w\" (UID: \"55440e9e-5d99-4244-8b5c-55e2d270313b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-njx6w" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284116 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/877de785-bc18-4c1c-970a-1e6533539467-registry-tls\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284136 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284368 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/01d2c3f9-778c-4cf0-b8a4-76583f62df3c-srv-cert\") pod \"olm-operator-6b444d44fb-jdcpn\" (UID: \"01d2c3f9-778c-4cf0-b8a4-76583f62df3c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jdcpn" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284430 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5276546c-f731-4bd0-bb93-b5cd19b0992c-serving-cert\") pod \"console-operator-58897d9998-npths\" (UID: \"5276546c-f731-4bd0-bb93-b5cd19b0992c\") " pod="openshift-console-operator/console-operator-58897d9998-npths" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284463 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwmzq\" (UniqueName: \"kubernetes.io/projected/17bdd07d-f7e5-47f8-b730-724d5cc8e3d2-kube-api-access-cwmzq\") pod \"downloads-7954f5f757-vswtw\" (UID: \"17bdd07d-f7e5-47f8-b730-724d5cc8e3d2\") " pod="openshift-console/downloads-7954f5f757-vswtw" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284491 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3-serving-cert\") pod \"etcd-operator-b45778765-pwmhd\" (UID: \"c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pwmhd" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284516 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6b54ae2-d365-4988-8e69-704574c7962a-secret-volume\") pod \"collect-profiles-29492100-r2zj6\" (UID: \"c6b54ae2-d365-4988-8e69-704574c7962a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492100-r2zj6" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284548 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/877de785-bc18-4c1c-970a-1e6533539467-installation-pull-secrets\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284574 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3150273d-63f7-4908-bcc5-2403e123d1e7-metrics-tls\") pod \"dns-default-jplbk\" (UID: \"3150273d-63f7-4908-bcc5-2403e123d1e7\") " pod="openshift-dns/dns-default-jplbk" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284597 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f9ab8468-b920-41e9-a5f2-1af70f1b5ffd-trusted-ca\") pod \"ingress-operator-5b745b69d9-w4j4h\" (UID: \"f9ab8468-b920-41e9-a5f2-1af70f1b5ffd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4j4h" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284648 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smmpv\" (UniqueName: \"kubernetes.io/projected/ee07e18c-9f40-41c3-b2fb-05fd325976e4-kube-api-access-smmpv\") pod \"openshift-config-operator-7777fb866f-bck4j\" (UID: \"ee07e18c-9f40-41c3-b2fb-05fd325976e4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bck4j" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284675 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/85774248-1879-439e-9dd2-0d8661c299d6-proxy-tls\") pod \"machine-config-controller-84d6567774-v7l4k\" (UID: \"85774248-1879-439e-9dd2-0d8661c299d6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v7l4k" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284703 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284729 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dd05c164-83f7-4ebe-bbe8-9db6707741c5-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-mzkn2\" (UID: \"dd05c164-83f7-4ebe-bbe8-9db6707741c5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzkn2" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284753 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/55440e9e-5d99-4244-8b5c-55e2d270313b-proxy-tls\") pod \"machine-config-operator-74547568cd-njx6w\" (UID: \"55440e9e-5d99-4244-8b5c-55e2d270313b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-njx6w" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284782 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8800c113-7b51-4554-8e52-c1d0df1a08be-auth-proxy-config\") pod \"machine-approver-56656f9798-klfsg\" (UID: \"8800c113-7b51-4554-8e52-c1d0df1a08be\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfsg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284807 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/06755863-1a8b-4f4d-a304-03bfd45725ec-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-g667l\" (UID: \"06755863-1a8b-4f4d-a304-03bfd45725ec\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g667l" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284834 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284863 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f469e02c-9404-422a-bff0-1b945d9c8768-cert\") pod \"ingress-canary-fw6bh\" (UID: \"f469e02c-9404-422a-bff0-1b945d9c8768\") " pod="openshift-ingress-canary/ingress-canary-fw6bh" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284903 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2scxc\" (UniqueName: \"kubernetes.io/projected/794cdbb9-3392-465a-8a0a-a78a465aee2b-kube-api-access-2scxc\") pod \"csi-hostpathplugin-hbbxh\" (UID: \"794cdbb9-3392-465a-8a0a-a78a465aee2b\") " pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284930 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg5gf\" (UniqueName: \"kubernetes.io/projected/f469e02c-9404-422a-bff0-1b945d9c8768-kube-api-access-xg5gf\") pod \"ingress-canary-fw6bh\" (UID: \"f469e02c-9404-422a-bff0-1b945d9c8768\") " pod="openshift-ingress-canary/ingress-canary-fw6bh" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284968 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrbkm\" (UniqueName: \"kubernetes.io/projected/4bab7259-25d5-4c53-9ebb-ef2787adf010-kube-api-access-vrbkm\") pod \"machine-config-server-58dhr\" (UID: \"4bab7259-25d5-4c53-9ebb-ef2787adf010\") " pod="openshift-machine-config-operator/machine-config-server-58dhr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.284995 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285018 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/4bab7259-25d5-4c53-9ebb-ef2787adf010-certs\") pod \"machine-config-server-58dhr\" (UID: \"4bab7259-25d5-4c53-9ebb-ef2787adf010\") " pod="openshift-machine-config-operator/machine-config-server-58dhr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285040 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6b54ae2-d365-4988-8e69-704574c7962a-config-volume\") pod \"collect-profiles-29492100-r2zj6\" (UID: \"c6b54ae2-d365-4988-8e69-704574c7962a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492100-r2zj6" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285049 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-service-ca\") pod \"console-f9d7485db-7qfrl\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285063 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72xbw\" (UniqueName: \"kubernetes.io/projected/b3eefbc3-6dc4-479c-93e4-94a70fda0f83-kube-api-access-72xbw\") pod \"package-server-manager-789f6589d5-cv2z7\" (UID: \"b3eefbc3-6dc4-479c-93e4-94a70fda0f83\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cv2z7" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285110 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/877de785-bc18-4c1c-970a-1e6533539467-bound-sa-token\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285134 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/794cdbb9-3392-465a-8a0a-a78a465aee2b-socket-dir\") pod \"csi-hostpathplugin-hbbxh\" (UID: \"794cdbb9-3392-465a-8a0a-a78a465aee2b\") " pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285157 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s25qm\" (UniqueName: \"kubernetes.io/projected/59cdf584-81d0-4d66-8fc2-da3a3f995f73-kube-api-access-s25qm\") pod \"migrator-59844c95c7-xqtff\" (UID: \"59cdf584-81d0-4d66-8fc2-da3a3f995f73\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xqtff" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285201 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffe33359-31f5-4a6c-93fc-6502d2516335-config\") pod \"kube-controller-manager-operator-78b949d7b-fbwkz\" (UID: \"ffe33359-31f5-4a6c-93fc-6502d2516335\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fbwkz" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285247 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/877de785-bc18-4c1c-970a-1e6533539467-trusted-ca\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285272 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c66d8c7d-2de8-492f-ba5e-7ff0e236bf64-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-zkmjj\" (UID: \"c66d8c7d-2de8-492f-ba5e-7ff0e236bf64\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zkmjj" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285303 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/06cdc094-b372-4016-bc5e-4c15a28e032e-audit-policies\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285325 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ffe33359-31f5-4a6c-93fc-6502d2516335-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fbwkz\" (UID: \"ffe33359-31f5-4a6c-93fc-6502d2516335\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fbwkz" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285349 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmnrs\" (UniqueName: \"kubernetes.io/projected/dbd05529-6d54-416a-8df0-5973ee3179b6-kube-api-access-kmnrs\") pod \"service-ca-operator-777779d784-djmb4\" (UID: \"dbd05529-6d54-416a-8df0-5973ee3179b6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-djmb4" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285374 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285398 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5f6v\" (UniqueName: \"kubernetes.io/projected/06cdc094-b372-4016-bc5e-4c15a28e032e-kube-api-access-l5f6v\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285419 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/4bab7259-25d5-4c53-9ebb-ef2787adf010-node-bootstrap-token\") pod \"machine-config-server-58dhr\" (UID: \"4bab7259-25d5-4c53-9ebb-ef2787adf010\") " pod="openshift-machine-config-operator/machine-config-server-58dhr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285441 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/01d08621-494b-4232-b678-9caa94e61085-tmpfs\") pod \"packageserver-d55dfcdfc-cf6v7\" (UID: \"01d08621-494b-4232-b678-9caa94e61085\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cf6v7" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285469 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285502 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4225ddc-bdcd-4158-811b-113234d0c3d0-service-ca-bundle\") pod \"router-default-5444994796-7k7sg\" (UID: \"c4225ddc-bdcd-4158-811b-113234d0c3d0\") " pod="openshift-ingress/router-default-5444994796-7k7sg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285526 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285553 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/79859b30-67ee-456b-82e5-f8806347a0b9-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-c6n8m\" (UID: \"79859b30-67ee-456b-82e5-f8806347a0b9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c6n8m" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285583 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5z5pj\" (UniqueName: \"kubernetes.io/projected/c4225ddc-bdcd-4158-811b-113234d0c3d0-kube-api-access-5z5pj\") pod \"router-default-5444994796-7k7sg\" (UID: \"c4225ddc-bdcd-4158-811b-113234d0c3d0\") " pod="openshift-ingress/router-default-5444994796-7k7sg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285609 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/ee07e18c-9f40-41c3-b2fb-05fd325976e4-available-featuregates\") pod \"openshift-config-operator-7777fb866f-bck4j\" (UID: \"ee07e18c-9f40-41c3-b2fb-05fd325976e4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bck4j" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285646 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg5jv\" (UniqueName: \"kubernetes.io/projected/3150273d-63f7-4908-bcc5-2403e123d1e7-kube-api-access-dg5jv\") pod \"dns-default-jplbk\" (UID: \"3150273d-63f7-4908-bcc5-2403e123d1e7\") " pod="openshift-dns/dns-default-jplbk" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285670 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j667k\" (UniqueName: \"kubernetes.io/projected/5276546c-f731-4bd0-bb93-b5cd19b0992c-kube-api-access-j667k\") pod \"console-operator-58897d9998-npths\" (UID: \"5276546c-f731-4bd0-bb93-b5cd19b0992c\") " pod="openshift-console-operator/console-operator-58897d9998-npths" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285695 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285719 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r6t5\" (UniqueName: \"kubernetes.io/projected/79859b30-67ee-456b-82e5-f8806347a0b9-kube-api-access-9r6t5\") pod \"openshift-controller-manager-operator-756b6f6bc6-c6n8m\" (UID: \"79859b30-67ee-456b-82e5-f8806347a0b9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c6n8m" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285750 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmgkc\" (UniqueName: \"kubernetes.io/projected/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-kube-api-access-vmgkc\") pod \"console-f9d7485db-7qfrl\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285775 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcp9p\" (UniqueName: \"kubernetes.io/projected/85774248-1879-439e-9dd2-0d8661c299d6-kube-api-access-gcp9p\") pod \"machine-config-controller-84d6567774-v7l4k\" (UID: \"85774248-1879-439e-9dd2-0d8661c299d6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v7l4k" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285801 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/877de785-bc18-4c1c-970a-1e6533539467-registry-certificates\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285826 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c4225ddc-bdcd-4158-811b-113234d0c3d0-stats-auth\") pod \"router-default-5444994796-7k7sg\" (UID: \"c4225ddc-bdcd-4158-811b-113234d0c3d0\") " pod="openshift-ingress/router-default-5444994796-7k7sg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285850 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2zfh\" (UniqueName: \"kubernetes.io/projected/ce4b4e2e-496b-4334-8736-db4f25473731-kube-api-access-k2zfh\") pod \"dns-operator-744455d44c-bhgv8\" (UID: \"ce4b4e2e-496b-4334-8736-db4f25473731\") " pod="openshift-dns-operator/dns-operator-744455d44c-bhgv8" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285874 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2c9eef7e-3996-45b6-ab7b-50d319dc1117-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-4v228\" (UID: \"2c9eef7e-3996-45b6-ab7b-50d319dc1117\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4v228" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285900 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llq59\" (UniqueName: \"kubernetes.io/projected/c6b54ae2-d365-4988-8e69-704574c7962a-kube-api-access-llq59\") pod \"collect-profiles-29492100-r2zj6\" (UID: \"c6b54ae2-d365-4988-8e69-704574c7962a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492100-r2zj6" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285924 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/794cdbb9-3392-465a-8a0a-a78a465aee2b-mountpoint-dir\") pod \"csi-hostpathplugin-hbbxh\" (UID: \"794cdbb9-3392-465a-8a0a-a78a465aee2b\") " pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285946 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/946946b3-f9e0-45e4-803f-edb3f7218489-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f9fqs\" (UID: \"946946b3-f9e0-45e4-803f-edb3f7218489\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f9fqs" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.285976 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/794cdbb9-3392-465a-8a0a-a78a465aee2b-csi-data-dir\") pod \"csi-hostpathplugin-hbbxh\" (UID: \"794cdbb9-3392-465a-8a0a-a78a465aee2b\") " pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286025 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/db4a3858-5afa-44c8-a435-2010f7e7340d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5vmlj\" (UID: \"db4a3858-5afa-44c8-a435-2010f7e7340d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5vmlj" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286050 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvbbk\" (UniqueName: \"kubernetes.io/projected/1f03647c-a3e9-4099-9780-e79e3a4d4cf2-kube-api-access-xvbbk\") pod \"catalog-operator-68c6474976-mnltb\" (UID: \"1f03647c-a3e9-4099-9780-e79e3a4d4cf2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mnltb" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286072 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/55440e9e-5d99-4244-8b5c-55e2d270313b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-njx6w\" (UID: \"55440e9e-5d99-4244-8b5c-55e2d270313b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-njx6w" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286110 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-console-oauth-config\") pod \"console-f9d7485db-7qfrl\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286135 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/55440e9e-5d99-4244-8b5c-55e2d270313b-images\") pod \"machine-config-operator-74547568cd-njx6w\" (UID: \"55440e9e-5d99-4244-8b5c-55e2d270313b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-njx6w" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286160 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8800c113-7b51-4554-8e52-c1d0df1a08be-machine-approver-tls\") pod \"machine-approver-56656f9798-klfsg\" (UID: \"8800c113-7b51-4554-8e52-c1d0df1a08be\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfsg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286202 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/06cdc094-b372-4016-bc5e-4c15a28e032e-audit-dir\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286228 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1f03647c-a3e9-4099-9780-e79e3a4d4cf2-srv-cert\") pod \"catalog-operator-68c6474976-mnltb\" (UID: \"1f03647c-a3e9-4099-9780-e79e3a4d4cf2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mnltb" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286251 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3-etcd-service-ca\") pod \"etcd-operator-b45778765-pwmhd\" (UID: \"c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pwmhd" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286272 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w88tz\" (UniqueName: \"kubernetes.io/projected/c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3-kube-api-access-w88tz\") pod \"etcd-operator-b45778765-pwmhd\" (UID: \"c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pwmhd" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286295 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/946946b3-f9e0-45e4-803f-edb3f7218489-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f9fqs\" (UID: \"946946b3-f9e0-45e4-803f-edb3f7218489\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f9fqs" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286320 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee07e18c-9f40-41c3-b2fb-05fd325976e4-serving-cert\") pod \"openshift-config-operator-7777fb866f-bck4j\" (UID: \"ee07e18c-9f40-41c3-b2fb-05fd325976e4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bck4j" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286342 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ffe33359-31f5-4a6c-93fc-6502d2516335-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fbwkz\" (UID: \"ffe33359-31f5-4a6c-93fc-6502d2516335\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fbwkz" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286365 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c8ebf890-c3b0-468e-bf7d-0ec590df084b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4lj2h\" (UID: \"c8ebf890-c3b0-468e-bf7d-0ec590df084b\") " pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286409 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ce4b4e2e-496b-4334-8736-db4f25473731-metrics-tls\") pod \"dns-operator-744455d44c-bhgv8\" (UID: \"ce4b4e2e-496b-4334-8736-db4f25473731\") " pod="openshift-dns-operator/dns-operator-744455d44c-bhgv8" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286438 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7s88q\" (UniqueName: \"kubernetes.io/projected/192d0f8f-10f9-43e2-a24a-2019aae0db44-kube-api-access-7s88q\") pod \"openshift-apiserver-operator-796bbdcf4f-zq27x\" (UID: \"192d0f8f-10f9-43e2-a24a-2019aae0db44\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zq27x" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286462 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2c9eef7e-3996-45b6-ab7b-50d319dc1117-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-4v228\" (UID: \"2c9eef7e-3996-45b6-ab7b-50d319dc1117\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4v228" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286486 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/01d08621-494b-4232-b678-9caa94e61085-apiservice-cert\") pod \"packageserver-d55dfcdfc-cf6v7\" (UID: \"01d08621-494b-4232-b678-9caa94e61085\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cf6v7" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286488 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f9ab8468-b920-41e9-a5f2-1af70f1b5ffd-trusted-ca\") pod \"ingress-operator-5b745b69d9-w4j4h\" (UID: \"f9ab8468-b920-41e9-a5f2-1af70f1b5ffd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4j4h" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286511 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h66t\" (UniqueName: \"kubernetes.io/projected/877de785-bc18-4c1c-970a-1e6533539467-kube-api-access-5h66t\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286535 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2c9eef7e-3996-45b6-ab7b-50d319dc1117-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-4v228\" (UID: \"2c9eef7e-3996-45b6-ab7b-50d319dc1117\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4v228" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286559 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjbpj\" (UniqueName: \"kubernetes.io/projected/01d08621-494b-4232-b678-9caa94e61085-kube-api-access-mjbpj\") pod \"packageserver-d55dfcdfc-cf6v7\" (UID: \"01d08621-494b-4232-b678-9caa94e61085\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cf6v7" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286584 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5276546c-f731-4bd0-bb93-b5cd19b0992c-config\") pod \"console-operator-58897d9998-npths\" (UID: \"5276546c-f731-4bd0-bb93-b5cd19b0992c\") " pod="openshift-console-operator/console-operator-58897d9998-npths" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286612 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/877de785-bc18-4c1c-970a-1e6533539467-ca-trust-extracted\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286635 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lbs2\" (UniqueName: \"kubernetes.io/projected/8800c113-7b51-4554-8e52-c1d0df1a08be-kube-api-access-4lbs2\") pod \"machine-approver-56656f9798-klfsg\" (UID: \"8800c113-7b51-4554-8e52-c1d0df1a08be\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfsg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286658 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-trusted-ca-bundle\") pod \"console-f9d7485db-7qfrl\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286681 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twqnj\" (UniqueName: \"kubernetes.io/projected/06755863-1a8b-4f4d-a304-03bfd45725ec-kube-api-access-twqnj\") pod \"cluster-samples-operator-665b6dd947-g667l\" (UID: \"06755863-1a8b-4f4d-a304-03bfd45725ec\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g667l" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286705 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd05c164-83f7-4ebe-bbe8-9db6707741c5-config\") pod \"kube-apiserver-operator-766d6c64bb-mzkn2\" (UID: \"dd05c164-83f7-4ebe-bbe8-9db6707741c5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzkn2" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286729 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3-etcd-client\") pod \"etcd-operator-b45778765-pwmhd\" (UID: \"c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pwmhd" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286731 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286754 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c4225ddc-bdcd-4158-811b-113234d0c3d0-default-certificate\") pod \"router-default-5444994796-7k7sg\" (UID: \"c4225ddc-bdcd-4158-811b-113234d0c3d0\") " pod="openshift-ingress/router-default-5444994796-7k7sg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286783 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286809 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3-config\") pod \"etcd-operator-b45778765-pwmhd\" (UID: \"c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pwmhd" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286833 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3745a211-9fa8-41a7-aa26-d733431bc9aa-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-24dmv\" (UID: \"3745a211-9fa8-41a7-aa26-d733431bc9aa\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-24dmv" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286859 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8800c113-7b51-4554-8e52-c1d0df1a08be-config\") pod \"machine-approver-56656f9798-klfsg\" (UID: \"8800c113-7b51-4554-8e52-c1d0df1a08be\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfsg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286882 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-console-config\") pod \"console-f9d7485db-7qfrl\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286905 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c66d8c7d-2de8-492f-ba5e-7ff0e236bf64-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-zkmjj\" (UID: \"c66d8c7d-2de8-492f-ba5e-7ff0e236bf64\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zkmjj" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.286976 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvb9h\" (UniqueName: \"kubernetes.io/projected/c8ebf890-c3b0-468e-bf7d-0ec590df084b-kube-api-access-vvb9h\") pod \"marketplace-operator-79b997595-4lj2h\" (UID: \"c8ebf890-c3b0-468e-bf7d-0ec590df084b\") " pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287006 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f9ab8468-b920-41e9-a5f2-1af70f1b5ffd-metrics-tls\") pod \"ingress-operator-5b745b69d9-w4j4h\" (UID: \"f9ab8468-b920-41e9-a5f2-1af70f1b5ffd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4j4h" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287030 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b3eefbc3-6dc4-479c-93e4-94a70fda0f83-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-cv2z7\" (UID: \"b3eefbc3-6dc4-479c-93e4-94a70fda0f83\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cv2z7" Jan 27 15:09:16 crc kubenswrapper[4772]: E0127 15:09:16.287053 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:16.78704049 +0000 UTC m=+142.767649588 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287081 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/85774248-1879-439e-9dd2-0d8661c299d6-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-v7l4k\" (UID: \"85774248-1879-439e-9dd2-0d8661c299d6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v7l4k" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287102 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1f03647c-a3e9-4099-9780-e79e3a4d4cf2-profile-collector-cert\") pod \"catalog-operator-68c6474976-mnltb\" (UID: \"1f03647c-a3e9-4099-9780-e79e3a4d4cf2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mnltb" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287122 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvzg8\" (UniqueName: \"kubernetes.io/projected/c66d8c7d-2de8-492f-ba5e-7ff0e236bf64-kube-api-access-jvzg8\") pod \"kube-storage-version-migrator-operator-b67b599dd-zkmjj\" (UID: \"c66d8c7d-2de8-492f-ba5e-7ff0e236bf64\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zkmjj" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287140 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c8ebf890-c3b0-468e-bf7d-0ec590df084b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4lj2h\" (UID: \"c8ebf890-c3b0-468e-bf7d-0ec590df084b\") " pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287159 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxscw\" (UniqueName: \"kubernetes.io/projected/3745a211-9fa8-41a7-aa26-d733431bc9aa-kube-api-access-sxscw\") pod \"multus-admission-controller-857f4d67dd-24dmv\" (UID: \"3745a211-9fa8-41a7-aa26-d733431bc9aa\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-24dmv" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287193 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287211 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f9ab8468-b920-41e9-a5f2-1af70f1b5ffd-bound-sa-token\") pod \"ingress-operator-5b745b69d9-w4j4h\" (UID: \"f9ab8468-b920-41e9-a5f2-1af70f1b5ffd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4j4h" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287230 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd05c164-83f7-4ebe-bbe8-9db6707741c5-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-mzkn2\" (UID: \"dd05c164-83f7-4ebe-bbe8-9db6707741c5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzkn2" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287248 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5cbs\" (UniqueName: \"kubernetes.io/projected/855486b0-11f8-4ff0-930d-75c7e9d790d3-kube-api-access-m5cbs\") pod \"service-ca-9c57cc56f-9wv77\" (UID: \"855486b0-11f8-4ff0-930d-75c7e9d790d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-9wv77" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287267 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/794cdbb9-3392-465a-8a0a-a78a465aee2b-registration-dir\") pod \"csi-hostpathplugin-hbbxh\" (UID: \"794cdbb9-3392-465a-8a0a-a78a465aee2b\") " pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287284 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9v6df\" (UniqueName: \"kubernetes.io/projected/db4a3858-5afa-44c8-a435-2010f7e7340d-kube-api-access-9v6df\") pod \"control-plane-machine-set-operator-78cbb6b69f-5vmlj\" (UID: \"db4a3858-5afa-44c8-a435-2010f7e7340d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5vmlj" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287303 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-oauth-serving-cert\") pod \"console-f9d7485db-7qfrl\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287326 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/192d0f8f-10f9-43e2-a24a-2019aae0db44-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zq27x\" (UID: \"192d0f8f-10f9-43e2-a24a-2019aae0db44\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zq27x" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287350 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c4225ddc-bdcd-4158-811b-113234d0c3d0-metrics-certs\") pod \"router-default-5444994796-7k7sg\" (UID: \"c4225ddc-bdcd-4158-811b-113234d0c3d0\") " pod="openshift-ingress/router-default-5444994796-7k7sg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287376 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/192d0f8f-10f9-43e2-a24a-2019aae0db44-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zq27x\" (UID: \"192d0f8f-10f9-43e2-a24a-2019aae0db44\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zq27x" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287398 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79859b30-67ee-456b-82e5-f8806347a0b9-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-c6n8m\" (UID: \"79859b30-67ee-456b-82e5-f8806347a0b9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c6n8m" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287423 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/01d2c3f9-778c-4cf0-b8a4-76583f62df3c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-jdcpn\" (UID: \"01d2c3f9-778c-4cf0-b8a4-76583f62df3c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jdcpn" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287449 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287478 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mswpx\" (UniqueName: \"kubernetes.io/projected/2c9eef7e-3996-45b6-ab7b-50d319dc1117-kube-api-access-mswpx\") pod \"cluster-image-registry-operator-dc59b4c8b-4v228\" (UID: \"2c9eef7e-3996-45b6-ab7b-50d319dc1117\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4v228" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287503 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/946946b3-f9e0-45e4-803f-edb3f7218489-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f9fqs\" (UID: \"946946b3-f9e0-45e4-803f-edb3f7218489\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f9fqs" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287527 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3150273d-63f7-4908-bcc5-2403e123d1e7-config-volume\") pod \"dns-default-jplbk\" (UID: \"3150273d-63f7-4908-bcc5-2403e123d1e7\") " pod="openshift-dns/dns-default-jplbk" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287547 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/855486b0-11f8-4ff0-930d-75c7e9d790d3-signing-key\") pod \"service-ca-9c57cc56f-9wv77\" (UID: \"855486b0-11f8-4ff0-930d-75c7e9d790d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-9wv77" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287570 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3-etcd-ca\") pod \"etcd-operator-b45778765-pwmhd\" (UID: \"c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pwmhd" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287590 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbd05529-6d54-416a-8df0-5973ee3179b6-config\") pod \"service-ca-operator-777779d784-djmb4\" (UID: \"dbd05529-6d54-416a-8df0-5973ee3179b6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-djmb4" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287618 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttqp4\" (UniqueName: \"kubernetes.io/projected/01d2c3f9-778c-4cf0-b8a4-76583f62df3c-kube-api-access-ttqp4\") pod \"olm-operator-6b444d44fb-jdcpn\" (UID: \"01d2c3f9-778c-4cf0-b8a4-76583f62df3c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jdcpn" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287642 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/01d08621-494b-4232-b678-9caa94e61085-webhook-cert\") pod \"packageserver-d55dfcdfc-cf6v7\" (UID: \"01d08621-494b-4232-b678-9caa94e61085\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cf6v7" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287667 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5276546c-f731-4bd0-bb93-b5cd19b0992c-trusted-ca\") pod \"console-operator-58897d9998-npths\" (UID: \"5276546c-f731-4bd0-bb93-b5cd19b0992c\") " pod="openshift-console-operator/console-operator-58897d9998-npths" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287703 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.287723 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/855486b0-11f8-4ff0-930d-75c7e9d790d3-signing-cabundle\") pod \"service-ca-9c57cc56f-9wv77\" (UID: \"855486b0-11f8-4ff0-930d-75c7e9d790d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-9wv77" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.288574 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4225ddc-bdcd-4158-811b-113234d0c3d0-service-ca-bundle\") pod \"router-default-5444994796-7k7sg\" (UID: \"c4225ddc-bdcd-4158-811b-113234d0c3d0\") " pod="openshift-ingress/router-default-5444994796-7k7sg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.289503 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.289787 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/06755863-1a8b-4f4d-a304-03bfd45725ec-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-g667l\" (UID: \"06755863-1a8b-4f4d-a304-03bfd45725ec\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g667l" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.290129 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8800c113-7b51-4554-8e52-c1d0df1a08be-auth-proxy-config\") pod \"machine-approver-56656f9798-klfsg\" (UID: \"8800c113-7b51-4554-8e52-c1d0df1a08be\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfsg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.290269 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/ee07e18c-9f40-41c3-b2fb-05fd325976e4-available-featuregates\") pod \"openshift-config-operator-7777fb866f-bck4j\" (UID: \"ee07e18c-9f40-41c3-b2fb-05fd325976e4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bck4j" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.290628 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.291437 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/06cdc094-b372-4016-bc5e-4c15a28e032e-audit-policies\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.291935 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-oauth-serving-cert\") pod \"console-f9d7485db-7qfrl\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.292047 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3150273d-63f7-4908-bcc5-2403e123d1e7-config-volume\") pod \"dns-default-jplbk\" (UID: \"3150273d-63f7-4908-bcc5-2403e123d1e7\") " pod="openshift-dns/dns-default-jplbk" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.292641 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/877de785-bc18-4c1c-970a-1e6533539467-trusted-ca\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.293353 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.293903 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/877de785-bc18-4c1c-970a-1e6533539467-registry-tls\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.294054 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/192d0f8f-10f9-43e2-a24a-2019aae0db44-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zq27x\" (UID: \"192d0f8f-10f9-43e2-a24a-2019aae0db44\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zq27x" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.294281 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3150273d-63f7-4908-bcc5-2403e123d1e7-metrics-tls\") pod \"dns-default-jplbk\" (UID: \"3150273d-63f7-4908-bcc5-2403e123d1e7\") " pod="openshift-dns/dns-default-jplbk" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.295181 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee07e18c-9f40-41c3-b2fb-05fd325976e4-serving-cert\") pod \"openshift-config-operator-7777fb866f-bck4j\" (UID: \"ee07e18c-9f40-41c3-b2fb-05fd325976e4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bck4j" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.297904 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/877de785-bc18-4c1c-970a-1e6533539467-registry-certificates\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.299274 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.299784 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.300237 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.300317 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/06cdc094-b372-4016-bc5e-4c15a28e032e-audit-dir\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.301089 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd05c164-83f7-4ebe-bbe8-9db6707741c5-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-mzkn2\" (UID: \"dd05c164-83f7-4ebe-bbe8-9db6707741c5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzkn2" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.301143 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2c9eef7e-3996-45b6-ab7b-50d319dc1117-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-4v228\" (UID: \"2c9eef7e-3996-45b6-ab7b-50d319dc1117\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4v228" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.301793 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c4225ddc-bdcd-4158-811b-113234d0c3d0-metrics-certs\") pod \"router-default-5444994796-7k7sg\" (UID: \"c4225ddc-bdcd-4158-811b-113234d0c3d0\") " pod="openshift-ingress/router-default-5444994796-7k7sg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.301939 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-console-config\") pod \"console-f9d7485db-7qfrl\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.302385 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c4225ddc-bdcd-4158-811b-113234d0c3d0-stats-auth\") pod \"router-default-5444994796-7k7sg\" (UID: \"c4225ddc-bdcd-4158-811b-113234d0c3d0\") " pod="openshift-ingress/router-default-5444994796-7k7sg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.302703 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/192d0f8f-10f9-43e2-a24a-2019aae0db44-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zq27x\" (UID: \"192d0f8f-10f9-43e2-a24a-2019aae0db44\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zq27x" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.303123 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/877de785-bc18-4c1c-970a-1e6533539467-ca-trust-extracted\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.303389 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.303414 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8800c113-7b51-4554-8e52-c1d0df1a08be-config\") pod \"machine-approver-56656f9798-klfsg\" (UID: \"8800c113-7b51-4554-8e52-c1d0df1a08be\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfsg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.304525 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd05c164-83f7-4ebe-bbe8-9db6707741c5-config\") pod \"kube-apiserver-operator-766d6c64bb-mzkn2\" (UID: \"dd05c164-83f7-4ebe-bbe8-9db6707741c5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzkn2" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.304763 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5276546c-f731-4bd0-bb93-b5cd19b0992c-config\") pod \"console-operator-58897d9998-npths\" (UID: \"5276546c-f731-4bd0-bb93-b5cd19b0992c\") " pod="openshift-console-operator/console-operator-58897d9998-npths" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.305727 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5276546c-f731-4bd0-bb93-b5cd19b0992c-serving-cert\") pod \"console-operator-58897d9998-npths\" (UID: \"5276546c-f731-4bd0-bb93-b5cd19b0992c\") " pod="openshift-console-operator/console-operator-58897d9998-npths" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.305741 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5276546c-f731-4bd0-bb93-b5cd19b0992c-trusted-ca\") pod \"console-operator-58897d9998-npths\" (UID: \"5276546c-f731-4bd0-bb93-b5cd19b0992c\") " pod="openshift-console-operator/console-operator-58897d9998-npths" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.306792 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8800c113-7b51-4554-8e52-c1d0df1a08be-machine-approver-tls\") pod \"machine-approver-56656f9798-klfsg\" (UID: \"8800c113-7b51-4554-8e52-c1d0df1a08be\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfsg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.307222 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c4225ddc-bdcd-4158-811b-113234d0c3d0-default-certificate\") pod \"router-default-5444994796-7k7sg\" (UID: \"c4225ddc-bdcd-4158-811b-113234d0c3d0\") " pod="openshift-ingress/router-default-5444994796-7k7sg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.308256 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2c9eef7e-3996-45b6-ab7b-50d319dc1117-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-4v228\" (UID: \"2c9eef7e-3996-45b6-ab7b-50d319dc1117\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4v228" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.308492 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ce4b4e2e-496b-4334-8736-db4f25473731-metrics-tls\") pod \"dns-operator-744455d44c-bhgv8\" (UID: \"ce4b4e2e-496b-4334-8736-db4f25473731\") " pod="openshift-dns-operator/dns-operator-744455d44c-bhgv8" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.308582 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f9ab8468-b920-41e9-a5f2-1af70f1b5ffd-metrics-tls\") pod \"ingress-operator-5b745b69d9-w4j4h\" (UID: \"f9ab8468-b920-41e9-a5f2-1af70f1b5ffd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4j4h" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.309019 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.309187 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-trusted-ca-bundle\") pod \"console-f9d7485db-7qfrl\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.309345 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/877de785-bc18-4c1c-970a-1e6533539467-installation-pull-secrets\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.310457 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-console-serving-cert\") pod \"console-f9d7485db-7qfrl\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.310694 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.314426 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.321778 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-console-oauth-config\") pod \"console-f9d7485db-7qfrl\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.323949 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfglr\" (UniqueName: \"kubernetes.io/projected/f9ab8468-b920-41e9-a5f2-1af70f1b5ffd-kube-api-access-cfglr\") pod \"ingress-operator-5b745b69d9-w4j4h\" (UID: \"f9ab8468-b920-41e9-a5f2-1af70f1b5ffd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4j4h" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.342062 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dd05c164-83f7-4ebe-bbe8-9db6707741c5-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-mzkn2\" (UID: \"dd05c164-83f7-4ebe-bbe8-9db6707741c5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzkn2" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.360533 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzkn2" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.366368 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwmzq\" (UniqueName: \"kubernetes.io/projected/17bdd07d-f7e5-47f8-b730-724d5cc8e3d2-kube-api-access-cwmzq\") pod \"downloads-7954f5f757-vswtw\" (UID: \"17bdd07d-f7e5-47f8-b730-724d5cc8e3d2\") " pod="openshift-console/downloads-7954f5f757-vswtw" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.379612 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smmpv\" (UniqueName: \"kubernetes.io/projected/ee07e18c-9f40-41c3-b2fb-05fd325976e4-kube-api-access-smmpv\") pod \"openshift-config-operator-7777fb866f-bck4j\" (UID: \"ee07e18c-9f40-41c3-b2fb-05fd325976e4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bck4j" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.390316 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:16 crc kubenswrapper[4772]: E0127 15:09:16.390474 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:16.890450807 +0000 UTC m=+142.871059905 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.390532 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5cbs\" (UniqueName: \"kubernetes.io/projected/855486b0-11f8-4ff0-930d-75c7e9d790d3-kube-api-access-m5cbs\") pod \"service-ca-9c57cc56f-9wv77\" (UID: \"855486b0-11f8-4ff0-930d-75c7e9d790d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-9wv77" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.390561 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/794cdbb9-3392-465a-8a0a-a78a465aee2b-registration-dir\") pod \"csi-hostpathplugin-hbbxh\" (UID: \"794cdbb9-3392-465a-8a0a-a78a465aee2b\") " pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.390582 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9v6df\" (UniqueName: \"kubernetes.io/projected/db4a3858-5afa-44c8-a435-2010f7e7340d-kube-api-access-9v6df\") pod \"control-plane-machine-set-operator-78cbb6b69f-5vmlj\" (UID: \"db4a3858-5afa-44c8-a435-2010f7e7340d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5vmlj" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.390605 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxscw\" (UniqueName: \"kubernetes.io/projected/3745a211-9fa8-41a7-aa26-d733431bc9aa-kube-api-access-sxscw\") pod \"multus-admission-controller-857f4d67dd-24dmv\" (UID: \"3745a211-9fa8-41a7-aa26-d733431bc9aa\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-24dmv" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.390630 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79859b30-67ee-456b-82e5-f8806347a0b9-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-c6n8m\" (UID: \"79859b30-67ee-456b-82e5-f8806347a0b9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c6n8m" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.390652 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/01d2c3f9-778c-4cf0-b8a4-76583f62df3c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-jdcpn\" (UID: \"01d2c3f9-778c-4cf0-b8a4-76583f62df3c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jdcpn" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.390681 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/946946b3-f9e0-45e4-803f-edb3f7218489-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f9fqs\" (UID: \"946946b3-f9e0-45e4-803f-edb3f7218489\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f9fqs" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.390703 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/855486b0-11f8-4ff0-930d-75c7e9d790d3-signing-key\") pod \"service-ca-9c57cc56f-9wv77\" (UID: \"855486b0-11f8-4ff0-930d-75c7e9d790d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-9wv77" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.390722 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3-etcd-ca\") pod \"etcd-operator-b45778765-pwmhd\" (UID: \"c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pwmhd" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.390742 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbd05529-6d54-416a-8df0-5973ee3179b6-config\") pod \"service-ca-operator-777779d784-djmb4\" (UID: \"dbd05529-6d54-416a-8df0-5973ee3179b6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-djmb4" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.390773 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttqp4\" (UniqueName: \"kubernetes.io/projected/01d2c3f9-778c-4cf0-b8a4-76583f62df3c-kube-api-access-ttqp4\") pod \"olm-operator-6b444d44fb-jdcpn\" (UID: \"01d2c3f9-778c-4cf0-b8a4-76583f62df3c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jdcpn" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.390793 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/01d08621-494b-4232-b678-9caa94e61085-webhook-cert\") pod \"packageserver-d55dfcdfc-cf6v7\" (UID: \"01d08621-494b-4232-b678-9caa94e61085\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cf6v7" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.390814 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/855486b0-11f8-4ff0-930d-75c7e9d790d3-signing-cabundle\") pod \"service-ca-9c57cc56f-9wv77\" (UID: \"855486b0-11f8-4ff0-930d-75c7e9d790d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-9wv77" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.390836 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dbd05529-6d54-416a-8df0-5973ee3179b6-serving-cert\") pod \"service-ca-operator-777779d784-djmb4\" (UID: \"dbd05529-6d54-416a-8df0-5973ee3179b6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-djmb4" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.390856 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/794cdbb9-3392-465a-8a0a-a78a465aee2b-plugins-dir\") pod \"csi-hostpathplugin-hbbxh\" (UID: \"794cdbb9-3392-465a-8a0a-a78a465aee2b\") " pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.390875 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8vjd\" (UniqueName: \"kubernetes.io/projected/55440e9e-5d99-4244-8b5c-55e2d270313b-kube-api-access-s8vjd\") pod \"machine-config-operator-74547568cd-njx6w\" (UID: \"55440e9e-5d99-4244-8b5c-55e2d270313b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-njx6w" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.390924 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/01d2c3f9-778c-4cf0-b8a4-76583f62df3c-srv-cert\") pod \"olm-operator-6b444d44fb-jdcpn\" (UID: \"01d2c3f9-778c-4cf0-b8a4-76583f62df3c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jdcpn" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.390944 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3-serving-cert\") pod \"etcd-operator-b45778765-pwmhd\" (UID: \"c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pwmhd" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.390964 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6b54ae2-d365-4988-8e69-704574c7962a-secret-volume\") pod \"collect-profiles-29492100-r2zj6\" (UID: \"c6b54ae2-d365-4988-8e69-704574c7962a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492100-r2zj6" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.390985 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/85774248-1879-439e-9dd2-0d8661c299d6-proxy-tls\") pod \"machine-config-controller-84d6567774-v7l4k\" (UID: \"85774248-1879-439e-9dd2-0d8661c299d6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v7l4k" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391011 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f469e02c-9404-422a-bff0-1b945d9c8768-cert\") pod \"ingress-canary-fw6bh\" (UID: \"f469e02c-9404-422a-bff0-1b945d9c8768\") " pod="openshift-ingress-canary/ingress-canary-fw6bh" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391031 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/55440e9e-5d99-4244-8b5c-55e2d270313b-proxy-tls\") pod \"machine-config-operator-74547568cd-njx6w\" (UID: \"55440e9e-5d99-4244-8b5c-55e2d270313b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-njx6w" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391053 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg5gf\" (UniqueName: \"kubernetes.io/projected/f469e02c-9404-422a-bff0-1b945d9c8768-kube-api-access-xg5gf\") pod \"ingress-canary-fw6bh\" (UID: \"f469e02c-9404-422a-bff0-1b945d9c8768\") " pod="openshift-ingress-canary/ingress-canary-fw6bh" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391086 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2scxc\" (UniqueName: \"kubernetes.io/projected/794cdbb9-3392-465a-8a0a-a78a465aee2b-kube-api-access-2scxc\") pod \"csi-hostpathplugin-hbbxh\" (UID: \"794cdbb9-3392-465a-8a0a-a78a465aee2b\") " pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391114 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrbkm\" (UniqueName: \"kubernetes.io/projected/4bab7259-25d5-4c53-9ebb-ef2787adf010-kube-api-access-vrbkm\") pod \"machine-config-server-58dhr\" (UID: \"4bab7259-25d5-4c53-9ebb-ef2787adf010\") " pod="openshift-machine-config-operator/machine-config-server-58dhr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391134 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/4bab7259-25d5-4c53-9ebb-ef2787adf010-certs\") pod \"machine-config-server-58dhr\" (UID: \"4bab7259-25d5-4c53-9ebb-ef2787adf010\") " pod="openshift-machine-config-operator/machine-config-server-58dhr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391154 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6b54ae2-d365-4988-8e69-704574c7962a-config-volume\") pod \"collect-profiles-29492100-r2zj6\" (UID: \"c6b54ae2-d365-4988-8e69-704574c7962a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492100-r2zj6" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391196 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/794cdbb9-3392-465a-8a0a-a78a465aee2b-socket-dir\") pod \"csi-hostpathplugin-hbbxh\" (UID: \"794cdbb9-3392-465a-8a0a-a78a465aee2b\") " pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391218 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s25qm\" (UniqueName: \"kubernetes.io/projected/59cdf584-81d0-4d66-8fc2-da3a3f995f73-kube-api-access-s25qm\") pod \"migrator-59844c95c7-xqtff\" (UID: \"59cdf584-81d0-4d66-8fc2-da3a3f995f73\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xqtff" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391240 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72xbw\" (UniqueName: \"kubernetes.io/projected/b3eefbc3-6dc4-479c-93e4-94a70fda0f83-kube-api-access-72xbw\") pod \"package-server-manager-789f6589d5-cv2z7\" (UID: \"b3eefbc3-6dc4-479c-93e4-94a70fda0f83\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cv2z7" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391269 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffe33359-31f5-4a6c-93fc-6502d2516335-config\") pod \"kube-controller-manager-operator-78b949d7b-fbwkz\" (UID: \"ffe33359-31f5-4a6c-93fc-6502d2516335\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fbwkz" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391291 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c66d8c7d-2de8-492f-ba5e-7ff0e236bf64-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-zkmjj\" (UID: \"c66d8c7d-2de8-492f-ba5e-7ff0e236bf64\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zkmjj" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391314 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ffe33359-31f5-4a6c-93fc-6502d2516335-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fbwkz\" (UID: \"ffe33359-31f5-4a6c-93fc-6502d2516335\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fbwkz" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391337 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmnrs\" (UniqueName: \"kubernetes.io/projected/dbd05529-6d54-416a-8df0-5973ee3179b6-kube-api-access-kmnrs\") pod \"service-ca-operator-777779d784-djmb4\" (UID: \"dbd05529-6d54-416a-8df0-5973ee3179b6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-djmb4" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391359 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/4bab7259-25d5-4c53-9ebb-ef2787adf010-node-bootstrap-token\") pod \"machine-config-server-58dhr\" (UID: \"4bab7259-25d5-4c53-9ebb-ef2787adf010\") " pod="openshift-machine-config-operator/machine-config-server-58dhr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391377 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/01d08621-494b-4232-b678-9caa94e61085-tmpfs\") pod \"packageserver-d55dfcdfc-cf6v7\" (UID: \"01d08621-494b-4232-b678-9caa94e61085\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cf6v7" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391408 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391430 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/79859b30-67ee-456b-82e5-f8806347a0b9-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-c6n8m\" (UID: \"79859b30-67ee-456b-82e5-f8806347a0b9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c6n8m" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391477 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r6t5\" (UniqueName: \"kubernetes.io/projected/79859b30-67ee-456b-82e5-f8806347a0b9-kube-api-access-9r6t5\") pod \"openshift-controller-manager-operator-756b6f6bc6-c6n8m\" (UID: \"79859b30-67ee-456b-82e5-f8806347a0b9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c6n8m" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391512 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcp9p\" (UniqueName: \"kubernetes.io/projected/85774248-1879-439e-9dd2-0d8661c299d6-kube-api-access-gcp9p\") pod \"machine-config-controller-84d6567774-v7l4k\" (UID: \"85774248-1879-439e-9dd2-0d8661c299d6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v7l4k" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391544 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llq59\" (UniqueName: \"kubernetes.io/projected/c6b54ae2-d365-4988-8e69-704574c7962a-kube-api-access-llq59\") pod \"collect-profiles-29492100-r2zj6\" (UID: \"c6b54ae2-d365-4988-8e69-704574c7962a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492100-r2zj6" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391565 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/794cdbb9-3392-465a-8a0a-a78a465aee2b-mountpoint-dir\") pod \"csi-hostpathplugin-hbbxh\" (UID: \"794cdbb9-3392-465a-8a0a-a78a465aee2b\") " pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391589 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/946946b3-f9e0-45e4-803f-edb3f7218489-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f9fqs\" (UID: \"946946b3-f9e0-45e4-803f-edb3f7218489\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f9fqs" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391609 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/794cdbb9-3392-465a-8a0a-a78a465aee2b-csi-data-dir\") pod \"csi-hostpathplugin-hbbxh\" (UID: \"794cdbb9-3392-465a-8a0a-a78a465aee2b\") " pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391633 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/db4a3858-5afa-44c8-a435-2010f7e7340d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5vmlj\" (UID: \"db4a3858-5afa-44c8-a435-2010f7e7340d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5vmlj" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391658 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvbbk\" (UniqueName: \"kubernetes.io/projected/1f03647c-a3e9-4099-9780-e79e3a4d4cf2-kube-api-access-xvbbk\") pod \"catalog-operator-68c6474976-mnltb\" (UID: \"1f03647c-a3e9-4099-9780-e79e3a4d4cf2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mnltb" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391679 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/55440e9e-5d99-4244-8b5c-55e2d270313b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-njx6w\" (UID: \"55440e9e-5d99-4244-8b5c-55e2d270313b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-njx6w" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391710 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/55440e9e-5d99-4244-8b5c-55e2d270313b-images\") pod \"machine-config-operator-74547568cd-njx6w\" (UID: \"55440e9e-5d99-4244-8b5c-55e2d270313b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-njx6w" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391733 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1f03647c-a3e9-4099-9780-e79e3a4d4cf2-srv-cert\") pod \"catalog-operator-68c6474976-mnltb\" (UID: \"1f03647c-a3e9-4099-9780-e79e3a4d4cf2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mnltb" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391754 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3-etcd-service-ca\") pod \"etcd-operator-b45778765-pwmhd\" (UID: \"c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pwmhd" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391775 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w88tz\" (UniqueName: \"kubernetes.io/projected/c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3-kube-api-access-w88tz\") pod \"etcd-operator-b45778765-pwmhd\" (UID: \"c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pwmhd" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391794 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/946946b3-f9e0-45e4-803f-edb3f7218489-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f9fqs\" (UID: \"946946b3-f9e0-45e4-803f-edb3f7218489\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f9fqs" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391815 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ffe33359-31f5-4a6c-93fc-6502d2516335-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fbwkz\" (UID: \"ffe33359-31f5-4a6c-93fc-6502d2516335\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fbwkz" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391838 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c8ebf890-c3b0-468e-bf7d-0ec590df084b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4lj2h\" (UID: \"c8ebf890-c3b0-468e-bf7d-0ec590df084b\") " pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391866 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/01d08621-494b-4232-b678-9caa94e61085-apiservice-cert\") pod \"packageserver-d55dfcdfc-cf6v7\" (UID: \"01d08621-494b-4232-b678-9caa94e61085\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cf6v7" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391897 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjbpj\" (UniqueName: \"kubernetes.io/projected/01d08621-494b-4232-b678-9caa94e61085-kube-api-access-mjbpj\") pod \"packageserver-d55dfcdfc-cf6v7\" (UID: \"01d08621-494b-4232-b678-9caa94e61085\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cf6v7" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391940 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3-config\") pod \"etcd-operator-b45778765-pwmhd\" (UID: \"c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pwmhd" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391960 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3-etcd-client\") pod \"etcd-operator-b45778765-pwmhd\" (UID: \"c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pwmhd" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.391983 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3745a211-9fa8-41a7-aa26-d733431bc9aa-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-24dmv\" (UID: \"3745a211-9fa8-41a7-aa26-d733431bc9aa\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-24dmv" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.392010 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c66d8c7d-2de8-492f-ba5e-7ff0e236bf64-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-zkmjj\" (UID: \"c66d8c7d-2de8-492f-ba5e-7ff0e236bf64\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zkmjj" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.392032 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvb9h\" (UniqueName: \"kubernetes.io/projected/c8ebf890-c3b0-468e-bf7d-0ec590df084b-kube-api-access-vvb9h\") pod \"marketplace-operator-79b997595-4lj2h\" (UID: \"c8ebf890-c3b0-468e-bf7d-0ec590df084b\") " pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.392066 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b3eefbc3-6dc4-479c-93e4-94a70fda0f83-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-cv2z7\" (UID: \"b3eefbc3-6dc4-479c-93e4-94a70fda0f83\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cv2z7" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.392088 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/85774248-1879-439e-9dd2-0d8661c299d6-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-v7l4k\" (UID: \"85774248-1879-439e-9dd2-0d8661c299d6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v7l4k" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.392112 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1f03647c-a3e9-4099-9780-e79e3a4d4cf2-profile-collector-cert\") pod \"catalog-operator-68c6474976-mnltb\" (UID: \"1f03647c-a3e9-4099-9780-e79e3a4d4cf2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mnltb" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.392132 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c8ebf890-c3b0-468e-bf7d-0ec590df084b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4lj2h\" (UID: \"c8ebf890-c3b0-468e-bf7d-0ec590df084b\") " pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.392157 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvzg8\" (UniqueName: \"kubernetes.io/projected/c66d8c7d-2de8-492f-ba5e-7ff0e236bf64-kube-api-access-jvzg8\") pod \"kube-storage-version-migrator-operator-b67b599dd-zkmjj\" (UID: \"c66d8c7d-2de8-492f-ba5e-7ff0e236bf64\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zkmjj" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.392343 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbd05529-6d54-416a-8df0-5973ee3179b6-config\") pod \"service-ca-operator-777779d784-djmb4\" (UID: \"dbd05529-6d54-416a-8df0-5973ee3179b6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-djmb4" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.392682 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/794cdbb9-3392-465a-8a0a-a78a465aee2b-registration-dir\") pod \"csi-hostpathplugin-hbbxh\" (UID: \"794cdbb9-3392-465a-8a0a-a78a465aee2b\") " pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.393007 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffe33359-31f5-4a6c-93fc-6502d2516335-config\") pod \"kube-controller-manager-operator-78b949d7b-fbwkz\" (UID: \"ffe33359-31f5-4a6c-93fc-6502d2516335\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fbwkz" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.393322 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79859b30-67ee-456b-82e5-f8806347a0b9-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-c6n8m\" (UID: \"79859b30-67ee-456b-82e5-f8806347a0b9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c6n8m" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.393625 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c66d8c7d-2de8-492f-ba5e-7ff0e236bf64-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-zkmjj\" (UID: \"c66d8c7d-2de8-492f-ba5e-7ff0e236bf64\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zkmjj" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.393964 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/794cdbb9-3392-465a-8a0a-a78a465aee2b-mountpoint-dir\") pod \"csi-hostpathplugin-hbbxh\" (UID: \"794cdbb9-3392-465a-8a0a-a78a465aee2b\") " pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.394094 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/794cdbb9-3392-465a-8a0a-a78a465aee2b-csi-data-dir\") pod \"csi-hostpathplugin-hbbxh\" (UID: \"794cdbb9-3392-465a-8a0a-a78a465aee2b\") " pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.394421 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/4bab7259-25d5-4c53-9ebb-ef2787adf010-certs\") pod \"machine-config-server-58dhr\" (UID: \"4bab7259-25d5-4c53-9ebb-ef2787adf010\") " pod="openshift-machine-config-operator/machine-config-server-58dhr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.395041 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6b54ae2-d365-4988-8e69-704574c7962a-config-volume\") pod \"collect-profiles-29492100-r2zj6\" (UID: \"c6b54ae2-d365-4988-8e69-704574c7962a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492100-r2zj6" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.395112 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/794cdbb9-3392-465a-8a0a-a78a465aee2b-socket-dir\") pod \"csi-hostpathplugin-hbbxh\" (UID: \"794cdbb9-3392-465a-8a0a-a78a465aee2b\") " pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.396519 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f469e02c-9404-422a-bff0-1b945d9c8768-cert\") pod \"ingress-canary-fw6bh\" (UID: \"f469e02c-9404-422a-bff0-1b945d9c8768\") " pod="openshift-ingress-canary/ingress-canary-fw6bh" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.396658 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/85774248-1879-439e-9dd2-0d8661c299d6-proxy-tls\") pod \"machine-config-controller-84d6567774-v7l4k\" (UID: \"85774248-1879-439e-9dd2-0d8661c299d6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v7l4k" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.397430 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/db4a3858-5afa-44c8-a435-2010f7e7340d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5vmlj\" (UID: \"db4a3858-5afa-44c8-a435-2010f7e7340d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5vmlj" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.397548 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/55440e9e-5d99-4244-8b5c-55e2d270313b-proxy-tls\") pod \"machine-config-operator-74547568cd-njx6w\" (UID: \"55440e9e-5d99-4244-8b5c-55e2d270313b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-njx6w" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.397893 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/855486b0-11f8-4ff0-930d-75c7e9d790d3-signing-cabundle\") pod \"service-ca-9c57cc56f-9wv77\" (UID: \"855486b0-11f8-4ff0-930d-75c7e9d790d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-9wv77" Jan 27 15:09:16 crc kubenswrapper[4772]: E0127 15:09:16.397976 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:16.897962754 +0000 UTC m=+142.878571852 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.398587 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/01d08621-494b-4232-b678-9caa94e61085-tmpfs\") pod \"packageserver-d55dfcdfc-cf6v7\" (UID: \"01d08621-494b-4232-b678-9caa94e61085\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cf6v7" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.399132 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/55440e9e-5d99-4244-8b5c-55e2d270313b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-njx6w\" (UID: \"55440e9e-5d99-4244-8b5c-55e2d270313b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-njx6w" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.399162 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5f6v\" (UniqueName: \"kubernetes.io/projected/06cdc094-b372-4016-bc5e-4c15a28e032e-kube-api-access-l5f6v\") pod \"oauth-openshift-558db77b4-fgw98\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.399225 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/794cdbb9-3392-465a-8a0a-a78a465aee2b-plugins-dir\") pod \"csi-hostpathplugin-hbbxh\" (UID: \"794cdbb9-3392-465a-8a0a-a78a465aee2b\") " pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.399405 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/855486b0-11f8-4ff0-930d-75c7e9d790d3-signing-key\") pod \"service-ca-9c57cc56f-9wv77\" (UID: \"855486b0-11f8-4ff0-930d-75c7e9d790d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-9wv77" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.399602 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/01d08621-494b-4232-b678-9caa94e61085-apiservice-cert\") pod \"packageserver-d55dfcdfc-cf6v7\" (UID: \"01d08621-494b-4232-b678-9caa94e61085\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cf6v7" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.399641 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/55440e9e-5d99-4244-8b5c-55e2d270313b-images\") pod \"machine-config-operator-74547568cd-njx6w\" (UID: \"55440e9e-5d99-4244-8b5c-55e2d270313b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-njx6w" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.399830 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/85774248-1879-439e-9dd2-0d8661c299d6-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-v7l4k\" (UID: \"85774248-1879-439e-9dd2-0d8661c299d6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v7l4k" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.400303 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/79859b30-67ee-456b-82e5-f8806347a0b9-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-c6n8m\" (UID: \"79859b30-67ee-456b-82e5-f8806347a0b9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c6n8m" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.400961 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/01d08621-494b-4232-b678-9caa94e61085-webhook-cert\") pod \"packageserver-d55dfcdfc-cf6v7\" (UID: \"01d08621-494b-4232-b678-9caa94e61085\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cf6v7" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.401299 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b3eefbc3-6dc4-479c-93e4-94a70fda0f83-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-cv2z7\" (UID: \"b3eefbc3-6dc4-479c-93e4-94a70fda0f83\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cv2z7" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.401399 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1f03647c-a3e9-4099-9780-e79e3a4d4cf2-srv-cert\") pod \"catalog-operator-68c6474976-mnltb\" (UID: \"1f03647c-a3e9-4099-9780-e79e3a4d4cf2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mnltb" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.401794 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c8ebf890-c3b0-468e-bf7d-0ec590df084b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4lj2h\" (UID: \"c8ebf890-c3b0-468e-bf7d-0ec590df084b\") " pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.402567 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/01d2c3f9-778c-4cf0-b8a4-76583f62df3c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-jdcpn\" (UID: \"01d2c3f9-778c-4cf0-b8a4-76583f62df3c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jdcpn" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.402647 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dbd05529-6d54-416a-8df0-5973ee3179b6-serving-cert\") pod \"service-ca-operator-777779d784-djmb4\" (UID: \"dbd05529-6d54-416a-8df0-5973ee3179b6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-djmb4" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.403055 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3-etcd-client\") pod \"etcd-operator-b45778765-pwmhd\" (UID: \"c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pwmhd" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.403220 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3-etcd-service-ca\") pod \"etcd-operator-b45778765-pwmhd\" (UID: \"c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pwmhd" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.403672 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3-config\") pod \"etcd-operator-b45778765-pwmhd\" (UID: \"c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pwmhd" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.404030 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/946946b3-f9e0-45e4-803f-edb3f7218489-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f9fqs\" (UID: \"946946b3-f9e0-45e4-803f-edb3f7218489\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f9fqs" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.404117 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/01d2c3f9-778c-4cf0-b8a4-76583f62df3c-srv-cert\") pod \"olm-operator-6b444d44fb-jdcpn\" (UID: \"01d2c3f9-778c-4cf0-b8a4-76583f62df3c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jdcpn" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.404325 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3-etcd-ca\") pod \"etcd-operator-b45778765-pwmhd\" (UID: \"c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pwmhd" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.404894 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1f03647c-a3e9-4099-9780-e79e3a4d4cf2-profile-collector-cert\") pod \"catalog-operator-68c6474976-mnltb\" (UID: \"1f03647c-a3e9-4099-9780-e79e3a4d4cf2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mnltb" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.405124 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3745a211-9fa8-41a7-aa26-d733431bc9aa-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-24dmv\" (UID: \"3745a211-9fa8-41a7-aa26-d733431bc9aa\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-24dmv" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.405361 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3-serving-cert\") pod \"etcd-operator-b45778765-pwmhd\" (UID: \"c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pwmhd" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.405909 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6b54ae2-d365-4988-8e69-704574c7962a-secret-volume\") pod \"collect-profiles-29492100-r2zj6\" (UID: \"c6b54ae2-d365-4988-8e69-704574c7962a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492100-r2zj6" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.406412 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c66d8c7d-2de8-492f-ba5e-7ff0e236bf64-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-zkmjj\" (UID: \"c66d8c7d-2de8-492f-ba5e-7ff0e236bf64\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zkmjj" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.406458 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c8ebf890-c3b0-468e-bf7d-0ec590df084b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4lj2h\" (UID: \"c8ebf890-c3b0-468e-bf7d-0ec590df084b\") " pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.407043 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ffe33359-31f5-4a6c-93fc-6502d2516335-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fbwkz\" (UID: \"ffe33359-31f5-4a6c-93fc-6502d2516335\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fbwkz" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.408322 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/4bab7259-25d5-4c53-9ebb-ef2787adf010-node-bootstrap-token\") pod \"machine-config-server-58dhr\" (UID: \"4bab7259-25d5-4c53-9ebb-ef2787adf010\") " pod="openshift-machine-config-operator/machine-config-server-58dhr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.408568 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/946946b3-f9e0-45e4-803f-edb3f7218489-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f9fqs\" (UID: \"946946b3-f9e0-45e4-803f-edb3f7218489\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f9fqs" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.422512 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5z5pj\" (UniqueName: \"kubernetes.io/projected/c4225ddc-bdcd-4158-811b-113234d0c3d0-kube-api-access-5z5pj\") pod \"router-default-5444994796-7k7sg\" (UID: \"c4225ddc-bdcd-4158-811b-113234d0c3d0\") " pod="openshift-ingress/router-default-5444994796-7k7sg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.438664 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f9ab8468-b920-41e9-a5f2-1af70f1b5ffd-bound-sa-token\") pod \"ingress-operator-5b745b69d9-w4j4h\" (UID: \"f9ab8468-b920-41e9-a5f2-1af70f1b5ffd\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4j4h" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.444116 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bck4j" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.472262 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg5jv\" (UniqueName: \"kubernetes.io/projected/3150273d-63f7-4908-bcc5-2403e123d1e7-kube-api-access-dg5jv\") pod \"dns-default-jplbk\" (UID: \"3150273d-63f7-4908-bcc5-2403e123d1e7\") " pod="openshift-dns/dns-default-jplbk" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.480366 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j667k\" (UniqueName: \"kubernetes.io/projected/5276546c-f731-4bd0-bb93-b5cd19b0992c-kube-api-access-j667k\") pod \"console-operator-58897d9998-npths\" (UID: \"5276546c-f731-4bd0-bb93-b5cd19b0992c\") " pod="openshift-console-operator/console-operator-58897d9998-npths" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.492985 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:16 crc kubenswrapper[4772]: E0127 15:09:16.493465 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:16.993450821 +0000 UTC m=+142.974059919 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.500827 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mswpx\" (UniqueName: \"kubernetes.io/projected/2c9eef7e-3996-45b6-ab7b-50d319dc1117-kube-api-access-mswpx\") pod \"cluster-image-registry-operator-dc59b4c8b-4v228\" (UID: \"2c9eef7e-3996-45b6-ab7b-50d319dc1117\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4v228" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.518545 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-npths" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.524154 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmgkc\" (UniqueName: \"kubernetes.io/projected/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-kube-api-access-vmgkc\") pod \"console-f9d7485db-7qfrl\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.531482 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzkn2"] Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.539550 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.543444 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2zfh\" (UniqueName: \"kubernetes.io/projected/ce4b4e2e-496b-4334-8736-db4f25473731-kube-api-access-k2zfh\") pod \"dns-operator-744455d44c-bhgv8\" (UID: \"ce4b4e2e-496b-4334-8736-db4f25473731\") " pod="openshift-dns-operator/dns-operator-744455d44c-bhgv8" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.561894 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2c9eef7e-3996-45b6-ab7b-50d319dc1117-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-4v228\" (UID: \"2c9eef7e-3996-45b6-ab7b-50d319dc1117\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4v228" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.576487 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4v228" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.581106 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s88q\" (UniqueName: \"kubernetes.io/projected/192d0f8f-10f9-43e2-a24a-2019aae0db44-kube-api-access-7s88q\") pod \"openshift-apiserver-operator-796bbdcf4f-zq27x\" (UID: \"192d0f8f-10f9-43e2-a24a-2019aae0db44\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zq27x" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.589660 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-vswtw" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.594963 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: E0127 15:09:16.595376 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:17.095359944 +0000 UTC m=+143.075969052 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.610763 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h66t\" (UniqueName: \"kubernetes.io/projected/877de785-bc18-4c1c-970a-1e6533539467-kube-api-access-5h66t\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.613493 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.619292 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4j4h" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.627199 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-jplbk" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.630441 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/877de785-bc18-4c1c-970a-1e6533539467-bound-sa-token\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.636014 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzkn2" event={"ID":"dd05c164-83f7-4ebe-bbe8-9db6707741c5","Type":"ContainerStarted","Data":"68665c93b188501aa2d7df5f5af86d404704d1403f6fe10a4093685c7a6ee1cf"} Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.638552 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-bhgv8" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.639580 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-mfh29" event={"ID":"625f7e2d-0e3f-4c2c-8f49-b09fc3638536","Type":"ContainerStarted","Data":"b28cd8c3fe65bb44ff5ed3705cd139aedd1d6522b0d06267a643bc732701558f"} Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.641480 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-tgmck" event={"ID":"39dd090e-b988-4c36-88f0-c0cb28a23e8b","Type":"ContainerStarted","Data":"257d4dbd74cf09cb130ad2e955eeb1266c017333a47677c8b512b2f5edde9667"} Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.641505 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-tgmck" event={"ID":"39dd090e-b988-4c36-88f0-c0cb28a23e8b","Type":"ContainerStarted","Data":"0a778ca4bceec9fd16b9c8f9ec95ee3fbacfad7d71ea4a992fe5cb953899bd2d"} Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.647057 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-7k7sg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.649340 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twqnj\" (UniqueName: \"kubernetes.io/projected/06755863-1a8b-4f4d-a304-03bfd45725ec-kube-api-access-twqnj\") pod \"cluster-samples-operator-665b6dd947-g667l\" (UID: \"06755863-1a8b-4f4d-a304-03bfd45725ec\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g667l" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.654431 4772 generic.go:334] "Generic (PLEG): container finished" podID="34e7a553-e424-472e-a143-76e7e08e57aa" containerID="d15bf5cf293ed09673e38bdec1e04909a7c8924975f8f59b2ddf45cd1d0c265e" exitCode=0 Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.654579 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" event={"ID":"34e7a553-e424-472e-a143-76e7e08e57aa","Type":"ContainerDied","Data":"d15bf5cf293ed09673e38bdec1e04909a7c8924975f8f59b2ddf45cd1d0c265e"} Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.664861 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lbs2\" (UniqueName: \"kubernetes.io/projected/8800c113-7b51-4554-8e52-c1d0df1a08be-kube-api-access-4lbs2\") pod \"machine-approver-56656f9798-klfsg\" (UID: \"8800c113-7b51-4554-8e52-c1d0df1a08be\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfsg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.696401 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:16 crc kubenswrapper[4772]: E0127 15:09:16.697016 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:17.196980169 +0000 UTC m=+143.177589267 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.697256 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.697274 4772 generic.go:334] "Generic (PLEG): container finished" podID="d1aba7eb-5916-4023-90f2-10152ad89b63" containerID="5a300fb78fa1e933a3df435d7e3452e503aa76c100098936d48e100eae063049" exitCode=0 Jan 27 15:09:16 crc kubenswrapper[4772]: E0127 15:09:16.697626 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:17.197614567 +0000 UTC m=+143.178223665 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.700348 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" event={"ID":"d1aba7eb-5916-4023-90f2-10152ad89b63","Type":"ContainerDied","Data":"5a300fb78fa1e933a3df435d7e3452e503aa76c100098936d48e100eae063049"} Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.700381 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-bck4j"] Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.711732 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.713454 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.720025 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72xbw\" (UniqueName: \"kubernetes.io/projected/b3eefbc3-6dc4-479c-93e4-94a70fda0f83-kube-api-access-72xbw\") pod \"package-server-manager-789f6589d5-cv2z7\" (UID: \"b3eefbc3-6dc4-479c-93e4-94a70fda0f83\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cv2z7" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.737595 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrbkm\" (UniqueName: \"kubernetes.io/projected/4bab7259-25d5-4c53-9ebb-ef2787adf010-kube-api-access-vrbkm\") pod \"machine-config-server-58dhr\" (UID: \"4bab7259-25d5-4c53-9ebb-ef2787adf010\") " pod="openshift-machine-config-operator/machine-config-server-58dhr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.737845 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g667l" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.746840 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvzg8\" (UniqueName: \"kubernetes.io/projected/c66d8c7d-2de8-492f-ba5e-7ff0e236bf64-kube-api-access-jvzg8\") pod \"kube-storage-version-migrator-operator-b67b599dd-zkmjj\" (UID: \"c66d8c7d-2de8-492f-ba5e-7ff0e236bf64\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zkmjj" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.752975 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zq27x" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.761627 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zkmjj" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.763274 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5cbs\" (UniqueName: \"kubernetes.io/projected/855486b0-11f8-4ff0-930d-75c7e9d790d3-kube-api-access-m5cbs\") pod \"service-ca-9c57cc56f-9wv77\" (UID: \"855486b0-11f8-4ff0-930d-75c7e9d790d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-9wv77" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.776572 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-9wv77" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.785482 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9v6df\" (UniqueName: \"kubernetes.io/projected/db4a3858-5afa-44c8-a435-2010f7e7340d-kube-api-access-9v6df\") pod \"control-plane-machine-set-operator-78cbb6b69f-5vmlj\" (UID: \"db4a3858-5afa-44c8-a435-2010f7e7340d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5vmlj" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.797964 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:16 crc kubenswrapper[4772]: E0127 15:09:16.799154 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:17.299133809 +0000 UTC m=+143.279742907 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.822099 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-58dhr" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.822752 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r6t5\" (UniqueName: \"kubernetes.io/projected/79859b30-67ee-456b-82e5-f8806347a0b9-kube-api-access-9r6t5\") pod \"openshift-controller-manager-operator-756b6f6bc6-c6n8m\" (UID: \"79859b30-67ee-456b-82e5-f8806347a0b9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c6n8m" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.833730 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxscw\" (UniqueName: \"kubernetes.io/projected/3745a211-9fa8-41a7-aa26-d733431bc9aa-kube-api-access-sxscw\") pod \"multus-admission-controller-857f4d67dd-24dmv\" (UID: \"3745a211-9fa8-41a7-aa26-d733431bc9aa\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-24dmv" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.864574 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-npths"] Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.865559 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfsg" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.892028 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcp9p\" (UniqueName: \"kubernetes.io/projected/85774248-1879-439e-9dd2-0d8661c299d6-kube-api-access-gcp9p\") pod \"machine-config-controller-84d6567774-v7l4k\" (UID: \"85774248-1879-439e-9dd2-0d8661c299d6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v7l4k" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.897128 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llq59\" (UniqueName: \"kubernetes.io/projected/c6b54ae2-d365-4988-8e69-704574c7962a-kube-api-access-llq59\") pod \"collect-profiles-29492100-r2zj6\" (UID: \"c6b54ae2-d365-4988-8e69-704574c7962a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492100-r2zj6" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.899725 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/946946b3-f9e0-45e4-803f-edb3f7218489-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f9fqs\" (UID: \"946946b3-f9e0-45e4-803f-edb3f7218489\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f9fqs" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.900300 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:16 crc kubenswrapper[4772]: E0127 15:09:16.900867 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:17.400853066 +0000 UTC m=+143.381462164 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.919206 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s25qm\" (UniqueName: \"kubernetes.io/projected/59cdf584-81d0-4d66-8fc2-da3a3f995f73-kube-api-access-s25qm\") pod \"migrator-59844c95c7-xqtff\" (UID: \"59cdf584-81d0-4d66-8fc2-da3a3f995f73\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xqtff" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.937248 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-7qfrl"] Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.954445 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg5gf\" (UniqueName: \"kubernetes.io/projected/f469e02c-9404-422a-bff0-1b945d9c8768-kube-api-access-xg5gf\") pod \"ingress-canary-fw6bh\" (UID: \"f469e02c-9404-422a-bff0-1b945d9c8768\") " pod="openshift-ingress-canary/ingress-canary-fw6bh" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.968002 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.977621 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cv2z7" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.981933 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2scxc\" (UniqueName: \"kubernetes.io/projected/794cdbb9-3392-465a-8a0a-a78a465aee2b-kube-api-access-2scxc\") pod \"csi-hostpathplugin-hbbxh\" (UID: \"794cdbb9-3392-465a-8a0a-a78a465aee2b\") " pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.985496 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c6n8m" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.988354 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8vjd\" (UniqueName: \"kubernetes.io/projected/55440e9e-5d99-4244-8b5c-55e2d270313b-kube-api-access-s8vjd\") pod \"machine-config-operator-74547568cd-njx6w\" (UID: \"55440e9e-5d99-4244-8b5c-55e2d270313b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-njx6w" Jan 27 15:09:16 crc kubenswrapper[4772]: I0127 15:09:16.988579 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttqp4\" (UniqueName: \"kubernetes.io/projected/01d2c3f9-778c-4cf0-b8a4-76583f62df3c-kube-api-access-ttqp4\") pod \"olm-operator-6b444d44fb-jdcpn\" (UID: \"01d2c3f9-778c-4cf0-b8a4-76583f62df3c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jdcpn" Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.001482 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:17 crc kubenswrapper[4772]: E0127 15:09:17.001853 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:17.501837623 +0000 UTC m=+143.482446731 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.001996 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f9fqs" Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.010218 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvbbk\" (UniqueName: \"kubernetes.io/projected/1f03647c-a3e9-4099-9780-e79e3a4d4cf2-kube-api-access-xvbbk\") pod \"catalog-operator-68c6474976-mnltb\" (UID: \"1f03647c-a3e9-4099-9780-e79e3a4d4cf2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mnltb" Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.021376 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvb9h\" (UniqueName: \"kubernetes.io/projected/c8ebf890-c3b0-468e-bf7d-0ec590df084b-kube-api-access-vvb9h\") pod \"marketplace-operator-79b997595-4lj2h\" (UID: \"c8ebf890-c3b0-468e-bf7d-0ec590df084b\") " pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.027764 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492100-r2zj6" Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.039522 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-fw6bh" Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.046270 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xqtff" Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.060665 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jdcpn" Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.067148 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjbpj\" (UniqueName: \"kubernetes.io/projected/01d08621-494b-4232-b678-9caa94e61085-kube-api-access-mjbpj\") pod \"packageserver-d55dfcdfc-cf6v7\" (UID: \"01d08621-494b-4232-b678-9caa94e61085\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cf6v7" Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.068537 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5vmlj" Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.063113 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w88tz\" (UniqueName: \"kubernetes.io/projected/c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3-kube-api-access-w88tz\") pod \"etcd-operator-b45778765-pwmhd\" (UID: \"c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pwmhd" Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.089216 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-24dmv" Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.092295 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-njx6w" Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.097221 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ffe33359-31f5-4a6c-93fc-6502d2516335-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fbwkz\" (UID: \"ffe33359-31f5-4a6c-93fc-6502d2516335\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fbwkz" Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.102807 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cf6v7" Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.104092 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:17 crc kubenswrapper[4772]: E0127 15:09:17.104644 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:17.604634351 +0000 UTC m=+143.585243449 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.114319 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mnltb" Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.114671 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v7l4k" Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.115255 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmnrs\" (UniqueName: \"kubernetes.io/projected/dbd05529-6d54-416a-8df0-5973ee3179b6-kube-api-access-kmnrs\") pod \"service-ca-operator-777779d784-djmb4\" (UID: \"dbd05529-6d54-416a-8df0-5973ee3179b6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-djmb4" Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.144378 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.191018 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-vswtw"] Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.204758 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:17 crc kubenswrapper[4772]: E0127 15:09:17.205071 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:17.705056981 +0000 UTC m=+143.685666079 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.267403 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-pwmhd" Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.296329 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fbwkz" Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.310937 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-djmb4" Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.311846 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:17 crc kubenswrapper[4772]: E0127 15:09:17.312207 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:17.812192595 +0000 UTC m=+143.792801693 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.317644 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.387108 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-bhgv8"] Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.411136 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fgw98"] Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.413203 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4v228"] Jan 27 15:09:17 crc kubenswrapper[4772]: E0127 15:09:17.413854 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:17.91383787 +0000 UTC m=+143.894446968 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.413684 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.414340 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:17 crc kubenswrapper[4772]: E0127 15:09:17.414620 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:17.914613233 +0000 UTC m=+143.895222331 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.435900 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-w4j4h"] Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.476859 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-jplbk"] Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.497365 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zq27x"] Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.517001 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:17 crc kubenswrapper[4772]: E0127 15:09:17.517564 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:18.017544655 +0000 UTC m=+143.998153753 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.629800 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:17 crc kubenswrapper[4772]: E0127 15:09:17.631822 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:18.131791204 +0000 UTC m=+144.112400292 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.632285 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9wv77"] Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.633288 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g667l"] Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.711677 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-bhgv8" event={"ID":"ce4b4e2e-496b-4334-8736-db4f25473731","Type":"ContainerStarted","Data":"591f6838813fcb88aeef0cd6093d0e0bcd3252c2d5c3d310e4d6567aed6039c2"} Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.714051 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" event={"ID":"06cdc094-b372-4016-bc5e-4c15a28e032e","Type":"ContainerStarted","Data":"3982e848b6f4ab4d0d2958e425dd1a480bb7b8b136363856076bf9ce68e097fb"} Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.722384 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-58dhr" event={"ID":"4bab7259-25d5-4c53-9ebb-ef2787adf010","Type":"ContainerStarted","Data":"e98f41d2510b33d0698dbd7ae4d039343cceadb5ae237cf925b7e7a8d5435dc6"} Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.737312 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.738874 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-7k7sg" event={"ID":"c4225ddc-bdcd-4158-811b-113234d0c3d0","Type":"ContainerStarted","Data":"25cd0105e9828132acd6b8f37dcf02f4c7acf6955ff8d08ef76b9c80aa284d6e"} Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.738923 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-7k7sg" event={"ID":"c4225ddc-bdcd-4158-811b-113234d0c3d0","Type":"ContainerStarted","Data":"ea4583bad816e9577db96aec74ce06766b98235a7e5195bb6afc2a37cf197924"} Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.744333 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4v228" event={"ID":"2c9eef7e-3996-45b6-ab7b-50d319dc1117","Type":"ContainerStarted","Data":"0771b389042915c69da1c41aa8f21b3569cbce61ac4e69d6456d7e18da8f069f"} Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.745377 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-npths" event={"ID":"5276546c-f731-4bd0-bb93-b5cd19b0992c","Type":"ContainerStarted","Data":"2bbe0cbce3f8ea6011ebf6dcbfd30782c40d8ce69552ccff0de447ef0b35edb8"} Jan 27 15:09:17 crc kubenswrapper[4772]: E0127 15:09:17.745563 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:18.245533799 +0000 UTC m=+144.226142907 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.746055 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-vswtw" event={"ID":"17bdd07d-f7e5-47f8-b730-724d5cc8e3d2","Type":"ContainerStarted","Data":"01a9119414207291aef3e1db49c29878d281b57cf66ccb2596ca13d4b5131174"} Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.748009 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7qfrl" event={"ID":"e2e31e5f-3a41-42f5-90b0-99c05a8033a6","Type":"ContainerStarted","Data":"689105dc82b6dcc122fad60678c44aee714f4e2b250e67f0c76903dd34d0b5c3"} Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.748700 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfsg" event={"ID":"8800c113-7b51-4554-8e52-c1d0df1a08be","Type":"ContainerStarted","Data":"5da0fe9c33333bb873c2a165867fd7f8badf46f60b030c054559d9f8586f985a"} Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.749674 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zq27x" event={"ID":"192d0f8f-10f9-43e2-a24a-2019aae0db44","Type":"ContainerStarted","Data":"f67ee9130f52502f0ec03d9b0c5b1493ac32c1219f9e65c73c2fc6b7e0fc3336"} Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.768235 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" event={"ID":"34e7a553-e424-472e-a143-76e7e08e57aa","Type":"ContainerStarted","Data":"2cae3b4aa67dc5e6a63a0bb18d2ab08e5b6efd969347371e8159f04131f1c67c"} Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.780203 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzkn2" event={"ID":"dd05c164-83f7-4ebe-bbe8-9db6707741c5","Type":"ContainerStarted","Data":"0093a9aed6c8a4572b3b2643fc6ead21c9a27ea9f74c29c40bfc2795ebe6034e"} Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.795305 4772 generic.go:334] "Generic (PLEG): container finished" podID="ee07e18c-9f40-41c3-b2fb-05fd325976e4" containerID="b7ce8aeca0ab242b2614d3530a13b679f4b6cbfc2d58f53b6ac82ac0eae6ad31" exitCode=0 Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.798375 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bck4j" event={"ID":"ee07e18c-9f40-41c3-b2fb-05fd325976e4","Type":"ContainerDied","Data":"b7ce8aeca0ab242b2614d3530a13b679f4b6cbfc2d58f53b6ac82ac0eae6ad31"} Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.798410 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bck4j" event={"ID":"ee07e18c-9f40-41c3-b2fb-05fd325976e4","Type":"ContainerStarted","Data":"401d72ccb1e730289431212d99b14989550a7ab785fccf9bb2d9a54765a06a2d"} Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.842945 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:17 crc kubenswrapper[4772]: E0127 15:09:17.843330 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:18.343314463 +0000 UTC m=+144.323923561 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.924485 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-mfh29" podStartSLOduration=120.924469446 podStartE2EDuration="2m0.924469446s" podCreationTimestamp="2026-01-27 15:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:17.924299351 +0000 UTC m=+143.904908449" watchObservedRunningTime="2026-01-27 15:09:17.924469446 +0000 UTC m=+143.905078544" Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.946013 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:17 crc kubenswrapper[4772]: E0127 15:09:17.946288 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:18.446231425 +0000 UTC m=+144.426840533 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:17 crc kubenswrapper[4772]: I0127 15:09:17.946922 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:17 crc kubenswrapper[4772]: E0127 15:09:17.953711 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:18.45369719 +0000 UTC m=+144.434306288 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.047906 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:18 crc kubenswrapper[4772]: E0127 15:09:18.048289 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:18.548273981 +0000 UTC m=+144.528883079 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.148973 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:18 crc kubenswrapper[4772]: E0127 15:09:18.149603 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:18.649589957 +0000 UTC m=+144.630199055 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.244027 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zkmjj"] Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.247232 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c6n8m"] Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.252977 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:18 crc kubenswrapper[4772]: E0127 15:09:18.253282 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:18.753264071 +0000 UTC m=+144.733873169 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.306508 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-v7l4k"] Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.361449 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cv2z7"] Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.362449 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:18 crc kubenswrapper[4772]: E0127 15:09:18.363532 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:18.863520514 +0000 UTC m=+144.844129612 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.433730 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-tgmck" podStartSLOduration=122.433709101 podStartE2EDuration="2m2.433709101s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:18.420047026 +0000 UTC m=+144.400656124" watchObservedRunningTime="2026-01-27 15:09:18.433709101 +0000 UTC m=+144.414318199" Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.465108 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:18 crc kubenswrapper[4772]: E0127 15:09:18.465257 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:18.965235312 +0000 UTC m=+144.945844410 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.465417 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:18 crc kubenswrapper[4772]: E0127 15:09:18.465741 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:18.965726106 +0000 UTC m=+144.946335204 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.505938 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" podStartSLOduration=121.505923997 podStartE2EDuration="2m1.505923997s" podCreationTimestamp="2026-01-27 15:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:18.505463453 +0000 UTC m=+144.486072551" watchObservedRunningTime="2026-01-27 15:09:18.505923997 +0000 UTC m=+144.486533095" Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.570702 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:18 crc kubenswrapper[4772]: E0127 15:09:18.571084 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:19.071068898 +0000 UTC m=+145.051677996 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.615498 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" podStartSLOduration=122.615480881 podStartE2EDuration="2m2.615480881s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:18.613609246 +0000 UTC m=+144.594218364" watchObservedRunningTime="2026-01-27 15:09:18.615480881 +0000 UTC m=+144.596089979" Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.651215 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-7k7sg" Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.671737 4772 patch_prober.go:28] interesting pod/router-default-5444994796-7k7sg container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.671798 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7k7sg" podUID="c4225ddc-bdcd-4158-811b-113234d0c3d0" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.673149 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:18 crc kubenswrapper[4772]: E0127 15:09:18.673508 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:19.173494126 +0000 UTC m=+145.154103224 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.782415 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:18 crc kubenswrapper[4772]: E0127 15:09:18.782872 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:19.282857724 +0000 UTC m=+145.263466832 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.895484 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:18 crc kubenswrapper[4772]: E0127 15:09:18.895925 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:19.395910739 +0000 UTC m=+145.376519837 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.945909 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cv2z7" event={"ID":"b3eefbc3-6dc4-479c-93e4-94a70fda0f83","Type":"ContainerStarted","Data":"6b86debb441021b003508513ead164bde4f10f62cc65635db9f5f4a1885e9223"} Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.946336 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f9fqs"] Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.946353 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v7l4k" event={"ID":"85774248-1879-439e-9dd2-0d8661c299d6","Type":"ContainerStarted","Data":"97aa3ed61e762c362dda1f72b62105466808bca994d182bf11015ff53163c53a"} Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.946365 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zkmjj" event={"ID":"c66d8c7d-2de8-492f-ba5e-7ff0e236bf64","Type":"ContainerStarted","Data":"ef588de4253172ca4c9ff7d80a68abed8f0bb2c9628e6b73579aaf45a29f1da7"} Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.946374 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4j4h" event={"ID":"f9ab8468-b920-41e9-a5f2-1af70f1b5ffd","Type":"ContainerStarted","Data":"e585c3d7dc35ef0921dfcf5d7a69da5c2136ec4e226763f0534e80b6e7fc26a1"} Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.946385 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-fw6bh"] Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.946396 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4lj2h"] Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.946409 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-58dhr" event={"ID":"4bab7259-25d5-4c53-9ebb-ef2787adf010","Type":"ContainerStarted","Data":"6686f6e3b763d8ff3782aa8904c0077068565aa5e1027885a9ec0f54ceffe277"} Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.946422 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-9wv77" event={"ID":"855486b0-11f8-4ff0-930d-75c7e9d790d3","Type":"ContainerStarted","Data":"2716f89b59875013e93f2fbf23b4377a3b3341f577d15221a25e0df7535fd5b6"} Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.946434 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fbwkz"] Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.946445 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g667l" event={"ID":"06755863-1a8b-4f4d-a304-03bfd45725ec","Type":"ContainerStarted","Data":"27109b40b600a3fae2adf6265056a955275795e1927236d4e8a26d80dedb3184"} Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.946457 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7qfrl" event={"ID":"e2e31e5f-3a41-42f5-90b0-99c05a8033a6","Type":"ContainerStarted","Data":"5fa14544d0f474c1dab5359f3cbcee7247c22e26d183454cd30d9fa3ab064652"} Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.946469 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-jplbk" event={"ID":"3150273d-63f7-4908-bcc5-2403e123d1e7","Type":"ContainerStarted","Data":"7f1ee6e3a83ddf43c8ad216851c662b4e9a3878b1d55de426bd25cf10e59ddb7"} Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.946483 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mnltb"] Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.952623 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jdcpn"] Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.964480 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c6n8m" event={"ID":"79859b30-67ee-456b-82e5-f8806347a0b9","Type":"ContainerStarted","Data":"e6839e71f42bf9813140ab9409d87011aad5df15e0b92baa26f0ae3dbd0708ac"} Jan 27 15:09:18 crc kubenswrapper[4772]: I0127 15:09:18.992710 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" event={"ID":"d1aba7eb-5916-4023-90f2-10152ad89b63","Type":"ContainerStarted","Data":"1cc65229ba685a5832f280c3c5752f78d182cfcc1da6b7cfa03ee16063a744c8"} Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.007246 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:19 crc kubenswrapper[4772]: E0127 15:09:19.028712 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:19.528691963 +0000 UTC m=+145.509301061 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.029798 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-pwmhd"] Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.054220 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-hbbxh"] Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.082398 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-24dmv"] Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.098458 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5vmlj"] Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.104216 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-7qfrl" podStartSLOduration=123.104194734 podStartE2EDuration="2m3.104194734s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:19.073973011 +0000 UTC m=+145.054582109" watchObservedRunningTime="2026-01-27 15:09:19.104194734 +0000 UTC m=+145.084803832" Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.108485 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:19 crc kubenswrapper[4772]: E0127 15:09:19.108962 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:19.608942171 +0000 UTC m=+145.589551309 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:19 crc kubenswrapper[4772]: W0127 15:09:19.119531 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb4a3858_5afa_44c8_a435_2010f7e7340d.slice/crio-ae81c92d45d0547ff142b72b22b85da95795638755056827220d1a9427feb720 WatchSource:0}: Error finding container ae81c92d45d0547ff142b72b22b85da95795638755056827220d1a9427feb720: Status 404 returned error can't find the container with id ae81c92d45d0547ff142b72b22b85da95795638755056827220d1a9427feb720 Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.120773 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-7k7sg" podStartSLOduration=123.120748842 podStartE2EDuration="2m3.120748842s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:19.106252673 +0000 UTC m=+145.086861781" watchObservedRunningTime="2026-01-27 15:09:19.120748842 +0000 UTC m=+145.101357940" Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.159064 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" podStartSLOduration=122.159043188 podStartE2EDuration="2m2.159043188s" podCreationTimestamp="2026-01-27 15:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:19.134722875 +0000 UTC m=+145.115331993" watchObservedRunningTime="2026-01-27 15:09:19.159043188 +0000 UTC m=+145.139652296" Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.161811 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492100-r2zj6"] Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.211567 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:19 crc kubenswrapper[4772]: E0127 15:09:19.211834 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:19.711807311 +0000 UTC m=+145.692416409 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.212052 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:19 crc kubenswrapper[4772]: E0127 15:09:19.212502 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:19.712486071 +0000 UTC m=+145.693095179 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:19 crc kubenswrapper[4772]: W0127 15:09:19.226902 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6b54ae2_d365_4988_8e69_704574c7962a.slice/crio-7de1df4376bfe65e7f653cd434cd6a00a28483c62ada492613581e939925776c WatchSource:0}: Error finding container 7de1df4376bfe65e7f653cd434cd6a00a28483c62ada492613581e939925776c: Status 404 returned error can't find the container with id 7de1df4376bfe65e7f653cd434cd6a00a28483c62ada492613581e939925776c Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.234138 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mzkn2" podStartSLOduration=123.234119656 podStartE2EDuration="2m3.234119656s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:19.23356676 +0000 UTC m=+145.214175878" watchObservedRunningTime="2026-01-27 15:09:19.234119656 +0000 UTC m=+145.214728764" Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.234644 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-58dhr" podStartSLOduration=6.23463554 podStartE2EDuration="6.23463554s" podCreationTimestamp="2026-01-27 15:09:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:19.212144421 +0000 UTC m=+145.192753519" watchObservedRunningTime="2026-01-27 15:09:19.23463554 +0000 UTC m=+145.215244638" Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.257482 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-njx6w"] Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.281048 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-xqtff"] Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.292400 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-djmb4"] Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.294876 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cf6v7"] Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.317044 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:19 crc kubenswrapper[4772]: E0127 15:09:19.317335 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:19.817322658 +0000 UTC m=+145.797931756 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:19 crc kubenswrapper[4772]: W0127 15:09:19.322764 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55440e9e_5d99_4244_8b5c_55e2d270313b.slice/crio-f58ff164a8a285569068e112a435cf8823b51149c0f63ce1fea966019e44a1dc WatchSource:0}: Error finding container f58ff164a8a285569068e112a435cf8823b51149c0f63ce1fea966019e44a1dc: Status 404 returned error can't find the container with id f58ff164a8a285569068e112a435cf8823b51149c0f63ce1fea966019e44a1dc Jan 27 15:09:19 crc kubenswrapper[4772]: W0127 15:09:19.334898 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59cdf584_81d0_4d66_8fc2_da3a3f995f73.slice/crio-9d1cdd41fa91ae41d42bf9383e73d200fdebf020e8f9e1969deb26535791e0cc WatchSource:0}: Error finding container 9d1cdd41fa91ae41d42bf9383e73d200fdebf020e8f9e1969deb26535791e0cc: Status 404 returned error can't find the container with id 9d1cdd41fa91ae41d42bf9383e73d200fdebf020e8f9e1969deb26535791e0cc Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.418268 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:19 crc kubenswrapper[4772]: E0127 15:09:19.418591 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:19.918579082 +0000 UTC m=+145.899188180 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.519025 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:19 crc kubenswrapper[4772]: E0127 15:09:19.519417 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:20.019395974 +0000 UTC m=+146.000005072 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.621240 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:19 crc kubenswrapper[4772]: E0127 15:09:19.622047 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:20.122026067 +0000 UTC m=+146.102635365 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.655217 4772 patch_prober.go:28] interesting pod/router-default-5444994796-7k7sg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 27 15:09:19 crc kubenswrapper[4772]: [-]has-synced failed: reason withheld Jan 27 15:09:19 crc kubenswrapper[4772]: [+]process-running ok Jan 27 15:09:19 crc kubenswrapper[4772]: healthz check failed Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.655294 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7k7sg" podUID="c4225ddc-bdcd-4158-811b-113234d0c3d0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.723484 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:19 crc kubenswrapper[4772]: E0127 15:09:19.724076 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:20.224055084 +0000 UTC m=+146.204664182 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.825480 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:19 crc kubenswrapper[4772]: E0127 15:09:19.825896 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:20.325878894 +0000 UTC m=+146.306487992 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:19 crc kubenswrapper[4772]: I0127 15:09:19.927001 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:19 crc kubenswrapper[4772]: E0127 15:09:19.927416 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:20.427398706 +0000 UTC m=+146.408007804 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.000707 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" event={"ID":"794cdbb9-3392-465a-8a0a-a78a465aee2b","Type":"ContainerStarted","Data":"cf6cfdcc00e037a8fc2f5bf5dd6d157915e559e9d062c1cb653a624ab52284d0"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.002148 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-bhgv8" event={"ID":"ce4b4e2e-496b-4334-8736-db4f25473731","Type":"ContainerStarted","Data":"60e71282f3e1dce71c46cf15179352066c7821472a93deaf7e3dc1987d0ad580"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.005602 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" event={"ID":"06cdc094-b372-4016-bc5e-4c15a28e032e","Type":"ContainerStarted","Data":"437c578755bfcacf0145c1b3dcede3b1938b4e11e6ad9c7db9d8ac6a8b6df37e"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.009798 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-fw6bh" event={"ID":"f469e02c-9404-422a-bff0-1b945d9c8768","Type":"ContainerStarted","Data":"e8e2cc78c92118a9abbd5bc3282c48c2f6dffed936de924a7357cb4bf4e29f56"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.010139 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-fw6bh" event={"ID":"f469e02c-9404-422a-bff0-1b945d9c8768","Type":"ContainerStarted","Data":"e386fe16c45d765b84619d4118b927f00e6e3264118ad353551e3583d8ea5056"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.013839 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g667l" event={"ID":"06755863-1a8b-4f4d-a304-03bfd45725ec","Type":"ContainerStarted","Data":"b7b77accaed2345fee295b630eb5adb6b9631c110c247f9ecf3d3fc9ff16886f"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.015294 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-24dmv" event={"ID":"3745a211-9fa8-41a7-aa26-d733431bc9aa","Type":"ContainerStarted","Data":"5e2bd2b743e8df88f7a26b59b03cc89fccfe5d20dd70c1895ce6c74553b1027b"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.016920 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xqtff" event={"ID":"59cdf584-81d0-4d66-8fc2-da3a3f995f73","Type":"ContainerStarted","Data":"9d1cdd41fa91ae41d42bf9383e73d200fdebf020e8f9e1969deb26535791e0cc"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.019160 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c6n8m" event={"ID":"79859b30-67ee-456b-82e5-f8806347a0b9","Type":"ContainerStarted","Data":"0e2ff311b3e5eef53c0624d4be740833a870996d5dce14243ce98f9b18306eda"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.020763 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mnltb" event={"ID":"1f03647c-a3e9-4099-9780-e79e3a4d4cf2","Type":"ContainerStarted","Data":"e0ecf6fe4fa34d98fd7c9648eec1e104d0f1c9d28c2d868a1d2c170c384ff0f0"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.020819 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mnltb" event={"ID":"1f03647c-a3e9-4099-9780-e79e3a4d4cf2","Type":"ContainerStarted","Data":"cabee604e502b0d0eec8b97ad150a591d062292394114ff343b9bbda957df6ac"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.021290 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mnltb" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.023464 4772 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-mnltb container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.023515 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mnltb" podUID="1f03647c-a3e9-4099-9780-e79e3a4d4cf2" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.024485 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-jplbk" event={"ID":"3150273d-63f7-4908-bcc5-2403e123d1e7","Type":"ContainerStarted","Data":"0c6c8c10410ade3e540c4fbbc835970cf958b2c46809a38faccfa0a9200de343"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.025773 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cf6v7" event={"ID":"01d08621-494b-4232-b678-9caa94e61085","Type":"ContainerStarted","Data":"e840d87b8a5ee61f7c7dde102858ec11b60d0afb9d35b2281b41290cf928fec4"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.032908 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:20 crc kubenswrapper[4772]: E0127 15:09:20.034112 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:20.533351216 +0000 UTC m=+146.513960314 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.035843 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfsg" event={"ID":"8800c113-7b51-4554-8e52-c1d0df1a08be","Type":"ContainerStarted","Data":"cb614bb5e97e5646988ab1d613a1a5c80cf9b21832653e0860dfbb43f8404b12"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.036048 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" podStartSLOduration=124.036029293 podStartE2EDuration="2m4.036029293s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:20.032032348 +0000 UTC m=+146.012641446" watchObservedRunningTime="2026-01-27 15:09:20.036029293 +0000 UTC m=+146.016638391" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.041037 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5vmlj" event={"ID":"db4a3858-5afa-44c8-a435-2010f7e7340d","Type":"ContainerStarted","Data":"ae81c92d45d0547ff142b72b22b85da95795638755056827220d1a9427feb720"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.044486 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zkmjj" event={"ID":"c66d8c7d-2de8-492f-ba5e-7ff0e236bf64","Type":"ContainerStarted","Data":"673ad9af094d138247ba57d0ec9f4236fa565d62b07383106835f25e54fd66b8"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.058365 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-pwmhd" event={"ID":"c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3","Type":"ContainerStarted","Data":"738fd09154f46306d4ba55689df083e507d469edb2551766b356700d6cb0156e"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.067666 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4j4h" event={"ID":"f9ab8468-b920-41e9-a5f2-1af70f1b5ffd","Type":"ContainerStarted","Data":"b930618ced9b6fc2ff90df7cee8a41329b4824295485280c8b28689467e39a1c"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.072330 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fbwkz" event={"ID":"ffe33359-31f5-4a6c-93fc-6502d2516335","Type":"ContainerStarted","Data":"53fdace7b5e2131fd348c2954060cf41815d5871bf72cef5636e2bab12ed0b7a"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.072985 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mnltb" podStartSLOduration=123.07297003 podStartE2EDuration="2m3.07297003s" podCreationTimestamp="2026-01-27 15:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:20.072189217 +0000 UTC m=+146.052798315" watchObservedRunningTime="2026-01-27 15:09:20.07297003 +0000 UTC m=+146.053579128" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.073089 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-c6n8m" podStartSLOduration=124.073083973 podStartE2EDuration="2m4.073083973s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:20.053755625 +0000 UTC m=+146.034364723" watchObservedRunningTime="2026-01-27 15:09:20.073083973 +0000 UTC m=+146.053693071" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.076616 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jdcpn" event={"ID":"01d2c3f9-778c-4cf0-b8a4-76583f62df3c","Type":"ContainerStarted","Data":"68ec70224d1a508567ee2ab655eea34f597e6e0cc299689fa5a399faf743e0ab"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.076657 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jdcpn" event={"ID":"01d2c3f9-778c-4cf0-b8a4-76583f62df3c","Type":"ContainerStarted","Data":"993555ddd9411a0fbc04bff6d61b92ed6d128e38ba89d44582647ac7eae8fc72"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.077098 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jdcpn" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.079494 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bck4j" event={"ID":"ee07e18c-9f40-41c3-b2fb-05fd325976e4","Type":"ContainerStarted","Data":"c2c3ba6783909759f21603251263d974ccddd9a3fab183c459ffd4bea1591850"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.079850 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bck4j" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.080499 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-njx6w" event={"ID":"55440e9e-5d99-4244-8b5c-55e2d270313b","Type":"ContainerStarted","Data":"f58ff164a8a285569068e112a435cf8823b51149c0f63ce1fea966019e44a1dc"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.081371 4772 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-jdcpn container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.081430 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jdcpn" podUID="01d2c3f9-778c-4cf0-b8a4-76583f62df3c" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.081927 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-npths" event={"ID":"5276546c-f731-4bd0-bb93-b5cd19b0992c","Type":"ContainerStarted","Data":"1d51339c275272b67e2e7b518c3e5c43f5ac5e12439635abbf3fb11f3545e804"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.082444 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-npths" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.084051 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cv2z7" event={"ID":"b3eefbc3-6dc4-479c-93e4-94a70fda0f83","Type":"ContainerStarted","Data":"1c65cf47eb5b99b895cec64ff223cf5872c35c6374b48f246c3d1c1fc1ecc64c"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.084997 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f9fqs" event={"ID":"946946b3-f9e0-45e4-803f-edb3f7218489","Type":"ContainerStarted","Data":"cba1cd7e1e663bc926502749eb9df09ac4203e7c5246c4a725f052b689f32b97"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.086397 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492100-r2zj6" event={"ID":"c6b54ae2-d365-4988-8e69-704574c7962a","Type":"ContainerStarted","Data":"7de1df4376bfe65e7f653cd434cd6a00a28483c62ada492613581e939925776c"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.087293 4772 patch_prober.go:28] interesting pod/console-operator-58897d9998-npths container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.087322 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-npths" podUID="5276546c-f731-4bd0-bb93-b5cd19b0992c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.089670 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" event={"ID":"34e7a553-e424-472e-a143-76e7e08e57aa","Type":"ContainerStarted","Data":"2907069c7dc18352c253a6dd8a614f2fcda2e6ef3cfec82559f2ecaf55c235bf"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.091989 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v7l4k" event={"ID":"85774248-1879-439e-9dd2-0d8661c299d6","Type":"ContainerStarted","Data":"616c4f9696aa8b033d29e5e8dc00d721a29a5606281dc1c5344744ca66877594"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.095771 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-vswtw" event={"ID":"17bdd07d-f7e5-47f8-b730-724d5cc8e3d2","Type":"ContainerStarted","Data":"ed60d396d7c416511dd8344b674f89e48671cdc264384930b775cd7370b999c1"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.096611 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-vswtw" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.099645 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" event={"ID":"c8ebf890-c3b0-468e-bf7d-0ec590df084b","Type":"ContainerStarted","Data":"8632589c7dbe4bb64d8d2a9e0983c8088c1ff445e316f1dd7c4e04e72fa148df"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.099685 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" event={"ID":"c8ebf890-c3b0-468e-bf7d-0ec590df084b","Type":"ContainerStarted","Data":"267d22366b6b80c120159c7b29d573289ed71a1b2d51c437b57f97f84c344fdc"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.100244 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.103150 4772 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4lj2h container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.103213 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.105754 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-9wv77" event={"ID":"855486b0-11f8-4ff0-930d-75c7e9d790d3","Type":"ContainerStarted","Data":"aea1a08e5aad39ed9a0a97079f3d1003246d905f1fda942ca2d0ca870f020eb3"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.109452 4772 patch_prober.go:28] interesting pod/downloads-7954f5f757-vswtw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.109705 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vswtw" podUID="17bdd07d-f7e5-47f8-b730-724d5cc8e3d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.123998 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-fw6bh" podStartSLOduration=7.123973693 podStartE2EDuration="7.123973693s" podCreationTimestamp="2026-01-27 15:09:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:20.120669007 +0000 UTC m=+146.101278105" watchObservedRunningTime="2026-01-27 15:09:20.123973693 +0000 UTC m=+146.104582791" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.129758 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-djmb4" event={"ID":"dbd05529-6d54-416a-8df0-5973ee3179b6","Type":"ContainerStarted","Data":"f8a87a1260878fbad151d7e44e315c2aed855c2914f6ae6266e3f184df9702f6"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.134484 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:20 crc kubenswrapper[4772]: E0127 15:09:20.136353 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:20.63633539 +0000 UTC m=+146.616944488 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.155640 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.155921 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.155995 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.156199 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.158053 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4v228" event={"ID":"2c9eef7e-3996-45b6-ab7b-50d319dc1117","Type":"ContainerStarted","Data":"79c4e74a245ca3f2ef71fe6a698ce5a12f2c8cc3d935736d847f9759a8d4904d"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.165563 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zq27x" event={"ID":"192d0f8f-10f9-43e2-a24a-2019aae0db44","Type":"ContainerStarted","Data":"d49a7fd9cabdd4167adee013f8da89e57a16016ffc4a14822acbc6007a368377"} Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.176333 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" podStartSLOduration=123.176317214 podStartE2EDuration="2m3.176317214s" podCreationTimestamp="2026-01-27 15:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:20.150646393 +0000 UTC m=+146.131255491" watchObservedRunningTime="2026-01-27 15:09:20.176317214 +0000 UTC m=+146.156926312" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.176703 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bck4j" podStartSLOduration=124.176697355 podStartE2EDuration="2m4.176697355s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:20.175879182 +0000 UTC m=+146.156488300" watchObservedRunningTime="2026-01-27 15:09:20.176697355 +0000 UTC m=+146.157306453" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.195852 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-npths" podStartSLOduration=124.195834658 podStartE2EDuration="2m4.195834658s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:20.194433218 +0000 UTC m=+146.175042336" watchObservedRunningTime="2026-01-27 15:09:20.195834658 +0000 UTC m=+146.176443746" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.216348 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jdcpn" podStartSLOduration=123.21631503 podStartE2EDuration="2m3.21631503s" podCreationTimestamp="2026-01-27 15:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:20.214106526 +0000 UTC m=+146.194715624" watchObservedRunningTime="2026-01-27 15:09:20.21631503 +0000 UTC m=+146.196924128" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.236306 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:20 crc kubenswrapper[4772]: E0127 15:09:20.239083 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:20.739041376 +0000 UTC m=+146.719650474 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.239846 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" podStartSLOduration=124.239817698 podStartE2EDuration="2m4.239817698s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:20.234377771 +0000 UTC m=+146.214986869" watchObservedRunningTime="2026-01-27 15:09:20.239817698 +0000 UTC m=+146.220426796" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.263672 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fbwkz" podStartSLOduration=124.263655197 podStartE2EDuration="2m4.263655197s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:20.262086751 +0000 UTC m=+146.242695849" watchObservedRunningTime="2026-01-27 15:09:20.263655197 +0000 UTC m=+146.244264295" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.268318 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.286802 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zkmjj" podStartSLOduration=123.286782174 podStartE2EDuration="2m3.286782174s" podCreationTimestamp="2026-01-27 15:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:20.284063866 +0000 UTC m=+146.264672964" watchObservedRunningTime="2026-01-27 15:09:20.286782174 +0000 UTC m=+146.267391272" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.322592 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-9wv77" podStartSLOduration=123.322578358 podStartE2EDuration="2m3.322578358s" podCreationTimestamp="2026-01-27 15:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:20.306100052 +0000 UTC m=+146.286709150" watchObservedRunningTime="2026-01-27 15:09:20.322578358 +0000 UTC m=+146.303187456" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.323740 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-vswtw" podStartSLOduration=124.323732012 podStartE2EDuration="2m4.323732012s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:20.319635463 +0000 UTC m=+146.300244561" watchObservedRunningTime="2026-01-27 15:09:20.323732012 +0000 UTC m=+146.304341110" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.337428 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:20 crc kubenswrapper[4772]: E0127 15:09:20.337778 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:20.837763847 +0000 UTC m=+146.818372945 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.348178 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zq27x" podStartSLOduration=124.348151017 podStartE2EDuration="2m4.348151017s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:20.346070757 +0000 UTC m=+146.326679865" watchObservedRunningTime="2026-01-27 15:09:20.348151017 +0000 UTC m=+146.328760115" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.426015 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4v228" podStartSLOduration=124.425995445 podStartE2EDuration="2m4.425995445s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:20.418834068 +0000 UTC m=+146.399443166" watchObservedRunningTime="2026-01-27 15:09:20.425995445 +0000 UTC m=+146.406604543" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.438701 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:20 crc kubenswrapper[4772]: E0127 15:09:20.439112 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:20.939096893 +0000 UTC m=+146.919705981 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.540370 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:20 crc kubenswrapper[4772]: E0127 15:09:20.542253 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:21.042232761 +0000 UTC m=+147.022841859 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.643257 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:20 crc kubenswrapper[4772]: E0127 15:09:20.643630 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:21.143603779 +0000 UTC m=+147.124212877 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.664406 4772 patch_prober.go:28] interesting pod/router-default-5444994796-7k7sg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 27 15:09:20 crc kubenswrapper[4772]: [-]has-synced failed: reason withheld Jan 27 15:09:20 crc kubenswrapper[4772]: [+]process-running ok Jan 27 15:09:20 crc kubenswrapper[4772]: healthz check failed Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.664789 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7k7sg" podUID="c4225ddc-bdcd-4158-811b-113234d0c3d0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.746646 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:20 crc kubenswrapper[4772]: E0127 15:09:20.746948 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:21.246932693 +0000 UTC m=+147.227541781 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.848922 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:20 crc kubenswrapper[4772]: E0127 15:09:20.849467 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:21.349447893 +0000 UTC m=+147.330057061 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.950556 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:20 crc kubenswrapper[4772]: E0127 15:09:20.950840 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:21.450815 +0000 UTC m=+147.431424088 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:20 crc kubenswrapper[4772]: I0127 15:09:20.950906 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:20 crc kubenswrapper[4772]: E0127 15:09:20.951416 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:21.451404097 +0000 UTC m=+147.432013185 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.052635 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:21 crc kubenswrapper[4772]: E0127 15:09:21.052820 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:21.552786745 +0000 UTC m=+147.533395843 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.052972 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:21 crc kubenswrapper[4772]: E0127 15:09:21.053363 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:21.553349801 +0000 UTC m=+147.533958899 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.153967 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:21 crc kubenswrapper[4772]: E0127 15:09:21.154209 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:21.654157392 +0000 UTC m=+147.634766490 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.154691 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:21 crc kubenswrapper[4772]: E0127 15:09:21.155039 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:21.655021137 +0000 UTC m=+147.635630235 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.170309 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v7l4k" event={"ID":"85774248-1879-439e-9dd2-0d8661c299d6","Type":"ContainerStarted","Data":"c77561dd43c0905e11cc9da17704b2525d4b7afef4f98f4af607bef043388148"} Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.172283 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f9fqs" event={"ID":"946946b3-f9e0-45e4-803f-edb3f7218489","Type":"ContainerStarted","Data":"0ff467fa052293b06237dea9bb855a2e878b19713b29159eec9015ee3fa68cd1"} Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.174352 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492100-r2zj6" event={"ID":"c6b54ae2-d365-4988-8e69-704574c7962a","Type":"ContainerStarted","Data":"4f5ed02624877f82608d4a7a5fead892a80497d0b63bf729eaa6c0d56cf6aac6"} Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.176365 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g667l" event={"ID":"06755863-1a8b-4f4d-a304-03bfd45725ec","Type":"ContainerStarted","Data":"83143922be92b5fa0adc3d1c04d2777313e5d1d785f87047a8d0acebd953a6ee"} Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.177914 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-24dmv" event={"ID":"3745a211-9fa8-41a7-aa26-d733431bc9aa","Type":"ContainerStarted","Data":"a5ebf3fad362d41f306cdc320239a12f9391ce6f794ea71a601b38d8cd31e4df"} Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.177957 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-24dmv" event={"ID":"3745a211-9fa8-41a7-aa26-d733431bc9aa","Type":"ContainerStarted","Data":"2bce162515e7b97ab026e1bb12a122d93bb3ceabcd5847641abc7df9d4da0ccd"} Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.179698 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4j4h" event={"ID":"f9ab8468-b920-41e9-a5f2-1af70f1b5ffd","Type":"ContainerStarted","Data":"22a70d5916ada3cb7b0f3cfc5aa6a47b5d556678fc4040abd3c926b168155a0a"} Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.181203 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-djmb4" event={"ID":"dbd05529-6d54-416a-8df0-5973ee3179b6","Type":"ContainerStarted","Data":"64ec27ae905d0cb73ec74a804453f8d9836f7bdc705f2062bf22fde875ba140a"} Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.182826 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cv2z7" event={"ID":"b3eefbc3-6dc4-479c-93e4-94a70fda0f83","Type":"ContainerStarted","Data":"e9b923018e39913453724269344c98b4510c0d3afe324798fdc10365d8b1c22b"} Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.182975 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cv2z7" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.183948 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-pwmhd" event={"ID":"c4354b42-c43f-45ce-b1e5-f1d6e0ed1bd3","Type":"ContainerStarted","Data":"48ed2cade3088ad9ef9e0aee7dc9a0a76326ddf212937d1b64722bb46b50b984"} Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.185442 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cf6v7" event={"ID":"01d08621-494b-4232-b678-9caa94e61085","Type":"ContainerStarted","Data":"082c258a684655d96d16487fead32cf89c30175a36b818bc16b83999d0295232"} Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.186111 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cf6v7" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.187836 4772 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-cf6v7 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:5443/healthz\": dial tcp 10.217.0.38:5443: connect: connection refused" start-of-body= Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.187870 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cf6v7" podUID="01d08621-494b-4232-b678-9caa94e61085" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.38:5443/healthz\": dial tcp 10.217.0.38:5443: connect: connection refused" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.188191 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-jplbk" event={"ID":"3150273d-63f7-4908-bcc5-2403e123d1e7","Type":"ContainerStarted","Data":"0fe2fd92bdfb6082b1eced78bf89302ffcb2851203a712765ffc9978cba71854"} Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.188594 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-jplbk" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.190019 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" event={"ID":"794cdbb9-3392-465a-8a0a-a78a465aee2b","Type":"ContainerStarted","Data":"3a614d57463b1f07afccc480ba9fd3a454c9119bb8aa40cd8763206f105d3d25"} Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.191309 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5vmlj" event={"ID":"db4a3858-5afa-44c8-a435-2010f7e7340d","Type":"ContainerStarted","Data":"dcc89b74c73c56621c851823bb624cba5d70cdd0a5cf057ba421dbea811cfd80"} Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.194327 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-njx6w" event={"ID":"55440e9e-5d99-4244-8b5c-55e2d270313b","Type":"ContainerStarted","Data":"bd13bc4d5640fefd1ee1b05a0f76dd4bd95062a26abbdf38f089811f6ef9d615"} Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.194362 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-njx6w" event={"ID":"55440e9e-5d99-4244-8b5c-55e2d270313b","Type":"ContainerStarted","Data":"15cc4626a580771e8008c7c13476968249bb59abaa3c5706dc853bb20066c974"} Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.200141 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfsg" event={"ID":"8800c113-7b51-4554-8e52-c1d0df1a08be","Type":"ContainerStarted","Data":"6af6dcdc0b1ff60e46ab4aaec43713a1460bc2484adfbbac9528161c3edb645a"} Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.201388 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xqtff" event={"ID":"59cdf584-81d0-4d66-8fc2-da3a3f995f73","Type":"ContainerStarted","Data":"a4f9e554de172277e02dd3d65d14bce9ec30394850fee45550108a3b8df8b44c"} Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.201424 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xqtff" event={"ID":"59cdf584-81d0-4d66-8fc2-da3a3f995f73","Type":"ContainerStarted","Data":"3f03f69ab90e8bff4b373cfa302b43affe0a5f955abea222fe2bbe3d950ac56d"} Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.202371 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fbwkz" event={"ID":"ffe33359-31f5-4a6c-93fc-6502d2516335","Type":"ContainerStarted","Data":"49e23824b145bf79c0d3e66cc5f10aad5c7da00f1e7e173bb52b34665a531f44"} Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.203911 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-bhgv8" event={"ID":"ce4b4e2e-496b-4334-8736-db4f25473731","Type":"ContainerStarted","Data":"ea62d58e4b98cc0c9d276cb1c41f819e098e87a78641d39d6c12c00942a8a45b"} Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.204484 4772 patch_prober.go:28] interesting pod/downloads-7954f5f757-vswtw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.204522 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vswtw" podUID="17bdd07d-f7e5-47f8-b730-724d5cc8e3d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.204485 4772 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-jdcpn container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.204734 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jdcpn" podUID="01d2c3f9-778c-4cf0-b8a4-76583f62df3c" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.204825 4772 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-mnltb container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.204856 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mnltb" podUID="1f03647c-a3e9-4099-9780-e79e3a4d4cf2" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.204993 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.205139 4772 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4lj2h container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.205185 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.207407 4772 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-fgw98 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.16:6443/healthz\": dial tcp 10.217.0.16:6443: connect: connection refused" start-of-body= Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.207448 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" podUID="06cdc094-b372-4016-bc5e-4c15a28e032e" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.16:6443/healthz\": dial tcp 10.217.0.16:6443: connect: connection refused" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.218988 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wk7gd" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.230796 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-djmb4" podStartSLOduration=124.230777425 podStartE2EDuration="2m4.230777425s" podCreationTimestamp="2026-01-27 15:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:21.230290161 +0000 UTC m=+147.210899269" watchObservedRunningTime="2026-01-27 15:09:21.230777425 +0000 UTC m=+147.211386523" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.232503 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v7l4k" podStartSLOduration=124.232496215 podStartE2EDuration="2m4.232496215s" podCreationTimestamp="2026-01-27 15:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:21.198292967 +0000 UTC m=+147.178902065" watchObservedRunningTime="2026-01-27 15:09:21.232496215 +0000 UTC m=+147.213105313" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.255327 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:21 crc kubenswrapper[4772]: E0127 15:09:21.256736 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:21.756712764 +0000 UTC m=+147.737321932 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.264623 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w4j4h" podStartSLOduration=125.264609632 podStartE2EDuration="2m5.264609632s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:21.262231953 +0000 UTC m=+147.242841051" watchObservedRunningTime="2026-01-27 15:09:21.264609632 +0000 UTC m=+147.245218730" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.286375 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f9fqs" podStartSLOduration=125.28636129 podStartE2EDuration="2m5.28636129s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:21.284814796 +0000 UTC m=+147.265423894" watchObservedRunningTime="2026-01-27 15:09:21.28636129 +0000 UTC m=+147.266970388" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.351063 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-24dmv" podStartSLOduration=124.351023228 podStartE2EDuration="2m4.351023228s" podCreationTimestamp="2026-01-27 15:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:21.310698853 +0000 UTC m=+147.291307951" watchObservedRunningTime="2026-01-27 15:09:21.351023228 +0000 UTC m=+147.331632336" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.354280 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29492100-r2zj6" podStartSLOduration=125.354265511 podStartE2EDuration="2m5.354265511s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:21.352950093 +0000 UTC m=+147.333559191" watchObservedRunningTime="2026-01-27 15:09:21.354265511 +0000 UTC m=+147.334874619" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.359375 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:21 crc kubenswrapper[4772]: E0127 15:09:21.359674 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:21.859662337 +0000 UTC m=+147.840271435 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.421109 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cv2z7" podStartSLOduration=124.421088001 podStartE2EDuration="2m4.421088001s" podCreationTimestamp="2026-01-27 15:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:21.389673474 +0000 UTC m=+147.370282572" watchObservedRunningTime="2026-01-27 15:09:21.421088001 +0000 UTC m=+147.401697099" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.422437 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-pwmhd" podStartSLOduration=125.42242856 podStartE2EDuration="2m5.42242856s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:21.419261308 +0000 UTC m=+147.399870416" watchObservedRunningTime="2026-01-27 15:09:21.42242856 +0000 UTC m=+147.403037658" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.437921 4772 patch_prober.go:28] interesting pod/apiserver-76f77b778f-2h2z8 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Jan 27 15:09:21 crc kubenswrapper[4772]: [+]log ok Jan 27 15:09:21 crc kubenswrapper[4772]: [+]etcd ok Jan 27 15:09:21 crc kubenswrapper[4772]: [+]poststarthook/start-apiserver-admission-initializer ok Jan 27 15:09:21 crc kubenswrapper[4772]: [+]poststarthook/generic-apiserver-start-informers ok Jan 27 15:09:21 crc kubenswrapper[4772]: [+]poststarthook/max-in-flight-filter ok Jan 27 15:09:21 crc kubenswrapper[4772]: [+]poststarthook/storage-object-count-tracker-hook ok Jan 27 15:09:21 crc kubenswrapper[4772]: [+]poststarthook/image.openshift.io-apiserver-caches ok Jan 27 15:09:21 crc kubenswrapper[4772]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Jan 27 15:09:21 crc kubenswrapper[4772]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Jan 27 15:09:21 crc kubenswrapper[4772]: [+]poststarthook/project.openshift.io-projectcache ok Jan 27 15:09:21 crc kubenswrapper[4772]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Jan 27 15:09:21 crc kubenswrapper[4772]: [+]poststarthook/openshift.io-startinformers ok Jan 27 15:09:21 crc kubenswrapper[4772]: [+]poststarthook/openshift.io-restmapperupdater ok Jan 27 15:09:21 crc kubenswrapper[4772]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Jan 27 15:09:21 crc kubenswrapper[4772]: livez check failed Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.438062 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" podUID="34e7a553-e424-472e-a143-76e7e08e57aa" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.448739 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-jplbk" podStartSLOduration=8.448716929 podStartE2EDuration="8.448716929s" podCreationTimestamp="2026-01-27 15:09:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:21.446089123 +0000 UTC m=+147.426698251" watchObservedRunningTime="2026-01-27 15:09:21.448716929 +0000 UTC m=+147.429326027" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.460290 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:21 crc kubenswrapper[4772]: E0127 15:09:21.460786 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:21.960767797 +0000 UTC m=+147.941376885 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.520981 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5vmlj" podStartSLOduration=124.520963215 podStartE2EDuration="2m4.520963215s" podCreationTimestamp="2026-01-27 15:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:21.472583868 +0000 UTC m=+147.453192966" watchObservedRunningTime="2026-01-27 15:09:21.520963215 +0000 UTC m=+147.501572313" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.522267 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-g667l" podStartSLOduration=125.522261913 podStartE2EDuration="2m5.522261913s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:21.508078663 +0000 UTC m=+147.488687761" watchObservedRunningTime="2026-01-27 15:09:21.522261913 +0000 UTC m=+147.502871011" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.564950 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:21 crc kubenswrapper[4772]: E0127 15:09:21.565341 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:22.065327186 +0000 UTC m=+148.045936284 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.587790 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-njx6w" podStartSLOduration=124.587772954 podStartE2EDuration="2m4.587772954s" podCreationTimestamp="2026-01-27 15:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:21.554440792 +0000 UTC m=+147.535049910" watchObservedRunningTime="2026-01-27 15:09:21.587772954 +0000 UTC m=+147.568382052" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.607351 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-klfsg" podStartSLOduration=125.607307619 podStartE2EDuration="2m5.607307619s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:21.607262377 +0000 UTC m=+147.587871485" watchObservedRunningTime="2026-01-27 15:09:21.607307619 +0000 UTC m=+147.587916717" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.608340 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cf6v7" podStartSLOduration=124.608333428 podStartE2EDuration="2m4.608333428s" podCreationTimestamp="2026-01-27 15:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:21.588625999 +0000 UTC m=+147.569235107" watchObservedRunningTime="2026-01-27 15:09:21.608333428 +0000 UTC m=+147.588942526" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.614534 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-npths" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.654021 4772 patch_prober.go:28] interesting pod/router-default-5444994796-7k7sg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 27 15:09:21 crc kubenswrapper[4772]: [-]has-synced failed: reason withheld Jan 27 15:09:21 crc kubenswrapper[4772]: [+]process-running ok Jan 27 15:09:21 crc kubenswrapper[4772]: healthz check failed Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.654371 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7k7sg" podUID="c4225ddc-bdcd-4158-811b-113234d0c3d0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.665855 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:21 crc kubenswrapper[4772]: E0127 15:09:21.666339 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:22.166321413 +0000 UTC m=+148.146930511 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.692799 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-xqtff" podStartSLOduration=124.69278125700001 podStartE2EDuration="2m4.692781257s" podCreationTimestamp="2026-01-27 15:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:21.692380245 +0000 UTC m=+147.672989353" watchObservedRunningTime="2026-01-27 15:09:21.692781257 +0000 UTC m=+147.673390355" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.758758 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-bhgv8" podStartSLOduration=125.758743022 podStartE2EDuration="2m5.758743022s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:21.730586089 +0000 UTC m=+147.711195187" watchObservedRunningTime="2026-01-27 15:09:21.758743022 +0000 UTC m=+147.739352120" Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.768044 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:21 crc kubenswrapper[4772]: E0127 15:09:21.768686 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:22.268669778 +0000 UTC m=+148.249278886 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.868987 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:21 crc kubenswrapper[4772]: E0127 15:09:21.869204 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:22.36915747 +0000 UTC m=+148.349766568 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.869357 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:21 crc kubenswrapper[4772]: E0127 15:09:21.869676 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:22.369661105 +0000 UTC m=+148.350270203 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:21 crc kubenswrapper[4772]: I0127 15:09:21.969950 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:21 crc kubenswrapper[4772]: E0127 15:09:21.970248 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:22.470233508 +0000 UTC m=+148.450842606 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.071740 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:22 crc kubenswrapper[4772]: E0127 15:09:22.072125 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:22.57210884 +0000 UTC m=+148.552717938 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.173276 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:22 crc kubenswrapper[4772]: E0127 15:09:22.173467 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:22.673439906 +0000 UTC m=+148.654049004 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.173544 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:22 crc kubenswrapper[4772]: E0127 15:09:22.173838 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:22.673827277 +0000 UTC m=+148.654436365 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.225065 4772 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4lj2h container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.225113 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.225433 4772 patch_prober.go:28] interesting pod/downloads-7954f5f757-vswtw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.225483 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vswtw" podUID="17bdd07d-f7e5-47f8-b730-724d5cc8e3d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.274958 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:22 crc kubenswrapper[4772]: E0127 15:09:22.275088 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:22.775069461 +0000 UTC m=+148.755678559 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.276955 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:22 crc kubenswrapper[4772]: E0127 15:09:22.282302 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:22.78228732 +0000 UTC m=+148.762896418 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.378803 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:22 crc kubenswrapper[4772]: E0127 15:09:22.379109 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:22.879093375 +0000 UTC m=+148.859702483 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.392275 4772 csr.go:261] certificate signing request csr-x7cq6 is approved, waiting to be issued Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.397050 4772 csr.go:257] certificate signing request csr-x7cq6 is issued Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.481712 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:22 crc kubenswrapper[4772]: E0127 15:09:22.482047 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:22.982036118 +0000 UTC m=+148.962645216 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.582277 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:22 crc kubenswrapper[4772]: E0127 15:09:22.582478 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:23.082462598 +0000 UTC m=+149.063071696 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.582781 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:22 crc kubenswrapper[4772]: E0127 15:09:22.583055 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:23.083047945 +0000 UTC m=+149.063657043 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.583328 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.653154 4772 patch_prober.go:28] interesting pod/router-default-5444994796-7k7sg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 27 15:09:22 crc kubenswrapper[4772]: [-]has-synced failed: reason withheld Jan 27 15:09:22 crc kubenswrapper[4772]: [+]process-running ok Jan 27 15:09:22 crc kubenswrapper[4772]: healthz check failed Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.653217 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7k7sg" podUID="c4225ddc-bdcd-4158-811b-113234d0c3d0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.689661 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.689963 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.690014 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.690054 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.690097 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:09:22 crc kubenswrapper[4772]: E0127 15:09:22.691210 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:23.191185208 +0000 UTC m=+149.171794306 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.691809 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.702207 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.702216 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.720683 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.767835 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bck4j" Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.791636 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:22 crc kubenswrapper[4772]: E0127 15:09:22.792030 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:23.292015029 +0000 UTC m=+149.272624127 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.800352 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.884856 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.894627 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:22 crc kubenswrapper[4772]: E0127 15:09:22.895072 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:23.395057355 +0000 UTC m=+149.375666453 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.897522 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 27 15:09:22 crc kubenswrapper[4772]: I0127 15:09:22.995831 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:23 crc kubenswrapper[4772]: E0127 15:09:23.020266 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:23.52023638 +0000 UTC m=+149.500845478 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.103897 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:23 crc kubenswrapper[4772]: E0127 15:09:23.104474 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:23.604452492 +0000 UTC m=+149.585061610 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.205915 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:23 crc kubenswrapper[4772]: E0127 15:09:23.206455 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:23.706442817 +0000 UTC m=+149.687051915 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.233706 4772 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-cf6v7 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.233758 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cf6v7" podUID="01d08621-494b-4232-b678-9caa94e61085" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.38:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.267354 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" event={"ID":"794cdbb9-3392-465a-8a0a-a78a465aee2b","Type":"ContainerStarted","Data":"458ba566568151ced1efe9b01df96546d245e5b95865522f4163f8c7b9dc4145"} Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.267393 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" event={"ID":"794cdbb9-3392-465a-8a0a-a78a465aee2b","Type":"ContainerStarted","Data":"12b644f033abf82b2aa10593fc8f1f5c89fe7b2f0adfc95ab3ba0235049af613"} Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.307657 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:23 crc kubenswrapper[4772]: E0127 15:09:23.309012 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:23.808997729 +0000 UTC m=+149.789606827 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.344799 4772 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.399267 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-01-27 15:04:22 +0000 UTC, rotation deadline is 2026-12-04 07:51:36.577008503 +0000 UTC Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.399306 4772 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7456h42m13.177705526s for next certificate rotation Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.410791 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:23 crc kubenswrapper[4772]: E0127 15:09:23.411078 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:23.911065496 +0000 UTC m=+149.891674594 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.489194 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-95rh9"] Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.490121 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-95rh9" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.499599 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.511863 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:23 crc kubenswrapper[4772]: E0127 15:09:23.512300 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:24.012284599 +0000 UTC m=+149.992893697 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.558585 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-95rh9"] Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.613853 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7-utilities\") pod \"community-operators-95rh9\" (UID: \"4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7\") " pod="openshift-marketplace/community-operators-95rh9" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.613893 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7-catalog-content\") pod \"community-operators-95rh9\" (UID: \"4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7\") " pod="openshift-marketplace/community-operators-95rh9" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.613944 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zvrk\" (UniqueName: \"kubernetes.io/projected/4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7-kube-api-access-4zvrk\") pod \"community-operators-95rh9\" (UID: \"4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7\") " pod="openshift-marketplace/community-operators-95rh9" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.613974 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:23 crc kubenswrapper[4772]: E0127 15:09:23.614271 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:24.114260424 +0000 UTC m=+150.094869522 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.656365 4772 patch_prober.go:28] interesting pod/router-default-5444994796-7k7sg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 27 15:09:23 crc kubenswrapper[4772]: [-]has-synced failed: reason withheld Jan 27 15:09:23 crc kubenswrapper[4772]: [+]process-running ok Jan 27 15:09:23 crc kubenswrapper[4772]: healthz check failed Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.656831 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7k7sg" podUID="c4225ddc-bdcd-4158-811b-113234d0c3d0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.682488 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9wdps"] Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.697968 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9wdps" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.701209 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.714589 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.714734 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7-utilities\") pod \"community-operators-95rh9\" (UID: \"4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7\") " pod="openshift-marketplace/community-operators-95rh9" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.714757 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7-catalog-content\") pod \"community-operators-95rh9\" (UID: \"4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7\") " pod="openshift-marketplace/community-operators-95rh9" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.714780 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zvrk\" (UniqueName: \"kubernetes.io/projected/4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7-kube-api-access-4zvrk\") pod \"community-operators-95rh9\" (UID: \"4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7\") " pod="openshift-marketplace/community-operators-95rh9" Jan 27 15:09:23 crc kubenswrapper[4772]: E0127 15:09:23.715099 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-27 15:09:24.215085716 +0000 UTC m=+150.195694814 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.715499 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7-utilities\") pod \"community-operators-95rh9\" (UID: \"4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7\") " pod="openshift-marketplace/community-operators-95rh9" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.715622 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7-catalog-content\") pod \"community-operators-95rh9\" (UID: \"4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7\") " pod="openshift-marketplace/community-operators-95rh9" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.721631 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9wdps"] Jan 27 15:09:23 crc kubenswrapper[4772]: W0127 15:09:23.752738 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-c6ced8694bd7595fee00b262ac36c45bd68f29fbefa1dc4a8eaae1198c1d0493 WatchSource:0}: Error finding container c6ced8694bd7595fee00b262ac36c45bd68f29fbefa1dc4a8eaae1198c1d0493: Status 404 returned error can't find the container with id c6ced8694bd7595fee00b262ac36c45bd68f29fbefa1dc4a8eaae1198c1d0493 Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.765196 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zvrk\" (UniqueName: \"kubernetes.io/projected/4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7-kube-api-access-4zvrk\") pod \"community-operators-95rh9\" (UID: \"4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7\") " pod="openshift-marketplace/community-operators-95rh9" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.815474 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96e88efd-1f25-4e44-b459-ab773db93656-catalog-content\") pod \"certified-operators-9wdps\" (UID: \"96e88efd-1f25-4e44-b459-ab773db93656\") " pod="openshift-marketplace/certified-operators-9wdps" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.815549 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96e88efd-1f25-4e44-b459-ab773db93656-utilities\") pod \"certified-operators-9wdps\" (UID: \"96e88efd-1f25-4e44-b459-ab773db93656\") " pod="openshift-marketplace/certified-operators-9wdps" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.815618 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.815645 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fhkv\" (UniqueName: \"kubernetes.io/projected/96e88efd-1f25-4e44-b459-ab773db93656-kube-api-access-7fhkv\") pod \"certified-operators-9wdps\" (UID: \"96e88efd-1f25-4e44-b459-ab773db93656\") " pod="openshift-marketplace/certified-operators-9wdps" Jan 27 15:09:23 crc kubenswrapper[4772]: E0127 15:09:23.816000 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-27 15:09:24.31598888 +0000 UTC m=+150.296597968 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-crlcr" (UID: "877de785-bc18-4c1c-970a-1e6533539467") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.844512 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-95rh9" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.859906 4772 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-01-27T15:09:23.344822003Z","Handler":null,"Name":""} Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.865894 4772 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.865939 4772 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.881803 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jwrpk"] Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.885015 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jwrpk" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.898401 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jwrpk"] Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.918181 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.918378 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fhkv\" (UniqueName: \"kubernetes.io/projected/96e88efd-1f25-4e44-b459-ab773db93656-kube-api-access-7fhkv\") pod \"certified-operators-9wdps\" (UID: \"96e88efd-1f25-4e44-b459-ab773db93656\") " pod="openshift-marketplace/certified-operators-9wdps" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.918430 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96e88efd-1f25-4e44-b459-ab773db93656-catalog-content\") pod \"certified-operators-9wdps\" (UID: \"96e88efd-1f25-4e44-b459-ab773db93656\") " pod="openshift-marketplace/certified-operators-9wdps" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.918483 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67q8x\" (UniqueName: \"kubernetes.io/projected/dd415ccf-2b4a-4797-962f-a464ef96bc22-kube-api-access-67q8x\") pod \"community-operators-jwrpk\" (UID: \"dd415ccf-2b4a-4797-962f-a464ef96bc22\") " pod="openshift-marketplace/community-operators-jwrpk" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.918513 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd415ccf-2b4a-4797-962f-a464ef96bc22-utilities\") pod \"community-operators-jwrpk\" (UID: \"dd415ccf-2b4a-4797-962f-a464ef96bc22\") " pod="openshift-marketplace/community-operators-jwrpk" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.918543 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96e88efd-1f25-4e44-b459-ab773db93656-utilities\") pod \"certified-operators-9wdps\" (UID: \"96e88efd-1f25-4e44-b459-ab773db93656\") " pod="openshift-marketplace/certified-operators-9wdps" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.918573 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd415ccf-2b4a-4797-962f-a464ef96bc22-catalog-content\") pod \"community-operators-jwrpk\" (UID: \"dd415ccf-2b4a-4797-962f-a464ef96bc22\") " pod="openshift-marketplace/community-operators-jwrpk" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.919451 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96e88efd-1f25-4e44-b459-ab773db93656-catalog-content\") pod \"certified-operators-9wdps\" (UID: \"96e88efd-1f25-4e44-b459-ab773db93656\") " pod="openshift-marketplace/certified-operators-9wdps" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.919493 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96e88efd-1f25-4e44-b459-ab773db93656-utilities\") pod \"certified-operators-9wdps\" (UID: \"96e88efd-1f25-4e44-b459-ab773db93656\") " pod="openshift-marketplace/certified-operators-9wdps" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.928253 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 27 15:09:23 crc kubenswrapper[4772]: I0127 15:09:23.951868 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fhkv\" (UniqueName: \"kubernetes.io/projected/96e88efd-1f25-4e44-b459-ab773db93656-kube-api-access-7fhkv\") pod \"certified-operators-9wdps\" (UID: \"96e88efd-1f25-4e44-b459-ab773db93656\") " pod="openshift-marketplace/certified-operators-9wdps" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.020323 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd415ccf-2b4a-4797-962f-a464ef96bc22-utilities\") pod \"community-operators-jwrpk\" (UID: \"dd415ccf-2b4a-4797-962f-a464ef96bc22\") " pod="openshift-marketplace/community-operators-jwrpk" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.020657 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd415ccf-2b4a-4797-962f-a464ef96bc22-catalog-content\") pod \"community-operators-jwrpk\" (UID: \"dd415ccf-2b4a-4797-962f-a464ef96bc22\") " pod="openshift-marketplace/community-operators-jwrpk" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.020707 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.020765 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67q8x\" (UniqueName: \"kubernetes.io/projected/dd415ccf-2b4a-4797-962f-a464ef96bc22-kube-api-access-67q8x\") pod \"community-operators-jwrpk\" (UID: \"dd415ccf-2b4a-4797-962f-a464ef96bc22\") " pod="openshift-marketplace/community-operators-jwrpk" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.020967 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cf6v7" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.021575 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd415ccf-2b4a-4797-962f-a464ef96bc22-utilities\") pod \"community-operators-jwrpk\" (UID: \"dd415ccf-2b4a-4797-962f-a464ef96bc22\") " pod="openshift-marketplace/community-operators-jwrpk" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.021594 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd415ccf-2b4a-4797-962f-a464ef96bc22-catalog-content\") pod \"community-operators-jwrpk\" (UID: \"dd415ccf-2b4a-4797-962f-a464ef96bc22\") " pod="openshift-marketplace/community-operators-jwrpk" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.024943 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9wdps" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.027819 4772 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.027853 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.044941 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67q8x\" (UniqueName: \"kubernetes.io/projected/dd415ccf-2b4a-4797-962f-a464ef96bc22-kube-api-access-67q8x\") pod \"community-operators-jwrpk\" (UID: \"dd415ccf-2b4a-4797-962f-a464ef96bc22\") " pod="openshift-marketplace/community-operators-jwrpk" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.094872 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-crlcr\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.114295 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wcldz"] Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.116415 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wcldz" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.162775 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wcldz"] Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.208986 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-95rh9"] Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.216007 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jwrpk" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.223806 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/987488b4-af4d-4b20-bb26-f433d4d1299a-utilities\") pod \"certified-operators-wcldz\" (UID: \"987488b4-af4d-4b20-bb26-f433d4d1299a\") " pod="openshift-marketplace/certified-operators-wcldz" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.223897 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp9qx\" (UniqueName: \"kubernetes.io/projected/987488b4-af4d-4b20-bb26-f433d4d1299a-kube-api-access-sp9qx\") pod \"certified-operators-wcldz\" (UID: \"987488b4-af4d-4b20-bb26-f433d4d1299a\") " pod="openshift-marketplace/certified-operators-wcldz" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.224013 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/987488b4-af4d-4b20-bb26-f433d4d1299a-catalog-content\") pod \"certified-operators-wcldz\" (UID: \"987488b4-af4d-4b20-bb26-f433d4d1299a\") " pod="openshift-marketplace/certified-operators-wcldz" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.288935 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"12b334f2ce60be8c17d4eadace35220018c7dd18ca877512f3960dae41a53451"} Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.288978 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"c6ced8694bd7595fee00b262ac36c45bd68f29fbefa1dc4a8eaae1198c1d0493"} Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.292055 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"065902ad26e91894adf3e424b61f04e80232696bd334ca7b0c2dadf70f3db0ea"} Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.292088 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"25ed67dd4363c69e0895784cd995321f5be2d7e99c48536df754953077e3f64b"} Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.292283 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.294523 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" event={"ID":"794cdbb9-3392-465a-8a0a-a78a465aee2b","Type":"ContainerStarted","Data":"00440aa18fc91af0074763f2776da1676a6e036d0e6f05e019e4e01bcb358a23"} Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.305614 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-95rh9" event={"ID":"4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7","Type":"ContainerStarted","Data":"ae198d4139eb016b136b591cf513d4e1d588e78f4cc1966851a08fad44048adb"} Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.311357 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"7f3417080ebdeac18b7818cfe03adc19876ba742c9b98f9f66e8089e83a729af"} Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.311403 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"600800860d1d7b700cbc19d04957ce86b9999ffdbabf36c53b619fc569f4ba28"} Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.324980 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/987488b4-af4d-4b20-bb26-f433d4d1299a-catalog-content\") pod \"certified-operators-wcldz\" (UID: \"987488b4-af4d-4b20-bb26-f433d4d1299a\") " pod="openshift-marketplace/certified-operators-wcldz" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.325049 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/987488b4-af4d-4b20-bb26-f433d4d1299a-utilities\") pod \"certified-operators-wcldz\" (UID: \"987488b4-af4d-4b20-bb26-f433d4d1299a\") " pod="openshift-marketplace/certified-operators-wcldz" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.325102 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp9qx\" (UniqueName: \"kubernetes.io/projected/987488b4-af4d-4b20-bb26-f433d4d1299a-kube-api-access-sp9qx\") pod \"certified-operators-wcldz\" (UID: \"987488b4-af4d-4b20-bb26-f433d4d1299a\") " pod="openshift-marketplace/certified-operators-wcldz" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.326314 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/987488b4-af4d-4b20-bb26-f433d4d1299a-catalog-content\") pod \"certified-operators-wcldz\" (UID: \"987488b4-af4d-4b20-bb26-f433d4d1299a\") " pod="openshift-marketplace/certified-operators-wcldz" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.326623 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/987488b4-af4d-4b20-bb26-f433d4d1299a-utilities\") pod \"certified-operators-wcldz\" (UID: \"987488b4-af4d-4b20-bb26-f433d4d1299a\") " pod="openshift-marketplace/certified-operators-wcldz" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.364250 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp9qx\" (UniqueName: \"kubernetes.io/projected/987488b4-af4d-4b20-bb26-f433d4d1299a-kube-api-access-sp9qx\") pod \"certified-operators-wcldz\" (UID: \"987488b4-af4d-4b20-bb26-f433d4d1299a\") " pod="openshift-marketplace/certified-operators-wcldz" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.382115 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-hbbxh" podStartSLOduration=11.382087837 podStartE2EDuration="11.382087837s" podCreationTimestamp="2026-01-27 15:09:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:24.351372211 +0000 UTC m=+150.331981299" watchObservedRunningTime="2026-01-27 15:09:24.382087837 +0000 UTC m=+150.362696945" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.394914 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.477543 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wcldz" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.519981 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9wdps"] Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.602803 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jwrpk"] Jan 27 15:09:24 crc kubenswrapper[4772]: W0127 15:09:24.630832 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd415ccf_2b4a_4797_962f_a464ef96bc22.slice/crio-625e8aba305d47657203f1b68ee02b451e267b67921c5b887eaa601500155d6c WatchSource:0}: Error finding container 625e8aba305d47657203f1b68ee02b451e267b67921c5b887eaa601500155d6c: Status 404 returned error can't find the container with id 625e8aba305d47657203f1b68ee02b451e267b67921c5b887eaa601500155d6c Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.652199 4772 patch_prober.go:28] interesting pod/router-default-5444994796-7k7sg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 27 15:09:24 crc kubenswrapper[4772]: [-]has-synced failed: reason withheld Jan 27 15:09:24 crc kubenswrapper[4772]: [+]process-running ok Jan 27 15:09:24 crc kubenswrapper[4772]: healthz check failed Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.652265 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7k7sg" podUID="c4225ddc-bdcd-4158-811b-113234d0c3d0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.676488 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.696216 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-crlcr"] Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.923469 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.924647 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.926100 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.931660 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.931826 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 27 15:09:24 crc kubenswrapper[4772]: I0127 15:09:24.999728 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wcldz"] Jan 27 15:09:25 crc kubenswrapper[4772]: W0127 15:09:25.001127 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod987488b4_af4d_4b20_bb26_f433d4d1299a.slice/crio-cca259e2810aba1b9b6e47d70088b4b43bd08d57432f8fbb61e7ddcd0a7abb94 WatchSource:0}: Error finding container cca259e2810aba1b9b6e47d70088b4b43bd08d57432f8fbb61e7ddcd0a7abb94: Status 404 returned error can't find the container with id cca259e2810aba1b9b6e47d70088b4b43bd08d57432f8fbb61e7ddcd0a7abb94 Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.046035 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53a6c4ad-816a-4d22-af98-0587a6a68304-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"53a6c4ad-816a-4d22-af98-0587a6a68304\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.046080 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53a6c4ad-816a-4d22-af98-0587a6a68304-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"53a6c4ad-816a-4d22-af98-0587a6a68304\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.147768 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53a6c4ad-816a-4d22-af98-0587a6a68304-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"53a6c4ad-816a-4d22-af98-0587a6a68304\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.147845 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53a6c4ad-816a-4d22-af98-0587a6a68304-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"53a6c4ad-816a-4d22-af98-0587a6a68304\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.147917 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53a6c4ad-816a-4d22-af98-0587a6a68304-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"53a6c4ad-816a-4d22-af98-0587a6a68304\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.168519 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.176934 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53a6c4ad-816a-4d22-af98-0587a6a68304-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"53a6c4ad-816a-4d22-af98-0587a6a68304\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.180825 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-2h2z8" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.248914 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.348251 4772 generic.go:334] "Generic (PLEG): container finished" podID="c6b54ae2-d365-4988-8e69-704574c7962a" containerID="4f5ed02624877f82608d4a7a5fead892a80497d0b63bf729eaa6c0d56cf6aac6" exitCode=0 Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.348365 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492100-r2zj6" event={"ID":"c6b54ae2-d365-4988-8e69-704574c7962a","Type":"ContainerDied","Data":"4f5ed02624877f82608d4a7a5fead892a80497d0b63bf729eaa6c0d56cf6aac6"} Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.363097 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wcldz" event={"ID":"987488b4-af4d-4b20-bb26-f433d4d1299a","Type":"ContainerStarted","Data":"cca259e2810aba1b9b6e47d70088b4b43bd08d57432f8fbb61e7ddcd0a7abb94"} Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.375259 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" event={"ID":"877de785-bc18-4c1c-970a-1e6533539467","Type":"ContainerStarted","Data":"a3585a039b9cbf60a67ac7ced2eaf947fce2a88abe7705503eb446ef5ad9fc74"} Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.376353 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.378177 4772 generic.go:334] "Generic (PLEG): container finished" podID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" containerID="16478a3399ae464f44f2335d6adb9e9f0cad87f3f55187da1906c5dcf87534dd" exitCode=0 Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.378228 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-95rh9" event={"ID":"4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7","Type":"ContainerDied","Data":"16478a3399ae464f44f2335d6adb9e9f0cad87f3f55187da1906c5dcf87534dd"} Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.379957 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.381337 4772 generic.go:334] "Generic (PLEG): container finished" podID="96e88efd-1f25-4e44-b459-ab773db93656" containerID="19447f4e7cf721e9d74fba0b607574c2c8311dd4da0f4f4a1ba431d8c2a3ca81" exitCode=0 Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.381388 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9wdps" event={"ID":"96e88efd-1f25-4e44-b459-ab773db93656","Type":"ContainerDied","Data":"19447f4e7cf721e9d74fba0b607574c2c8311dd4da0f4f4a1ba431d8c2a3ca81"} Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.381411 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9wdps" event={"ID":"96e88efd-1f25-4e44-b459-ab773db93656","Type":"ContainerStarted","Data":"806bc56d016ac75a91f0a1effbd4a1494b65e83f8a957a34cd38d253ad927cc3"} Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.387251 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwrpk" event={"ID":"dd415ccf-2b4a-4797-962f-a464ef96bc22","Type":"ContainerStarted","Data":"625e8aba305d47657203f1b68ee02b451e267b67921c5b887eaa601500155d6c"} Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.419540 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" podStartSLOduration=129.419512406 podStartE2EDuration="2m9.419512406s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:25.39607262 +0000 UTC m=+151.376681718" watchObservedRunningTime="2026-01-27 15:09:25.419512406 +0000 UTC m=+151.400121504" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.443072 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.445154 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.448674 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.452901 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.477704 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.555883 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/545a0fd1-38f1-4cbc-9f37-1870b1673589-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"545a0fd1-38f1-4cbc-9f37-1870b1673589\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.555961 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/545a0fd1-38f1-4cbc-9f37-1870b1673589-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"545a0fd1-38f1-4cbc-9f37-1870b1673589\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.595088 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.656320 4772 patch_prober.go:28] interesting pod/router-default-5444994796-7k7sg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 27 15:09:25 crc kubenswrapper[4772]: [-]has-synced failed: reason withheld Jan 27 15:09:25 crc kubenswrapper[4772]: [+]process-running ok Jan 27 15:09:25 crc kubenswrapper[4772]: healthz check failed Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.656527 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7k7sg" podUID="c4225ddc-bdcd-4158-811b-113234d0c3d0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.657294 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/545a0fd1-38f1-4cbc-9f37-1870b1673589-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"545a0fd1-38f1-4cbc-9f37-1870b1673589\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.657374 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/545a0fd1-38f1-4cbc-9f37-1870b1673589-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"545a0fd1-38f1-4cbc-9f37-1870b1673589\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.657436 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/545a0fd1-38f1-4cbc-9f37-1870b1673589-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"545a0fd1-38f1-4cbc-9f37-1870b1673589\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.691610 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xp8ph"] Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.698460 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xp8ph" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.704118 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/545a0fd1-38f1-4cbc-9f37-1870b1673589-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"545a0fd1-38f1-4cbc-9f37-1870b1673589\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.706766 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.707375 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xp8ph"] Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.758532 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac2b5800-ce98-4847-bfcd-67a97375aa1b-utilities\") pod \"redhat-marketplace-xp8ph\" (UID: \"ac2b5800-ce98-4847-bfcd-67a97375aa1b\") " pod="openshift-marketplace/redhat-marketplace-xp8ph" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.758586 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac2b5800-ce98-4847-bfcd-67a97375aa1b-catalog-content\") pod \"redhat-marketplace-xp8ph\" (UID: \"ac2b5800-ce98-4847-bfcd-67a97375aa1b\") " pod="openshift-marketplace/redhat-marketplace-xp8ph" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.758616 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g6kk\" (UniqueName: \"kubernetes.io/projected/ac2b5800-ce98-4847-bfcd-67a97375aa1b-kube-api-access-7g6kk\") pod \"redhat-marketplace-xp8ph\" (UID: \"ac2b5800-ce98-4847-bfcd-67a97375aa1b\") " pod="openshift-marketplace/redhat-marketplace-xp8ph" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.789553 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.859814 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac2b5800-ce98-4847-bfcd-67a97375aa1b-utilities\") pod \"redhat-marketplace-xp8ph\" (UID: \"ac2b5800-ce98-4847-bfcd-67a97375aa1b\") " pod="openshift-marketplace/redhat-marketplace-xp8ph" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.859882 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac2b5800-ce98-4847-bfcd-67a97375aa1b-catalog-content\") pod \"redhat-marketplace-xp8ph\" (UID: \"ac2b5800-ce98-4847-bfcd-67a97375aa1b\") " pod="openshift-marketplace/redhat-marketplace-xp8ph" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.859923 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g6kk\" (UniqueName: \"kubernetes.io/projected/ac2b5800-ce98-4847-bfcd-67a97375aa1b-kube-api-access-7g6kk\") pod \"redhat-marketplace-xp8ph\" (UID: \"ac2b5800-ce98-4847-bfcd-67a97375aa1b\") " pod="openshift-marketplace/redhat-marketplace-xp8ph" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.860682 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac2b5800-ce98-4847-bfcd-67a97375aa1b-utilities\") pod \"redhat-marketplace-xp8ph\" (UID: \"ac2b5800-ce98-4847-bfcd-67a97375aa1b\") " pod="openshift-marketplace/redhat-marketplace-xp8ph" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.860884 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac2b5800-ce98-4847-bfcd-67a97375aa1b-catalog-content\") pod \"redhat-marketplace-xp8ph\" (UID: \"ac2b5800-ce98-4847-bfcd-67a97375aa1b\") " pod="openshift-marketplace/redhat-marketplace-xp8ph" Jan 27 15:09:25 crc kubenswrapper[4772]: I0127 15:09:25.887825 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g6kk\" (UniqueName: \"kubernetes.io/projected/ac2b5800-ce98-4847-bfcd-67a97375aa1b-kube-api-access-7g6kk\") pod \"redhat-marketplace-xp8ph\" (UID: \"ac2b5800-ce98-4847-bfcd-67a97375aa1b\") " pod="openshift-marketplace/redhat-marketplace-xp8ph" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.024809 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xp8ph" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.070959 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dfgjh"] Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.072095 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dfgjh" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.120628 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dfgjh"] Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.163981 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cbabfa8-79d8-4b23-b186-b40ba8b3017e-utilities\") pod \"redhat-marketplace-dfgjh\" (UID: \"8cbabfa8-79d8-4b23-b186-b40ba8b3017e\") " pod="openshift-marketplace/redhat-marketplace-dfgjh" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.164097 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cbabfa8-79d8-4b23-b186-b40ba8b3017e-catalog-content\") pod \"redhat-marketplace-dfgjh\" (UID: \"8cbabfa8-79d8-4b23-b186-b40ba8b3017e\") " pod="openshift-marketplace/redhat-marketplace-dfgjh" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.164153 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87w6q\" (UniqueName: \"kubernetes.io/projected/8cbabfa8-79d8-4b23-b186-b40ba8b3017e-kube-api-access-87w6q\") pod \"redhat-marketplace-dfgjh\" (UID: \"8cbabfa8-79d8-4b23-b186-b40ba8b3017e\") " pod="openshift-marketplace/redhat-marketplace-dfgjh" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.206044 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.238205 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xp8ph"] Jan 27 15:09:26 crc kubenswrapper[4772]: W0127 15:09:26.251242 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac2b5800_ce98_4847_bfcd_67a97375aa1b.slice/crio-d8c8cf461f8b99ac3badb881b8d4938b4c0d57c4110c08a8b732cf718594c112 WatchSource:0}: Error finding container d8c8cf461f8b99ac3badb881b8d4938b4c0d57c4110c08a8b732cf718594c112: Status 404 returned error can't find the container with id d8c8cf461f8b99ac3badb881b8d4938b4c0d57c4110c08a8b732cf718594c112 Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.265096 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cbabfa8-79d8-4b23-b186-b40ba8b3017e-utilities\") pod \"redhat-marketplace-dfgjh\" (UID: \"8cbabfa8-79d8-4b23-b186-b40ba8b3017e\") " pod="openshift-marketplace/redhat-marketplace-dfgjh" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.265179 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cbabfa8-79d8-4b23-b186-b40ba8b3017e-catalog-content\") pod \"redhat-marketplace-dfgjh\" (UID: \"8cbabfa8-79d8-4b23-b186-b40ba8b3017e\") " pod="openshift-marketplace/redhat-marketplace-dfgjh" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.265244 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87w6q\" (UniqueName: \"kubernetes.io/projected/8cbabfa8-79d8-4b23-b186-b40ba8b3017e-kube-api-access-87w6q\") pod \"redhat-marketplace-dfgjh\" (UID: \"8cbabfa8-79d8-4b23-b186-b40ba8b3017e\") " pod="openshift-marketplace/redhat-marketplace-dfgjh" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.265620 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cbabfa8-79d8-4b23-b186-b40ba8b3017e-utilities\") pod \"redhat-marketplace-dfgjh\" (UID: \"8cbabfa8-79d8-4b23-b186-b40ba8b3017e\") " pod="openshift-marketplace/redhat-marketplace-dfgjh" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.274343 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cbabfa8-79d8-4b23-b186-b40ba8b3017e-catalog-content\") pod \"redhat-marketplace-dfgjh\" (UID: \"8cbabfa8-79d8-4b23-b186-b40ba8b3017e\") " pod="openshift-marketplace/redhat-marketplace-dfgjh" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.283406 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87w6q\" (UniqueName: \"kubernetes.io/projected/8cbabfa8-79d8-4b23-b186-b40ba8b3017e-kube-api-access-87w6q\") pod \"redhat-marketplace-dfgjh\" (UID: \"8cbabfa8-79d8-4b23-b186-b40ba8b3017e\") " pod="openshift-marketplace/redhat-marketplace-dfgjh" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.389070 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dfgjh" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.396650 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xp8ph" event={"ID":"ac2b5800-ce98-4847-bfcd-67a97375aa1b","Type":"ContainerStarted","Data":"d8c8cf461f8b99ac3badb881b8d4938b4c0d57c4110c08a8b732cf718594c112"} Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.398777 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" event={"ID":"877de785-bc18-4c1c-970a-1e6533539467","Type":"ContainerStarted","Data":"228e6fd0668bf433c1f6aa09021f79564dfe5e7bb750301de0ab0cbfce9f1ef2"} Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.400904 4772 generic.go:334] "Generic (PLEG): container finished" podID="dd415ccf-2b4a-4797-962f-a464ef96bc22" containerID="759437d05ef598aec5d4669f7ffea07fc52730444984e390ed6235fe2f84e271" exitCode=0 Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.401084 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwrpk" event={"ID":"dd415ccf-2b4a-4797-962f-a464ef96bc22","Type":"ContainerDied","Data":"759437d05ef598aec5d4669f7ffea07fc52730444984e390ed6235fe2f84e271"} Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.407981 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"53a6c4ad-816a-4d22-af98-0587a6a68304","Type":"ContainerStarted","Data":"63104e6a47ff51e625e55ab363bafa9155dc299b2880304fb3771d251df37ba5"} Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.408018 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"53a6c4ad-816a-4d22-af98-0587a6a68304","Type":"ContainerStarted","Data":"4e5fb5c0a561c986217a03a765b2818013365a6cf31c0cb2523e728110829fe1"} Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.412917 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"545a0fd1-38f1-4cbc-9f37-1870b1673589","Type":"ContainerStarted","Data":"fb9d01e0f3553308e4bfd0ea1eb0d0086a446ca7d54e19f2aad89776bdcb2b2b"} Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.419024 4772 generic.go:334] "Generic (PLEG): container finished" podID="987488b4-af4d-4b20-bb26-f433d4d1299a" containerID="05b1bd7fef5819af5a4449e0a65b00c0b94c405308dc0ea03120ac4091e22a7a" exitCode=0 Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.420767 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wcldz" event={"ID":"987488b4-af4d-4b20-bb26-f433d4d1299a","Type":"ContainerDied","Data":"05b1bd7fef5819af5a4449e0a65b00c0b94c405308dc0ea03120ac4091e22a7a"} Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.439254 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.439229263 podStartE2EDuration="2.439229263s" podCreationTimestamp="2026-01-27 15:09:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:09:26.431342785 +0000 UTC m=+152.411951903" watchObservedRunningTime="2026-01-27 15:09:26.439229263 +0000 UTC m=+152.419838361" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.540614 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.541001 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.559324 4772 patch_prober.go:28] interesting pod/console-f9d7485db-7qfrl container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.559387 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-7qfrl" podUID="e2e31e5f-3a41-42f5-90b0-99c05a8033a6" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.592221 4772 patch_prober.go:28] interesting pod/downloads-7954f5f757-vswtw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.592296 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vswtw" podUID="17bdd07d-f7e5-47f8-b730-724d5cc8e3d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.592480 4772 patch_prober.go:28] interesting pod/downloads-7954f5f757-vswtw container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.592533 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-vswtw" podUID="17bdd07d-f7e5-47f8-b730-724d5cc8e3d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.647627 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-7k7sg" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.672694 4772 patch_prober.go:28] interesting pod/router-default-5444994796-7k7sg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 27 15:09:26 crc kubenswrapper[4772]: [-]has-synced failed: reason withheld Jan 27 15:09:26 crc kubenswrapper[4772]: [+]process-running ok Jan 27 15:09:26 crc kubenswrapper[4772]: healthz check failed Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.672752 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7k7sg" podUID="c4225ddc-bdcd-4158-811b-113234d0c3d0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.684854 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-75jrg"] Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.687421 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-75jrg" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.693621 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.696402 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-75jrg"] Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.761776 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492100-r2zj6" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.807565 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dfgjh"] Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.878577 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6b54ae2-d365-4988-8e69-704574c7962a-config-volume\") pod \"c6b54ae2-d365-4988-8e69-704574c7962a\" (UID: \"c6b54ae2-d365-4988-8e69-704574c7962a\") " Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.878652 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6b54ae2-d365-4988-8e69-704574c7962a-secret-volume\") pod \"c6b54ae2-d365-4988-8e69-704574c7962a\" (UID: \"c6b54ae2-d365-4988-8e69-704574c7962a\") " Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.878704 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llq59\" (UniqueName: \"kubernetes.io/projected/c6b54ae2-d365-4988-8e69-704574c7962a-kube-api-access-llq59\") pod \"c6b54ae2-d365-4988-8e69-704574c7962a\" (UID: \"c6b54ae2-d365-4988-8e69-704574c7962a\") " Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.879083 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvxxl\" (UniqueName: \"kubernetes.io/projected/f637b998-b13b-486d-9042-4cd40a01c833-kube-api-access-vvxxl\") pod \"redhat-operators-75jrg\" (UID: \"f637b998-b13b-486d-9042-4cd40a01c833\") " pod="openshift-marketplace/redhat-operators-75jrg" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.879158 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f637b998-b13b-486d-9042-4cd40a01c833-utilities\") pod \"redhat-operators-75jrg\" (UID: \"f637b998-b13b-486d-9042-4cd40a01c833\") " pod="openshift-marketplace/redhat-operators-75jrg" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.879262 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f637b998-b13b-486d-9042-4cd40a01c833-catalog-content\") pod \"redhat-operators-75jrg\" (UID: \"f637b998-b13b-486d-9042-4cd40a01c833\") " pod="openshift-marketplace/redhat-operators-75jrg" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.881497 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6b54ae2-d365-4988-8e69-704574c7962a-config-volume" (OuterVolumeSpecName: "config-volume") pod "c6b54ae2-d365-4988-8e69-704574c7962a" (UID: "c6b54ae2-d365-4988-8e69-704574c7962a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.886523 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6b54ae2-d365-4988-8e69-704574c7962a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c6b54ae2-d365-4988-8e69-704574c7962a" (UID: "c6b54ae2-d365-4988-8e69-704574c7962a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.886675 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6b54ae2-d365-4988-8e69-704574c7962a-kube-api-access-llq59" (OuterVolumeSpecName: "kube-api-access-llq59") pod "c6b54ae2-d365-4988-8e69-704574c7962a" (UID: "c6b54ae2-d365-4988-8e69-704574c7962a"). InnerVolumeSpecName "kube-api-access-llq59". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.981376 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f637b998-b13b-486d-9042-4cd40a01c833-catalog-content\") pod \"redhat-operators-75jrg\" (UID: \"f637b998-b13b-486d-9042-4cd40a01c833\") " pod="openshift-marketplace/redhat-operators-75jrg" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.981439 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvxxl\" (UniqueName: \"kubernetes.io/projected/f637b998-b13b-486d-9042-4cd40a01c833-kube-api-access-vvxxl\") pod \"redhat-operators-75jrg\" (UID: \"f637b998-b13b-486d-9042-4cd40a01c833\") " pod="openshift-marketplace/redhat-operators-75jrg" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.981487 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f637b998-b13b-486d-9042-4cd40a01c833-utilities\") pod \"redhat-operators-75jrg\" (UID: \"f637b998-b13b-486d-9042-4cd40a01c833\") " pod="openshift-marketplace/redhat-operators-75jrg" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.981532 4772 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6b54ae2-d365-4988-8e69-704574c7962a-config-volume\") on node \"crc\" DevicePath \"\"" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.981544 4772 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c6b54ae2-d365-4988-8e69-704574c7962a-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.981553 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llq59\" (UniqueName: \"kubernetes.io/projected/c6b54ae2-d365-4988-8e69-704574c7962a-kube-api-access-llq59\") on node \"crc\" DevicePath \"\"" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.981927 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f637b998-b13b-486d-9042-4cd40a01c833-utilities\") pod \"redhat-operators-75jrg\" (UID: \"f637b998-b13b-486d-9042-4cd40a01c833\") " pod="openshift-marketplace/redhat-operators-75jrg" Jan 27 15:09:26 crc kubenswrapper[4772]: I0127 15:09:26.982142 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f637b998-b13b-486d-9042-4cd40a01c833-catalog-content\") pod \"redhat-operators-75jrg\" (UID: \"f637b998-b13b-486d-9042-4cd40a01c833\") " pod="openshift-marketplace/redhat-operators-75jrg" Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.002421 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvxxl\" (UniqueName: \"kubernetes.io/projected/f637b998-b13b-486d-9042-4cd40a01c833-kube-api-access-vvxxl\") pod \"redhat-operators-75jrg\" (UID: \"f637b998-b13b-486d-9042-4cd40a01c833\") " pod="openshift-marketplace/redhat-operators-75jrg" Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.029108 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-75jrg" Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.069245 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jdcpn" Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.073392 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-k7pfr"] Jan 27 15:09:27 crc kubenswrapper[4772]: E0127 15:09:27.073613 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6b54ae2-d365-4988-8e69-704574c7962a" containerName="collect-profiles" Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.073627 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6b54ae2-d365-4988-8e69-704574c7962a" containerName="collect-profiles" Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.073734 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6b54ae2-d365-4988-8e69-704574c7962a" containerName="collect-profiles" Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.074538 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k7pfr" Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.083335 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnhcz\" (UniqueName: \"kubernetes.io/projected/d0b33686-8107-4caf-b67f-3c608119a049-kube-api-access-fnhcz\") pod \"redhat-operators-k7pfr\" (UID: \"d0b33686-8107-4caf-b67f-3c608119a049\") " pod="openshift-marketplace/redhat-operators-k7pfr" Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.083401 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0b33686-8107-4caf-b67f-3c608119a049-utilities\") pod \"redhat-operators-k7pfr\" (UID: \"d0b33686-8107-4caf-b67f-3c608119a049\") " pod="openshift-marketplace/redhat-operators-k7pfr" Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.083425 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0b33686-8107-4caf-b67f-3c608119a049-catalog-content\") pod \"redhat-operators-k7pfr\" (UID: \"d0b33686-8107-4caf-b67f-3c608119a049\") " pod="openshift-marketplace/redhat-operators-k7pfr" Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.095201 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k7pfr"] Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.148492 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-mnltb" Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.184109 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0b33686-8107-4caf-b67f-3c608119a049-catalog-content\") pod \"redhat-operators-k7pfr\" (UID: \"d0b33686-8107-4caf-b67f-3c608119a049\") " pod="openshift-marketplace/redhat-operators-k7pfr" Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.184501 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnhcz\" (UniqueName: \"kubernetes.io/projected/d0b33686-8107-4caf-b67f-3c608119a049-kube-api-access-fnhcz\") pod \"redhat-operators-k7pfr\" (UID: \"d0b33686-8107-4caf-b67f-3c608119a049\") " pod="openshift-marketplace/redhat-operators-k7pfr" Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.184570 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0b33686-8107-4caf-b67f-3c608119a049-utilities\") pod \"redhat-operators-k7pfr\" (UID: \"d0b33686-8107-4caf-b67f-3c608119a049\") " pod="openshift-marketplace/redhat-operators-k7pfr" Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.184958 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0b33686-8107-4caf-b67f-3c608119a049-utilities\") pod \"redhat-operators-k7pfr\" (UID: \"d0b33686-8107-4caf-b67f-3c608119a049\") " pod="openshift-marketplace/redhat-operators-k7pfr" Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.185161 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0b33686-8107-4caf-b67f-3c608119a049-catalog-content\") pod \"redhat-operators-k7pfr\" (UID: \"d0b33686-8107-4caf-b67f-3c608119a049\") " pod="openshift-marketplace/redhat-operators-k7pfr" Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.222372 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnhcz\" (UniqueName: \"kubernetes.io/projected/d0b33686-8107-4caf-b67f-3c608119a049-kube-api-access-fnhcz\") pod \"redhat-operators-k7pfr\" (UID: \"d0b33686-8107-4caf-b67f-3c608119a049\") " pod="openshift-marketplace/redhat-operators-k7pfr" Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.345598 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.391849 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k7pfr" Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.427979 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492100-r2zj6" event={"ID":"c6b54ae2-d365-4988-8e69-704574c7962a","Type":"ContainerDied","Data":"7de1df4376bfe65e7f653cd434cd6a00a28483c62ada492613581e939925776c"} Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.428020 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7de1df4376bfe65e7f653cd434cd6a00a28483c62ada492613581e939925776c" Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.428127 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492100-r2zj6" Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.444772 4772 generic.go:334] "Generic (PLEG): container finished" podID="53a6c4ad-816a-4d22-af98-0587a6a68304" containerID="63104e6a47ff51e625e55ab363bafa9155dc299b2880304fb3771d251df37ba5" exitCode=0 Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.444830 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"53a6c4ad-816a-4d22-af98-0587a6a68304","Type":"ContainerDied","Data":"63104e6a47ff51e625e55ab363bafa9155dc299b2880304fb3771d251df37ba5"} Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.447350 4772 generic.go:334] "Generic (PLEG): container finished" podID="545a0fd1-38f1-4cbc-9f37-1870b1673589" containerID="4635120e4ed4dd9ae851ebf881fd2e955918cdec5e9f2fb5b5236a34009790a9" exitCode=0 Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.447377 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"545a0fd1-38f1-4cbc-9f37-1870b1673589","Type":"ContainerDied","Data":"4635120e4ed4dd9ae851ebf881fd2e955918cdec5e9f2fb5b5236a34009790a9"} Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.475426 4772 generic.go:334] "Generic (PLEG): container finished" podID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" containerID="2b642630dd7f7b63f30ba841d8958c4ae79e62858aee5dff568bed444b47b036" exitCode=0 Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.475558 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xp8ph" event={"ID":"ac2b5800-ce98-4847-bfcd-67a97375aa1b","Type":"ContainerDied","Data":"2b642630dd7f7b63f30ba841d8958c4ae79e62858aee5dff568bed444b47b036"} Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.490405 4772 generic.go:334] "Generic (PLEG): container finished" podID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" containerID="e90d644c15f4a502d49563577bfa11dc77829d65c3be871a6762542c2dc18bcb" exitCode=0 Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.491612 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dfgjh" event={"ID":"8cbabfa8-79d8-4b23-b186-b40ba8b3017e","Type":"ContainerDied","Data":"e90d644c15f4a502d49563577bfa11dc77829d65c3be871a6762542c2dc18bcb"} Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.491653 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dfgjh" event={"ID":"8cbabfa8-79d8-4b23-b186-b40ba8b3017e","Type":"ContainerStarted","Data":"a43ec03d3d30e58e4b1f455c9ffe4cf357362515c513668efc22ad8bedf315c3"} Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.650835 4772 patch_prober.go:28] interesting pod/router-default-5444994796-7k7sg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 27 15:09:27 crc kubenswrapper[4772]: [-]has-synced failed: reason withheld Jan 27 15:09:27 crc kubenswrapper[4772]: [+]process-running ok Jan 27 15:09:27 crc kubenswrapper[4772]: healthz check failed Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.651239 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7k7sg" podUID="c4225ddc-bdcd-4158-811b-113234d0c3d0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.652306 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-75jrg"] Jan 27 15:09:27 crc kubenswrapper[4772]: W0127 15:09:27.697560 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf637b998_b13b_486d_9042_4cd40a01c833.slice/crio-e6520bb81fcb94747f96d0bde0c4f99cc40f51207c4752c3b45e79ef35a202b3 WatchSource:0}: Error finding container e6520bb81fcb94747f96d0bde0c4f99cc40f51207c4752c3b45e79ef35a202b3: Status 404 returned error can't find the container with id e6520bb81fcb94747f96d0bde0c4f99cc40f51207c4752c3b45e79ef35a202b3 Jan 27 15:09:27 crc kubenswrapper[4772]: I0127 15:09:27.700827 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k7pfr"] Jan 27 15:09:28 crc kubenswrapper[4772]: I0127 15:09:28.499156 4772 generic.go:334] "Generic (PLEG): container finished" podID="d0b33686-8107-4caf-b67f-3c608119a049" containerID="142c7ce0c1e97146a8a91a5cb46adbe2ee5537497547ac756fc38abfd7afe96c" exitCode=0 Jan 27 15:09:28 crc kubenswrapper[4772]: I0127 15:09:28.499219 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7pfr" event={"ID":"d0b33686-8107-4caf-b67f-3c608119a049","Type":"ContainerDied","Data":"142c7ce0c1e97146a8a91a5cb46adbe2ee5537497547ac756fc38abfd7afe96c"} Jan 27 15:09:28 crc kubenswrapper[4772]: I0127 15:09:28.499487 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7pfr" event={"ID":"d0b33686-8107-4caf-b67f-3c608119a049","Type":"ContainerStarted","Data":"73b204aeb508c9b52f9299efc082b371c9a1b4cb36fb76b5cd3cb0d31f443821"} Jan 27 15:09:28 crc kubenswrapper[4772]: I0127 15:09:28.501654 4772 generic.go:334] "Generic (PLEG): container finished" podID="f637b998-b13b-486d-9042-4cd40a01c833" containerID="980163ea21706e95ea0803e7c47d4cf7427d498f1351e03fac04539f250bddc6" exitCode=0 Jan 27 15:09:28 crc kubenswrapper[4772]: I0127 15:09:28.501737 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75jrg" event={"ID":"f637b998-b13b-486d-9042-4cd40a01c833","Type":"ContainerDied","Data":"980163ea21706e95ea0803e7c47d4cf7427d498f1351e03fac04539f250bddc6"} Jan 27 15:09:28 crc kubenswrapper[4772]: I0127 15:09:28.501763 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75jrg" event={"ID":"f637b998-b13b-486d-9042-4cd40a01c833","Type":"ContainerStarted","Data":"e6520bb81fcb94747f96d0bde0c4f99cc40f51207c4752c3b45e79ef35a202b3"} Jan 27 15:09:28 crc kubenswrapper[4772]: I0127 15:09:28.651218 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-7k7sg" Jan 27 15:09:28 crc kubenswrapper[4772]: I0127 15:09:28.653525 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-7k7sg" Jan 27 15:09:28 crc kubenswrapper[4772]: I0127 15:09:28.852080 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 27 15:09:28 crc kubenswrapper[4772]: I0127 15:09:28.988469 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 27 15:09:29 crc kubenswrapper[4772]: I0127 15:09:29.034559 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53a6c4ad-816a-4d22-af98-0587a6a68304-kubelet-dir\") pod \"53a6c4ad-816a-4d22-af98-0587a6a68304\" (UID: \"53a6c4ad-816a-4d22-af98-0587a6a68304\") " Jan 27 15:09:29 crc kubenswrapper[4772]: I0127 15:09:29.034774 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53a6c4ad-816a-4d22-af98-0587a6a68304-kube-api-access\") pod \"53a6c4ad-816a-4d22-af98-0587a6a68304\" (UID: \"53a6c4ad-816a-4d22-af98-0587a6a68304\") " Jan 27 15:09:29 crc kubenswrapper[4772]: I0127 15:09:29.035282 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53a6c4ad-816a-4d22-af98-0587a6a68304-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "53a6c4ad-816a-4d22-af98-0587a6a68304" (UID: "53a6c4ad-816a-4d22-af98-0587a6a68304"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:09:29 crc kubenswrapper[4772]: I0127 15:09:29.043556 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53a6c4ad-816a-4d22-af98-0587a6a68304-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "53a6c4ad-816a-4d22-af98-0587a6a68304" (UID: "53a6c4ad-816a-4d22-af98-0587a6a68304"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:09:29 crc kubenswrapper[4772]: I0127 15:09:29.135565 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/545a0fd1-38f1-4cbc-9f37-1870b1673589-kube-api-access\") pod \"545a0fd1-38f1-4cbc-9f37-1870b1673589\" (UID: \"545a0fd1-38f1-4cbc-9f37-1870b1673589\") " Jan 27 15:09:29 crc kubenswrapper[4772]: I0127 15:09:29.135754 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/545a0fd1-38f1-4cbc-9f37-1870b1673589-kubelet-dir\") pod \"545a0fd1-38f1-4cbc-9f37-1870b1673589\" (UID: \"545a0fd1-38f1-4cbc-9f37-1870b1673589\") " Jan 27 15:09:29 crc kubenswrapper[4772]: I0127 15:09:29.136055 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/53a6c4ad-816a-4d22-af98-0587a6a68304-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 27 15:09:29 crc kubenswrapper[4772]: I0127 15:09:29.136080 4772 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/53a6c4ad-816a-4d22-af98-0587a6a68304-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 27 15:09:29 crc kubenswrapper[4772]: I0127 15:09:29.136144 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/545a0fd1-38f1-4cbc-9f37-1870b1673589-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "545a0fd1-38f1-4cbc-9f37-1870b1673589" (UID: "545a0fd1-38f1-4cbc-9f37-1870b1673589"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:09:29 crc kubenswrapper[4772]: I0127 15:09:29.141514 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/545a0fd1-38f1-4cbc-9f37-1870b1673589-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "545a0fd1-38f1-4cbc-9f37-1870b1673589" (UID: "545a0fd1-38f1-4cbc-9f37-1870b1673589"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:09:29 crc kubenswrapper[4772]: I0127 15:09:29.237601 4772 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/545a0fd1-38f1-4cbc-9f37-1870b1673589-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 27 15:09:29 crc kubenswrapper[4772]: I0127 15:09:29.237676 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/545a0fd1-38f1-4cbc-9f37-1870b1673589-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 27 15:09:29 crc kubenswrapper[4772]: I0127 15:09:29.536831 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"53a6c4ad-816a-4d22-af98-0587a6a68304","Type":"ContainerDied","Data":"4e5fb5c0a561c986217a03a765b2818013365a6cf31c0cb2523e728110829fe1"} Jan 27 15:09:29 crc kubenswrapper[4772]: I0127 15:09:29.537137 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e5fb5c0a561c986217a03a765b2818013365a6cf31c0cb2523e728110829fe1" Jan 27 15:09:29 crc kubenswrapper[4772]: I0127 15:09:29.537214 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 27 15:09:29 crc kubenswrapper[4772]: I0127 15:09:29.566513 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 27 15:09:29 crc kubenswrapper[4772]: I0127 15:09:29.566860 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"545a0fd1-38f1-4cbc-9f37-1870b1673589","Type":"ContainerDied","Data":"fb9d01e0f3553308e4bfd0ea1eb0d0086a446ca7d54e19f2aad89776bdcb2b2b"} Jan 27 15:09:29 crc kubenswrapper[4772]: I0127 15:09:29.566913 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb9d01e0f3553308e4bfd0ea1eb0d0086a446ca7d54e19f2aad89776bdcb2b2b" Jan 27 15:09:31 crc kubenswrapper[4772]: I0127 15:09:31.630337 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-jplbk" Jan 27 15:09:36 crc kubenswrapper[4772]: I0127 15:09:36.557298 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:36 crc kubenswrapper[4772]: I0127 15:09:36.561620 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:09:36 crc kubenswrapper[4772]: I0127 15:09:36.591739 4772 patch_prober.go:28] interesting pod/downloads-7954f5f757-vswtw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 27 15:09:36 crc kubenswrapper[4772]: I0127 15:09:36.591808 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vswtw" podUID="17bdd07d-f7e5-47f8-b730-724d5cc8e3d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 27 15:09:36 crc kubenswrapper[4772]: I0127 15:09:36.591998 4772 patch_prober.go:28] interesting pod/downloads-7954f5f757-vswtw container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 27 15:09:36 crc kubenswrapper[4772]: I0127 15:09:36.604562 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-vswtw" podUID="17bdd07d-f7e5-47f8-b730-724d5cc8e3d2" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 27 15:09:38 crc kubenswrapper[4772]: I0127 15:09:38.430203 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs\") pod \"network-metrics-daemon-ql2vx\" (UID: \"371016c8-5a23-427d-aa0a-0faa241d86a7\") " pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:09:38 crc kubenswrapper[4772]: I0127 15:09:38.435838 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/371016c8-5a23-427d-aa0a-0faa241d86a7-metrics-certs\") pod \"network-metrics-daemon-ql2vx\" (UID: \"371016c8-5a23-427d-aa0a-0faa241d86a7\") " pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:09:38 crc kubenswrapper[4772]: I0127 15:09:38.684687 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-ql2vx" Jan 27 15:09:42 crc kubenswrapper[4772]: I0127 15:09:42.058540 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:09:42 crc kubenswrapper[4772]: I0127 15:09:42.058618 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:09:44 crc kubenswrapper[4772]: I0127 15:09:44.399067 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:09:46 crc kubenswrapper[4772]: I0127 15:09:46.595013 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-vswtw" Jan 27 15:09:56 crc kubenswrapper[4772]: I0127 15:09:56.981945 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cv2z7" Jan 27 15:10:02 crc kubenswrapper[4772]: I0127 15:10:02.303548 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 27 15:10:02 crc kubenswrapper[4772]: E0127 15:10:02.304076 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="545a0fd1-38f1-4cbc-9f37-1870b1673589" containerName="pruner" Jan 27 15:10:02 crc kubenswrapper[4772]: I0127 15:10:02.304088 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="545a0fd1-38f1-4cbc-9f37-1870b1673589" containerName="pruner" Jan 27 15:10:02 crc kubenswrapper[4772]: E0127 15:10:02.304104 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53a6c4ad-816a-4d22-af98-0587a6a68304" containerName="pruner" Jan 27 15:10:02 crc kubenswrapper[4772]: I0127 15:10:02.304110 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="53a6c4ad-816a-4d22-af98-0587a6a68304" containerName="pruner" Jan 27 15:10:02 crc kubenswrapper[4772]: I0127 15:10:02.304254 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="545a0fd1-38f1-4cbc-9f37-1870b1673589" containerName="pruner" Jan 27 15:10:02 crc kubenswrapper[4772]: I0127 15:10:02.304264 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="53a6c4ad-816a-4d22-af98-0587a6a68304" containerName="pruner" Jan 27 15:10:02 crc kubenswrapper[4772]: I0127 15:10:02.304719 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 27 15:10:02 crc kubenswrapper[4772]: I0127 15:10:02.307413 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 27 15:10:02 crc kubenswrapper[4772]: I0127 15:10:02.308293 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 27 15:10:02 crc kubenswrapper[4772]: I0127 15:10:02.310769 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 27 15:10:02 crc kubenswrapper[4772]: I0127 15:10:02.453974 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b62e8603-1857-4f92-9e4f-b3eab5be12ed-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b62e8603-1857-4f92-9e4f-b3eab5be12ed\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 27 15:10:02 crc kubenswrapper[4772]: I0127 15:10:02.454047 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b62e8603-1857-4f92-9e4f-b3eab5be12ed-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b62e8603-1857-4f92-9e4f-b3eab5be12ed\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 27 15:10:02 crc kubenswrapper[4772]: I0127 15:10:02.555117 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b62e8603-1857-4f92-9e4f-b3eab5be12ed-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b62e8603-1857-4f92-9e4f-b3eab5be12ed\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 27 15:10:02 crc kubenswrapper[4772]: I0127 15:10:02.555339 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b62e8603-1857-4f92-9e4f-b3eab5be12ed-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b62e8603-1857-4f92-9e4f-b3eab5be12ed\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 27 15:10:02 crc kubenswrapper[4772]: I0127 15:10:02.555533 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b62e8603-1857-4f92-9e4f-b3eab5be12ed-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b62e8603-1857-4f92-9e4f-b3eab5be12ed\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 27 15:10:02 crc kubenswrapper[4772]: I0127 15:10:02.579669 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b62e8603-1857-4f92-9e4f-b3eab5be12ed-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b62e8603-1857-4f92-9e4f-b3eab5be12ed\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 27 15:10:02 crc kubenswrapper[4772]: I0127 15:10:02.626834 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 27 15:10:02 crc kubenswrapper[4772]: I0127 15:10:02.905566 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 27 15:10:06 crc kubenswrapper[4772]: E0127 15:10:06.076433 4772 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 27 15:10:06 crc kubenswrapper[4772]: E0127 15:10:06.077123 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sp9qx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-wcldz_openshift-marketplace(987488b4-af4d-4b20-bb26-f433d4d1299a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 27 15:10:06 crc kubenswrapper[4772]: E0127 15:10:06.078285 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-wcldz" podUID="987488b4-af4d-4b20-bb26-f433d4d1299a" Jan 27 15:10:06 crc kubenswrapper[4772]: I0127 15:10:06.698112 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 27 15:10:06 crc kubenswrapper[4772]: I0127 15:10:06.698752 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 27 15:10:06 crc kubenswrapper[4772]: I0127 15:10:06.707822 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 27 15:10:06 crc kubenswrapper[4772]: I0127 15:10:06.733358 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7bd0c383-7376-4e95-9919-863297cbd807-kubelet-dir\") pod \"installer-9-crc\" (UID: \"7bd0c383-7376-4e95-9919-863297cbd807\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 27 15:10:06 crc kubenswrapper[4772]: I0127 15:10:06.733440 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7bd0c383-7376-4e95-9919-863297cbd807-var-lock\") pod \"installer-9-crc\" (UID: \"7bd0c383-7376-4e95-9919-863297cbd807\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 27 15:10:06 crc kubenswrapper[4772]: I0127 15:10:06.733471 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7bd0c383-7376-4e95-9919-863297cbd807-kube-api-access\") pod \"installer-9-crc\" (UID: \"7bd0c383-7376-4e95-9919-863297cbd807\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 27 15:10:06 crc kubenswrapper[4772]: I0127 15:10:06.834352 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7bd0c383-7376-4e95-9919-863297cbd807-kubelet-dir\") pod \"installer-9-crc\" (UID: \"7bd0c383-7376-4e95-9919-863297cbd807\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 27 15:10:06 crc kubenswrapper[4772]: I0127 15:10:06.834430 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7bd0c383-7376-4e95-9919-863297cbd807-var-lock\") pod \"installer-9-crc\" (UID: \"7bd0c383-7376-4e95-9919-863297cbd807\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 27 15:10:06 crc kubenswrapper[4772]: I0127 15:10:06.834459 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7bd0c383-7376-4e95-9919-863297cbd807-kube-api-access\") pod \"installer-9-crc\" (UID: \"7bd0c383-7376-4e95-9919-863297cbd807\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 27 15:10:06 crc kubenswrapper[4772]: I0127 15:10:06.834889 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7bd0c383-7376-4e95-9919-863297cbd807-kubelet-dir\") pod \"installer-9-crc\" (UID: \"7bd0c383-7376-4e95-9919-863297cbd807\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 27 15:10:06 crc kubenswrapper[4772]: I0127 15:10:06.834938 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7bd0c383-7376-4e95-9919-863297cbd807-var-lock\") pod \"installer-9-crc\" (UID: \"7bd0c383-7376-4e95-9919-863297cbd807\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 27 15:10:06 crc kubenswrapper[4772]: I0127 15:10:06.858029 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7bd0c383-7376-4e95-9919-863297cbd807-kube-api-access\") pod \"installer-9-crc\" (UID: \"7bd0c383-7376-4e95-9919-863297cbd807\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 27 15:10:07 crc kubenswrapper[4772]: I0127 15:10:07.039704 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 27 15:10:09 crc kubenswrapper[4772]: E0127 15:10:09.683741 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-wcldz" podUID="987488b4-af4d-4b20-bb26-f433d4d1299a" Jan 27 15:10:09 crc kubenswrapper[4772]: E0127 15:10:09.784005 4772 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 27 15:10:09 crc kubenswrapper[4772]: E0127 15:10:09.784395 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vvxxl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-75jrg_openshift-marketplace(f637b998-b13b-486d-9042-4cd40a01c833): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 27 15:10:09 crc kubenswrapper[4772]: E0127 15:10:09.785566 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-75jrg" podUID="f637b998-b13b-486d-9042-4cd40a01c833" Jan 27 15:10:09 crc kubenswrapper[4772]: E0127 15:10:09.790273 4772 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 27 15:10:09 crc kubenswrapper[4772]: E0127 15:10:09.790406 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fnhcz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-k7pfr_openshift-marketplace(d0b33686-8107-4caf-b67f-3c608119a049): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 27 15:10:09 crc kubenswrapper[4772]: E0127 15:10:09.791775 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-k7pfr" podUID="d0b33686-8107-4caf-b67f-3c608119a049" Jan 27 15:10:10 crc kubenswrapper[4772]: E0127 15:10:10.887120 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-75jrg" podUID="f637b998-b13b-486d-9042-4cd40a01c833" Jan 27 15:10:10 crc kubenswrapper[4772]: E0127 15:10:10.887125 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-k7pfr" podUID="d0b33686-8107-4caf-b67f-3c608119a049" Jan 27 15:10:10 crc kubenswrapper[4772]: E0127 15:10:10.949844 4772 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 27 15:10:10 crc kubenswrapper[4772]: E0127 15:10:10.950024 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-67q8x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-jwrpk_openshift-marketplace(dd415ccf-2b4a-4797-962f-a464ef96bc22): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 27 15:10:10 crc kubenswrapper[4772]: E0127 15:10:10.951310 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-jwrpk" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" Jan 27 15:10:12 crc kubenswrapper[4772]: I0127 15:10:12.058642 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:10:12 crc kubenswrapper[4772]: I0127 15:10:12.059026 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:10:13 crc kubenswrapper[4772]: E0127 15:10:13.287137 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-jwrpk" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" Jan 27 15:10:13 crc kubenswrapper[4772]: E0127 15:10:13.526840 4772 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 27 15:10:13 crc kubenswrapper[4772]: E0127 15:10:13.527136 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-87w6q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-dfgjh_openshift-marketplace(8cbabfa8-79d8-4b23-b186-b40ba8b3017e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 27 15:10:13 crc kubenswrapper[4772]: E0127 15:10:13.528367 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-dfgjh" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" Jan 27 15:10:13 crc kubenswrapper[4772]: E0127 15:10:13.666530 4772 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 27 15:10:13 crc kubenswrapper[4772]: E0127 15:10:13.667307 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4zvrk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-95rh9_openshift-marketplace(4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 27 15:10:13 crc kubenswrapper[4772]: E0127 15:10:13.668532 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-95rh9" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" Jan 27 15:10:13 crc kubenswrapper[4772]: I0127 15:10:13.726387 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 27 15:10:13 crc kubenswrapper[4772]: W0127 15:10:13.733596 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podb62e8603_1857_4f92_9e4f_b3eab5be12ed.slice/crio-61dfd2fab37415adeb86c2e4199d19720dbf229c31a5352f47a4c6a31dab51bb WatchSource:0}: Error finding container 61dfd2fab37415adeb86c2e4199d19720dbf229c31a5352f47a4c6a31dab51bb: Status 404 returned error can't find the container with id 61dfd2fab37415adeb86c2e4199d19720dbf229c31a5352f47a4c6a31dab51bb Jan 27 15:10:13 crc kubenswrapper[4772]: E0127 15:10:13.763315 4772 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 27 15:10:13 crc kubenswrapper[4772]: E0127 15:10:13.763491 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7g6kk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-xp8ph_openshift-marketplace(ac2b5800-ce98-4847-bfcd-67a97375aa1b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 27 15:10:13 crc kubenswrapper[4772]: E0127 15:10:13.770973 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-xp8ph" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" Jan 27 15:10:13 crc kubenswrapper[4772]: I0127 15:10:13.775257 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 27 15:10:13 crc kubenswrapper[4772]: I0127 15:10:13.786025 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-ql2vx"] Jan 27 15:10:13 crc kubenswrapper[4772]: W0127 15:10:13.803086 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod371016c8_5a23_427d_aa0a_0faa241d86a7.slice/crio-39b983878e98594f803a2a5d18787411ce8df5bebcdb7feb44bab66559d18d93 WatchSource:0}: Error finding container 39b983878e98594f803a2a5d18787411ce8df5bebcdb7feb44bab66559d18d93: Status 404 returned error can't find the container with id 39b983878e98594f803a2a5d18787411ce8df5bebcdb7feb44bab66559d18d93 Jan 27 15:10:13 crc kubenswrapper[4772]: E0127 15:10:13.870864 4772 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 27 15:10:13 crc kubenswrapper[4772]: E0127 15:10:13.871381 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7fhkv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-9wdps_openshift-marketplace(96e88efd-1f25-4e44-b459-ab773db93656): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 27 15:10:13 crc kubenswrapper[4772]: E0127 15:10:13.872719 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-9wdps" podUID="96e88efd-1f25-4e44-b459-ab773db93656" Jan 27 15:10:13 crc kubenswrapper[4772]: I0127 15:10:13.936896 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-ql2vx" event={"ID":"371016c8-5a23-427d-aa0a-0faa241d86a7","Type":"ContainerStarted","Data":"39b983878e98594f803a2a5d18787411ce8df5bebcdb7feb44bab66559d18d93"} Jan 27 15:10:13 crc kubenswrapper[4772]: I0127 15:10:13.938105 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"b62e8603-1857-4f92-9e4f-b3eab5be12ed","Type":"ContainerStarted","Data":"61dfd2fab37415adeb86c2e4199d19720dbf229c31a5352f47a4c6a31dab51bb"} Jan 27 15:10:13 crc kubenswrapper[4772]: I0127 15:10:13.939673 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7bd0c383-7376-4e95-9919-863297cbd807","Type":"ContainerStarted","Data":"27a7557244977e8f16265740d543848969b7bcbc8b87db7409dc58f332f68492"} Jan 27 15:10:13 crc kubenswrapper[4772]: E0127 15:10:13.941874 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-dfgjh" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" Jan 27 15:10:13 crc kubenswrapper[4772]: E0127 15:10:13.943096 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-95rh9" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" Jan 27 15:10:13 crc kubenswrapper[4772]: E0127 15:10:13.943152 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-xp8ph" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" Jan 27 15:10:13 crc kubenswrapper[4772]: E0127 15:10:13.943319 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-9wdps" podUID="96e88efd-1f25-4e44-b459-ab773db93656" Jan 27 15:10:14 crc kubenswrapper[4772]: I0127 15:10:14.951179 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-ql2vx" event={"ID":"371016c8-5a23-427d-aa0a-0faa241d86a7","Type":"ContainerStarted","Data":"4d8d625089c13d4ad5b6cee5beecca8d34e5a17f31b33f041a0a842cab85604c"} Jan 27 15:10:14 crc kubenswrapper[4772]: I0127 15:10:14.951503 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-ql2vx" event={"ID":"371016c8-5a23-427d-aa0a-0faa241d86a7","Type":"ContainerStarted","Data":"bf3f4ce0b98e4dc2124a6a0ca30c113726e0a65db7e268a10567ef09bb041136"} Jan 27 15:10:14 crc kubenswrapper[4772]: I0127 15:10:14.954709 4772 generic.go:334] "Generic (PLEG): container finished" podID="b62e8603-1857-4f92-9e4f-b3eab5be12ed" containerID="1f7d72c416be4fb3e74c5e402c3e0877f29d27571b78b1c7dc5eefbf979e7d8b" exitCode=0 Jan 27 15:10:14 crc kubenswrapper[4772]: I0127 15:10:14.954865 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"b62e8603-1857-4f92-9e4f-b3eab5be12ed","Type":"ContainerDied","Data":"1f7d72c416be4fb3e74c5e402c3e0877f29d27571b78b1c7dc5eefbf979e7d8b"} Jan 27 15:10:14 crc kubenswrapper[4772]: I0127 15:10:14.956954 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7bd0c383-7376-4e95-9919-863297cbd807","Type":"ContainerStarted","Data":"0eeaf2bcc5f54216d999847c8ecf3f795fa45776d35017701579dff468e8db9c"} Jan 27 15:10:14 crc kubenswrapper[4772]: I0127 15:10:14.971470 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-ql2vx" podStartSLOduration=178.971447894 podStartE2EDuration="2m58.971447894s" podCreationTimestamp="2026-01-27 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:10:14.96496303 +0000 UTC m=+200.945572128" watchObservedRunningTime="2026-01-27 15:10:14.971447894 +0000 UTC m=+200.952056992" Jan 27 15:10:14 crc kubenswrapper[4772]: I0127 15:10:14.992815 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=8.992796131 podStartE2EDuration="8.992796131s" podCreationTimestamp="2026-01-27 15:10:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:10:14.992220684 +0000 UTC m=+200.972829782" watchObservedRunningTime="2026-01-27 15:10:14.992796131 +0000 UTC m=+200.973405229" Jan 27 15:10:16 crc kubenswrapper[4772]: I0127 15:10:16.201403 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 27 15:10:16 crc kubenswrapper[4772]: I0127 15:10:16.272598 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b62e8603-1857-4f92-9e4f-b3eab5be12ed-kube-api-access\") pod \"b62e8603-1857-4f92-9e4f-b3eab5be12ed\" (UID: \"b62e8603-1857-4f92-9e4f-b3eab5be12ed\") " Jan 27 15:10:16 crc kubenswrapper[4772]: I0127 15:10:16.272715 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b62e8603-1857-4f92-9e4f-b3eab5be12ed-kubelet-dir\") pod \"b62e8603-1857-4f92-9e4f-b3eab5be12ed\" (UID: \"b62e8603-1857-4f92-9e4f-b3eab5be12ed\") " Jan 27 15:10:16 crc kubenswrapper[4772]: I0127 15:10:16.272990 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b62e8603-1857-4f92-9e4f-b3eab5be12ed-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b62e8603-1857-4f92-9e4f-b3eab5be12ed" (UID: "b62e8603-1857-4f92-9e4f-b3eab5be12ed"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:10:16 crc kubenswrapper[4772]: I0127 15:10:16.279363 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b62e8603-1857-4f92-9e4f-b3eab5be12ed-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b62e8603-1857-4f92-9e4f-b3eab5be12ed" (UID: "b62e8603-1857-4f92-9e4f-b3eab5be12ed"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:10:16 crc kubenswrapper[4772]: I0127 15:10:16.374309 4772 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b62e8603-1857-4f92-9e4f-b3eab5be12ed-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:16 crc kubenswrapper[4772]: I0127 15:10:16.374598 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b62e8603-1857-4f92-9e4f-b3eab5be12ed-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:16 crc kubenswrapper[4772]: I0127 15:10:16.972563 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"b62e8603-1857-4f92-9e4f-b3eab5be12ed","Type":"ContainerDied","Data":"61dfd2fab37415adeb86c2e4199d19720dbf229c31a5352f47a4c6a31dab51bb"} Jan 27 15:10:16 crc kubenswrapper[4772]: I0127 15:10:16.972610 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61dfd2fab37415adeb86c2e4199d19720dbf229c31a5352f47a4c6a31dab51bb" Jan 27 15:10:16 crc kubenswrapper[4772]: I0127 15:10:16.972670 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 27 15:10:25 crc kubenswrapper[4772]: I0127 15:10:25.015647 4772 generic.go:334] "Generic (PLEG): container finished" podID="987488b4-af4d-4b20-bb26-f433d4d1299a" containerID="f473a9dfe0a18cc21db70fde482752286c2a5587ef7f32cf2144d6c58d4053e6" exitCode=0 Jan 27 15:10:25 crc kubenswrapper[4772]: I0127 15:10:25.016243 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wcldz" event={"ID":"987488b4-af4d-4b20-bb26-f433d4d1299a","Type":"ContainerDied","Data":"f473a9dfe0a18cc21db70fde482752286c2a5587ef7f32cf2144d6c58d4053e6"} Jan 27 15:10:27 crc kubenswrapper[4772]: I0127 15:10:27.029409 4772 generic.go:334] "Generic (PLEG): container finished" podID="f637b998-b13b-486d-9042-4cd40a01c833" containerID="2c0b59089ec0a9f3a0a19eaaa3c657c89533289d4c824c8863bf9ca1e0f2856b" exitCode=0 Jan 27 15:10:27 crc kubenswrapper[4772]: I0127 15:10:27.029474 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75jrg" event={"ID":"f637b998-b13b-486d-9042-4cd40a01c833","Type":"ContainerDied","Data":"2c0b59089ec0a9f3a0a19eaaa3c657c89533289d4c824c8863bf9ca1e0f2856b"} Jan 27 15:10:27 crc kubenswrapper[4772]: I0127 15:10:27.033819 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wcldz" event={"ID":"987488b4-af4d-4b20-bb26-f433d4d1299a","Type":"ContainerStarted","Data":"2524cad9551fd6cd1e12590205a9cfa2493995c09e8d8b75c74bdbd4fbd9dab8"} Jan 27 15:10:27 crc kubenswrapper[4772]: I0127 15:10:27.035697 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7pfr" event={"ID":"d0b33686-8107-4caf-b67f-3c608119a049","Type":"ContainerStarted","Data":"35038329828ccd832c938fb7caf96024c35dc820e86dd3e9aadcf5ac8ef257b5"} Jan 27 15:10:27 crc kubenswrapper[4772]: I0127 15:10:27.095033 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wcldz" podStartSLOduration=3.849487516 podStartE2EDuration="1m3.095012728s" podCreationTimestamp="2026-01-27 15:09:24 +0000 UTC" firstStartedPulling="2026-01-27 15:09:26.435160755 +0000 UTC m=+152.415769863" lastFinishedPulling="2026-01-27 15:10:25.680685977 +0000 UTC m=+211.661295075" observedRunningTime="2026-01-27 15:10:27.092365043 +0000 UTC m=+213.072974161" watchObservedRunningTime="2026-01-27 15:10:27.095012728 +0000 UTC m=+213.075621826" Jan 27 15:10:28 crc kubenswrapper[4772]: I0127 15:10:28.042970 4772 generic.go:334] "Generic (PLEG): container finished" podID="d0b33686-8107-4caf-b67f-3c608119a049" containerID="35038329828ccd832c938fb7caf96024c35dc820e86dd3e9aadcf5ac8ef257b5" exitCode=0 Jan 27 15:10:28 crc kubenswrapper[4772]: I0127 15:10:28.043038 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7pfr" event={"ID":"d0b33686-8107-4caf-b67f-3c608119a049","Type":"ContainerDied","Data":"35038329828ccd832c938fb7caf96024c35dc820e86dd3e9aadcf5ac8ef257b5"} Jan 27 15:10:34 crc kubenswrapper[4772]: I0127 15:10:34.478274 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wcldz" Jan 27 15:10:34 crc kubenswrapper[4772]: I0127 15:10:34.478343 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wcldz" Jan 27 15:10:35 crc kubenswrapper[4772]: I0127 15:10:35.570816 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wcldz" Jan 27 15:10:35 crc kubenswrapper[4772]: I0127 15:10:35.605433 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wcldz" Jan 27 15:10:35 crc kubenswrapper[4772]: I0127 15:10:35.800433 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wcldz"] Jan 27 15:10:37 crc kubenswrapper[4772]: I0127 15:10:37.090618 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wcldz" podUID="987488b4-af4d-4b20-bb26-f433d4d1299a" containerName="registry-server" containerID="cri-o://2524cad9551fd6cd1e12590205a9cfa2493995c09e8d8b75c74bdbd4fbd9dab8" gracePeriod=2 Jan 27 15:10:38 crc kubenswrapper[4772]: I0127 15:10:38.097595 4772 generic.go:334] "Generic (PLEG): container finished" podID="987488b4-af4d-4b20-bb26-f433d4d1299a" containerID="2524cad9551fd6cd1e12590205a9cfa2493995c09e8d8b75c74bdbd4fbd9dab8" exitCode=0 Jan 27 15:10:38 crc kubenswrapper[4772]: I0127 15:10:38.097669 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wcldz" event={"ID":"987488b4-af4d-4b20-bb26-f433d4d1299a","Type":"ContainerDied","Data":"2524cad9551fd6cd1e12590205a9cfa2493995c09e8d8b75c74bdbd4fbd9dab8"} Jan 27 15:10:41 crc kubenswrapper[4772]: I0127 15:10:41.770927 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wcldz" Jan 27 15:10:41 crc kubenswrapper[4772]: I0127 15:10:41.852579 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/987488b4-af4d-4b20-bb26-f433d4d1299a-utilities\") pod \"987488b4-af4d-4b20-bb26-f433d4d1299a\" (UID: \"987488b4-af4d-4b20-bb26-f433d4d1299a\") " Jan 27 15:10:41 crc kubenswrapper[4772]: I0127 15:10:41.852649 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/987488b4-af4d-4b20-bb26-f433d4d1299a-catalog-content\") pod \"987488b4-af4d-4b20-bb26-f433d4d1299a\" (UID: \"987488b4-af4d-4b20-bb26-f433d4d1299a\") " Jan 27 15:10:41 crc kubenswrapper[4772]: I0127 15:10:41.852684 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sp9qx\" (UniqueName: \"kubernetes.io/projected/987488b4-af4d-4b20-bb26-f433d4d1299a-kube-api-access-sp9qx\") pod \"987488b4-af4d-4b20-bb26-f433d4d1299a\" (UID: \"987488b4-af4d-4b20-bb26-f433d4d1299a\") " Jan 27 15:10:41 crc kubenswrapper[4772]: I0127 15:10:41.853750 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/987488b4-af4d-4b20-bb26-f433d4d1299a-utilities" (OuterVolumeSpecName: "utilities") pod "987488b4-af4d-4b20-bb26-f433d4d1299a" (UID: "987488b4-af4d-4b20-bb26-f433d4d1299a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:10:41 crc kubenswrapper[4772]: I0127 15:10:41.863403 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/987488b4-af4d-4b20-bb26-f433d4d1299a-kube-api-access-sp9qx" (OuterVolumeSpecName: "kube-api-access-sp9qx") pod "987488b4-af4d-4b20-bb26-f433d4d1299a" (UID: "987488b4-af4d-4b20-bb26-f433d4d1299a"). InnerVolumeSpecName "kube-api-access-sp9qx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:10:41 crc kubenswrapper[4772]: I0127 15:10:41.906360 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/987488b4-af4d-4b20-bb26-f433d4d1299a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "987488b4-af4d-4b20-bb26-f433d4d1299a" (UID: "987488b4-af4d-4b20-bb26-f433d4d1299a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:10:41 crc kubenswrapper[4772]: I0127 15:10:41.954652 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/987488b4-af4d-4b20-bb26-f433d4d1299a-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:41 crc kubenswrapper[4772]: I0127 15:10:41.954992 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/987488b4-af4d-4b20-bb26-f433d4d1299a-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:41 crc kubenswrapper[4772]: I0127 15:10:41.955004 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sp9qx\" (UniqueName: \"kubernetes.io/projected/987488b4-af4d-4b20-bb26-f433d4d1299a-kube-api-access-sp9qx\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:42 crc kubenswrapper[4772]: I0127 15:10:42.058593 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:10:42 crc kubenswrapper[4772]: I0127 15:10:42.058669 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:10:42 crc kubenswrapper[4772]: I0127 15:10:42.058757 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:10:42 crc kubenswrapper[4772]: I0127 15:10:42.059412 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 15:10:42 crc kubenswrapper[4772]: I0127 15:10:42.059568 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19" gracePeriod=600 Jan 27 15:10:42 crc kubenswrapper[4772]: I0127 15:10:42.137709 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wcldz" event={"ID":"987488b4-af4d-4b20-bb26-f433d4d1299a","Type":"ContainerDied","Data":"cca259e2810aba1b9b6e47d70088b4b43bd08d57432f8fbb61e7ddcd0a7abb94"} Jan 27 15:10:42 crc kubenswrapper[4772]: I0127 15:10:42.137769 4772 scope.go:117] "RemoveContainer" containerID="2524cad9551fd6cd1e12590205a9cfa2493995c09e8d8b75c74bdbd4fbd9dab8" Jan 27 15:10:42 crc kubenswrapper[4772]: I0127 15:10:42.137901 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wcldz" Jan 27 15:10:42 crc kubenswrapper[4772]: I0127 15:10:42.168785 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wcldz"] Jan 27 15:10:42 crc kubenswrapper[4772]: I0127 15:10:42.168849 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wcldz"] Jan 27 15:10:42 crc kubenswrapper[4772]: I0127 15:10:42.670517 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="987488b4-af4d-4b20-bb26-f433d4d1299a" path="/var/lib/kubelet/pods/987488b4-af4d-4b20-bb26-f433d4d1299a/volumes" Jan 27 15:10:43 crc kubenswrapper[4772]: I0127 15:10:43.143745 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19" exitCode=0 Jan 27 15:10:43 crc kubenswrapper[4772]: I0127 15:10:43.143800 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19"} Jan 27 15:10:44 crc kubenswrapper[4772]: I0127 15:10:44.541865 4772 scope.go:117] "RemoveContainer" containerID="f473a9dfe0a18cc21db70fde482752286c2a5587ef7f32cf2144d6c58d4053e6" Jan 27 15:10:46 crc kubenswrapper[4772]: I0127 15:10:46.520193 4772 scope.go:117] "RemoveContainer" containerID="05b1bd7fef5819af5a4449e0a65b00c0b94c405308dc0ea03120ac4091e22a7a" Jan 27 15:10:47 crc kubenswrapper[4772]: I0127 15:10:47.176713 4772 generic.go:334] "Generic (PLEG): container finished" podID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" containerID="7958617c1c11ed30a9375bc289d465de0afd6e9db9de8e33b0d1ef509c9e2adb" exitCode=0 Jan 27 15:10:47 crc kubenswrapper[4772]: I0127 15:10:47.176800 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dfgjh" event={"ID":"8cbabfa8-79d8-4b23-b186-b40ba8b3017e","Type":"ContainerDied","Data":"7958617c1c11ed30a9375bc289d465de0afd6e9db9de8e33b0d1ef509c9e2adb"} Jan 27 15:10:47 crc kubenswrapper[4772]: I0127 15:10:47.185897 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75jrg" event={"ID":"f637b998-b13b-486d-9042-4cd40a01c833","Type":"ContainerStarted","Data":"0085f838645429f4fe1db48da5f434fde0158d222d9039bcf74e2bc9cea6bf7f"} Jan 27 15:10:47 crc kubenswrapper[4772]: I0127 15:10:47.187784 4772 generic.go:334] "Generic (PLEG): container finished" podID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" containerID="b837145fe89058e7be062b6cdd2bde2c15ab5a2a27d1b5c341bb196fff256ac4" exitCode=0 Jan 27 15:10:47 crc kubenswrapper[4772]: I0127 15:10:47.187838 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xp8ph" event={"ID":"ac2b5800-ce98-4847-bfcd-67a97375aa1b","Type":"ContainerDied","Data":"b837145fe89058e7be062b6cdd2bde2c15ab5a2a27d1b5c341bb196fff256ac4"} Jan 27 15:10:47 crc kubenswrapper[4772]: I0127 15:10:47.193139 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"8e72007caa5160368d39dc40b9c7f95a9beba3bef9f9e290eac1d112ef6eeb10"} Jan 27 15:10:47 crc kubenswrapper[4772]: I0127 15:10:47.198409 4772 generic.go:334] "Generic (PLEG): container finished" podID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" containerID="a2e75315afb920925e12db40d3166aec37363a00b9620db15e2cf976b7c362c8" exitCode=0 Jan 27 15:10:47 crc kubenswrapper[4772]: I0127 15:10:47.198453 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-95rh9" event={"ID":"4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7","Type":"ContainerDied","Data":"a2e75315afb920925e12db40d3166aec37363a00b9620db15e2cf976b7c362c8"} Jan 27 15:10:47 crc kubenswrapper[4772]: I0127 15:10:47.204659 4772 generic.go:334] "Generic (PLEG): container finished" podID="96e88efd-1f25-4e44-b459-ab773db93656" containerID="7127f22fb066327a7e4cfc38f52653579bcf1558a1847106ba8e7a3945fa02a6" exitCode=0 Jan 27 15:10:47 crc kubenswrapper[4772]: I0127 15:10:47.204758 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9wdps" event={"ID":"96e88efd-1f25-4e44-b459-ab773db93656","Type":"ContainerDied","Data":"7127f22fb066327a7e4cfc38f52653579bcf1558a1847106ba8e7a3945fa02a6"} Jan 27 15:10:47 crc kubenswrapper[4772]: I0127 15:10:47.207600 4772 generic.go:334] "Generic (PLEG): container finished" podID="dd415ccf-2b4a-4797-962f-a464ef96bc22" containerID="aff635156e5d675ca2fa44615b92754120942925bd9551591ea562f79911f6af" exitCode=0 Jan 27 15:10:47 crc kubenswrapper[4772]: I0127 15:10:47.207652 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwrpk" event={"ID":"dd415ccf-2b4a-4797-962f-a464ef96bc22","Type":"ContainerDied","Data":"aff635156e5d675ca2fa44615b92754120942925bd9551591ea562f79911f6af"} Jan 27 15:10:47 crc kubenswrapper[4772]: I0127 15:10:47.213515 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7pfr" event={"ID":"d0b33686-8107-4caf-b67f-3c608119a049","Type":"ContainerStarted","Data":"1c52c9067d3a0dfe5bf38e17654a83a4d2211b850c4ac05e58ed278ee0de4d7e"} Jan 27 15:10:47 crc kubenswrapper[4772]: I0127 15:10:47.221781 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-75jrg" podStartSLOduration=3.211485278 podStartE2EDuration="1m21.221756806s" podCreationTimestamp="2026-01-27 15:09:26 +0000 UTC" firstStartedPulling="2026-01-27 15:09:28.503281858 +0000 UTC m=+154.483890956" lastFinishedPulling="2026-01-27 15:10:46.513553386 +0000 UTC m=+232.494162484" observedRunningTime="2026-01-27 15:10:47.220307575 +0000 UTC m=+233.200916693" watchObservedRunningTime="2026-01-27 15:10:47.221756806 +0000 UTC m=+233.202365904" Jan 27 15:10:47 crc kubenswrapper[4772]: I0127 15:10:47.293114 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-k7pfr" podStartSLOduration=2.275063455 podStartE2EDuration="1m20.293098254s" podCreationTimestamp="2026-01-27 15:09:27 +0000 UTC" firstStartedPulling="2026-01-27 15:09:28.502325511 +0000 UTC m=+154.482934599" lastFinishedPulling="2026-01-27 15:10:46.5203603 +0000 UTC m=+232.500969398" observedRunningTime="2026-01-27 15:10:47.290427428 +0000 UTC m=+233.271036526" watchObservedRunningTime="2026-01-27 15:10:47.293098254 +0000 UTC m=+233.273707362" Jan 27 15:10:47 crc kubenswrapper[4772]: I0127 15:10:47.392433 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-k7pfr" Jan 27 15:10:47 crc kubenswrapper[4772]: I0127 15:10:47.392658 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-k7pfr" Jan 27 15:10:48 crc kubenswrapper[4772]: I0127 15:10:48.223098 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-95rh9" event={"ID":"4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7","Type":"ContainerStarted","Data":"891a196fc4fbfd6fd254de7d8acfa840ad89166efe2aae3755959d4061df327e"} Jan 27 15:10:48 crc kubenswrapper[4772]: I0127 15:10:48.226646 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9wdps" event={"ID":"96e88efd-1f25-4e44-b459-ab773db93656","Type":"ContainerStarted","Data":"321d45f45e6bfbd7d001b6cc4114cba4c366b13901f8014ac260d8f8a05b7c8d"} Jan 27 15:10:48 crc kubenswrapper[4772]: I0127 15:10:48.229206 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwrpk" event={"ID":"dd415ccf-2b4a-4797-962f-a464ef96bc22","Type":"ContainerStarted","Data":"7cdcc72178b424cbd1356a3055a7eccdb609ddd782039c152e1c33a7dc48ebfd"} Jan 27 15:10:48 crc kubenswrapper[4772]: I0127 15:10:48.231251 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dfgjh" event={"ID":"8cbabfa8-79d8-4b23-b186-b40ba8b3017e","Type":"ContainerStarted","Data":"932559e335376251fa378d1d6f007b100323207571225373c52e6683753426ad"} Jan 27 15:10:48 crc kubenswrapper[4772]: I0127 15:10:48.243280 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-95rh9" podStartSLOduration=2.927046524 podStartE2EDuration="1m25.243260121s" podCreationTimestamp="2026-01-27 15:09:23 +0000 UTC" firstStartedPulling="2026-01-27 15:09:25.379733918 +0000 UTC m=+151.360343006" lastFinishedPulling="2026-01-27 15:10:47.695947505 +0000 UTC m=+233.676556603" observedRunningTime="2026-01-27 15:10:48.240891624 +0000 UTC m=+234.221500722" watchObservedRunningTime="2026-01-27 15:10:48.243260121 +0000 UTC m=+234.223869219" Jan 27 15:10:48 crc kubenswrapper[4772]: I0127 15:10:48.263415 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9wdps" podStartSLOduration=2.931746939 podStartE2EDuration="1m25.263396294s" podCreationTimestamp="2026-01-27 15:09:23 +0000 UTC" firstStartedPulling="2026-01-27 15:09:25.382702473 +0000 UTC m=+151.363311571" lastFinishedPulling="2026-01-27 15:10:47.714351828 +0000 UTC m=+233.694960926" observedRunningTime="2026-01-27 15:10:48.260376568 +0000 UTC m=+234.240985696" watchObservedRunningTime="2026-01-27 15:10:48.263396294 +0000 UTC m=+234.244005392" Jan 27 15:10:48 crc kubenswrapper[4772]: I0127 15:10:48.280683 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dfgjh" podStartSLOduration=2.2119710599999998 podStartE2EDuration="1m22.280664564s" podCreationTimestamp="2026-01-27 15:09:26 +0000 UTC" firstStartedPulling="2026-01-27 15:09:27.495602569 +0000 UTC m=+153.476211667" lastFinishedPulling="2026-01-27 15:10:47.564296073 +0000 UTC m=+233.544905171" observedRunningTime="2026-01-27 15:10:48.278518013 +0000 UTC m=+234.259127131" watchObservedRunningTime="2026-01-27 15:10:48.280664564 +0000 UTC m=+234.261273662" Jan 27 15:10:48 crc kubenswrapper[4772]: I0127 15:10:48.296786 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jwrpk" podStartSLOduration=2.957311848 podStartE2EDuration="1m25.296771232s" podCreationTimestamp="2026-01-27 15:09:23 +0000 UTC" firstStartedPulling="2026-01-27 15:09:25.387805521 +0000 UTC m=+151.368414619" lastFinishedPulling="2026-01-27 15:10:47.727264905 +0000 UTC m=+233.707874003" observedRunningTime="2026-01-27 15:10:48.292642735 +0000 UTC m=+234.273251833" watchObservedRunningTime="2026-01-27 15:10:48.296771232 +0000 UTC m=+234.277380330" Jan 27 15:10:48 crc kubenswrapper[4772]: I0127 15:10:48.439862 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-k7pfr" podUID="d0b33686-8107-4caf-b67f-3c608119a049" containerName="registry-server" probeResult="failure" output=< Jan 27 15:10:48 crc kubenswrapper[4772]: timeout: failed to connect service ":50051" within 1s Jan 27 15:10:48 crc kubenswrapper[4772]: > Jan 27 15:10:49 crc kubenswrapper[4772]: I0127 15:10:49.240073 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xp8ph" event={"ID":"ac2b5800-ce98-4847-bfcd-67a97375aa1b","Type":"ContainerStarted","Data":"f36658ad464a5804d99326d4821347a5bf28ef6ab12d7aacdef462094deb1db8"} Jan 27 15:10:49 crc kubenswrapper[4772]: I0127 15:10:49.262095 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xp8ph" podStartSLOduration=3.125388608 podStartE2EDuration="1m24.262076841s" podCreationTimestamp="2026-01-27 15:09:25 +0000 UTC" firstStartedPulling="2026-01-27 15:09:27.491286984 +0000 UTC m=+153.471896082" lastFinishedPulling="2026-01-27 15:10:48.627975217 +0000 UTC m=+234.608584315" observedRunningTime="2026-01-27 15:10:49.257226513 +0000 UTC m=+235.237835621" watchObservedRunningTime="2026-01-27 15:10:49.262076841 +0000 UTC m=+235.242685949" Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.578105 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9wdps"] Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.578841 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9wdps" podUID="96e88efd-1f25-4e44-b459-ab773db93656" containerName="registry-server" containerID="cri-o://321d45f45e6bfbd7d001b6cc4114cba4c366b13901f8014ac260d8f8a05b7c8d" gracePeriod=30 Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.589460 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-95rh9"] Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.589733 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-95rh9" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" containerName="registry-server" containerID="cri-o://891a196fc4fbfd6fd254de7d8acfa840ad89166efe2aae3755959d4061df327e" gracePeriod=30 Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.599775 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jwrpk"] Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.599980 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jwrpk" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" containerName="registry-server" containerID="cri-o://7cdcc72178b424cbd1356a3055a7eccdb609ddd782039c152e1c33a7dc48ebfd" gracePeriod=30 Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.606683 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4lj2h"] Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.606895 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" containerName="marketplace-operator" containerID="cri-o://8632589c7dbe4bb64d8d2a9e0983c8088c1ff445e316f1dd7c4e04e72fa148df" gracePeriod=30 Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.617897 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dfgjh"] Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.618141 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dfgjh" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" containerName="registry-server" containerID="cri-o://932559e335376251fa378d1d6f007b100323207571225373c52e6683753426ad" gracePeriod=30 Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.627578 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xp8ph"] Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.627840 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xp8ph" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" containerName="registry-server" containerID="cri-o://f36658ad464a5804d99326d4821347a5bf28ef6ab12d7aacdef462094deb1db8" gracePeriod=30 Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.638534 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2glnd"] Jan 27 15:10:50 crc kubenswrapper[4772]: E0127 15:10:50.638828 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="987488b4-af4d-4b20-bb26-f433d4d1299a" containerName="extract-utilities" Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.638858 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="987488b4-af4d-4b20-bb26-f433d4d1299a" containerName="extract-utilities" Jan 27 15:10:50 crc kubenswrapper[4772]: E0127 15:10:50.638876 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="987488b4-af4d-4b20-bb26-f433d4d1299a" containerName="registry-server" Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.638883 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="987488b4-af4d-4b20-bb26-f433d4d1299a" containerName="registry-server" Jan 27 15:10:50 crc kubenswrapper[4772]: E0127 15:10:50.638897 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="987488b4-af4d-4b20-bb26-f433d4d1299a" containerName="extract-content" Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.638909 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="987488b4-af4d-4b20-bb26-f433d4d1299a" containerName="extract-content" Jan 27 15:10:50 crc kubenswrapper[4772]: E0127 15:10:50.638926 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b62e8603-1857-4f92-9e4f-b3eab5be12ed" containerName="pruner" Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.638933 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="b62e8603-1857-4f92-9e4f-b3eab5be12ed" containerName="pruner" Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.639053 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="987488b4-af4d-4b20-bb26-f433d4d1299a" containerName="registry-server" Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.639066 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="b62e8603-1857-4f92-9e4f-b3eab5be12ed" containerName="pruner" Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.639500 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.644281 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-75jrg"] Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.644564 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-75jrg" podUID="f637b998-b13b-486d-9042-4cd40a01c833" containerName="registry-server" containerID="cri-o://0085f838645429f4fe1db48da5f434fde0158d222d9039bcf74e2bc9cea6bf7f" gracePeriod=30 Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.657703 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2glnd"] Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.661135 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k7pfr"] Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.666523 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-k7pfr" podUID="d0b33686-8107-4caf-b67f-3c608119a049" containerName="registry-server" containerID="cri-o://1c52c9067d3a0dfe5bf38e17654a83a4d2211b850c4ac05e58ed278ee0de4d7e" gracePeriod=30 Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.671105 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqkv6\" (UniqueName: \"kubernetes.io/projected/d8591d45-25d0-47ea-a856-9cd5334e4a8c-kube-api-access-zqkv6\") pod \"marketplace-operator-79b997595-2glnd\" (UID: \"d8591d45-25d0-47ea-a856-9cd5334e4a8c\") " pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.671160 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d8591d45-25d0-47ea-a856-9cd5334e4a8c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2glnd\" (UID: \"d8591d45-25d0-47ea-a856-9cd5334e4a8c\") " pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.671263 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d8591d45-25d0-47ea-a856-9cd5334e4a8c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2glnd\" (UID: \"d8591d45-25d0-47ea-a856-9cd5334e4a8c\") " pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.786360 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d8591d45-25d0-47ea-a856-9cd5334e4a8c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2glnd\" (UID: \"d8591d45-25d0-47ea-a856-9cd5334e4a8c\") " pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.787434 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqkv6\" (UniqueName: \"kubernetes.io/projected/d8591d45-25d0-47ea-a856-9cd5334e4a8c-kube-api-access-zqkv6\") pod \"marketplace-operator-79b997595-2glnd\" (UID: \"d8591d45-25d0-47ea-a856-9cd5334e4a8c\") " pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.787507 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d8591d45-25d0-47ea-a856-9cd5334e4a8c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2glnd\" (UID: \"d8591d45-25d0-47ea-a856-9cd5334e4a8c\") " pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.789129 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d8591d45-25d0-47ea-a856-9cd5334e4a8c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2glnd\" (UID: \"d8591d45-25d0-47ea-a856-9cd5334e4a8c\") " pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.799836 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d8591d45-25d0-47ea-a856-9cd5334e4a8c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2glnd\" (UID: \"d8591d45-25d0-47ea-a856-9cd5334e4a8c\") " pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.817078 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqkv6\" (UniqueName: \"kubernetes.io/projected/d8591d45-25d0-47ea-a856-9cd5334e4a8c-kube-api-access-zqkv6\") pod \"marketplace-operator-79b997595-2glnd\" (UID: \"d8591d45-25d0-47ea-a856-9cd5334e4a8c\") " pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" Jan 27 15:10:50 crc kubenswrapper[4772]: I0127 15:10:50.962573 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.203543 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-95rh9" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.204670 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9wdps" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.260661 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fgw98"] Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.269938 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-k7pfr_d0b33686-8107-4caf-b67f-3c608119a049/registry-server/0.log" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.280412 4772 generic.go:334] "Generic (PLEG): container finished" podID="d0b33686-8107-4caf-b67f-3c608119a049" containerID="1c52c9067d3a0dfe5bf38e17654a83a4d2211b850c4ac05e58ed278ee0de4d7e" exitCode=1 Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.280483 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7pfr" event={"ID":"d0b33686-8107-4caf-b67f-3c608119a049","Type":"ContainerDied","Data":"1c52c9067d3a0dfe5bf38e17654a83a4d2211b850c4ac05e58ed278ee0de4d7e"} Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.309585 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jwrpk" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.310058 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-75jrg_f637b998-b13b-486d-9042-4cd40a01c833/registry-server/0.log" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.329392 4772 generic.go:334] "Generic (PLEG): container finished" podID="f637b998-b13b-486d-9042-4cd40a01c833" containerID="0085f838645429f4fe1db48da5f434fde0158d222d9039bcf74e2bc9cea6bf7f" exitCode=1 Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.329518 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75jrg" event={"ID":"f637b998-b13b-486d-9042-4cd40a01c833","Type":"ContainerDied","Data":"0085f838645429f4fe1db48da5f434fde0158d222d9039bcf74e2bc9cea6bf7f"} Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.349588 4772 generic.go:334] "Generic (PLEG): container finished" podID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" containerID="f36658ad464a5804d99326d4821347a5bf28ef6ab12d7aacdef462094deb1db8" exitCode=0 Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.349853 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xp8ph" event={"ID":"ac2b5800-ce98-4847-bfcd-67a97375aa1b","Type":"ContainerDied","Data":"f36658ad464a5804d99326d4821347a5bf28ef6ab12d7aacdef462094deb1db8"} Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.353093 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-75jrg_f637b998-b13b-486d-9042-4cd40a01c833/registry-server/0.log" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.356122 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-75jrg" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.356629 4772 generic.go:334] "Generic (PLEG): container finished" podID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" containerID="891a196fc4fbfd6fd254de7d8acfa840ad89166efe2aae3755959d4061df327e" exitCode=0 Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.356766 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-95rh9" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.356805 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-95rh9" event={"ID":"4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7","Type":"ContainerDied","Data":"891a196fc4fbfd6fd254de7d8acfa840ad89166efe2aae3755959d4061df327e"} Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.356864 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-95rh9" event={"ID":"4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7","Type":"ContainerDied","Data":"ae198d4139eb016b136b591cf513d4e1d588e78f4cc1966851a08fad44048adb"} Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.356887 4772 scope.go:117] "RemoveContainer" containerID="891a196fc4fbfd6fd254de7d8acfa840ad89166efe2aae3755959d4061df327e" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.370352 4772 generic.go:334] "Generic (PLEG): container finished" podID="96e88efd-1f25-4e44-b459-ab773db93656" containerID="321d45f45e6bfbd7d001b6cc4114cba4c366b13901f8014ac260d8f8a05b7c8d" exitCode=0 Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.370410 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9wdps" event={"ID":"96e88efd-1f25-4e44-b459-ab773db93656","Type":"ContainerDied","Data":"321d45f45e6bfbd7d001b6cc4114cba4c366b13901f8014ac260d8f8a05b7c8d"} Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.370437 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9wdps" event={"ID":"96e88efd-1f25-4e44-b459-ab773db93656","Type":"ContainerDied","Data":"806bc56d016ac75a91f0a1effbd4a1494b65e83f8a957a34cd38d253ad927cc3"} Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.370513 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9wdps" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.370928 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-k7pfr_d0b33686-8107-4caf-b67f-3c608119a049/registry-server/0.log" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.371467 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k7pfr" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.376576 4772 generic.go:334] "Generic (PLEG): container finished" podID="dd415ccf-2b4a-4797-962f-a464ef96bc22" containerID="7cdcc72178b424cbd1356a3055a7eccdb609ddd782039c152e1c33a7dc48ebfd" exitCode=0 Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.376638 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwrpk" event={"ID":"dd415ccf-2b4a-4797-962f-a464ef96bc22","Type":"ContainerDied","Data":"7cdcc72178b424cbd1356a3055a7eccdb609ddd782039c152e1c33a7dc48ebfd"} Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.376710 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jwrpk" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.377457 4772 scope.go:117] "RemoveContainer" containerID="a2e75315afb920925e12db40d3166aec37363a00b9620db15e2cf976b7c362c8" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.388117 4772 generic.go:334] "Generic (PLEG): container finished" podID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" containerID="932559e335376251fa378d1d6f007b100323207571225373c52e6683753426ad" exitCode=0 Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.388180 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dfgjh" event={"ID":"8cbabfa8-79d8-4b23-b186-b40ba8b3017e","Type":"ContainerDied","Data":"932559e335376251fa378d1d6f007b100323207571225373c52e6683753426ad"} Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.389736 4772 generic.go:334] "Generic (PLEG): container finished" podID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" containerID="8632589c7dbe4bb64d8d2a9e0983c8088c1ff445e316f1dd7c4e04e72fa148df" exitCode=0 Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.389762 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" event={"ID":"c8ebf890-c3b0-468e-bf7d-0ec590df084b","Type":"ContainerDied","Data":"8632589c7dbe4bb64d8d2a9e0983c8088c1ff445e316f1dd7c4e04e72fa148df"} Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.402688 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7-utilities\") pod \"4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7\" (UID: \"4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7\") " Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.402782 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96e88efd-1f25-4e44-b459-ab773db93656-catalog-content\") pod \"96e88efd-1f25-4e44-b459-ab773db93656\" (UID: \"96e88efd-1f25-4e44-b459-ab773db93656\") " Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.402810 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zvrk\" (UniqueName: \"kubernetes.io/projected/4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7-kube-api-access-4zvrk\") pod \"4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7\" (UID: \"4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7\") " Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.402843 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fhkv\" (UniqueName: \"kubernetes.io/projected/96e88efd-1f25-4e44-b459-ab773db93656-kube-api-access-7fhkv\") pod \"96e88efd-1f25-4e44-b459-ab773db93656\" (UID: \"96e88efd-1f25-4e44-b459-ab773db93656\") " Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.402880 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7-catalog-content\") pod \"4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7\" (UID: \"4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7\") " Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.403021 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96e88efd-1f25-4e44-b459-ab773db93656-utilities\") pod \"96e88efd-1f25-4e44-b459-ab773db93656\" (UID: \"96e88efd-1f25-4e44-b459-ab773db93656\") " Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.404390 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96e88efd-1f25-4e44-b459-ab773db93656-utilities" (OuterVolumeSpecName: "utilities") pod "96e88efd-1f25-4e44-b459-ab773db93656" (UID: "96e88efd-1f25-4e44-b459-ab773db93656"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.405626 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7-utilities" (OuterVolumeSpecName: "utilities") pod "4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" (UID: "4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.421742 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96e88efd-1f25-4e44-b459-ab773db93656-kube-api-access-7fhkv" (OuterVolumeSpecName: "kube-api-access-7fhkv") pod "96e88efd-1f25-4e44-b459-ab773db93656" (UID: "96e88efd-1f25-4e44-b459-ab773db93656"). InnerVolumeSpecName "kube-api-access-7fhkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.422786 4772 scope.go:117] "RemoveContainer" containerID="16478a3399ae464f44f2335d6adb9e9f0cad87f3f55187da1906c5dcf87534dd" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.453296 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7-kube-api-access-4zvrk" (OuterVolumeSpecName: "kube-api-access-4zvrk") pod "4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" (UID: "4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7"). InnerVolumeSpecName "kube-api-access-4zvrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.473637 4772 scope.go:117] "RemoveContainer" containerID="891a196fc4fbfd6fd254de7d8acfa840ad89166efe2aae3755959d4061df327e" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.475281 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"891a196fc4fbfd6fd254de7d8acfa840ad89166efe2aae3755959d4061df327e\": container with ID starting with 891a196fc4fbfd6fd254de7d8acfa840ad89166efe2aae3755959d4061df327e not found: ID does not exist" containerID="891a196fc4fbfd6fd254de7d8acfa840ad89166efe2aae3755959d4061df327e" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.475317 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"891a196fc4fbfd6fd254de7d8acfa840ad89166efe2aae3755959d4061df327e"} err="failed to get container status \"891a196fc4fbfd6fd254de7d8acfa840ad89166efe2aae3755959d4061df327e\": rpc error: code = NotFound desc = could not find container \"891a196fc4fbfd6fd254de7d8acfa840ad89166efe2aae3755959d4061df327e\": container with ID starting with 891a196fc4fbfd6fd254de7d8acfa840ad89166efe2aae3755959d4061df327e not found: ID does not exist" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.475343 4772 scope.go:117] "RemoveContainer" containerID="a2e75315afb920925e12db40d3166aec37363a00b9620db15e2cf976b7c362c8" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.477964 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2e75315afb920925e12db40d3166aec37363a00b9620db15e2cf976b7c362c8\": container with ID starting with a2e75315afb920925e12db40d3166aec37363a00b9620db15e2cf976b7c362c8 not found: ID does not exist" containerID="a2e75315afb920925e12db40d3166aec37363a00b9620db15e2cf976b7c362c8" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.477999 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2e75315afb920925e12db40d3166aec37363a00b9620db15e2cf976b7c362c8"} err="failed to get container status \"a2e75315afb920925e12db40d3166aec37363a00b9620db15e2cf976b7c362c8\": rpc error: code = NotFound desc = could not find container \"a2e75315afb920925e12db40d3166aec37363a00b9620db15e2cf976b7c362c8\": container with ID starting with a2e75315afb920925e12db40d3166aec37363a00b9620db15e2cf976b7c362c8 not found: ID does not exist" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.478021 4772 scope.go:117] "RemoveContainer" containerID="16478a3399ae464f44f2335d6adb9e9f0cad87f3f55187da1906c5dcf87534dd" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.478292 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.482078 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16478a3399ae464f44f2335d6adb9e9f0cad87f3f55187da1906c5dcf87534dd\": container with ID starting with 16478a3399ae464f44f2335d6adb9e9f0cad87f3f55187da1906c5dcf87534dd not found: ID does not exist" containerID="16478a3399ae464f44f2335d6adb9e9f0cad87f3f55187da1906c5dcf87534dd" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.482124 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16478a3399ae464f44f2335d6adb9e9f0cad87f3f55187da1906c5dcf87534dd"} err="failed to get container status \"16478a3399ae464f44f2335d6adb9e9f0cad87f3f55187da1906c5dcf87534dd\": rpc error: code = NotFound desc = could not find container \"16478a3399ae464f44f2335d6adb9e9f0cad87f3f55187da1906c5dcf87534dd\": container with ID starting with 16478a3399ae464f44f2335d6adb9e9f0cad87f3f55187da1906c5dcf87534dd not found: ID does not exist" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.482155 4772 scope.go:117] "RemoveContainer" containerID="321d45f45e6bfbd7d001b6cc4114cba4c366b13901f8014ac260d8f8a05b7c8d" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.484053 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" (UID: "4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.487043 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dfgjh" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.509023 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0b33686-8107-4caf-b67f-3c608119a049-catalog-content\") pod \"d0b33686-8107-4caf-b67f-3c608119a049\" (UID: \"d0b33686-8107-4caf-b67f-3c608119a049\") " Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.509082 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0b33686-8107-4caf-b67f-3c608119a049-utilities\") pod \"d0b33686-8107-4caf-b67f-3c608119a049\" (UID: \"d0b33686-8107-4caf-b67f-3c608119a049\") " Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.509191 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67q8x\" (UniqueName: \"kubernetes.io/projected/dd415ccf-2b4a-4797-962f-a464ef96bc22-kube-api-access-67q8x\") pod \"dd415ccf-2b4a-4797-962f-a464ef96bc22\" (UID: \"dd415ccf-2b4a-4797-962f-a464ef96bc22\") " Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.509252 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnhcz\" (UniqueName: \"kubernetes.io/projected/d0b33686-8107-4caf-b67f-3c608119a049-kube-api-access-fnhcz\") pod \"d0b33686-8107-4caf-b67f-3c608119a049\" (UID: \"d0b33686-8107-4caf-b67f-3c608119a049\") " Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.509277 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd415ccf-2b4a-4797-962f-a464ef96bc22-utilities\") pod \"dd415ccf-2b4a-4797-962f-a464ef96bc22\" (UID: \"dd415ccf-2b4a-4797-962f-a464ef96bc22\") " Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.509318 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f637b998-b13b-486d-9042-4cd40a01c833-catalog-content\") pod \"f637b998-b13b-486d-9042-4cd40a01c833\" (UID: \"f637b998-b13b-486d-9042-4cd40a01c833\") " Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.509348 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd415ccf-2b4a-4797-962f-a464ef96bc22-catalog-content\") pod \"dd415ccf-2b4a-4797-962f-a464ef96bc22\" (UID: \"dd415ccf-2b4a-4797-962f-a464ef96bc22\") " Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.509380 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvxxl\" (UniqueName: \"kubernetes.io/projected/f637b998-b13b-486d-9042-4cd40a01c833-kube-api-access-vvxxl\") pod \"f637b998-b13b-486d-9042-4cd40a01c833\" (UID: \"f637b998-b13b-486d-9042-4cd40a01c833\") " Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.509418 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f637b998-b13b-486d-9042-4cd40a01c833-utilities\") pod \"f637b998-b13b-486d-9042-4cd40a01c833\" (UID: \"f637b998-b13b-486d-9042-4cd40a01c833\") " Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.509660 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96e88efd-1f25-4e44-b459-ab773db93656-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.509677 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.509690 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zvrk\" (UniqueName: \"kubernetes.io/projected/4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7-kube-api-access-4zvrk\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.509705 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fhkv\" (UniqueName: \"kubernetes.io/projected/96e88efd-1f25-4e44-b459-ab773db93656-kube-api-access-7fhkv\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.509718 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.523531 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f637b998-b13b-486d-9042-4cd40a01c833-utilities" (OuterVolumeSpecName: "utilities") pod "f637b998-b13b-486d-9042-4cd40a01c833" (UID: "f637b998-b13b-486d-9042-4cd40a01c833"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.527825 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd415ccf-2b4a-4797-962f-a464ef96bc22-kube-api-access-67q8x" (OuterVolumeSpecName: "kube-api-access-67q8x") pod "dd415ccf-2b4a-4797-962f-a464ef96bc22" (UID: "dd415ccf-2b4a-4797-962f-a464ef96bc22"). InnerVolumeSpecName "kube-api-access-67q8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.532185 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f637b998-b13b-486d-9042-4cd40a01c833-kube-api-access-vvxxl" (OuterVolumeSpecName: "kube-api-access-vvxxl") pod "f637b998-b13b-486d-9042-4cd40a01c833" (UID: "f637b998-b13b-486d-9042-4cd40a01c833"). InnerVolumeSpecName "kube-api-access-vvxxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.532298 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0b33686-8107-4caf-b67f-3c608119a049-kube-api-access-fnhcz" (OuterVolumeSpecName: "kube-api-access-fnhcz") pod "d0b33686-8107-4caf-b67f-3c608119a049" (UID: "d0b33686-8107-4caf-b67f-3c608119a049"). InnerVolumeSpecName "kube-api-access-fnhcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.551328 4772 scope.go:117] "RemoveContainer" containerID="7127f22fb066327a7e4cfc38f52653579bcf1558a1847106ba8e7a3945fa02a6" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.581203 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd415ccf-2b4a-4797-962f-a464ef96bc22-utilities" (OuterVolumeSpecName: "utilities") pod "dd415ccf-2b4a-4797-962f-a464ef96bc22" (UID: "dd415ccf-2b4a-4797-962f-a464ef96bc22"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.592037 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96e88efd-1f25-4e44-b459-ab773db93656-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "96e88efd-1f25-4e44-b459-ab773db93656" (UID: "96e88efd-1f25-4e44-b459-ab773db93656"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.604262 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0b33686-8107-4caf-b67f-3c608119a049-utilities" (OuterVolumeSpecName: "utilities") pod "d0b33686-8107-4caf-b67f-3c608119a049" (UID: "d0b33686-8107-4caf-b67f-3c608119a049"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.617497 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cbabfa8-79d8-4b23-b186-b40ba8b3017e-catalog-content\") pod \"8cbabfa8-79d8-4b23-b186-b40ba8b3017e\" (UID: \"8cbabfa8-79d8-4b23-b186-b40ba8b3017e\") " Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.617598 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cbabfa8-79d8-4b23-b186-b40ba8b3017e-utilities\") pod \"8cbabfa8-79d8-4b23-b186-b40ba8b3017e\" (UID: \"8cbabfa8-79d8-4b23-b186-b40ba8b3017e\") " Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.617630 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvb9h\" (UniqueName: \"kubernetes.io/projected/c8ebf890-c3b0-468e-bf7d-0ec590df084b-kube-api-access-vvb9h\") pod \"c8ebf890-c3b0-468e-bf7d-0ec590df084b\" (UID: \"c8ebf890-c3b0-468e-bf7d-0ec590df084b\") " Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.617659 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c8ebf890-c3b0-468e-bf7d-0ec590df084b-marketplace-operator-metrics\") pod \"c8ebf890-c3b0-468e-bf7d-0ec590df084b\" (UID: \"c8ebf890-c3b0-468e-bf7d-0ec590df084b\") " Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.617765 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87w6q\" (UniqueName: \"kubernetes.io/projected/8cbabfa8-79d8-4b23-b186-b40ba8b3017e-kube-api-access-87w6q\") pod \"8cbabfa8-79d8-4b23-b186-b40ba8b3017e\" (UID: \"8cbabfa8-79d8-4b23-b186-b40ba8b3017e\") " Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.617810 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c8ebf890-c3b0-468e-bf7d-0ec590df084b-marketplace-trusted-ca\") pod \"c8ebf890-c3b0-468e-bf7d-0ec590df084b\" (UID: \"c8ebf890-c3b0-468e-bf7d-0ec590df084b\") " Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.618031 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnhcz\" (UniqueName: \"kubernetes.io/projected/d0b33686-8107-4caf-b67f-3c608119a049-kube-api-access-fnhcz\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.618046 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd415ccf-2b4a-4797-962f-a464ef96bc22-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.618058 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvxxl\" (UniqueName: \"kubernetes.io/projected/f637b998-b13b-486d-9042-4cd40a01c833-kube-api-access-vvxxl\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.618070 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f637b998-b13b-486d-9042-4cd40a01c833-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.618083 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96e88efd-1f25-4e44-b459-ab773db93656-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.618094 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0b33686-8107-4caf-b67f-3c608119a049-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.618105 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67q8x\" (UniqueName: \"kubernetes.io/projected/dd415ccf-2b4a-4797-962f-a464ef96bc22-kube-api-access-67q8x\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.619334 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8ebf890-c3b0-468e-bf7d-0ec590df084b-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "c8ebf890-c3b0-468e-bf7d-0ec590df084b" (UID: "c8ebf890-c3b0-468e-bf7d-0ec590df084b"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.622029 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cbabfa8-79d8-4b23-b186-b40ba8b3017e-utilities" (OuterVolumeSpecName: "utilities") pod "8cbabfa8-79d8-4b23-b186-b40ba8b3017e" (UID: "8cbabfa8-79d8-4b23-b186-b40ba8b3017e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.624670 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8ebf890-c3b0-468e-bf7d-0ec590df084b-kube-api-access-vvb9h" (OuterVolumeSpecName: "kube-api-access-vvb9h") pod "c8ebf890-c3b0-468e-bf7d-0ec590df084b" (UID: "c8ebf890-c3b0-468e-bf7d-0ec590df084b"). InnerVolumeSpecName "kube-api-access-vvb9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.625998 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cbabfa8-79d8-4b23-b186-b40ba8b3017e-kube-api-access-87w6q" (OuterVolumeSpecName: "kube-api-access-87w6q") pod "8cbabfa8-79d8-4b23-b186-b40ba8b3017e" (UID: "8cbabfa8-79d8-4b23-b186-b40ba8b3017e"). InnerVolumeSpecName "kube-api-access-87w6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.627393 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2glnd"] Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.630199 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd415ccf-2b4a-4797-962f-a464ef96bc22-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dd415ccf-2b4a-4797-962f-a464ef96bc22" (UID: "dd415ccf-2b4a-4797-962f-a464ef96bc22"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.632389 4772 scope.go:117] "RemoveContainer" containerID="19447f4e7cf721e9d74fba0b607574c2c8311dd4da0f4f4a1ba431d8c2a3ca81" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.638285 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8ebf890-c3b0-468e-bf7d-0ec590df084b-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "c8ebf890-c3b0-468e-bf7d-0ec590df084b" (UID: "c8ebf890-c3b0-468e-bf7d-0ec590df084b"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.660617 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xp8ph" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.673494 4772 scope.go:117] "RemoveContainer" containerID="321d45f45e6bfbd7d001b6cc4114cba4c366b13901f8014ac260d8f8a05b7c8d" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.674241 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"321d45f45e6bfbd7d001b6cc4114cba4c366b13901f8014ac260d8f8a05b7c8d\": container with ID starting with 321d45f45e6bfbd7d001b6cc4114cba4c366b13901f8014ac260d8f8a05b7c8d not found: ID does not exist" containerID="321d45f45e6bfbd7d001b6cc4114cba4c366b13901f8014ac260d8f8a05b7c8d" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.674283 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"321d45f45e6bfbd7d001b6cc4114cba4c366b13901f8014ac260d8f8a05b7c8d"} err="failed to get container status \"321d45f45e6bfbd7d001b6cc4114cba4c366b13901f8014ac260d8f8a05b7c8d\": rpc error: code = NotFound desc = could not find container \"321d45f45e6bfbd7d001b6cc4114cba4c366b13901f8014ac260d8f8a05b7c8d\": container with ID starting with 321d45f45e6bfbd7d001b6cc4114cba4c366b13901f8014ac260d8f8a05b7c8d not found: ID does not exist" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.674313 4772 scope.go:117] "RemoveContainer" containerID="7127f22fb066327a7e4cfc38f52653579bcf1558a1847106ba8e7a3945fa02a6" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.680306 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7127f22fb066327a7e4cfc38f52653579bcf1558a1847106ba8e7a3945fa02a6\": container with ID starting with 7127f22fb066327a7e4cfc38f52653579bcf1558a1847106ba8e7a3945fa02a6 not found: ID does not exist" containerID="7127f22fb066327a7e4cfc38f52653579bcf1558a1847106ba8e7a3945fa02a6" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.680347 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7127f22fb066327a7e4cfc38f52653579bcf1558a1847106ba8e7a3945fa02a6"} err="failed to get container status \"7127f22fb066327a7e4cfc38f52653579bcf1558a1847106ba8e7a3945fa02a6\": rpc error: code = NotFound desc = could not find container \"7127f22fb066327a7e4cfc38f52653579bcf1558a1847106ba8e7a3945fa02a6\": container with ID starting with 7127f22fb066327a7e4cfc38f52653579bcf1558a1847106ba8e7a3945fa02a6 not found: ID does not exist" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.680374 4772 scope.go:117] "RemoveContainer" containerID="19447f4e7cf721e9d74fba0b607574c2c8311dd4da0f4f4a1ba431d8c2a3ca81" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.685473 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19447f4e7cf721e9d74fba0b607574c2c8311dd4da0f4f4a1ba431d8c2a3ca81\": container with ID starting with 19447f4e7cf721e9d74fba0b607574c2c8311dd4da0f4f4a1ba431d8c2a3ca81 not found: ID does not exist" containerID="19447f4e7cf721e9d74fba0b607574c2c8311dd4da0f4f4a1ba431d8c2a3ca81" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.685629 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19447f4e7cf721e9d74fba0b607574c2c8311dd4da0f4f4a1ba431d8c2a3ca81"} err="failed to get container status \"19447f4e7cf721e9d74fba0b607574c2c8311dd4da0f4f4a1ba431d8c2a3ca81\": rpc error: code = NotFound desc = could not find container \"19447f4e7cf721e9d74fba0b607574c2c8311dd4da0f4f4a1ba431d8c2a3ca81\": container with ID starting with 19447f4e7cf721e9d74fba0b607574c2c8311dd4da0f4f4a1ba431d8c2a3ca81 not found: ID does not exist" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.685730 4772 scope.go:117] "RemoveContainer" containerID="7cdcc72178b424cbd1356a3055a7eccdb609ddd782039c152e1c33a7dc48ebfd" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.719036 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd415ccf-2b4a-4797-962f-a464ef96bc22-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.719096 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87w6q\" (UniqueName: \"kubernetes.io/projected/8cbabfa8-79d8-4b23-b186-b40ba8b3017e-kube-api-access-87w6q\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.719109 4772 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c8ebf890-c3b0-468e-bf7d-0ec590df084b-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.719117 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cbabfa8-79d8-4b23-b186-b40ba8b3017e-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.719128 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvb9h\" (UniqueName: \"kubernetes.io/projected/c8ebf890-c3b0-468e-bf7d-0ec590df084b-kube-api-access-vvb9h\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.719152 4772 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c8ebf890-c3b0-468e-bf7d-0ec590df084b-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.722737 4772 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.722900 4772 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723303 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" containerName="extract-content" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723332 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" containerName="extract-content" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723355 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723364 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723375 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f637b998-b13b-486d-9042-4cd40a01c833" containerName="extract-content" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723383 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f637b998-b13b-486d-9042-4cd40a01c833" containerName="extract-content" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723390 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723399 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723411 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" containerName="registry-server" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723420 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" containerName="registry-server" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723433 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723443 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723455 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96e88efd-1f25-4e44-b459-ab773db93656" containerName="extract-content" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723463 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="96e88efd-1f25-4e44-b459-ab773db93656" containerName="extract-content" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723473 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" containerName="extract-utilities" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723481 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" containerName="extract-utilities" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723494 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96e88efd-1f25-4e44-b459-ab773db93656" containerName="registry-server" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723501 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="96e88efd-1f25-4e44-b459-ab773db93656" containerName="registry-server" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723508 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" containerName="registry-server" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723515 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" containerName="registry-server" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723525 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" containerName="extract-utilities" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723532 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" containerName="extract-utilities" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723541 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" containerName="extract-content" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723548 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" containerName="extract-content" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723558 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723565 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723577 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" containerName="extract-content" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723585 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" containerName="extract-content" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723595 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f637b998-b13b-486d-9042-4cd40a01c833" containerName="extract-utilities" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723602 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f637b998-b13b-486d-9042-4cd40a01c833" containerName="extract-utilities" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723615 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" containerName="registry-server" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723623 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" containerName="registry-server" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723633 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0b33686-8107-4caf-b67f-3c608119a049" containerName="registry-server" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723642 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0b33686-8107-4caf-b67f-3c608119a049" containerName="registry-server" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723654 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0b33686-8107-4caf-b67f-3c608119a049" containerName="extract-content" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723663 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0b33686-8107-4caf-b67f-3c608119a049" containerName="extract-content" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723675 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96e88efd-1f25-4e44-b459-ab773db93656" containerName="extract-utilities" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723683 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="96e88efd-1f25-4e44-b459-ab773db93656" containerName="extract-utilities" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723693 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723700 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723711 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" containerName="extract-utilities" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723718 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" containerName="extract-utilities" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723731 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" containerName="extract-content" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723739 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" containerName="extract-content" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723753 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0b33686-8107-4caf-b67f-3c608119a049" containerName="extract-utilities" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723764 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0b33686-8107-4caf-b67f-3c608119a049" containerName="extract-utilities" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723775 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f637b998-b13b-486d-9042-4cd40a01c833" containerName="registry-server" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723784 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f637b998-b13b-486d-9042-4cd40a01c833" containerName="registry-server" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723795 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723807 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723817 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723826 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723837 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723845 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723853 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" containerName="marketplace-operator" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723864 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" containerName="marketplace-operator" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723875 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" containerName="registry-server" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723883 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" containerName="registry-server" Jan 27 15:10:51 crc kubenswrapper[4772]: E0127 15:10:51.723893 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" containerName="extract-utilities" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.723901 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" containerName="extract-utilities" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.726672 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08" gracePeriod=15 Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.726895 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d" gracePeriod=15 Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.727064 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a" gracePeriod=15 Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.727143 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59" gracePeriod=15 Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.726850 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" containerName="registry-server" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.727244 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.727255 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.727266 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" containerName="marketplace-operator" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.727277 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="96e88efd-1f25-4e44-b459-ab773db93656" containerName="registry-server" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.727287 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" containerName="registry-server" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.727295 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.727301 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="f637b998-b13b-486d-9042-4cd40a01c833" containerName="registry-server" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.727308 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.727317 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" containerName="registry-server" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.727325 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.727333 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" containerName="registry-server" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.727341 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0b33686-8107-4caf-b67f-3c608119a049" containerName="registry-server" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.727352 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.727526 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.727929 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b" gracePeriod=15 Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.728827 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cbabfa8-79d8-4b23-b186-b40ba8b3017e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8cbabfa8-79d8-4b23-b186-b40ba8b3017e" (UID: "8cbabfa8-79d8-4b23-b186-b40ba8b3017e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.730808 4772 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.734010 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.746123 4772 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.820519 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac2b5800-ce98-4847-bfcd-67a97375aa1b-catalog-content\") pod \"ac2b5800-ce98-4847-bfcd-67a97375aa1b\" (UID: \"ac2b5800-ce98-4847-bfcd-67a97375aa1b\") " Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.820975 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.821079 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac2b5800-ce98-4847-bfcd-67a97375aa1b-utilities\") pod \"ac2b5800-ce98-4847-bfcd-67a97375aa1b\" (UID: \"ac2b5800-ce98-4847-bfcd-67a97375aa1b\") " Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.821253 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7g6kk\" (UniqueName: \"kubernetes.io/projected/ac2b5800-ce98-4847-bfcd-67a97375aa1b-kube-api-access-7g6kk\") pod \"ac2b5800-ce98-4847-bfcd-67a97375aa1b\" (UID: \"ac2b5800-ce98-4847-bfcd-67a97375aa1b\") " Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.821544 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cbabfa8-79d8-4b23-b186-b40ba8b3017e-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.822567 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac2b5800-ce98-4847-bfcd-67a97375aa1b-utilities" (OuterVolumeSpecName: "utilities") pod "ac2b5800-ce98-4847-bfcd-67a97375aa1b" (UID: "ac2b5800-ce98-4847-bfcd-67a97375aa1b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.828299 4772 scope.go:117] "RemoveContainer" containerID="aff635156e5d675ca2fa44615b92754120942925bd9551591ea562f79911f6af" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.845903 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac2b5800-ce98-4847-bfcd-67a97375aa1b-kube-api-access-7g6kk" (OuterVolumeSpecName: "kube-api-access-7g6kk") pod "ac2b5800-ce98-4847-bfcd-67a97375aa1b" (UID: "ac2b5800-ce98-4847-bfcd-67a97375aa1b"). InnerVolumeSpecName "kube-api-access-7g6kk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.853509 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0b33686-8107-4caf-b67f-3c608119a049-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d0b33686-8107-4caf-b67f-3c608119a049" (UID: "d0b33686-8107-4caf-b67f-3c608119a049"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.862204 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f637b998-b13b-486d-9042-4cd40a01c833-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f637b998-b13b-486d-9042-4cd40a01c833" (UID: "f637b998-b13b-486d-9042-4cd40a01c833"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.882471 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac2b5800-ce98-4847-bfcd-67a97375aa1b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ac2b5800-ce98-4847-bfcd-67a97375aa1b" (UID: "ac2b5800-ce98-4847-bfcd-67a97375aa1b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.904918 4772 scope.go:117] "RemoveContainer" containerID="759437d05ef598aec5d4669f7ffea07fc52730444984e390ed6235fe2f84e271" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.923947 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.923989 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.924016 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.924135 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.924179 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.924208 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.924243 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.924285 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.924350 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac2b5800-ce98-4847-bfcd-67a97375aa1b-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.924364 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac2b5800-ce98-4847-bfcd-67a97375aa1b-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.924376 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0b33686-8107-4caf-b67f-3c608119a049-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.924390 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7g6kk\" (UniqueName: \"kubernetes.io/projected/ac2b5800-ce98-4847-bfcd-67a97375aa1b-kube-api-access-7g6kk\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:51 crc kubenswrapper[4772]: I0127 15:10:51.924403 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f637b998-b13b-486d-9042-4cd40a01c833-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.013822 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.014193 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.014606 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.015007 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.015376 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.015641 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: E0127 15:10:52.016871 4772 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.129.56.134:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.025900 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.025983 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.026009 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.025997 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.026030 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.026068 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.026085 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.026111 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.026133 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.026152 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.026216 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.026246 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.026281 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.026313 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.026312 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.026390 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:10:52 crc kubenswrapper[4772]: E0127 15:10:52.066526 4772 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.129.56.134:6443: connect: connection refused" event="&Event{ObjectMeta:{marketplace-operator-79b997595-2glnd.188e9f1d5cf52191 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:marketplace-operator-79b997595-2glnd,UID:d8591d45-25d0-47ea-a856-9cd5334e4a8c,APIVersion:v1,ResourceVersion:29454,FieldPath:spec.containers{marketplace-operator},},Reason:Created,Message:Created container marketplace-operator,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-27 15:10:52.065620369 +0000 UTC m=+238.046229467,LastTimestamp:2026-01-27 15:10:52.065620369 +0000 UTC m=+238.046229467,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.318123 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 27 15:10:52 crc kubenswrapper[4772]: W0127 15:10:52.332720 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-c8e5981d023622831660010c517a0b33cd1e36ddfcc17aac8b21646f3366db9e WatchSource:0}: Error finding container c8e5981d023622831660010c517a0b33cd1e36ddfcc17aac8b21646f3366db9e: Status 404 returned error can't find the container with id c8e5981d023622831660010c517a0b33cd1e36ddfcc17aac8b21646f3366db9e Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.409440 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.411852 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.413138 4772 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08" exitCode=0 Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.413188 4772 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a" exitCode=0 Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.413202 4772 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d" exitCode=0 Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.413212 4772 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59" exitCode=2 Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.413225 4772 scope.go:117] "RemoveContainer" containerID="a6f8b4fa9f839939910224ff95f7788a5cdb3f9ff233a0621e06efdad5c3fa67" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.415742 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-k7pfr_d0b33686-8107-4caf-b67f-3c608119a049/registry-server/0.log" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.416703 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7pfr" event={"ID":"d0b33686-8107-4caf-b67f-3c608119a049","Type":"ContainerDied","Data":"73b204aeb508c9b52f9299efc082b371c9a1b4cb36fb76b5cd3cb0d31f443821"} Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.416755 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k7pfr" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.417428 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.417735 4772 status_manager.go:851] "Failed to get status for pod" podUID="d0b33686-8107-4caf-b67f-3c608119a049" pod="openshift-marketplace/redhat-operators-k7pfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-k7pfr\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.418129 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.418361 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.419982 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jwrpk" event={"ID":"dd415ccf-2b4a-4797-962f-a464ef96bc22","Type":"ContainerDied","Data":"625e8aba305d47657203f1b68ee02b451e267b67921c5b887eaa601500155d6c"} Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.423076 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dfgjh" event={"ID":"8cbabfa8-79d8-4b23-b186-b40ba8b3017e","Type":"ContainerDied","Data":"a43ec03d3d30e58e4b1f455c9ffe4cf357362515c513668efc22ad8bedf315c3"} Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.423153 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dfgjh" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.423990 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.424397 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.424751 4772 status_manager.go:851] "Failed to get status for pod" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" pod="openshift-marketplace/redhat-marketplace-dfgjh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dfgjh\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.425385 4772 status_manager.go:851] "Failed to get status for pod" podUID="d0b33686-8107-4caf-b67f-3c608119a049" pod="openshift-marketplace/redhat-operators-k7pfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-k7pfr\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.425711 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.427014 4772 generic.go:334] "Generic (PLEG): container finished" podID="7bd0c383-7376-4e95-9919-863297cbd807" containerID="0eeaf2bcc5f54216d999847c8ecf3f795fa45776d35017701579dff468e8db9c" exitCode=0 Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.427063 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7bd0c383-7376-4e95-9919-863297cbd807","Type":"ContainerDied","Data":"0eeaf2bcc5f54216d999847c8ecf3f795fa45776d35017701579dff468e8db9c"} Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.428349 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.428625 4772 status_manager.go:851] "Failed to get status for pod" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" pod="openshift-marketplace/redhat-marketplace-dfgjh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dfgjh\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.428900 4772 status_manager.go:851] "Failed to get status for pod" podUID="d0b33686-8107-4caf-b67f-3c608119a049" pod="openshift-marketplace/redhat-operators-k7pfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-k7pfr\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.429144 4772 status_manager.go:851] "Failed to get status for pod" podUID="7bd0c383-7376-4e95-9919-863297cbd807" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.429547 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.430027 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.431805 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-2glnd_d8591d45-25d0-47ea-a856-9cd5334e4a8c/marketplace-operator/0.log" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.431875 4772 generic.go:334] "Generic (PLEG): container finished" podID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" containerID="473257dde7457f1aee0bf59bde3a2b0a19d1fd13941a08f767066393bccd2ab9" exitCode=1 Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.431979 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" event={"ID":"d8591d45-25d0-47ea-a856-9cd5334e4a8c","Type":"ContainerDied","Data":"473257dde7457f1aee0bf59bde3a2b0a19d1fd13941a08f767066393bccd2ab9"} Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.432014 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" event={"ID":"d8591d45-25d0-47ea-a856-9cd5334e4a8c","Type":"ContainerStarted","Data":"6d8062e53fa2b66b9595b12b076dd98687a7682c2e18c22690a2da9d229c8552"} Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.432068 4772 status_manager.go:851] "Failed to get status for pod" podUID="d0b33686-8107-4caf-b67f-3c608119a049" pod="openshift-marketplace/redhat-operators-k7pfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-k7pfr\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.432150 4772 scope.go:117] "RemoveContainer" containerID="473257dde7457f1aee0bf59bde3a2b0a19d1fd13941a08f767066393bccd2ab9" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.432377 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.432806 4772 status_manager.go:851] "Failed to get status for pod" podUID="7bd0c383-7376-4e95-9919-863297cbd807" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.433211 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.433857 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.434515 4772 status_manager.go:851] "Failed to get status for pod" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" pod="openshift-marketplace/redhat-marketplace-dfgjh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dfgjh\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.435018 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.435204 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" event={"ID":"c8ebf890-c3b0-468e-bf7d-0ec590df084b","Type":"ContainerDied","Data":"267d22366b6b80c120159c7b29d573289ed71a1b2d51c437b57f97f84c344fdc"} Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.435273 4772 status_manager.go:851] "Failed to get status for pod" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-2glnd\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.435294 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.435568 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.436000 4772 status_manager.go:851] "Failed to get status for pod" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" pod="openshift-marketplace/redhat-marketplace-dfgjh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dfgjh\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.436378 4772 status_manager.go:851] "Failed to get status for pod" podUID="d0b33686-8107-4caf-b67f-3c608119a049" pod="openshift-marketplace/redhat-operators-k7pfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-k7pfr\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.436576 4772 status_manager.go:851] "Failed to get status for pod" podUID="7bd0c383-7376-4e95-9919-863297cbd807" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.436762 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.437083 4772 status_manager.go:851] "Failed to get status for pod" podUID="d0b33686-8107-4caf-b67f-3c608119a049" pod="openshift-marketplace/redhat-operators-k7pfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-k7pfr\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.437340 4772 status_manager.go:851] "Failed to get status for pod" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4lj2h\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.437533 4772 status_manager.go:851] "Failed to get status for pod" podUID="7bd0c383-7376-4e95-9919-863297cbd807" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.437812 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.438576 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.438775 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-75jrg_f637b998-b13b-486d-9042-4cd40a01c833/registry-server/0.log" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.438806 4772 status_manager.go:851] "Failed to get status for pod" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-2glnd\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.439020 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.439199 4772 status_manager.go:851] "Failed to get status for pod" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" pod="openshift-marketplace/redhat-marketplace-dfgjh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dfgjh\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.439413 4772 status_manager.go:851] "Failed to get status for pod" podUID="d0b33686-8107-4caf-b67f-3c608119a049" pod="openshift-marketplace/redhat-operators-k7pfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-k7pfr\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.439528 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75jrg" event={"ID":"f637b998-b13b-486d-9042-4cd40a01c833","Type":"ContainerDied","Data":"e6520bb81fcb94747f96d0bde0c4f99cc40f51207c4752c3b45e79ef35a202b3"} Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.439600 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-75jrg" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.439686 4772 status_manager.go:851] "Failed to get status for pod" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4lj2h\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.439967 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.440229 4772 status_manager.go:851] "Failed to get status for pod" podUID="7bd0c383-7376-4e95-9919-863297cbd807" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.440442 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.440644 4772 status_manager.go:851] "Failed to get status for pod" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-2glnd\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.440830 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.441008 4772 status_manager.go:851] "Failed to get status for pod" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" pod="openshift-marketplace/redhat-marketplace-dfgjh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dfgjh\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.441291 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.441494 4772 status_manager.go:851] "Failed to get status for pod" podUID="f637b998-b13b-486d-9042-4cd40a01c833" pod="openshift-marketplace/redhat-operators-75jrg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-75jrg\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.441597 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"c8e5981d023622831660010c517a0b33cd1e36ddfcc17aac8b21646f3366db9e"} Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.441699 4772 status_manager.go:851] "Failed to get status for pod" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-2glnd\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.442234 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.442435 4772 status_manager.go:851] "Failed to get status for pod" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" pod="openshift-marketplace/redhat-marketplace-dfgjh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dfgjh\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.443771 4772 status_manager.go:851] "Failed to get status for pod" podUID="d0b33686-8107-4caf-b67f-3c608119a049" pod="openshift-marketplace/redhat-operators-k7pfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-k7pfr\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.444055 4772 status_manager.go:851] "Failed to get status for pod" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4lj2h\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.445074 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.445748 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xp8ph" event={"ID":"ac2b5800-ce98-4847-bfcd-67a97375aa1b","Type":"ContainerDied","Data":"d8c8cf461f8b99ac3badb881b8d4938b4c0d57c4110c08a8b732cf718594c112"} Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.445834 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xp8ph" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.446051 4772 status_manager.go:851] "Failed to get status for pod" podUID="7bd0c383-7376-4e95-9919-863297cbd807" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.454405 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.455373 4772 status_manager.go:851] "Failed to get status for pod" podUID="f637b998-b13b-486d-9042-4cd40a01c833" pod="openshift-marketplace/redhat-operators-75jrg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-75jrg\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.455577 4772 status_manager.go:851] "Failed to get status for pod" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" pod="openshift-marketplace/redhat-marketplace-xp8ph" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xp8ph\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.455806 4772 status_manager.go:851] "Failed to get status for pod" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-2glnd\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.456204 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.456388 4772 status_manager.go:851] "Failed to get status for pod" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" pod="openshift-marketplace/redhat-marketplace-dfgjh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dfgjh\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.456598 4772 status_manager.go:851] "Failed to get status for pod" podUID="d0b33686-8107-4caf-b67f-3c608119a049" pod="openshift-marketplace/redhat-operators-k7pfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-k7pfr\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.456784 4772 status_manager.go:851] "Failed to get status for pod" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4lj2h\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.456964 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.457130 4772 status_manager.go:851] "Failed to get status for pod" podUID="7bd0c383-7376-4e95-9919-863297cbd807" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.472619 4772 scope.go:117] "RemoveContainer" containerID="1c52c9067d3a0dfe5bf38e17654a83a4d2211b850c4ac05e58ed278ee0de4d7e" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.488388 4772 scope.go:117] "RemoveContainer" containerID="35038329828ccd832c938fb7caf96024c35dc820e86dd3e9aadcf5ac8ef257b5" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.503279 4772 status_manager.go:851] "Failed to get status for pod" podUID="f637b998-b13b-486d-9042-4cd40a01c833" pod="openshift-marketplace/redhat-operators-75jrg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-75jrg\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.503445 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.503589 4772 status_manager.go:851] "Failed to get status for pod" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" pod="openshift-marketplace/redhat-marketplace-xp8ph" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xp8ph\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.503727 4772 status_manager.go:851] "Failed to get status for pod" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-2glnd\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.503856 4772 status_manager.go:851] "Failed to get status for pod" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" pod="openshift-marketplace/redhat-marketplace-dfgjh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dfgjh\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.503985 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.504132 4772 status_manager.go:851] "Failed to get status for pod" podUID="d0b33686-8107-4caf-b67f-3c608119a049" pod="openshift-marketplace/redhat-operators-k7pfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-k7pfr\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.504309 4772 status_manager.go:851] "Failed to get status for pod" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4lj2h\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.504447 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.504579 4772 status_manager.go:851] "Failed to get status for pod" podUID="7bd0c383-7376-4e95-9919-863297cbd807" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.504738 4772 status_manager.go:851] "Failed to get status for pod" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4lj2h\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.504926 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.505253 4772 status_manager.go:851] "Failed to get status for pod" podUID="7bd0c383-7376-4e95-9919-863297cbd807" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.506160 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.506358 4772 status_manager.go:851] "Failed to get status for pod" podUID="f637b998-b13b-486d-9042-4cd40a01c833" pod="openshift-marketplace/redhat-operators-75jrg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-75jrg\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.506550 4772 status_manager.go:851] "Failed to get status for pod" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" pod="openshift-marketplace/redhat-marketplace-xp8ph" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xp8ph\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.506722 4772 status_manager.go:851] "Failed to get status for pod" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-2glnd\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.506893 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.507084 4772 status_manager.go:851] "Failed to get status for pod" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" pod="openshift-marketplace/redhat-marketplace-dfgjh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dfgjh\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.507349 4772 status_manager.go:851] "Failed to get status for pod" podUID="d0b33686-8107-4caf-b67f-3c608119a049" pod="openshift-marketplace/redhat-operators-k7pfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-k7pfr\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.514308 4772 scope.go:117] "RemoveContainer" containerID="142c7ce0c1e97146a8a91a5cb46adbe2ee5537497547ac756fc38abfd7afe96c" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.532057 4772 scope.go:117] "RemoveContainer" containerID="932559e335376251fa378d1d6f007b100323207571225373c52e6683753426ad" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.553630 4772 scope.go:117] "RemoveContainer" containerID="7958617c1c11ed30a9375bc289d465de0afd6e9db9de8e33b0d1ef509c9e2adb" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.575352 4772 scope.go:117] "RemoveContainer" containerID="e90d644c15f4a502d49563577bfa11dc77829d65c3be871a6762542c2dc18bcb" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.589508 4772 scope.go:117] "RemoveContainer" containerID="8632589c7dbe4bb64d8d2a9e0983c8088c1ff445e316f1dd7c4e04e72fa148df" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.604495 4772 scope.go:117] "RemoveContainer" containerID="0085f838645429f4fe1db48da5f434fde0158d222d9039bcf74e2bc9cea6bf7f" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.623023 4772 scope.go:117] "RemoveContainer" containerID="2c0b59089ec0a9f3a0a19eaaa3c657c89533289d4c824c8863bf9ca1e0f2856b" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.640306 4772 scope.go:117] "RemoveContainer" containerID="980163ea21706e95ea0803e7c47d4cf7427d498f1351e03fac04539f250bddc6" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.657783 4772 scope.go:117] "RemoveContainer" containerID="f36658ad464a5804d99326d4821347a5bf28ef6ab12d7aacdef462094deb1db8" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.671827 4772 scope.go:117] "RemoveContainer" containerID="b837145fe89058e7be062b6cdd2bde2c15ab5a2a27d1b5c341bb196fff256ac4" Jan 27 15:10:52 crc kubenswrapper[4772]: I0127 15:10:52.715069 4772 scope.go:117] "RemoveContainer" containerID="2b642630dd7f7b63f30ba841d8958c4ae79e62858aee5dff568bed444b47b036" Jan 27 15:10:52 crc kubenswrapper[4772]: E0127 15:10:52.760498 4772 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.129.56.134:6443: connect: connection refused" event="&Event{ObjectMeta:{marketplace-operator-79b997595-2glnd.188e9f1d5cf52191 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:marketplace-operator-79b997595-2glnd,UID:d8591d45-25d0-47ea-a856-9cd5334e4a8c,APIVersion:v1,ResourceVersion:29454,FieldPath:spec.containers{marketplace-operator},},Reason:Created,Message:Created container marketplace-operator,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-27 15:10:52.065620369 +0000 UTC m=+238.046229467,LastTimestamp:2026-01-27 15:10:52.065620369 +0000 UTC m=+238.046229467,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.476918 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"6f6210876fa329e0bc46c9fcfeb492e4200a121e5f183839eb54f27fb32b52f4"} Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.477440 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: E0127 15:10:53.477565 4772 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.129.56.134:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.477721 4772 status_manager.go:851] "Failed to get status for pod" podUID="f637b998-b13b-486d-9042-4cd40a01c833" pod="openshift-marketplace/redhat-operators-75jrg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-75jrg\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.478249 4772 status_manager.go:851] "Failed to get status for pod" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" pod="openshift-marketplace/redhat-marketplace-xp8ph" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xp8ph\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.478597 4772 status_manager.go:851] "Failed to get status for pod" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-2glnd\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.478875 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.479257 4772 status_manager.go:851] "Failed to get status for pod" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" pod="openshift-marketplace/redhat-marketplace-dfgjh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dfgjh\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.479482 4772 status_manager.go:851] "Failed to get status for pod" podUID="d0b33686-8107-4caf-b67f-3c608119a049" pod="openshift-marketplace/redhat-operators-k7pfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-k7pfr\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.479721 4772 status_manager.go:851] "Failed to get status for pod" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4lj2h\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.480050 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.480361 4772 status_manager.go:851] "Failed to get status for pod" podUID="7bd0c383-7376-4e95-9919-863297cbd807" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.481128 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-2glnd_d8591d45-25d0-47ea-a856-9cd5334e4a8c/marketplace-operator/1.log" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.481587 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-2glnd_d8591d45-25d0-47ea-a856-9cd5334e4a8c/marketplace-operator/0.log" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.481631 4772 generic.go:334] "Generic (PLEG): container finished" podID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" containerID="c22226245670f0b13e58561f751beb68f90896b450c95a9a05a5d73430ee3957" exitCode=1 Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.481689 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" event={"ID":"d8591d45-25d0-47ea-a856-9cd5334e4a8c","Type":"ContainerDied","Data":"c22226245670f0b13e58561f751beb68f90896b450c95a9a05a5d73430ee3957"} Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.481792 4772 scope.go:117] "RemoveContainer" containerID="473257dde7457f1aee0bf59bde3a2b0a19d1fd13941a08f767066393bccd2ab9" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.482078 4772 scope.go:117] "RemoveContainer" containerID="c22226245670f0b13e58561f751beb68f90896b450c95a9a05a5d73430ee3957" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.482231 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: E0127 15:10:53.482287 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-2glnd_openshift-marketplace(d8591d45-25d0-47ea-a856-9cd5334e4a8c)\"" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.482542 4772 status_manager.go:851] "Failed to get status for pod" podUID="f637b998-b13b-486d-9042-4cd40a01c833" pod="openshift-marketplace/redhat-operators-75jrg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-75jrg\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.482820 4772 status_manager.go:851] "Failed to get status for pod" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" pod="openshift-marketplace/redhat-marketplace-xp8ph" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xp8ph\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.483039 4772 status_manager.go:851] "Failed to get status for pod" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-2glnd\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.483308 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.483532 4772 status_manager.go:851] "Failed to get status for pod" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" pod="openshift-marketplace/redhat-marketplace-dfgjh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dfgjh\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.483760 4772 status_manager.go:851] "Failed to get status for pod" podUID="d0b33686-8107-4caf-b67f-3c608119a049" pod="openshift-marketplace/redhat-operators-k7pfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-k7pfr\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.484016 4772 status_manager.go:851] "Failed to get status for pod" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4lj2h\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.484331 4772 status_manager.go:851] "Failed to get status for pod" podUID="7bd0c383-7376-4e95-9919-863297cbd807" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.484621 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.486299 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.703211 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.704270 4772 status_manager.go:851] "Failed to get status for pod" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4lj2h\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.704524 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.704836 4772 status_manager.go:851] "Failed to get status for pod" podUID="7bd0c383-7376-4e95-9919-863297cbd807" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.705060 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.705312 4772 status_manager.go:851] "Failed to get status for pod" podUID="f637b998-b13b-486d-9042-4cd40a01c833" pod="openshift-marketplace/redhat-operators-75jrg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-75jrg\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.705801 4772 status_manager.go:851] "Failed to get status for pod" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" pod="openshift-marketplace/redhat-marketplace-xp8ph" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xp8ph\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.706393 4772 status_manager.go:851] "Failed to get status for pod" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-2glnd\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.706680 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.706908 4772 status_manager.go:851] "Failed to get status for pod" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" pod="openshift-marketplace/redhat-marketplace-dfgjh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dfgjh\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.707192 4772 status_manager.go:851] "Failed to get status for pod" podUID="d0b33686-8107-4caf-b67f-3c608119a049" pod="openshift-marketplace/redhat-operators-k7pfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-k7pfr\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.849243 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7bd0c383-7376-4e95-9919-863297cbd807-kube-api-access\") pod \"7bd0c383-7376-4e95-9919-863297cbd807\" (UID: \"7bd0c383-7376-4e95-9919-863297cbd807\") " Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.849291 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7bd0c383-7376-4e95-9919-863297cbd807-kubelet-dir\") pod \"7bd0c383-7376-4e95-9919-863297cbd807\" (UID: \"7bd0c383-7376-4e95-9919-863297cbd807\") " Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.849337 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7bd0c383-7376-4e95-9919-863297cbd807-var-lock\") pod \"7bd0c383-7376-4e95-9919-863297cbd807\" (UID: \"7bd0c383-7376-4e95-9919-863297cbd807\") " Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.849605 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7bd0c383-7376-4e95-9919-863297cbd807-var-lock" (OuterVolumeSpecName: "var-lock") pod "7bd0c383-7376-4e95-9919-863297cbd807" (UID: "7bd0c383-7376-4e95-9919-863297cbd807"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.849641 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7bd0c383-7376-4e95-9919-863297cbd807-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7bd0c383-7376-4e95-9919-863297cbd807" (UID: "7bd0c383-7376-4e95-9919-863297cbd807"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.858616 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bd0c383-7376-4e95-9919-863297cbd807-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7bd0c383-7376-4e95-9919-863297cbd807" (UID: "7bd0c383-7376-4e95-9919-863297cbd807"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.950373 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7bd0c383-7376-4e95-9919-863297cbd807-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.950646 4772 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7bd0c383-7376-4e95-9919-863297cbd807-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:53 crc kubenswrapper[4772]: I0127 15:10:53.950658 4772 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7bd0c383-7376-4e95-9919-863297cbd807-var-lock\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.102394 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.103130 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.103740 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.104177 4772 status_manager.go:851] "Failed to get status for pod" podUID="f637b998-b13b-486d-9042-4cd40a01c833" pod="openshift-marketplace/redhat-operators-75jrg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-75jrg\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.104451 4772 status_manager.go:851] "Failed to get status for pod" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" pod="openshift-marketplace/redhat-marketplace-xp8ph" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xp8ph\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.104708 4772 status_manager.go:851] "Failed to get status for pod" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-2glnd\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.105202 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.105412 4772 status_manager.go:851] "Failed to get status for pod" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" pod="openshift-marketplace/redhat-marketplace-dfgjh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dfgjh\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.105575 4772 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.105810 4772 status_manager.go:851] "Failed to get status for pod" podUID="d0b33686-8107-4caf-b67f-3c608119a049" pod="openshift-marketplace/redhat-operators-k7pfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-k7pfr\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.106063 4772 status_manager.go:851] "Failed to get status for pod" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4lj2h\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.106269 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.106498 4772 status_manager.go:851] "Failed to get status for pod" podUID="7bd0c383-7376-4e95-9919-863297cbd807" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.253685 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.253838 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.253878 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.253912 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.253986 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.253992 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.254328 4772 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.254348 4772 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.254358 4772 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.495962 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7bd0c383-7376-4e95-9919-863297cbd807","Type":"ContainerDied","Data":"27a7557244977e8f16265740d543848969b7bcbc8b87db7409dc58f332f68492"} Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.496009 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27a7557244977e8f16265740d543848969b7bcbc8b87db7409dc58f332f68492" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.495988 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.503619 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-2glnd_d8591d45-25d0-47ea-a856-9cd5334e4a8c/marketplace-operator/1.log" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.504551 4772 scope.go:117] "RemoveContainer" containerID="c22226245670f0b13e58561f751beb68f90896b450c95a9a05a5d73430ee3957" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.505094 4772 status_manager.go:851] "Failed to get status for pod" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" pod="openshift-marketplace/redhat-marketplace-xp8ph" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xp8ph\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: E0127 15:10:54.505101 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-2glnd_openshift-marketplace(d8591d45-25d0-47ea-a856-9cd5334e4a8c)\"" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.505502 4772 status_manager.go:851] "Failed to get status for pod" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-2glnd\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.506023 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.506588 4772 status_manager.go:851] "Failed to get status for pod" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" pod="openshift-marketplace/redhat-marketplace-dfgjh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dfgjh\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.506879 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.506948 4772 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.507417 4772 status_manager.go:851] "Failed to get status for pod" podUID="d0b33686-8107-4caf-b67f-3c608119a049" pod="openshift-marketplace/redhat-operators-k7pfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-k7pfr\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.507737 4772 status_manager.go:851] "Failed to get status for pod" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4lj2h\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.508027 4772 status_manager.go:851] "Failed to get status for pod" podUID="7bd0c383-7376-4e95-9919-863297cbd807" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.508398 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.508449 4772 scope.go:117] "RemoveContainer" containerID="d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.508490 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.508417 4772 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b" exitCode=0 Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.508789 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.509093 4772 status_manager.go:851] "Failed to get status for pod" podUID="f637b998-b13b-486d-9042-4cd40a01c833" pod="openshift-marketplace/redhat-operators-75jrg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-75jrg\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: E0127 15:10:54.509255 4772 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.129.56.134:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.509580 4772 status_manager.go:851] "Failed to get status for pod" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4lj2h\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.509868 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.510357 4772 status_manager.go:851] "Failed to get status for pod" podUID="7bd0c383-7376-4e95-9919-863297cbd807" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.510656 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.510953 4772 status_manager.go:851] "Failed to get status for pod" podUID="f637b998-b13b-486d-9042-4cd40a01c833" pod="openshift-marketplace/redhat-operators-75jrg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-75jrg\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.511448 4772 status_manager.go:851] "Failed to get status for pod" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" pod="openshift-marketplace/redhat-marketplace-xp8ph" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xp8ph\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.511873 4772 status_manager.go:851] "Failed to get status for pod" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-2glnd\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.512404 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.512761 4772 status_manager.go:851] "Failed to get status for pod" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" pod="openshift-marketplace/redhat-marketplace-dfgjh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dfgjh\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.513204 4772 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.513667 4772 status_manager.go:851] "Failed to get status for pod" podUID="d0b33686-8107-4caf-b67f-3c608119a049" pod="openshift-marketplace/redhat-operators-k7pfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-k7pfr\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.520550 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.520974 4772 status_manager.go:851] "Failed to get status for pod" podUID="f637b998-b13b-486d-9042-4cd40a01c833" pod="openshift-marketplace/redhat-operators-75jrg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-75jrg\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.521333 4772 status_manager.go:851] "Failed to get status for pod" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" pod="openshift-marketplace/redhat-marketplace-xp8ph" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xp8ph\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.521807 4772 status_manager.go:851] "Failed to get status for pod" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-2glnd\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.522179 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.522493 4772 status_manager.go:851] "Failed to get status for pod" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" pod="openshift-marketplace/redhat-marketplace-dfgjh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dfgjh\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.522923 4772 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.523384 4772 status_manager.go:851] "Failed to get status for pod" podUID="d0b33686-8107-4caf-b67f-3c608119a049" pod="openshift-marketplace/redhat-operators-k7pfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-k7pfr\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.523557 4772 status_manager.go:851] "Failed to get status for pod" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4lj2h\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.523713 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.523991 4772 status_manager.go:851] "Failed to get status for pod" podUID="7bd0c383-7376-4e95-9919-863297cbd807" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.532047 4772 scope.go:117] "RemoveContainer" containerID="ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.547028 4772 scope.go:117] "RemoveContainer" containerID="cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.558741 4772 scope.go:117] "RemoveContainer" containerID="013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.571971 4772 scope.go:117] "RemoveContainer" containerID="f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.585340 4772 scope.go:117] "RemoveContainer" containerID="6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.603075 4772 scope.go:117] "RemoveContainer" containerID="d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08" Jan 27 15:10:54 crc kubenswrapper[4772]: E0127 15:10:54.606071 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08\": container with ID starting with d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08 not found: ID does not exist" containerID="d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.606104 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08"} err="failed to get container status \"d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08\": rpc error: code = NotFound desc = could not find container \"d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08\": container with ID starting with d1c727e444b798a9f19bb20f2a43ab26b74c929e7fc72824b497ade9bcc2ac08 not found: ID does not exist" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.606127 4772 scope.go:117] "RemoveContainer" containerID="ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a" Jan 27 15:10:54 crc kubenswrapper[4772]: E0127 15:10:54.606409 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\": container with ID starting with ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a not found: ID does not exist" containerID="ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.606433 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a"} err="failed to get container status \"ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\": rpc error: code = NotFound desc = could not find container \"ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a\": container with ID starting with ed0afec69304057b659922d98e91e37f07c44d3ad4b4e6e2e5633f394164ae4a not found: ID does not exist" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.606449 4772 scope.go:117] "RemoveContainer" containerID="cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d" Jan 27 15:10:54 crc kubenswrapper[4772]: E0127 15:10:54.606692 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\": container with ID starting with cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d not found: ID does not exist" containerID="cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.606715 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d"} err="failed to get container status \"cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\": rpc error: code = NotFound desc = could not find container \"cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d\": container with ID starting with cbbc7ef9a19ac21602529b8c46914ec727c2c7517f03372ab9aa3c823d315f4d not found: ID does not exist" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.606734 4772 scope.go:117] "RemoveContainer" containerID="013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59" Jan 27 15:10:54 crc kubenswrapper[4772]: E0127 15:10:54.606951 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\": container with ID starting with 013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59 not found: ID does not exist" containerID="013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.606974 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59"} err="failed to get container status \"013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\": rpc error: code = NotFound desc = could not find container \"013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59\": container with ID starting with 013829a83ab4749028d5a3020a9bab5621cef37de23ed39e25524caf3b340a59 not found: ID does not exist" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.606989 4772 scope.go:117] "RemoveContainer" containerID="f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b" Jan 27 15:10:54 crc kubenswrapper[4772]: E0127 15:10:54.607189 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\": container with ID starting with f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b not found: ID does not exist" containerID="f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.607207 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b"} err="failed to get container status \"f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\": rpc error: code = NotFound desc = could not find container \"f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b\": container with ID starting with f8200e14f2cf3fdae1549435b9f62f8588bd36d23201077561bba97795684d9b not found: ID does not exist" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.607220 4772 scope.go:117] "RemoveContainer" containerID="6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5" Jan 27 15:10:54 crc kubenswrapper[4772]: E0127 15:10:54.607411 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\": container with ID starting with 6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5 not found: ID does not exist" containerID="6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.607431 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5"} err="failed to get container status \"6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\": rpc error: code = NotFound desc = could not find container \"6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5\": container with ID starting with 6b3609eb7fbaf4e8741fc38683a8eb92729ac16475b035eb0476a9546b007bd5 not found: ID does not exist" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.666131 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.666471 4772 status_manager.go:851] "Failed to get status for pod" podUID="f637b998-b13b-486d-9042-4cd40a01c833" pod="openshift-marketplace/redhat-operators-75jrg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-75jrg\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.666752 4772 status_manager.go:851] "Failed to get status for pod" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" pod="openshift-marketplace/redhat-marketplace-xp8ph" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xp8ph\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.666975 4772 status_manager.go:851] "Failed to get status for pod" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-2glnd\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.667272 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.673320 4772 status_manager.go:851] "Failed to get status for pod" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" pod="openshift-marketplace/redhat-marketplace-dfgjh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dfgjh\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.673699 4772 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.674941 4772 status_manager.go:851] "Failed to get status for pod" podUID="d0b33686-8107-4caf-b67f-3c608119a049" pod="openshift-marketplace/redhat-operators-k7pfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-k7pfr\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.675147 4772 status_manager.go:851] "Failed to get status for pod" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4lj2h\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.675445 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.675662 4772 status_manager.go:851] "Failed to get status for pod" podUID="7bd0c383-7376-4e95-9919-863297cbd807" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.677374 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Jan 27 15:10:54 crc kubenswrapper[4772]: E0127 15:10:54.938846 4772 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: E0127 15:10:54.939118 4772 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: E0127 15:10:54.939748 4772 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: E0127 15:10:54.939945 4772 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: E0127 15:10:54.940126 4772 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:54 crc kubenswrapper[4772]: I0127 15:10:54.940158 4772 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Jan 27 15:10:54 crc kubenswrapper[4772]: E0127 15:10:54.940353 4772 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.134:6443: connect: connection refused" interval="200ms" Jan 27 15:10:55 crc kubenswrapper[4772]: E0127 15:10:55.141818 4772 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.134:6443: connect: connection refused" interval="400ms" Jan 27 15:10:55 crc kubenswrapper[4772]: E0127 15:10:55.542493 4772 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.134:6443: connect: connection refused" interval="800ms" Jan 27 15:10:55 crc kubenswrapper[4772]: E0127 15:10:55.950363 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:10:55Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:10:55Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:10:55Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-27T15:10:55Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:024b1ed0676c2e11f6a319392c82e7acd0ceeae31ca00b202307c4d86a796b20\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:ada03173793960eaa0e4263282fcbf5af3dea8aaf2c3b0d864906108db062e8a\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1672061160},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[],\\\"sizeBytes\\\":1201988853},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:19cccb48b9fd18a6ae02a77aeef83cf3d8e0bbde057c41c2a818afab51c859be\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:965650b0707047c6697952f57d2544e475608b828d1a3638867a50a7cdaf87b8\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1186361067},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:6d91aecdb391dd0cbb56f2b6335674bd2b4a25c63f0b9e893ba8977a71be3c0d\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:98739198606db13baf3fa39b12298669778a619dff80b9b5d51987d7f76056c9\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1180173538},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:55 crc kubenswrapper[4772]: E0127 15:10:55.951499 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:55 crc kubenswrapper[4772]: E0127 15:10:55.952029 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:55 crc kubenswrapper[4772]: E0127 15:10:55.952480 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:55 crc kubenswrapper[4772]: E0127 15:10:55.952955 4772 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:10:55 crc kubenswrapper[4772]: E0127 15:10:55.952984 4772 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 27 15:10:56 crc kubenswrapper[4772]: E0127 15:10:56.343219 4772 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.134:6443: connect: connection refused" interval="1.6s" Jan 27 15:10:57 crc kubenswrapper[4772]: E0127 15:10:57.943977 4772 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.134:6443: connect: connection refused" interval="3.2s" Jan 27 15:10:59 crc kubenswrapper[4772]: E0127 15:10:59.737268 4772 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.129.56.134:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" volumeName="registry-storage" Jan 27 15:11:00 crc kubenswrapper[4772]: I0127 15:11:00.963949 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" Jan 27 15:11:00 crc kubenswrapper[4772]: I0127 15:11:00.965023 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" Jan 27 15:11:00 crc kubenswrapper[4772]: I0127 15:11:00.966098 4772 scope.go:117] "RemoveContainer" containerID="c22226245670f0b13e58561f751beb68f90896b450c95a9a05a5d73430ee3957" Jan 27 15:11:00 crc kubenswrapper[4772]: E0127 15:11:00.966381 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-2glnd_openshift-marketplace(d8591d45-25d0-47ea-a856-9cd5334e4a8c)\"" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" Jan 27 15:11:01 crc kubenswrapper[4772]: E0127 15:11:01.145152 4772 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.134:6443: connect: connection refused" interval="6.4s" Jan 27 15:11:01 crc kubenswrapper[4772]: I0127 15:11:01.553921 4772 scope.go:117] "RemoveContainer" containerID="c22226245670f0b13e58561f751beb68f90896b450c95a9a05a5d73430ee3957" Jan 27 15:11:01 crc kubenswrapper[4772]: E0127 15:11:01.554310 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-2glnd_openshift-marketplace(d8591d45-25d0-47ea-a856-9cd5334e4a8c)\"" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" Jan 27 15:11:02 crc kubenswrapper[4772]: E0127 15:11:02.761324 4772 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.129.56.134:6443: connect: connection refused" event="&Event{ObjectMeta:{marketplace-operator-79b997595-2glnd.188e9f1d5cf52191 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:marketplace-operator-79b997595-2glnd,UID:d8591d45-25d0-47ea-a856-9cd5334e4a8c,APIVersion:v1,ResourceVersion:29454,FieldPath:spec.containers{marketplace-operator},},Reason:Created,Message:Created container marketplace-operator,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-27 15:10:52.065620369 +0000 UTC m=+238.046229467,LastTimestamp:2026-01-27 15:10:52.065620369 +0000 UTC m=+238.046229467,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 27 15:11:04 crc kubenswrapper[4772]: I0127 15:11:04.662752 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:11:04 crc kubenswrapper[4772]: I0127 15:11:04.666759 4772 status_manager.go:851] "Failed to get status for pod" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4lj2h\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:04 crc kubenswrapper[4772]: I0127 15:11:04.667315 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:04 crc kubenswrapper[4772]: I0127 15:11:04.667696 4772 status_manager.go:851] "Failed to get status for pod" podUID="7bd0c383-7376-4e95-9919-863297cbd807" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:04 crc kubenswrapper[4772]: I0127 15:11:04.668038 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:04 crc kubenswrapper[4772]: I0127 15:11:04.668475 4772 status_manager.go:851] "Failed to get status for pod" podUID="f637b998-b13b-486d-9042-4cd40a01c833" pod="openshift-marketplace/redhat-operators-75jrg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-75jrg\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:04 crc kubenswrapper[4772]: I0127 15:11:04.669256 4772 status_manager.go:851] "Failed to get status for pod" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" pod="openshift-marketplace/redhat-marketplace-xp8ph" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xp8ph\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:04 crc kubenswrapper[4772]: I0127 15:11:04.669827 4772 status_manager.go:851] "Failed to get status for pod" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-2glnd\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:04 crc kubenswrapper[4772]: I0127 15:11:04.670471 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:04 crc kubenswrapper[4772]: I0127 15:11:04.672876 4772 status_manager.go:851] "Failed to get status for pod" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" pod="openshift-marketplace/redhat-marketplace-dfgjh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dfgjh\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:04 crc kubenswrapper[4772]: I0127 15:11:04.673590 4772 status_manager.go:851] "Failed to get status for pod" podUID="d0b33686-8107-4caf-b67f-3c608119a049" pod="openshift-marketplace/redhat-operators-k7pfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-k7pfr\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:04 crc kubenswrapper[4772]: I0127 15:11:04.674131 4772 status_manager.go:851] "Failed to get status for pod" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4lj2h\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:04 crc kubenswrapper[4772]: I0127 15:11:04.674595 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:04 crc kubenswrapper[4772]: I0127 15:11:04.675005 4772 status_manager.go:851] "Failed to get status for pod" podUID="7bd0c383-7376-4e95-9919-863297cbd807" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:04 crc kubenswrapper[4772]: I0127 15:11:04.675516 4772 status_manager.go:851] "Failed to get status for pod" podUID="f637b998-b13b-486d-9042-4cd40a01c833" pod="openshift-marketplace/redhat-operators-75jrg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-75jrg\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:04 crc kubenswrapper[4772]: I0127 15:11:04.676841 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:04 crc kubenswrapper[4772]: I0127 15:11:04.677220 4772 status_manager.go:851] "Failed to get status for pod" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" pod="openshift-marketplace/redhat-marketplace-xp8ph" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xp8ph\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:04 crc kubenswrapper[4772]: I0127 15:11:04.677643 4772 status_manager.go:851] "Failed to get status for pod" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-2glnd\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:04 crc kubenswrapper[4772]: I0127 15:11:04.678240 4772 status_manager.go:851] "Failed to get status for pod" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" pod="openshift-marketplace/redhat-marketplace-dfgjh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dfgjh\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:04 crc kubenswrapper[4772]: I0127 15:11:04.678786 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:04 crc kubenswrapper[4772]: I0127 15:11:04.679648 4772 status_manager.go:851] "Failed to get status for pod" podUID="d0b33686-8107-4caf-b67f-3c608119a049" pod="openshift-marketplace/redhat-operators-k7pfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-k7pfr\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:04 crc kubenswrapper[4772]: I0127 15:11:04.697636 4772 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="11f71341-1cdc-430d-8d90-a87af2a493f1" Jan 27 15:11:04 crc kubenswrapper[4772]: I0127 15:11:04.697669 4772 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="11f71341-1cdc-430d-8d90-a87af2a493f1" Jan 27 15:11:04 crc kubenswrapper[4772]: E0127 15:11:04.698120 4772 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:11:04 crc kubenswrapper[4772]: I0127 15:11:04.698546 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.576615 4772 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="d1f019d2285511058b91ca49aeeeb1f9b79aa95fe8f3a01d9d191fa07365d0b1" exitCode=0 Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.576728 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"d1f019d2285511058b91ca49aeeeb1f9b79aa95fe8f3a01d9d191fa07365d0b1"} Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.576985 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"487453b2cce3e10be13ca110bb948d070e993f627cca8085aaa057a5ea8d371e"} Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.577548 4772 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="11f71341-1cdc-430d-8d90-a87af2a493f1" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.577564 4772 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="11f71341-1cdc-430d-8d90-a87af2a493f1" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.578028 4772 status_manager.go:851] "Failed to get status for pod" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-2glnd\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:05 crc kubenswrapper[4772]: E0127 15:11:05.578062 4772 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.578284 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.578632 4772 status_manager.go:851] "Failed to get status for pod" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" pod="openshift-marketplace/redhat-marketplace-dfgjh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dfgjh\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.579027 4772 status_manager.go:851] "Failed to get status for pod" podUID="d0b33686-8107-4caf-b67f-3c608119a049" pod="openshift-marketplace/redhat-operators-k7pfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-k7pfr\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.579291 4772 status_manager.go:851] "Failed to get status for pod" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4lj2h\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.579574 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.579814 4772 status_manager.go:851] "Failed to get status for pod" podUID="7bd0c383-7376-4e95-9919-863297cbd807" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.580157 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.580776 4772 status_manager.go:851] "Failed to get status for pod" podUID="f637b998-b13b-486d-9042-4cd40a01c833" pod="openshift-marketplace/redhat-operators-75jrg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-75jrg\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.581121 4772 status_manager.go:851] "Failed to get status for pod" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" pod="openshift-marketplace/redhat-marketplace-xp8ph" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xp8ph\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.581373 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.581434 4772 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c" exitCode=1 Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.581474 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c"} Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.582110 4772 scope.go:117] "RemoveContainer" containerID="573a9989fcd89f53d26d43e7b4c495cd9a4bbe98ebf1ed0a0dfd0e63875d5b8c" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.582184 4772 status_manager.go:851] "Failed to get status for pod" podUID="96e88efd-1f25-4e44-b459-ab773db93656" pod="openshift-marketplace/certified-operators-9wdps" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-9wdps\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.582761 4772 status_manager.go:851] "Failed to get status for pod" podUID="f637b998-b13b-486d-9042-4cd40a01c833" pod="openshift-marketplace/redhat-operators-75jrg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-75jrg\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.583239 4772 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.583586 4772 status_manager.go:851] "Failed to get status for pod" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" pod="openshift-marketplace/redhat-marketplace-xp8ph" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-xp8ph\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.583912 4772 status_manager.go:851] "Failed to get status for pod" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-2glnd\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.584211 4772 status_manager.go:851] "Failed to get status for pod" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" pod="openshift-marketplace/community-operators-95rh9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-95rh9\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.584722 4772 status_manager.go:851] "Failed to get status for pod" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" pod="openshift-marketplace/redhat-marketplace-dfgjh" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-dfgjh\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.585061 4772 status_manager.go:851] "Failed to get status for pod" podUID="d0b33686-8107-4caf-b67f-3c608119a049" pod="openshift-marketplace/redhat-operators-k7pfr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-k7pfr\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.585449 4772 status_manager.go:851] "Failed to get status for pod" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" pod="openshift-marketplace/marketplace-operator-79b997595-4lj2h" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4lj2h\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.585821 4772 status_manager.go:851] "Failed to get status for pod" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" pod="openshift-marketplace/community-operators-jwrpk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-jwrpk\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:05 crc kubenswrapper[4772]: I0127 15:11:05.586097 4772 status_manager.go:851] "Failed to get status for pod" podUID="7bd0c383-7376-4e95-9919-863297cbd807" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.134:6443: connect: connection refused" Jan 27 15:11:06 crc kubenswrapper[4772]: I0127 15:11:06.598383 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 27 15:11:06 crc kubenswrapper[4772]: I0127 15:11:06.598703 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"13aecdb9184e78557877d3f0d60ede8ad7b6734ec979e1825d5e4d812b031a46"} Jan 27 15:11:06 crc kubenswrapper[4772]: I0127 15:11:06.603025 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"19e50c2fa00f22230547bd9cba2a6de870dac135d7760fc3b8c311abd01dc913"} Jan 27 15:11:06 crc kubenswrapper[4772]: I0127 15:11:06.603072 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fe1a327d404ff115772dd864052396e082122483670a12515575f32b9375f496"} Jan 27 15:11:06 crc kubenswrapper[4772]: I0127 15:11:06.603083 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c38a0ebd0db1446c81c4731b72138976a46da7274531d03dbdb5286a6a7936ea"} Jan 27 15:11:06 crc kubenswrapper[4772]: I0127 15:11:06.603093 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9687b41b79ca543950ce1b639f10e2b19aea7dc77c2fa6716ac495daadca5b55"} Jan 27 15:11:06 crc kubenswrapper[4772]: I0127 15:11:06.603100 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5b3a1335d3569ec7885da4d368e7826c73c087ba95e7872fb2892ab273e7a880"} Jan 27 15:11:06 crc kubenswrapper[4772]: I0127 15:11:06.603301 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:11:06 crc kubenswrapper[4772]: I0127 15:11:06.603368 4772 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="11f71341-1cdc-430d-8d90-a87af2a493f1" Jan 27 15:11:06 crc kubenswrapper[4772]: I0127 15:11:06.603387 4772 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="11f71341-1cdc-430d-8d90-a87af2a493f1" Jan 27 15:11:07 crc kubenswrapper[4772]: I0127 15:11:07.023728 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 27 15:11:07 crc kubenswrapper[4772]: I0127 15:11:07.029505 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 27 15:11:07 crc kubenswrapper[4772]: I0127 15:11:07.607204 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 27 15:11:09 crc kubenswrapper[4772]: I0127 15:11:09.699278 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:11:09 crc kubenswrapper[4772]: I0127 15:11:09.699639 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:11:09 crc kubenswrapper[4772]: I0127 15:11:09.704252 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:11:12 crc kubenswrapper[4772]: I0127 15:11:12.422712 4772 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:11:12 crc kubenswrapper[4772]: I0127 15:11:12.628907 4772 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="11f71341-1cdc-430d-8d90-a87af2a493f1" Jan 27 15:11:12 crc kubenswrapper[4772]: I0127 15:11:12.628941 4772 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="11f71341-1cdc-430d-8d90-a87af2a493f1" Jan 27 15:11:12 crc kubenswrapper[4772]: I0127 15:11:12.632937 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:11:13 crc kubenswrapper[4772]: I0127 15:11:13.632407 4772 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="11f71341-1cdc-430d-8d90-a87af2a493f1" Jan 27 15:11:13 crc kubenswrapper[4772]: I0127 15:11:13.632438 4772 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="11f71341-1cdc-430d-8d90-a87af2a493f1" Jan 27 15:11:14 crc kubenswrapper[4772]: I0127 15:11:14.674621 4772 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="8fa90b23-89b2-4953-9e5c-23d515bc2224" Jan 27 15:11:15 crc kubenswrapper[4772]: I0127 15:11:15.664703 4772 scope.go:117] "RemoveContainer" containerID="c22226245670f0b13e58561f751beb68f90896b450c95a9a05a5d73430ee3957" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.312617 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" podUID="06cdc094-b372-4016-bc5e-4c15a28e032e" containerName="oauth-openshift" containerID="cri-o://437c578755bfcacf0145c1b3dcede3b1938b4e11e6ad9c7db9d8ac6a8b6df37e" gracePeriod=15 Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.652024 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-2glnd_d8591d45-25d0-47ea-a856-9cd5334e4a8c/marketplace-operator/2.log" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.654520 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-2glnd_d8591d45-25d0-47ea-a856-9cd5334e4a8c/marketplace-operator/1.log" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.654572 4772 generic.go:334] "Generic (PLEG): container finished" podID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" containerID="89be95f2c0621af885c1da302ff7e0b3e84dda04f9608b6e8af398107e1e9399" exitCode=1 Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.654630 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" event={"ID":"d8591d45-25d0-47ea-a856-9cd5334e4a8c","Type":"ContainerDied","Data":"89be95f2c0621af885c1da302ff7e0b3e84dda04f9608b6e8af398107e1e9399"} Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.654749 4772 scope.go:117] "RemoveContainer" containerID="c22226245670f0b13e58561f751beb68f90896b450c95a9a05a5d73430ee3957" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.655658 4772 scope.go:117] "RemoveContainer" containerID="89be95f2c0621af885c1da302ff7e0b3e84dda04f9608b6e8af398107e1e9399" Jan 27 15:11:16 crc kubenswrapper[4772]: E0127 15:11:16.656015 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-2glnd_openshift-marketplace(d8591d45-25d0-47ea-a856-9cd5334e4a8c)\"" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.656937 4772 generic.go:334] "Generic (PLEG): container finished" podID="06cdc094-b372-4016-bc5e-4c15a28e032e" containerID="437c578755bfcacf0145c1b3dcede3b1938b4e11e6ad9c7db9d8ac6a8b6df37e" exitCode=0 Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.657012 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" event={"ID":"06cdc094-b372-4016-bc5e-4c15a28e032e","Type":"ContainerDied","Data":"437c578755bfcacf0145c1b3dcede3b1938b4e11e6ad9c7db9d8ac6a8b6df37e"} Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.657055 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" event={"ID":"06cdc094-b372-4016-bc5e-4c15a28e032e","Type":"ContainerDied","Data":"3982e848b6f4ab4d0d2958e425dd1a480bb7b8b136363856076bf9ce68e097fb"} Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.657075 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3982e848b6f4ab4d0d2958e425dd1a480bb7b8b136363856076bf9ce68e097fb" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.682371 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.839682 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "06cdc094-b372-4016-bc5e-4c15a28e032e" (UID: "06cdc094-b372-4016-bc5e-4c15a28e032e"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.838282 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-cliconfig\") pod \"06cdc094-b372-4016-bc5e-4c15a28e032e\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.839889 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-user-idp-0-file-data\") pod \"06cdc094-b372-4016-bc5e-4c15a28e032e\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.841132 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-trusted-ca-bundle\") pod \"06cdc094-b372-4016-bc5e-4c15a28e032e\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.841493 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-ocp-branding-template\") pod \"06cdc094-b372-4016-bc5e-4c15a28e032e\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.841564 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-session\") pod \"06cdc094-b372-4016-bc5e-4c15a28e032e\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.841617 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/06cdc094-b372-4016-bc5e-4c15a28e032e-audit-policies\") pod \"06cdc094-b372-4016-bc5e-4c15a28e032e\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.841677 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-user-template-provider-selection\") pod \"06cdc094-b372-4016-bc5e-4c15a28e032e\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.841723 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-serving-cert\") pod \"06cdc094-b372-4016-bc5e-4c15a28e032e\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.841762 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/06cdc094-b372-4016-bc5e-4c15a28e032e-audit-dir\") pod \"06cdc094-b372-4016-bc5e-4c15a28e032e\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.841825 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-service-ca\") pod \"06cdc094-b372-4016-bc5e-4c15a28e032e\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.841859 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06cdc094-b372-4016-bc5e-4c15a28e032e-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "06cdc094-b372-4016-bc5e-4c15a28e032e" (UID: "06cdc094-b372-4016-bc5e-4c15a28e032e"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.841962 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5f6v\" (UniqueName: \"kubernetes.io/projected/06cdc094-b372-4016-bc5e-4c15a28e032e-kube-api-access-l5f6v\") pod \"06cdc094-b372-4016-bc5e-4c15a28e032e\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.842011 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-user-template-error\") pod \"06cdc094-b372-4016-bc5e-4c15a28e032e\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.842078 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-router-certs\") pod \"06cdc094-b372-4016-bc5e-4c15a28e032e\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.842115 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-user-template-login\") pod \"06cdc094-b372-4016-bc5e-4c15a28e032e\" (UID: \"06cdc094-b372-4016-bc5e-4c15a28e032e\") " Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.842076 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "06cdc094-b372-4016-bc5e-4c15a28e032e" (UID: "06cdc094-b372-4016-bc5e-4c15a28e032e"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.842560 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06cdc094-b372-4016-bc5e-4c15a28e032e-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "06cdc094-b372-4016-bc5e-4c15a28e032e" (UID: "06cdc094-b372-4016-bc5e-4c15a28e032e"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.842766 4772 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.842804 4772 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/06cdc094-b372-4016-bc5e-4c15a28e032e-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.842830 4772 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/06cdc094-b372-4016-bc5e-4c15a28e032e-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.842855 4772 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.842985 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "06cdc094-b372-4016-bc5e-4c15a28e032e" (UID: "06cdc094-b372-4016-bc5e-4c15a28e032e"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.847069 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "06cdc094-b372-4016-bc5e-4c15a28e032e" (UID: "06cdc094-b372-4016-bc5e-4c15a28e032e"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.848622 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "06cdc094-b372-4016-bc5e-4c15a28e032e" (UID: "06cdc094-b372-4016-bc5e-4c15a28e032e"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.848916 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06cdc094-b372-4016-bc5e-4c15a28e032e-kube-api-access-l5f6v" (OuterVolumeSpecName: "kube-api-access-l5f6v") pod "06cdc094-b372-4016-bc5e-4c15a28e032e" (UID: "06cdc094-b372-4016-bc5e-4c15a28e032e"). InnerVolumeSpecName "kube-api-access-l5f6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.849049 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "06cdc094-b372-4016-bc5e-4c15a28e032e" (UID: "06cdc094-b372-4016-bc5e-4c15a28e032e"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.852049 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "06cdc094-b372-4016-bc5e-4c15a28e032e" (UID: "06cdc094-b372-4016-bc5e-4c15a28e032e"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.852209 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "06cdc094-b372-4016-bc5e-4c15a28e032e" (UID: "06cdc094-b372-4016-bc5e-4c15a28e032e"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.854532 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "06cdc094-b372-4016-bc5e-4c15a28e032e" (UID: "06cdc094-b372-4016-bc5e-4c15a28e032e"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.855277 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "06cdc094-b372-4016-bc5e-4c15a28e032e" (UID: "06cdc094-b372-4016-bc5e-4c15a28e032e"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.859493 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "06cdc094-b372-4016-bc5e-4c15a28e032e" (UID: "06cdc094-b372-4016-bc5e-4c15a28e032e"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.943400 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5f6v\" (UniqueName: \"kubernetes.io/projected/06cdc094-b372-4016-bc5e-4c15a28e032e-kube-api-access-l5f6v\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.943429 4772 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.943439 4772 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.943448 4772 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.943456 4772 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.943465 4772 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.943474 4772 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.943482 4772 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.943492 4772 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:16 crc kubenswrapper[4772]: I0127 15:11:16.943503 4772 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/06cdc094-b372-4016-bc5e-4c15a28e032e-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:17 crc kubenswrapper[4772]: I0127 15:11:17.615345 4772 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-fgw98 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.16:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 27 15:11:17 crc kubenswrapper[4772]: I0127 15:11:17.615425 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" podUID="06cdc094-b372-4016-bc5e-4c15a28e032e" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.16:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 27 15:11:17 crc kubenswrapper[4772]: I0127 15:11:17.664361 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-2glnd_d8591d45-25d0-47ea-a856-9cd5334e4a8c/marketplace-operator/2.log" Jan 27 15:11:17 crc kubenswrapper[4772]: I0127 15:11:17.664488 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fgw98" Jan 27 15:11:20 crc kubenswrapper[4772]: I0127 15:11:20.964151 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" Jan 27 15:11:20 crc kubenswrapper[4772]: I0127 15:11:20.964758 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" Jan 27 15:11:20 crc kubenswrapper[4772]: I0127 15:11:20.965640 4772 scope.go:117] "RemoveContainer" containerID="89be95f2c0621af885c1da302ff7e0b3e84dda04f9608b6e8af398107e1e9399" Jan 27 15:11:20 crc kubenswrapper[4772]: E0127 15:11:20.966121 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-2glnd_openshift-marketplace(d8591d45-25d0-47ea-a856-9cd5334e4a8c)\"" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" Jan 27 15:11:22 crc kubenswrapper[4772]: I0127 15:11:22.082590 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 27 15:11:22 crc kubenswrapper[4772]: I0127 15:11:22.485101 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 27 15:11:22 crc kubenswrapper[4772]: I0127 15:11:22.643013 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 27 15:11:22 crc kubenswrapper[4772]: I0127 15:11:22.783862 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 27 15:11:23 crc kubenswrapper[4772]: I0127 15:11:23.226885 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 27 15:11:23 crc kubenswrapper[4772]: I0127 15:11:23.501532 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 27 15:11:23 crc kubenswrapper[4772]: I0127 15:11:23.862073 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 27 15:11:23 crc kubenswrapper[4772]: I0127 15:11:23.916778 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 27 15:11:24 crc kubenswrapper[4772]: I0127 15:11:24.099214 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 27 15:11:24 crc kubenswrapper[4772]: I0127 15:11:24.216571 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 27 15:11:24 crc kubenswrapper[4772]: I0127 15:11:24.364821 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 27 15:11:24 crc kubenswrapper[4772]: I0127 15:11:24.463046 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 27 15:11:24 crc kubenswrapper[4772]: I0127 15:11:24.610407 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 27 15:11:24 crc kubenswrapper[4772]: I0127 15:11:24.735616 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 27 15:11:24 crc kubenswrapper[4772]: I0127 15:11:24.874786 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 27 15:11:24 crc kubenswrapper[4772]: I0127 15:11:24.888307 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 27 15:11:24 crc kubenswrapper[4772]: I0127 15:11:24.967797 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 27 15:11:25 crc kubenswrapper[4772]: I0127 15:11:25.052508 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 27 15:11:25 crc kubenswrapper[4772]: I0127 15:11:25.136806 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 27 15:11:25 crc kubenswrapper[4772]: I0127 15:11:25.159949 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 27 15:11:25 crc kubenswrapper[4772]: I0127 15:11:25.296741 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 27 15:11:25 crc kubenswrapper[4772]: I0127 15:11:25.351823 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 27 15:11:25 crc kubenswrapper[4772]: I0127 15:11:25.385071 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 27 15:11:25 crc kubenswrapper[4772]: I0127 15:11:25.388723 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 27 15:11:25 crc kubenswrapper[4772]: I0127 15:11:25.398662 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 27 15:11:25 crc kubenswrapper[4772]: I0127 15:11:25.506273 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 27 15:11:25 crc kubenswrapper[4772]: I0127 15:11:25.619042 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 27 15:11:25 crc kubenswrapper[4772]: I0127 15:11:25.906679 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 27 15:11:25 crc kubenswrapper[4772]: I0127 15:11:25.912628 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 27 15:11:26 crc kubenswrapper[4772]: I0127 15:11:26.001666 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 27 15:11:26 crc kubenswrapper[4772]: I0127 15:11:26.054000 4772 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 27 15:11:26 crc kubenswrapper[4772]: I0127 15:11:26.068432 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 27 15:11:26 crc kubenswrapper[4772]: I0127 15:11:26.239238 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 27 15:11:26 crc kubenswrapper[4772]: I0127 15:11:26.359145 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 27 15:11:26 crc kubenswrapper[4772]: I0127 15:11:26.389409 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 27 15:11:26 crc kubenswrapper[4772]: I0127 15:11:26.394888 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 27 15:11:26 crc kubenswrapper[4772]: I0127 15:11:26.579143 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 27 15:11:26 crc kubenswrapper[4772]: I0127 15:11:26.642957 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 27 15:11:26 crc kubenswrapper[4772]: I0127 15:11:26.783988 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 27 15:11:26 crc kubenswrapper[4772]: I0127 15:11:26.838934 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 27 15:11:26 crc kubenswrapper[4772]: I0127 15:11:26.849643 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 27 15:11:26 crc kubenswrapper[4772]: I0127 15:11:26.950666 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 27 15:11:27 crc kubenswrapper[4772]: I0127 15:11:27.017913 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 27 15:11:27 crc kubenswrapper[4772]: I0127 15:11:27.030189 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 27 15:11:27 crc kubenswrapper[4772]: I0127 15:11:27.049930 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 27 15:11:27 crc kubenswrapper[4772]: I0127 15:11:27.249809 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 27 15:11:27 crc kubenswrapper[4772]: I0127 15:11:27.262598 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 27 15:11:27 crc kubenswrapper[4772]: I0127 15:11:27.286411 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 27 15:11:27 crc kubenswrapper[4772]: I0127 15:11:27.308484 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 27 15:11:27 crc kubenswrapper[4772]: I0127 15:11:27.316667 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 27 15:11:27 crc kubenswrapper[4772]: I0127 15:11:27.377741 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 27 15:11:27 crc kubenswrapper[4772]: I0127 15:11:27.397868 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 27 15:11:27 crc kubenswrapper[4772]: I0127 15:11:27.414360 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 27 15:11:27 crc kubenswrapper[4772]: I0127 15:11:27.453731 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 27 15:11:27 crc kubenswrapper[4772]: I0127 15:11:27.622724 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 27 15:11:27 crc kubenswrapper[4772]: I0127 15:11:27.678397 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 27 15:11:27 crc kubenswrapper[4772]: I0127 15:11:27.838518 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 27 15:11:27 crc kubenswrapper[4772]: I0127 15:11:27.902301 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 27 15:11:27 crc kubenswrapper[4772]: I0127 15:11:27.924390 4772 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 27 15:11:27 crc kubenswrapper[4772]: I0127 15:11:27.968679 4772 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 27 15:11:27 crc kubenswrapper[4772]: I0127 15:11:27.984353 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 27 15:11:27 crc kubenswrapper[4772]: I0127 15:11:27.990392 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 27 15:11:28 crc kubenswrapper[4772]: I0127 15:11:28.106421 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 27 15:11:28 crc kubenswrapper[4772]: I0127 15:11:28.113094 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 27 15:11:28 crc kubenswrapper[4772]: I0127 15:11:28.184221 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 27 15:11:28 crc kubenswrapper[4772]: I0127 15:11:28.196542 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 27 15:11:28 crc kubenswrapper[4772]: I0127 15:11:28.217276 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 27 15:11:28 crc kubenswrapper[4772]: I0127 15:11:28.314707 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 27 15:11:28 crc kubenswrapper[4772]: I0127 15:11:28.450594 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 27 15:11:28 crc kubenswrapper[4772]: I0127 15:11:28.494190 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 27 15:11:28 crc kubenswrapper[4772]: I0127 15:11:28.587310 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 27 15:11:28 crc kubenswrapper[4772]: I0127 15:11:28.695781 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 27 15:11:28 crc kubenswrapper[4772]: I0127 15:11:28.760340 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 27 15:11:28 crc kubenswrapper[4772]: I0127 15:11:28.824698 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 27 15:11:28 crc kubenswrapper[4772]: I0127 15:11:28.825087 4772 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 27 15:11:28 crc kubenswrapper[4772]: I0127 15:11:28.837194 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 27 15:11:28 crc kubenswrapper[4772]: I0127 15:11:28.853047 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 27 15:11:28 crc kubenswrapper[4772]: I0127 15:11:28.883801 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 27 15:11:28 crc kubenswrapper[4772]: I0127 15:11:28.900090 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 27 15:11:29 crc kubenswrapper[4772]: I0127 15:11:29.010415 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 27 15:11:29 crc kubenswrapper[4772]: I0127 15:11:29.146540 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 27 15:11:29 crc kubenswrapper[4772]: I0127 15:11:29.194445 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 27 15:11:29 crc kubenswrapper[4772]: I0127 15:11:29.298674 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 27 15:11:29 crc kubenswrapper[4772]: I0127 15:11:29.373946 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 27 15:11:29 crc kubenswrapper[4772]: I0127 15:11:29.395969 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 27 15:11:29 crc kubenswrapper[4772]: I0127 15:11:29.488963 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 27 15:11:29 crc kubenswrapper[4772]: I0127 15:11:29.515908 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 27 15:11:29 crc kubenswrapper[4772]: I0127 15:11:29.559087 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 27 15:11:29 crc kubenswrapper[4772]: I0127 15:11:29.597994 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 27 15:11:29 crc kubenswrapper[4772]: I0127 15:11:29.637976 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 27 15:11:29 crc kubenswrapper[4772]: I0127 15:11:29.699925 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 27 15:11:29 crc kubenswrapper[4772]: I0127 15:11:29.747878 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 27 15:11:30 crc kubenswrapper[4772]: I0127 15:11:30.050114 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 27 15:11:30 crc kubenswrapper[4772]: I0127 15:11:30.126822 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 27 15:11:30 crc kubenswrapper[4772]: I0127 15:11:30.213462 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 27 15:11:30 crc kubenswrapper[4772]: I0127 15:11:30.220569 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 27 15:11:30 crc kubenswrapper[4772]: I0127 15:11:30.226026 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 27 15:11:30 crc kubenswrapper[4772]: I0127 15:11:30.295887 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 27 15:11:30 crc kubenswrapper[4772]: I0127 15:11:30.296945 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 27 15:11:30 crc kubenswrapper[4772]: I0127 15:11:30.427872 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 27 15:11:30 crc kubenswrapper[4772]: I0127 15:11:30.434692 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 27 15:11:30 crc kubenswrapper[4772]: I0127 15:11:30.438553 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 27 15:11:30 crc kubenswrapper[4772]: I0127 15:11:30.465539 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 27 15:11:30 crc kubenswrapper[4772]: I0127 15:11:30.496036 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 27 15:11:30 crc kubenswrapper[4772]: I0127 15:11:30.566288 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 27 15:11:30 crc kubenswrapper[4772]: I0127 15:11:30.579846 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 27 15:11:30 crc kubenswrapper[4772]: I0127 15:11:30.580093 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 27 15:11:30 crc kubenswrapper[4772]: I0127 15:11:30.694267 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 27 15:11:30 crc kubenswrapper[4772]: I0127 15:11:30.747060 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 27 15:11:30 crc kubenswrapper[4772]: I0127 15:11:30.761444 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 27 15:11:30 crc kubenswrapper[4772]: I0127 15:11:30.784862 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 27 15:11:30 crc kubenswrapper[4772]: I0127 15:11:30.816135 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 27 15:11:30 crc kubenswrapper[4772]: I0127 15:11:30.867348 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 27 15:11:30 crc kubenswrapper[4772]: I0127 15:11:30.970914 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 27 15:11:31 crc kubenswrapper[4772]: I0127 15:11:31.029660 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 27 15:11:31 crc kubenswrapper[4772]: I0127 15:11:31.083394 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 27 15:11:31 crc kubenswrapper[4772]: I0127 15:11:31.136827 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 27 15:11:31 crc kubenswrapper[4772]: I0127 15:11:31.164289 4772 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 27 15:11:31 crc kubenswrapper[4772]: I0127 15:11:31.170248 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-75jrg","openshift-marketplace/community-operators-jwrpk","openshift-marketplace/redhat-marketplace-dfgjh","openshift-marketplace/marketplace-operator-79b997595-4lj2h","openshift-marketplace/redhat-operators-k7pfr","openshift-marketplace/redhat-marketplace-xp8ph","openshift-authentication/oauth-openshift-558db77b4-fgw98","openshift-kube-apiserver/kube-apiserver-crc","openshift-marketplace/certified-operators-9wdps","openshift-marketplace/community-operators-95rh9"] Jan 27 15:11:31 crc kubenswrapper[4772]: I0127 15:11:31.170354 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 27 15:11:31 crc kubenswrapper[4772]: I0127 15:11:31.174369 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 27 15:11:31 crc kubenswrapper[4772]: I0127 15:11:31.174396 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 27 15:11:31 crc kubenswrapper[4772]: I0127 15:11:31.192476 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=19.192458385 podStartE2EDuration="19.192458385s" podCreationTimestamp="2026-01-27 15:11:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:11:31.191643951 +0000 UTC m=+277.172253059" watchObservedRunningTime="2026-01-27 15:11:31.192458385 +0000 UTC m=+277.173067483" Jan 27 15:11:31 crc kubenswrapper[4772]: I0127 15:11:31.292548 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 27 15:11:31 crc kubenswrapper[4772]: I0127 15:11:31.325600 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 27 15:11:31 crc kubenswrapper[4772]: I0127 15:11:31.370133 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 27 15:11:31 crc kubenswrapper[4772]: I0127 15:11:31.472335 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 27 15:11:31 crc kubenswrapper[4772]: I0127 15:11:31.603223 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 27 15:11:31 crc kubenswrapper[4772]: I0127 15:11:31.675950 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 27 15:11:31 crc kubenswrapper[4772]: I0127 15:11:31.804479 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 27 15:11:31 crc kubenswrapper[4772]: I0127 15:11:31.896321 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 27 15:11:31 crc kubenswrapper[4772]: I0127 15:11:31.951358 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 27 15:11:31 crc kubenswrapper[4772]: I0127 15:11:31.982569 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.165492 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.173665 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.191954 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.259742 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.407566 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.426073 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.440510 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.477291 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.524091 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.602783 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.604476 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.655187 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.668407 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06cdc094-b372-4016-bc5e-4c15a28e032e" path="/var/lib/kubelet/pods/06cdc094-b372-4016-bc5e-4c15a28e032e/volumes" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.669081 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7" path="/var/lib/kubelet/pods/4232ddc2-0fc9-45c7-b52a-ee96e2e3cef7/volumes" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.669730 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cbabfa8-79d8-4b23-b186-b40ba8b3017e" path="/var/lib/kubelet/pods/8cbabfa8-79d8-4b23-b186-b40ba8b3017e/volumes" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.670782 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96e88efd-1f25-4e44-b459-ab773db93656" path="/var/lib/kubelet/pods/96e88efd-1f25-4e44-b459-ab773db93656/volumes" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.671384 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac2b5800-ce98-4847-bfcd-67a97375aa1b" path="/var/lib/kubelet/pods/ac2b5800-ce98-4847-bfcd-67a97375aa1b/volumes" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.672466 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8ebf890-c3b0-468e-bf7d-0ec590df084b" path="/var/lib/kubelet/pods/c8ebf890-c3b0-468e-bf7d-0ec590df084b/volumes" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.672910 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0b33686-8107-4caf-b67f-3c608119a049" path="/var/lib/kubelet/pods/d0b33686-8107-4caf-b67f-3c608119a049/volumes" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.673563 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd415ccf-2b4a-4797-962f-a464ef96bc22" path="/var/lib/kubelet/pods/dd415ccf-2b4a-4797-962f-a464ef96bc22/volumes" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.674154 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.674633 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f637b998-b13b-486d-9042-4cd40a01c833" path="/var/lib/kubelet/pods/f637b998-b13b-486d-9042-4cd40a01c833/volumes" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.752578 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.767977 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.792162 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.875010 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.933234 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 27 15:11:32 crc kubenswrapper[4772]: I0127 15:11:32.970333 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 27 15:11:33 crc kubenswrapper[4772]: I0127 15:11:33.067784 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 27 15:11:33 crc kubenswrapper[4772]: I0127 15:11:33.191915 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 27 15:11:33 crc kubenswrapper[4772]: I0127 15:11:33.395245 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 27 15:11:33 crc kubenswrapper[4772]: I0127 15:11:33.395475 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 27 15:11:33 crc kubenswrapper[4772]: I0127 15:11:33.532335 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 27 15:11:33 crc kubenswrapper[4772]: I0127 15:11:33.616080 4772 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 27 15:11:33 crc kubenswrapper[4772]: I0127 15:11:33.627220 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 27 15:11:33 crc kubenswrapper[4772]: I0127 15:11:33.700959 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 27 15:11:33 crc kubenswrapper[4772]: I0127 15:11:33.747469 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 27 15:11:33 crc kubenswrapper[4772]: I0127 15:11:33.855534 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 27 15:11:33 crc kubenswrapper[4772]: I0127 15:11:33.857404 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 27 15:11:33 crc kubenswrapper[4772]: I0127 15:11:33.953785 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 27 15:11:34 crc kubenswrapper[4772]: I0127 15:11:34.024999 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 27 15:11:34 crc kubenswrapper[4772]: I0127 15:11:34.064452 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 27 15:11:34 crc kubenswrapper[4772]: I0127 15:11:34.175366 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 27 15:11:34 crc kubenswrapper[4772]: I0127 15:11:34.176002 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 27 15:11:34 crc kubenswrapper[4772]: I0127 15:11:34.198225 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 27 15:11:34 crc kubenswrapper[4772]: I0127 15:11:34.245625 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 27 15:11:34 crc kubenswrapper[4772]: I0127 15:11:34.246927 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 27 15:11:34 crc kubenswrapper[4772]: I0127 15:11:34.321025 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 27 15:11:34 crc kubenswrapper[4772]: I0127 15:11:34.438120 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 27 15:11:34 crc kubenswrapper[4772]: I0127 15:11:34.464725 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 27 15:11:34 crc kubenswrapper[4772]: I0127 15:11:34.501228 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 27 15:11:34 crc kubenswrapper[4772]: I0127 15:11:34.507607 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 27 15:11:34 crc kubenswrapper[4772]: I0127 15:11:34.572610 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 27 15:11:34 crc kubenswrapper[4772]: I0127 15:11:34.577603 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 27 15:11:34 crc kubenswrapper[4772]: I0127 15:11:34.585505 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 27 15:11:34 crc kubenswrapper[4772]: I0127 15:11:34.602096 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 27 15:11:34 crc kubenswrapper[4772]: I0127 15:11:34.666271 4772 scope.go:117] "RemoveContainer" containerID="89be95f2c0621af885c1da302ff7e0b3e84dda04f9608b6e8af398107e1e9399" Jan 27 15:11:34 crc kubenswrapper[4772]: E0127 15:11:34.666445 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-2glnd_openshift-marketplace(d8591d45-25d0-47ea-a856-9cd5334e4a8c)\"" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" Jan 27 15:11:34 crc kubenswrapper[4772]: I0127 15:11:34.696660 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 27 15:11:34 crc kubenswrapper[4772]: I0127 15:11:34.699647 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 27 15:11:34 crc kubenswrapper[4772]: I0127 15:11:34.725995 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 27 15:11:34 crc kubenswrapper[4772]: I0127 15:11:34.814743 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 27 15:11:34 crc kubenswrapper[4772]: I0127 15:11:34.842465 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 27 15:11:34 crc kubenswrapper[4772]: I0127 15:11:34.855760 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 27 15:11:35 crc kubenswrapper[4772]: I0127 15:11:35.017629 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 27 15:11:35 crc kubenswrapper[4772]: I0127 15:11:35.096913 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 27 15:11:35 crc kubenswrapper[4772]: I0127 15:11:35.105450 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 27 15:11:35 crc kubenswrapper[4772]: I0127 15:11:35.148686 4772 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 27 15:11:35 crc kubenswrapper[4772]: I0127 15:11:35.148910 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://6f6210876fa329e0bc46c9fcfeb492e4200a121e5f183839eb54f27fb32b52f4" gracePeriod=5 Jan 27 15:11:35 crc kubenswrapper[4772]: I0127 15:11:35.239092 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 27 15:11:35 crc kubenswrapper[4772]: I0127 15:11:35.278976 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 27 15:11:35 crc kubenswrapper[4772]: I0127 15:11:35.544712 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 27 15:11:35 crc kubenswrapper[4772]: I0127 15:11:35.656215 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 27 15:11:35 crc kubenswrapper[4772]: I0127 15:11:35.830433 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 27 15:11:35 crc kubenswrapper[4772]: I0127 15:11:35.909845 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 27 15:11:35 crc kubenswrapper[4772]: I0127 15:11:35.938775 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.021724 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.037533 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.143354 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.172355 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.197478 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.245631 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.301222 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.353943 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.399598 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.451068 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.690651 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.698306 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7687c8778f-hqrqm"] Jan 27 15:11:36 crc kubenswrapper[4772]: E0127 15:11:36.698551 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.698575 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 27 15:11:36 crc kubenswrapper[4772]: E0127 15:11:36.698596 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06cdc094-b372-4016-bc5e-4c15a28e032e" containerName="oauth-openshift" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.698604 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="06cdc094-b372-4016-bc5e-4c15a28e032e" containerName="oauth-openshift" Jan 27 15:11:36 crc kubenswrapper[4772]: E0127 15:11:36.698620 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bd0c383-7376-4e95-9919-863297cbd807" containerName="installer" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.698627 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bd0c383-7376-4e95-9919-863297cbd807" containerName="installer" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.698751 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bd0c383-7376-4e95-9919-863297cbd807" containerName="installer" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.698770 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.698782 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="06cdc094-b372-4016-bc5e-4c15a28e032e" containerName="oauth-openshift" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.699274 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.705553 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.706138 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.706343 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.706475 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.706589 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.707027 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.707854 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.708078 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.708255 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.708436 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.710586 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.705453 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.712884 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.713720 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7687c8778f-hqrqm"] Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.728279 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.732842 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.782231 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.782298 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4bcd7451-5572-4b94-a244-746f5c7145a2-audit-policies\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.782322 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-system-session\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.782353 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-user-template-login\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.782380 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.782396 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b9xq\" (UniqueName: \"kubernetes.io/projected/4bcd7451-5572-4b94-a244-746f5c7145a2-kube-api-access-4b9xq\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.782419 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-system-service-ca\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.782442 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-user-template-error\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.782463 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.782483 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.782498 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.782516 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-system-router-certs\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.782546 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.782565 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4bcd7451-5572-4b94-a244-746f5c7145a2-audit-dir\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.884180 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4bcd7451-5572-4b94-a244-746f5c7145a2-audit-policies\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.884242 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-system-session\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.884271 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-user-template-login\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.884306 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.884377 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b9xq\" (UniqueName: \"kubernetes.io/projected/4bcd7451-5572-4b94-a244-746f5c7145a2-kube-api-access-4b9xq\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.884417 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-system-service-ca\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.884460 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-user-template-error\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.884487 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.884516 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.884542 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.884571 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-system-router-certs\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.884602 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.884626 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4bcd7451-5572-4b94-a244-746f5c7145a2-audit-dir\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.884659 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.886880 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.886938 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4bcd7451-5572-4b94-a244-746f5c7145a2-audit-dir\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.886935 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4bcd7451-5572-4b94-a244-746f5c7145a2-audit-policies\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.887432 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-system-service-ca\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.888229 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.890988 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.891543 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.891908 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-user-template-login\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.893544 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-system-router-certs\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.893876 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.903496 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-system-session\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.905370 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.905751 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4bcd7451-5572-4b94-a244-746f5c7145a2-v4-0-config-user-template-error\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.906086 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.907987 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b9xq\" (UniqueName: \"kubernetes.io/projected/4bcd7451-5572-4b94-a244-746f5c7145a2-kube-api-access-4b9xq\") pod \"oauth-openshift-7687c8778f-hqrqm\" (UID: \"4bcd7451-5572-4b94-a244-746f5c7145a2\") " pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:36 crc kubenswrapper[4772]: I0127 15:11:36.934663 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 27 15:11:37 crc kubenswrapper[4772]: I0127 15:11:37.032058 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:37 crc kubenswrapper[4772]: I0127 15:11:37.115534 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 27 15:11:37 crc kubenswrapper[4772]: I0127 15:11:37.137748 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 27 15:11:37 crc kubenswrapper[4772]: I0127 15:11:37.351375 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 27 15:11:37 crc kubenswrapper[4772]: I0127 15:11:37.361659 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 27 15:11:37 crc kubenswrapper[4772]: I0127 15:11:37.429470 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 27 15:11:37 crc kubenswrapper[4772]: I0127 15:11:37.452053 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7687c8778f-hqrqm"] Jan 27 15:11:37 crc kubenswrapper[4772]: I0127 15:11:37.469864 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 27 15:11:37 crc kubenswrapper[4772]: I0127 15:11:37.565699 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 27 15:11:37 crc kubenswrapper[4772]: I0127 15:11:37.619666 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 27 15:11:37 crc kubenswrapper[4772]: I0127 15:11:37.656654 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 27 15:11:37 crc kubenswrapper[4772]: I0127 15:11:37.661004 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 27 15:11:37 crc kubenswrapper[4772]: I0127 15:11:37.766659 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" event={"ID":"4bcd7451-5572-4b94-a244-746f5c7145a2","Type":"ContainerStarted","Data":"7c7eff52d5c75bc856315b147ca80c5389960202b855ba1e1bbca7f221717152"} Jan 27 15:11:37 crc kubenswrapper[4772]: I0127 15:11:37.766716 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" event={"ID":"4bcd7451-5572-4b94-a244-746f5c7145a2","Type":"ContainerStarted","Data":"da64ed2d13939519724584d6676d260ea792a791943df0b2581f3127f25a5a56"} Jan 27 15:11:37 crc kubenswrapper[4772]: I0127 15:11:37.816809 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 27 15:11:38 crc kubenswrapper[4772]: I0127 15:11:38.015773 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 27 15:11:38 crc kubenswrapper[4772]: I0127 15:11:38.024299 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 27 15:11:38 crc kubenswrapper[4772]: I0127 15:11:38.032560 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 27 15:11:38 crc kubenswrapper[4772]: I0127 15:11:38.089886 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 27 15:11:38 crc kubenswrapper[4772]: I0127 15:11:38.237057 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 27 15:11:38 crc kubenswrapper[4772]: I0127 15:11:38.245548 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 27 15:11:38 crc kubenswrapper[4772]: I0127 15:11:38.368687 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 27 15:11:38 crc kubenswrapper[4772]: I0127 15:11:38.772498 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:38 crc kubenswrapper[4772]: I0127 15:11:38.777856 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" Jan 27 15:11:38 crc kubenswrapper[4772]: I0127 15:11:38.803294 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7687c8778f-hqrqm" podStartSLOduration=47.803262461 podStartE2EDuration="47.803262461s" podCreationTimestamp="2026-01-27 15:10:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:11:37.794011996 +0000 UTC m=+283.774621114" watchObservedRunningTime="2026-01-27 15:11:38.803262461 +0000 UTC m=+284.783871599" Jan 27 15:11:39 crc kubenswrapper[4772]: I0127 15:11:39.166636 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 27 15:11:39 crc kubenswrapper[4772]: I0127 15:11:39.241750 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 27 15:11:39 crc kubenswrapper[4772]: I0127 15:11:39.363085 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 27 15:11:39 crc kubenswrapper[4772]: I0127 15:11:39.378753 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 27 15:11:39 crc kubenswrapper[4772]: I0127 15:11:39.443761 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 27 15:11:39 crc kubenswrapper[4772]: I0127 15:11:39.476843 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 27 15:11:39 crc kubenswrapper[4772]: I0127 15:11:39.898459 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 27 15:11:39 crc kubenswrapper[4772]: I0127 15:11:39.935300 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.179849 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.223691 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.723504 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.723632 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.785261 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.785318 4772 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="6f6210876fa329e0bc46c9fcfeb492e4200a121e5f183839eb54f27fb32b52f4" exitCode=137 Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.785386 4772 scope.go:117] "RemoveContainer" containerID="6f6210876fa329e0bc46c9fcfeb492e4200a121e5f183839eb54f27fb32b52f4" Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.785386 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.800880 4772 scope.go:117] "RemoveContainer" containerID="6f6210876fa329e0bc46c9fcfeb492e4200a121e5f183839eb54f27fb32b52f4" Jan 27 15:11:40 crc kubenswrapper[4772]: E0127 15:11:40.801316 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f6210876fa329e0bc46c9fcfeb492e4200a121e5f183839eb54f27fb32b52f4\": container with ID starting with 6f6210876fa329e0bc46c9fcfeb492e4200a121e5f183839eb54f27fb32b52f4 not found: ID does not exist" containerID="6f6210876fa329e0bc46c9fcfeb492e4200a121e5f183839eb54f27fb32b52f4" Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.801359 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f6210876fa329e0bc46c9fcfeb492e4200a121e5f183839eb54f27fb32b52f4"} err="failed to get container status \"6f6210876fa329e0bc46c9fcfeb492e4200a121e5f183839eb54f27fb32b52f4\": rpc error: code = NotFound desc = could not find container \"6f6210876fa329e0bc46c9fcfeb492e4200a121e5f183839eb54f27fb32b52f4\": container with ID starting with 6f6210876fa329e0bc46c9fcfeb492e4200a121e5f183839eb54f27fb32b52f4 not found: ID does not exist" Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.832780 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.832863 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.832873 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.832897 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.832941 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.832985 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.833027 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.833104 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.833200 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.833317 4772 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.833334 4772 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.833345 4772 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.833356 4772 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.841116 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:11:40 crc kubenswrapper[4772]: I0127 15:11:40.934891 4772 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:41 crc kubenswrapper[4772]: I0127 15:11:41.303499 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 27 15:11:42 crc kubenswrapper[4772]: I0127 15:11:42.669384 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Jan 27 15:11:47 crc kubenswrapper[4772]: I0127 15:11:47.663237 4772 scope.go:117] "RemoveContainer" containerID="89be95f2c0621af885c1da302ff7e0b3e84dda04f9608b6e8af398107e1e9399" Jan 27 15:11:47 crc kubenswrapper[4772]: I0127 15:11:47.824614 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-2glnd_d8591d45-25d0-47ea-a856-9cd5334e4a8c/marketplace-operator/2.log" Jan 27 15:11:47 crc kubenswrapper[4772]: I0127 15:11:47.824674 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" event={"ID":"d8591d45-25d0-47ea-a856-9cd5334e4a8c","Type":"ContainerStarted","Data":"fed559ed1ffbdad5f1dd35dfce6ce0006386e7520831ccdcc78b5bc1005d87b8"} Jan 27 15:11:47 crc kubenswrapper[4772]: I0127 15:11:47.825202 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" Jan 27 15:11:47 crc kubenswrapper[4772]: I0127 15:11:47.826829 4772 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2glnd container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.56:8080/healthz\": dial tcp 10.217.0.56:8080: connect: connection refused" start-of-body= Jan 27 15:11:47 crc kubenswrapper[4772]: I0127 15:11:47.826894 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" podUID="d8591d45-25d0-47ea-a856-9cd5334e4a8c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.56:8080/healthz\": dial tcp 10.217.0.56:8080: connect: connection refused" Jan 27 15:11:47 crc kubenswrapper[4772]: I0127 15:11:47.847729 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" podStartSLOduration=57.847712173 podStartE2EDuration="57.847712173s" podCreationTimestamp="2026-01-27 15:10:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:11:47.842144599 +0000 UTC m=+293.822753697" watchObservedRunningTime="2026-01-27 15:11:47.847712173 +0000 UTC m=+293.828321271" Jan 27 15:11:48 crc kubenswrapper[4772]: I0127 15:11:48.831838 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-2glnd" Jan 27 15:11:54 crc kubenswrapper[4772]: I0127 15:11:54.465484 4772 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.069513 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6pclx"] Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.070344 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" podUID="3dfd9a91-e760-4c80-96e6-ca6525aa86b8" containerName="controller-manager" containerID="cri-o://bb01029e32299fb52d56d061afc654aa573880622860ea04caf54ad26b9a84eb" gracePeriod=30 Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.131577 4772 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-6pclx container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.131626 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" podUID="3dfd9a91-e760-4c80-96e6-ca6525aa86b8" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.193083 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz"] Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.193437 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" podUID="8d519648-7eaa-49bb-9a09-bd91d09d98c0" containerName="route-controller-manager" containerID="cri-o://3506ad5b58f850018e8ca4a14f82aae1d0b2f9ec52328d0669f3f49efb696d0f" gracePeriod=30 Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.414443 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.514108 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-serving-cert\") pod \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\" (UID: \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\") " Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.514159 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-client-ca\") pod \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\" (UID: \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\") " Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.514234 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-config\") pod \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\" (UID: \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\") " Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.514389 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-proxy-ca-bundles\") pod \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\" (UID: \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\") " Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.514429 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8c7gp\" (UniqueName: \"kubernetes.io/projected/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-kube-api-access-8c7gp\") pod \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\" (UID: \"3dfd9a91-e760-4c80-96e6-ca6525aa86b8\") " Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.514996 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-client-ca" (OuterVolumeSpecName: "client-ca") pod "3dfd9a91-e760-4c80-96e6-ca6525aa86b8" (UID: "3dfd9a91-e760-4c80-96e6-ca6525aa86b8"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.515621 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-config" (OuterVolumeSpecName: "config") pod "3dfd9a91-e760-4c80-96e6-ca6525aa86b8" (UID: "3dfd9a91-e760-4c80-96e6-ca6525aa86b8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.516239 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "3dfd9a91-e760-4c80-96e6-ca6525aa86b8" (UID: "3dfd9a91-e760-4c80-96e6-ca6525aa86b8"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.516645 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.519561 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3dfd9a91-e760-4c80-96e6-ca6525aa86b8" (UID: "3dfd9a91-e760-4c80-96e6-ca6525aa86b8"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.519931 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-kube-api-access-8c7gp" (OuterVolumeSpecName: "kube-api-access-8c7gp") pod "3dfd9a91-e760-4c80-96e6-ca6525aa86b8" (UID: "3dfd9a91-e760-4c80-96e6-ca6525aa86b8"). InnerVolumeSpecName "kube-api-access-8c7gp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.615502 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m24gk\" (UniqueName: \"kubernetes.io/projected/8d519648-7eaa-49bb-9a09-bd91d09d98c0-kube-api-access-m24gk\") pod \"8d519648-7eaa-49bb-9a09-bd91d09d98c0\" (UID: \"8d519648-7eaa-49bb-9a09-bd91d09d98c0\") " Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.615553 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d519648-7eaa-49bb-9a09-bd91d09d98c0-config\") pod \"8d519648-7eaa-49bb-9a09-bd91d09d98c0\" (UID: \"8d519648-7eaa-49bb-9a09-bd91d09d98c0\") " Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.615621 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d519648-7eaa-49bb-9a09-bd91d09d98c0-serving-cert\") pod \"8d519648-7eaa-49bb-9a09-bd91d09d98c0\" (UID: \"8d519648-7eaa-49bb-9a09-bd91d09d98c0\") " Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.615654 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8d519648-7eaa-49bb-9a09-bd91d09d98c0-client-ca\") pod \"8d519648-7eaa-49bb-9a09-bd91d09d98c0\" (UID: \"8d519648-7eaa-49bb-9a09-bd91d09d98c0\") " Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.615827 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.615837 4772 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.615847 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8c7gp\" (UniqueName: \"kubernetes.io/projected/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-kube-api-access-8c7gp\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.615855 4772 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.615862 4772 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3dfd9a91-e760-4c80-96e6-ca6525aa86b8-client-ca\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.616425 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d519648-7eaa-49bb-9a09-bd91d09d98c0-client-ca" (OuterVolumeSpecName: "client-ca") pod "8d519648-7eaa-49bb-9a09-bd91d09d98c0" (UID: "8d519648-7eaa-49bb-9a09-bd91d09d98c0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.616438 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d519648-7eaa-49bb-9a09-bd91d09d98c0-config" (OuterVolumeSpecName: "config") pod "8d519648-7eaa-49bb-9a09-bd91d09d98c0" (UID: "8d519648-7eaa-49bb-9a09-bd91d09d98c0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.618093 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d519648-7eaa-49bb-9a09-bd91d09d98c0-kube-api-access-m24gk" (OuterVolumeSpecName: "kube-api-access-m24gk") pod "8d519648-7eaa-49bb-9a09-bd91d09d98c0" (UID: "8d519648-7eaa-49bb-9a09-bd91d09d98c0"). InnerVolumeSpecName "kube-api-access-m24gk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.618332 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d519648-7eaa-49bb-9a09-bd91d09d98c0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8d519648-7eaa-49bb-9a09-bd91d09d98c0" (UID: "8d519648-7eaa-49bb-9a09-bd91d09d98c0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.717037 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m24gk\" (UniqueName: \"kubernetes.io/projected/8d519648-7eaa-49bb-9a09-bd91d09d98c0-kube-api-access-m24gk\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.717091 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d519648-7eaa-49bb-9a09-bd91d09d98c0-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.717112 4772 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d519648-7eaa-49bb-9a09-bd91d09d98c0-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.717131 4772 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8d519648-7eaa-49bb-9a09-bd91d09d98c0-client-ca\") on node \"crc\" DevicePath \"\"" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.867584 4772 generic.go:334] "Generic (PLEG): container finished" podID="3dfd9a91-e760-4c80-96e6-ca6525aa86b8" containerID="bb01029e32299fb52d56d061afc654aa573880622860ea04caf54ad26b9a84eb" exitCode=0 Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.867630 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" event={"ID":"3dfd9a91-e760-4c80-96e6-ca6525aa86b8","Type":"ContainerDied","Data":"bb01029e32299fb52d56d061afc654aa573880622860ea04caf54ad26b9a84eb"} Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.867675 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" event={"ID":"3dfd9a91-e760-4c80-96e6-ca6525aa86b8","Type":"ContainerDied","Data":"380eb12e1295d0270de3d27b76c2692262e194cd8678145268c2765050e2b23e"} Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.867695 4772 scope.go:117] "RemoveContainer" containerID="bb01029e32299fb52d56d061afc654aa573880622860ea04caf54ad26b9a84eb" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.867640 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-6pclx" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.869565 4772 generic.go:334] "Generic (PLEG): container finished" podID="8d519648-7eaa-49bb-9a09-bd91d09d98c0" containerID="3506ad5b58f850018e8ca4a14f82aae1d0b2f9ec52328d0669f3f49efb696d0f" exitCode=0 Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.869613 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" event={"ID":"8d519648-7eaa-49bb-9a09-bd91d09d98c0","Type":"ContainerDied","Data":"3506ad5b58f850018e8ca4a14f82aae1d0b2f9ec52328d0669f3f49efb696d0f"} Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.869651 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" event={"ID":"8d519648-7eaa-49bb-9a09-bd91d09d98c0","Type":"ContainerDied","Data":"326c33fb529962a602f8dfd5dbe7dcbd0ebb132fe4709244f27812007b261a68"} Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.869727 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.884862 4772 scope.go:117] "RemoveContainer" containerID="bb01029e32299fb52d56d061afc654aa573880622860ea04caf54ad26b9a84eb" Jan 27 15:11:55 crc kubenswrapper[4772]: E0127 15:11:55.885358 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb01029e32299fb52d56d061afc654aa573880622860ea04caf54ad26b9a84eb\": container with ID starting with bb01029e32299fb52d56d061afc654aa573880622860ea04caf54ad26b9a84eb not found: ID does not exist" containerID="bb01029e32299fb52d56d061afc654aa573880622860ea04caf54ad26b9a84eb" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.885413 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb01029e32299fb52d56d061afc654aa573880622860ea04caf54ad26b9a84eb"} err="failed to get container status \"bb01029e32299fb52d56d061afc654aa573880622860ea04caf54ad26b9a84eb\": rpc error: code = NotFound desc = could not find container \"bb01029e32299fb52d56d061afc654aa573880622860ea04caf54ad26b9a84eb\": container with ID starting with bb01029e32299fb52d56d061afc654aa573880622860ea04caf54ad26b9a84eb not found: ID does not exist" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.885448 4772 scope.go:117] "RemoveContainer" containerID="3506ad5b58f850018e8ca4a14f82aae1d0b2f9ec52328d0669f3f49efb696d0f" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.903643 4772 scope.go:117] "RemoveContainer" containerID="3506ad5b58f850018e8ca4a14f82aae1d0b2f9ec52328d0669f3f49efb696d0f" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.903800 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6pclx"] Jan 27 15:11:55 crc kubenswrapper[4772]: E0127 15:11:55.904557 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3506ad5b58f850018e8ca4a14f82aae1d0b2f9ec52328d0669f3f49efb696d0f\": container with ID starting with 3506ad5b58f850018e8ca4a14f82aae1d0b2f9ec52328d0669f3f49efb696d0f not found: ID does not exist" containerID="3506ad5b58f850018e8ca4a14f82aae1d0b2f9ec52328d0669f3f49efb696d0f" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.904611 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3506ad5b58f850018e8ca4a14f82aae1d0b2f9ec52328d0669f3f49efb696d0f"} err="failed to get container status \"3506ad5b58f850018e8ca4a14f82aae1d0b2f9ec52328d0669f3f49efb696d0f\": rpc error: code = NotFound desc = could not find container \"3506ad5b58f850018e8ca4a14f82aae1d0b2f9ec52328d0669f3f49efb696d0f\": container with ID starting with 3506ad5b58f850018e8ca4a14f82aae1d0b2f9ec52328d0669f3f49efb696d0f not found: ID does not exist" Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.906873 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6pclx"] Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.915263 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz"] Jan 27 15:11:55 crc kubenswrapper[4772]: I0127 15:11:55.918109 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-r9glz"] Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.670239 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dfd9a91-e760-4c80-96e6-ca6525aa86b8" path="/var/lib/kubelet/pods/3dfd9a91-e760-4c80-96e6-ca6525aa86b8/volumes" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.670781 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d519648-7eaa-49bb-9a09-bd91d09d98c0" path="/var/lib/kubelet/pods/8d519648-7eaa-49bb-9a09-bd91d09d98c0/volumes" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.707661 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq"] Jan 27 15:11:56 crc kubenswrapper[4772]: E0127 15:11:56.708100 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dfd9a91-e760-4c80-96e6-ca6525aa86b8" containerName="controller-manager" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.708125 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dfd9a91-e760-4c80-96e6-ca6525aa86b8" containerName="controller-manager" Jan 27 15:11:56 crc kubenswrapper[4772]: E0127 15:11:56.708139 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d519648-7eaa-49bb-9a09-bd91d09d98c0" containerName="route-controller-manager" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.708181 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d519648-7eaa-49bb-9a09-bd91d09d98c0" containerName="route-controller-manager" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.708348 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dfd9a91-e760-4c80-96e6-ca6525aa86b8" containerName="controller-manager" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.708372 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d519648-7eaa-49bb-9a09-bd91d09d98c0" containerName="route-controller-manager" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.709001 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.711681 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55985dff9-whh87"] Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.712106 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.712823 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-55985dff9-whh87" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.718498 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.718731 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.718785 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.718731 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.719217 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.719245 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.719368 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.719508 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.719733 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55985dff9-whh87"] Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.719772 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.719788 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.722977 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.735472 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.743229 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq"] Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.835096 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4vbvz"] Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.836272 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4vbvz" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.837480 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5748bb4e-846d-457a-af17-b1d6f0a36431-serving-cert\") pod \"controller-manager-9d7ff4cd6-9jdqq\" (UID: \"5748bb4e-846d-457a-af17-b1d6f0a36431\") " pod="openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.837526 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11-config\") pod \"route-controller-manager-55985dff9-whh87\" (UID: \"ce1b4cd5-019c-41b7-a994-e98eb4fd3b11\") " pod="openshift-route-controller-manager/route-controller-manager-55985dff9-whh87" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.837555 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11-client-ca\") pod \"route-controller-manager-55985dff9-whh87\" (UID: \"ce1b4cd5-019c-41b7-a994-e98eb4fd3b11\") " pod="openshift-route-controller-manager/route-controller-manager-55985dff9-whh87" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.837593 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5748bb4e-846d-457a-af17-b1d6f0a36431-client-ca\") pod \"controller-manager-9d7ff4cd6-9jdqq\" (UID: \"5748bb4e-846d-457a-af17-b1d6f0a36431\") " pod="openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.837615 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11-serving-cert\") pod \"route-controller-manager-55985dff9-whh87\" (UID: \"ce1b4cd5-019c-41b7-a994-e98eb4fd3b11\") " pod="openshift-route-controller-manager/route-controller-manager-55985dff9-whh87" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.837637 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5748bb4e-846d-457a-af17-b1d6f0a36431-proxy-ca-bundles\") pod \"controller-manager-9d7ff4cd6-9jdqq\" (UID: \"5748bb4e-846d-457a-af17-b1d6f0a36431\") " pod="openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.837668 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-799x8\" (UniqueName: \"kubernetes.io/projected/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11-kube-api-access-799x8\") pod \"route-controller-manager-55985dff9-whh87\" (UID: \"ce1b4cd5-019c-41b7-a994-e98eb4fd3b11\") " pod="openshift-route-controller-manager/route-controller-manager-55985dff9-whh87" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.837692 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5748bb4e-846d-457a-af17-b1d6f0a36431-config\") pod \"controller-manager-9d7ff4cd6-9jdqq\" (UID: \"5748bb4e-846d-457a-af17-b1d6f0a36431\") " pod="openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.837711 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkgwr\" (UniqueName: \"kubernetes.io/projected/5748bb4e-846d-457a-af17-b1d6f0a36431-kube-api-access-fkgwr\") pod \"controller-manager-9d7ff4cd6-9jdqq\" (UID: \"5748bb4e-846d-457a-af17-b1d6f0a36431\") " pod="openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.839342 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.853262 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4vbvz"] Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.938927 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11-config\") pod \"route-controller-manager-55985dff9-whh87\" (UID: \"ce1b4cd5-019c-41b7-a994-e98eb4fd3b11\") " pod="openshift-route-controller-manager/route-controller-manager-55985dff9-whh87" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.939406 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11-client-ca\") pod \"route-controller-manager-55985dff9-whh87\" (UID: \"ce1b4cd5-019c-41b7-a994-e98eb4fd3b11\") " pod="openshift-route-controller-manager/route-controller-manager-55985dff9-whh87" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.939556 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20e4371a-8bd2-4405-bb18-861923bfd37e-catalog-content\") pod \"community-operators-4vbvz\" (UID: \"20e4371a-8bd2-4405-bb18-861923bfd37e\") " pod="openshift-marketplace/community-operators-4vbvz" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.939643 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5748bb4e-846d-457a-af17-b1d6f0a36431-client-ca\") pod \"controller-manager-9d7ff4cd6-9jdqq\" (UID: \"5748bb4e-846d-457a-af17-b1d6f0a36431\") " pod="openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.939732 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11-serving-cert\") pod \"route-controller-manager-55985dff9-whh87\" (UID: \"ce1b4cd5-019c-41b7-a994-e98eb4fd3b11\") " pod="openshift-route-controller-manager/route-controller-manager-55985dff9-whh87" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.939854 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csq5c\" (UniqueName: \"kubernetes.io/projected/20e4371a-8bd2-4405-bb18-861923bfd37e-kube-api-access-csq5c\") pod \"community-operators-4vbvz\" (UID: \"20e4371a-8bd2-4405-bb18-861923bfd37e\") " pod="openshift-marketplace/community-operators-4vbvz" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.939964 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5748bb4e-846d-457a-af17-b1d6f0a36431-proxy-ca-bundles\") pod \"controller-manager-9d7ff4cd6-9jdqq\" (UID: \"5748bb4e-846d-457a-af17-b1d6f0a36431\") " pod="openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.940139 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-799x8\" (UniqueName: \"kubernetes.io/projected/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11-kube-api-access-799x8\") pod \"route-controller-manager-55985dff9-whh87\" (UID: \"ce1b4cd5-019c-41b7-a994-e98eb4fd3b11\") " pod="openshift-route-controller-manager/route-controller-manager-55985dff9-whh87" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.940303 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5748bb4e-846d-457a-af17-b1d6f0a36431-config\") pod \"controller-manager-9d7ff4cd6-9jdqq\" (UID: \"5748bb4e-846d-457a-af17-b1d6f0a36431\") " pod="openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.940408 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkgwr\" (UniqueName: \"kubernetes.io/projected/5748bb4e-846d-457a-af17-b1d6f0a36431-kube-api-access-fkgwr\") pod \"controller-manager-9d7ff4cd6-9jdqq\" (UID: \"5748bb4e-846d-457a-af17-b1d6f0a36431\") " pod="openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.940564 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5748bb4e-846d-457a-af17-b1d6f0a36431-serving-cert\") pod \"controller-manager-9d7ff4cd6-9jdqq\" (UID: \"5748bb4e-846d-457a-af17-b1d6f0a36431\") " pod="openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.942371 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20e4371a-8bd2-4405-bb18-861923bfd37e-utilities\") pod \"community-operators-4vbvz\" (UID: \"20e4371a-8bd2-4405-bb18-861923bfd37e\") " pod="openshift-marketplace/community-operators-4vbvz" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.942408 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5748bb4e-846d-457a-af17-b1d6f0a36431-config\") pod \"controller-manager-9d7ff4cd6-9jdqq\" (UID: \"5748bb4e-846d-457a-af17-b1d6f0a36431\") " pod="openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.940756 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11-client-ca\") pod \"route-controller-manager-55985dff9-whh87\" (UID: \"ce1b4cd5-019c-41b7-a994-e98eb4fd3b11\") " pod="openshift-route-controller-manager/route-controller-manager-55985dff9-whh87" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.941817 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11-config\") pod \"route-controller-manager-55985dff9-whh87\" (UID: \"ce1b4cd5-019c-41b7-a994-e98eb4fd3b11\") " pod="openshift-route-controller-manager/route-controller-manager-55985dff9-whh87" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.940756 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5748bb4e-846d-457a-af17-b1d6f0a36431-client-ca\") pod \"controller-manager-9d7ff4cd6-9jdqq\" (UID: \"5748bb4e-846d-457a-af17-b1d6f0a36431\") " pod="openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.941634 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5748bb4e-846d-457a-af17-b1d6f0a36431-proxy-ca-bundles\") pod \"controller-manager-9d7ff4cd6-9jdqq\" (UID: \"5748bb4e-846d-457a-af17-b1d6f0a36431\") " pod="openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.945276 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5748bb4e-846d-457a-af17-b1d6f0a36431-serving-cert\") pod \"controller-manager-9d7ff4cd6-9jdqq\" (UID: \"5748bb4e-846d-457a-af17-b1d6f0a36431\") " pod="openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.945734 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11-serving-cert\") pod \"route-controller-manager-55985dff9-whh87\" (UID: \"ce1b4cd5-019c-41b7-a994-e98eb4fd3b11\") " pod="openshift-route-controller-manager/route-controller-manager-55985dff9-whh87" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.960324 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkgwr\" (UniqueName: \"kubernetes.io/projected/5748bb4e-846d-457a-af17-b1d6f0a36431-kube-api-access-fkgwr\") pod \"controller-manager-9d7ff4cd6-9jdqq\" (UID: \"5748bb4e-846d-457a-af17-b1d6f0a36431\") " pod="openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq" Jan 27 15:11:56 crc kubenswrapper[4772]: I0127 15:11:56.960404 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-799x8\" (UniqueName: \"kubernetes.io/projected/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11-kube-api-access-799x8\") pod \"route-controller-manager-55985dff9-whh87\" (UID: \"ce1b4cd5-019c-41b7-a994-e98eb4fd3b11\") " pod="openshift-route-controller-manager/route-controller-manager-55985dff9-whh87" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.032639 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dfvcs"] Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.033769 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dfvcs" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.036767 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.037833 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.041281 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dfvcs"] Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.046450 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-55985dff9-whh87" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.047123 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20e4371a-8bd2-4405-bb18-861923bfd37e-utilities\") pod \"community-operators-4vbvz\" (UID: \"20e4371a-8bd2-4405-bb18-861923bfd37e\") " pod="openshift-marketplace/community-operators-4vbvz" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.047238 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20e4371a-8bd2-4405-bb18-861923bfd37e-catalog-content\") pod \"community-operators-4vbvz\" (UID: \"20e4371a-8bd2-4405-bb18-861923bfd37e\") " pod="openshift-marketplace/community-operators-4vbvz" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.047282 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csq5c\" (UniqueName: \"kubernetes.io/projected/20e4371a-8bd2-4405-bb18-861923bfd37e-kube-api-access-csq5c\") pod \"community-operators-4vbvz\" (UID: \"20e4371a-8bd2-4405-bb18-861923bfd37e\") " pod="openshift-marketplace/community-operators-4vbvz" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.049643 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20e4371a-8bd2-4405-bb18-861923bfd37e-utilities\") pod \"community-operators-4vbvz\" (UID: \"20e4371a-8bd2-4405-bb18-861923bfd37e\") " pod="openshift-marketplace/community-operators-4vbvz" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.049703 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20e4371a-8bd2-4405-bb18-861923bfd37e-catalog-content\") pod \"community-operators-4vbvz\" (UID: \"20e4371a-8bd2-4405-bb18-861923bfd37e\") " pod="openshift-marketplace/community-operators-4vbvz" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.080503 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csq5c\" (UniqueName: \"kubernetes.io/projected/20e4371a-8bd2-4405-bb18-861923bfd37e-kube-api-access-csq5c\") pod \"community-operators-4vbvz\" (UID: \"20e4371a-8bd2-4405-bb18-861923bfd37e\") " pod="openshift-marketplace/community-operators-4vbvz" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.164979 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4vbvz" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.165940 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/881b071c-048c-4f66-96e7-fd1f91ca23f8-catalog-content\") pod \"certified-operators-dfvcs\" (UID: \"881b071c-048c-4f66-96e7-fd1f91ca23f8\") " pod="openshift-marketplace/certified-operators-dfvcs" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.165969 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfvh7\" (UniqueName: \"kubernetes.io/projected/881b071c-048c-4f66-96e7-fd1f91ca23f8-kube-api-access-tfvh7\") pod \"certified-operators-dfvcs\" (UID: \"881b071c-048c-4f66-96e7-fd1f91ca23f8\") " pod="openshift-marketplace/certified-operators-dfvcs" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.166006 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/881b071c-048c-4f66-96e7-fd1f91ca23f8-utilities\") pod \"certified-operators-dfvcs\" (UID: \"881b071c-048c-4f66-96e7-fd1f91ca23f8\") " pod="openshift-marketplace/certified-operators-dfvcs" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.267220 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/881b071c-048c-4f66-96e7-fd1f91ca23f8-utilities\") pod \"certified-operators-dfvcs\" (UID: \"881b071c-048c-4f66-96e7-fd1f91ca23f8\") " pod="openshift-marketplace/certified-operators-dfvcs" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.267652 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/881b071c-048c-4f66-96e7-fd1f91ca23f8-catalog-content\") pod \"certified-operators-dfvcs\" (UID: \"881b071c-048c-4f66-96e7-fd1f91ca23f8\") " pod="openshift-marketplace/certified-operators-dfvcs" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.267688 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfvh7\" (UniqueName: \"kubernetes.io/projected/881b071c-048c-4f66-96e7-fd1f91ca23f8-kube-api-access-tfvh7\") pod \"certified-operators-dfvcs\" (UID: \"881b071c-048c-4f66-96e7-fd1f91ca23f8\") " pod="openshift-marketplace/certified-operators-dfvcs" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.268615 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/881b071c-048c-4f66-96e7-fd1f91ca23f8-utilities\") pod \"certified-operators-dfvcs\" (UID: \"881b071c-048c-4f66-96e7-fd1f91ca23f8\") " pod="openshift-marketplace/certified-operators-dfvcs" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.268878 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/881b071c-048c-4f66-96e7-fd1f91ca23f8-catalog-content\") pod \"certified-operators-dfvcs\" (UID: \"881b071c-048c-4f66-96e7-fd1f91ca23f8\") " pod="openshift-marketplace/certified-operators-dfvcs" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.276460 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55985dff9-whh87"] Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.292041 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfvh7\" (UniqueName: \"kubernetes.io/projected/881b071c-048c-4f66-96e7-fd1f91ca23f8-kube-api-access-tfvh7\") pod \"certified-operators-dfvcs\" (UID: \"881b071c-048c-4f66-96e7-fd1f91ca23f8\") " pod="openshift-marketplace/certified-operators-dfvcs" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.332386 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq"] Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.362319 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4vbvz"] Jan 27 15:11:57 crc kubenswrapper[4772]: W0127 15:11:57.369551 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20e4371a_8bd2_4405_bb18_861923bfd37e.slice/crio-f541cf85cdae422e49f9b3df32c8dd0416ecc007711177dc8c1d7d7680929e43 WatchSource:0}: Error finding container f541cf85cdae422e49f9b3df32c8dd0416ecc007711177dc8c1d7d7680929e43: Status 404 returned error can't find the container with id f541cf85cdae422e49f9b3df32c8dd0416ecc007711177dc8c1d7d7680929e43 Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.383313 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dfvcs" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.606912 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dfvcs"] Jan 27 15:11:57 crc kubenswrapper[4772]: W0127 15:11:57.655240 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod881b071c_048c_4f66_96e7_fd1f91ca23f8.slice/crio-fb988077c86d56b0ecd59d7c83bf17183c0cb096ca767c2c7c33983cb62018dd WatchSource:0}: Error finding container fb988077c86d56b0ecd59d7c83bf17183c0cb096ca767c2c7c33983cb62018dd: Status 404 returned error can't find the container with id fb988077c86d56b0ecd59d7c83bf17183c0cb096ca767c2c7c33983cb62018dd Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.887559 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-55985dff9-whh87" event={"ID":"ce1b4cd5-019c-41b7-a994-e98eb4fd3b11","Type":"ContainerStarted","Data":"d424ca2f80c2d25b1c4a792e0cbbcc57ee25bfd64b22c04930189bc1f4e34858"} Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.887620 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-55985dff9-whh87" event={"ID":"ce1b4cd5-019c-41b7-a994-e98eb4fd3b11","Type":"ContainerStarted","Data":"fe97b2d600f6d6cc4091b116b23735caf871c4e2a0580b348ce94443cc624fcf"} Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.887937 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-55985dff9-whh87" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.888893 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dfvcs" event={"ID":"881b071c-048c-4f66-96e7-fd1f91ca23f8","Type":"ContainerStarted","Data":"fe3fede8510b6c4bb0f1ffacbe3d0ec0288f88b5a6f5f6a793e5f3dd2eeff90b"} Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.888934 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dfvcs" event={"ID":"881b071c-048c-4f66-96e7-fd1f91ca23f8","Type":"ContainerStarted","Data":"fb988077c86d56b0ecd59d7c83bf17183c0cb096ca767c2c7c33983cb62018dd"} Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.891115 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq" event={"ID":"5748bb4e-846d-457a-af17-b1d6f0a36431","Type":"ContainerStarted","Data":"b33e07e1bbdb8f68554172eecced3999514bd01390467f8165d94835e2937303"} Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.891188 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq" event={"ID":"5748bb4e-846d-457a-af17-b1d6f0a36431","Type":"ContainerStarted","Data":"702d8de62fe8dceb9e42a32a6d4bbb6520eefa3da0fef6914772f56056d17cc4"} Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.891834 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.893638 4772 generic.go:334] "Generic (PLEG): container finished" podID="20e4371a-8bd2-4405-bb18-861923bfd37e" containerID="a913cea196ab4b53872fe2376a7200e2beda9ee644b2571a89e275314690904e" exitCode=0 Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.893672 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4vbvz" event={"ID":"20e4371a-8bd2-4405-bb18-861923bfd37e","Type":"ContainerDied","Data":"a913cea196ab4b53872fe2376a7200e2beda9ee644b2571a89e275314690904e"} Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.893701 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4vbvz" event={"ID":"20e4371a-8bd2-4405-bb18-861923bfd37e","Type":"ContainerStarted","Data":"f541cf85cdae422e49f9b3df32c8dd0416ecc007711177dc8c1d7d7680929e43"} Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.895797 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-55985dff9-whh87" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.904875 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.919443 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-55985dff9-whh87" podStartSLOduration=2.919420792 podStartE2EDuration="2.919420792s" podCreationTimestamp="2026-01-27 15:11:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:11:57.916687461 +0000 UTC m=+303.897296559" watchObservedRunningTime="2026-01-27 15:11:57.919420792 +0000 UTC m=+303.900029890" Jan 27 15:11:57 crc kubenswrapper[4772]: I0127 15:11:57.945141 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-9d7ff4cd6-9jdqq" podStartSLOduration=2.945102248 podStartE2EDuration="2.945102248s" podCreationTimestamp="2026-01-27 15:11:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:11:57.940247955 +0000 UTC m=+303.920857053" watchObservedRunningTime="2026-01-27 15:11:57.945102248 +0000 UTC m=+303.925711346" Jan 27 15:11:58 crc kubenswrapper[4772]: I0127 15:11:58.903208 4772 generic.go:334] "Generic (PLEG): container finished" podID="881b071c-048c-4f66-96e7-fd1f91ca23f8" containerID="fe3fede8510b6c4bb0f1ffacbe3d0ec0288f88b5a6f5f6a793e5f3dd2eeff90b" exitCode=0 Jan 27 15:11:58 crc kubenswrapper[4772]: I0127 15:11:58.903407 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dfvcs" event={"ID":"881b071c-048c-4f66-96e7-fd1f91ca23f8","Type":"ContainerDied","Data":"fe3fede8510b6c4bb0f1ffacbe3d0ec0288f88b5a6f5f6a793e5f3dd2eeff90b"} Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.238325 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f5shj"] Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.239940 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f5shj" Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.242279 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.249912 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5shj"] Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.412869 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72dcc284-e96b-4605-a428-176ca549eeb2-utilities\") pod \"redhat-marketplace-f5shj\" (UID: \"72dcc284-e96b-4605-a428-176ca549eeb2\") " pod="openshift-marketplace/redhat-marketplace-f5shj" Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.412937 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzc2r\" (UniqueName: \"kubernetes.io/projected/72dcc284-e96b-4605-a428-176ca549eeb2-kube-api-access-pzc2r\") pod \"redhat-marketplace-f5shj\" (UID: \"72dcc284-e96b-4605-a428-176ca549eeb2\") " pod="openshift-marketplace/redhat-marketplace-f5shj" Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.413044 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72dcc284-e96b-4605-a428-176ca549eeb2-catalog-content\") pod \"redhat-marketplace-f5shj\" (UID: \"72dcc284-e96b-4605-a428-176ca549eeb2\") " pod="openshift-marketplace/redhat-marketplace-f5shj" Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.435380 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5whzm"] Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.436655 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5whzm" Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.440206 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.447915 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5whzm"] Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.514146 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzc2r\" (UniqueName: \"kubernetes.io/projected/72dcc284-e96b-4605-a428-176ca549eeb2-kube-api-access-pzc2r\") pod \"redhat-marketplace-f5shj\" (UID: \"72dcc284-e96b-4605-a428-176ca549eeb2\") " pod="openshift-marketplace/redhat-marketplace-f5shj" Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.514254 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72dcc284-e96b-4605-a428-176ca549eeb2-catalog-content\") pod \"redhat-marketplace-f5shj\" (UID: \"72dcc284-e96b-4605-a428-176ca549eeb2\") " pod="openshift-marketplace/redhat-marketplace-f5shj" Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.514296 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72dcc284-e96b-4605-a428-176ca549eeb2-utilities\") pod \"redhat-marketplace-f5shj\" (UID: \"72dcc284-e96b-4605-a428-176ca549eeb2\") " pod="openshift-marketplace/redhat-marketplace-f5shj" Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.515073 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72dcc284-e96b-4605-a428-176ca549eeb2-utilities\") pod \"redhat-marketplace-f5shj\" (UID: \"72dcc284-e96b-4605-a428-176ca549eeb2\") " pod="openshift-marketplace/redhat-marketplace-f5shj" Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.515079 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72dcc284-e96b-4605-a428-176ca549eeb2-catalog-content\") pod \"redhat-marketplace-f5shj\" (UID: \"72dcc284-e96b-4605-a428-176ca549eeb2\") " pod="openshift-marketplace/redhat-marketplace-f5shj" Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.534148 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzc2r\" (UniqueName: \"kubernetes.io/projected/72dcc284-e96b-4605-a428-176ca549eeb2-kube-api-access-pzc2r\") pod \"redhat-marketplace-f5shj\" (UID: \"72dcc284-e96b-4605-a428-176ca549eeb2\") " pod="openshift-marketplace/redhat-marketplace-f5shj" Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.562926 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f5shj" Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.616124 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vr9pf\" (UniqueName: \"kubernetes.io/projected/79b85747-dcbc-462d-85d1-3d00801b5106-kube-api-access-vr9pf\") pod \"redhat-operators-5whzm\" (UID: \"79b85747-dcbc-462d-85d1-3d00801b5106\") " pod="openshift-marketplace/redhat-operators-5whzm" Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.623766 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79b85747-dcbc-462d-85d1-3d00801b5106-catalog-content\") pod \"redhat-operators-5whzm\" (UID: \"79b85747-dcbc-462d-85d1-3d00801b5106\") " pod="openshift-marketplace/redhat-operators-5whzm" Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.623868 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79b85747-dcbc-462d-85d1-3d00801b5106-utilities\") pod \"redhat-operators-5whzm\" (UID: \"79b85747-dcbc-462d-85d1-3d00801b5106\") " pod="openshift-marketplace/redhat-operators-5whzm" Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.725547 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79b85747-dcbc-462d-85d1-3d00801b5106-catalog-content\") pod \"redhat-operators-5whzm\" (UID: \"79b85747-dcbc-462d-85d1-3d00801b5106\") " pod="openshift-marketplace/redhat-operators-5whzm" Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.725600 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79b85747-dcbc-462d-85d1-3d00801b5106-utilities\") pod \"redhat-operators-5whzm\" (UID: \"79b85747-dcbc-462d-85d1-3d00801b5106\") " pod="openshift-marketplace/redhat-operators-5whzm" Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.725667 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vr9pf\" (UniqueName: \"kubernetes.io/projected/79b85747-dcbc-462d-85d1-3d00801b5106-kube-api-access-vr9pf\") pod \"redhat-operators-5whzm\" (UID: \"79b85747-dcbc-462d-85d1-3d00801b5106\") " pod="openshift-marketplace/redhat-operators-5whzm" Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.726911 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79b85747-dcbc-462d-85d1-3d00801b5106-catalog-content\") pod \"redhat-operators-5whzm\" (UID: \"79b85747-dcbc-462d-85d1-3d00801b5106\") " pod="openshift-marketplace/redhat-operators-5whzm" Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.728869 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79b85747-dcbc-462d-85d1-3d00801b5106-utilities\") pod \"redhat-operators-5whzm\" (UID: \"79b85747-dcbc-462d-85d1-3d00801b5106\") " pod="openshift-marketplace/redhat-operators-5whzm" Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.744310 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vr9pf\" (UniqueName: \"kubernetes.io/projected/79b85747-dcbc-462d-85d1-3d00801b5106-kube-api-access-vr9pf\") pod \"redhat-operators-5whzm\" (UID: \"79b85747-dcbc-462d-85d1-3d00801b5106\") " pod="openshift-marketplace/redhat-operators-5whzm" Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.750884 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5whzm" Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.910890 4772 generic.go:334] "Generic (PLEG): container finished" podID="20e4371a-8bd2-4405-bb18-861923bfd37e" containerID="3e6786fd605112a5af975912e2a7080680ebac9d9bf41ad3486e573e6d6d2f4c" exitCode=0 Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.911713 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4vbvz" event={"ID":"20e4371a-8bd2-4405-bb18-861923bfd37e","Type":"ContainerDied","Data":"3e6786fd605112a5af975912e2a7080680ebac9d9bf41ad3486e573e6d6d2f4c"} Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.956720 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5whzm"] Jan 27 15:11:59 crc kubenswrapper[4772]: W0127 15:11:59.960910 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79b85747_dcbc_462d_85d1_3d00801b5106.slice/crio-8e4bc519d0ced9952d6857ff31015675ce4705ed12ce2547ebdba49c33fc4d62 WatchSource:0}: Error finding container 8e4bc519d0ced9952d6857ff31015675ce4705ed12ce2547ebdba49c33fc4d62: Status 404 returned error can't find the container with id 8e4bc519d0ced9952d6857ff31015675ce4705ed12ce2547ebdba49c33fc4d62 Jan 27 15:11:59 crc kubenswrapper[4772]: I0127 15:11:59.965081 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5shj"] Jan 27 15:11:59 crc kubenswrapper[4772]: W0127 15:11:59.966101 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72dcc284_e96b_4605_a428_176ca549eeb2.slice/crio-55e61e49763a06e11c87abfce2babaac3a236dfbde7e77c897ef5f83bb000985 WatchSource:0}: Error finding container 55e61e49763a06e11c87abfce2babaac3a236dfbde7e77c897ef5f83bb000985: Status 404 returned error can't find the container with id 55e61e49763a06e11c87abfce2babaac3a236dfbde7e77c897ef5f83bb000985 Jan 27 15:12:00 crc kubenswrapper[4772]: I0127 15:12:00.916803 4772 generic.go:334] "Generic (PLEG): container finished" podID="79b85747-dcbc-462d-85d1-3d00801b5106" containerID="d2c52e57f45d6596ec80d58984736144fe9463319975ceee95db9f926e0b38f1" exitCode=0 Jan 27 15:12:00 crc kubenswrapper[4772]: I0127 15:12:00.916916 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5whzm" event={"ID":"79b85747-dcbc-462d-85d1-3d00801b5106","Type":"ContainerDied","Data":"d2c52e57f45d6596ec80d58984736144fe9463319975ceee95db9f926e0b38f1"} Jan 27 15:12:00 crc kubenswrapper[4772]: I0127 15:12:00.917262 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5whzm" event={"ID":"79b85747-dcbc-462d-85d1-3d00801b5106","Type":"ContainerStarted","Data":"8e4bc519d0ced9952d6857ff31015675ce4705ed12ce2547ebdba49c33fc4d62"} Jan 27 15:12:00 crc kubenswrapper[4772]: I0127 15:12:00.920527 4772 generic.go:334] "Generic (PLEG): container finished" podID="881b071c-048c-4f66-96e7-fd1f91ca23f8" containerID="42303aea9268a5a7664e56d12a4e4ea8c3524b3434d07f5f965bd4218953e4a5" exitCode=0 Jan 27 15:12:00 crc kubenswrapper[4772]: I0127 15:12:00.921015 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dfvcs" event={"ID":"881b071c-048c-4f66-96e7-fd1f91ca23f8","Type":"ContainerDied","Data":"42303aea9268a5a7664e56d12a4e4ea8c3524b3434d07f5f965bd4218953e4a5"} Jan 27 15:12:00 crc kubenswrapper[4772]: I0127 15:12:00.924937 4772 generic.go:334] "Generic (PLEG): container finished" podID="72dcc284-e96b-4605-a428-176ca549eeb2" containerID="a30e7f4b8f02ced5a831edb30ea63fa4ce910f9dd257e80f6edbb087e09ccd37" exitCode=0 Jan 27 15:12:00 crc kubenswrapper[4772]: I0127 15:12:00.925008 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5shj" event={"ID":"72dcc284-e96b-4605-a428-176ca549eeb2","Type":"ContainerDied","Data":"a30e7f4b8f02ced5a831edb30ea63fa4ce910f9dd257e80f6edbb087e09ccd37"} Jan 27 15:12:00 crc kubenswrapper[4772]: I0127 15:12:00.925042 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5shj" event={"ID":"72dcc284-e96b-4605-a428-176ca549eeb2","Type":"ContainerStarted","Data":"55e61e49763a06e11c87abfce2babaac3a236dfbde7e77c897ef5f83bb000985"} Jan 27 15:12:00 crc kubenswrapper[4772]: I0127 15:12:00.927419 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4vbvz" event={"ID":"20e4371a-8bd2-4405-bb18-861923bfd37e","Type":"ContainerStarted","Data":"facc4381b6ccbc800e2ee9d053a4ca1a05dc9ad0f250dcd95e8e1ea0e652318f"} Jan 27 15:12:00 crc kubenswrapper[4772]: I0127 15:12:00.970459 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4vbvz" podStartSLOduration=2.334685243 podStartE2EDuration="4.970440495s" podCreationTimestamp="2026-01-27 15:11:56 +0000 UTC" firstStartedPulling="2026-01-27 15:11:57.894706354 +0000 UTC m=+303.875315452" lastFinishedPulling="2026-01-27 15:12:00.530461606 +0000 UTC m=+306.511070704" observedRunningTime="2026-01-27 15:12:00.967327213 +0000 UTC m=+306.947936321" watchObservedRunningTime="2026-01-27 15:12:00.970440495 +0000 UTC m=+306.951049593" Jan 27 15:12:01 crc kubenswrapper[4772]: I0127 15:12:01.939025 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dfvcs" event={"ID":"881b071c-048c-4f66-96e7-fd1f91ca23f8","Type":"ContainerStarted","Data":"bebeaa0da80cc73fcfa950011bea7962115e29aeaf6343634912c781ccb9d6a8"} Jan 27 15:12:01 crc kubenswrapper[4772]: I0127 15:12:01.949257 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5shj" event={"ID":"72dcc284-e96b-4605-a428-176ca549eeb2","Type":"ContainerStarted","Data":"317f0c9fa8dd2e57bedd32005f0d5d616e9ee8287f35270140e4402ef9d2fd3c"} Jan 27 15:12:01 crc kubenswrapper[4772]: I0127 15:12:01.959068 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dfvcs" podStartSLOduration=2.312772231 podStartE2EDuration="4.959047383s" podCreationTimestamp="2026-01-27 15:11:57 +0000 UTC" firstStartedPulling="2026-01-27 15:11:58.982752311 +0000 UTC m=+304.963361409" lastFinishedPulling="2026-01-27 15:12:01.629027473 +0000 UTC m=+307.609636561" observedRunningTime="2026-01-27 15:12:01.957867069 +0000 UTC m=+307.938476177" watchObservedRunningTime="2026-01-27 15:12:01.959047383 +0000 UTC m=+307.939656491" Jan 27 15:12:02 crc kubenswrapper[4772]: I0127 15:12:02.955599 4772 generic.go:334] "Generic (PLEG): container finished" podID="79b85747-dcbc-462d-85d1-3d00801b5106" containerID="eb48a00f974d6ceae0f8e67b809e0845663f47278e4609e58f35d13935712ff7" exitCode=0 Jan 27 15:12:02 crc kubenswrapper[4772]: I0127 15:12:02.955733 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5whzm" event={"ID":"79b85747-dcbc-462d-85d1-3d00801b5106","Type":"ContainerDied","Data":"eb48a00f974d6ceae0f8e67b809e0845663f47278e4609e58f35d13935712ff7"} Jan 27 15:12:02 crc kubenswrapper[4772]: I0127 15:12:02.958585 4772 generic.go:334] "Generic (PLEG): container finished" podID="72dcc284-e96b-4605-a428-176ca549eeb2" containerID="317f0c9fa8dd2e57bedd32005f0d5d616e9ee8287f35270140e4402ef9d2fd3c" exitCode=0 Jan 27 15:12:02 crc kubenswrapper[4772]: I0127 15:12:02.959383 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5shj" event={"ID":"72dcc284-e96b-4605-a428-176ca549eeb2","Type":"ContainerDied","Data":"317f0c9fa8dd2e57bedd32005f0d5d616e9ee8287f35270140e4402ef9d2fd3c"} Jan 27 15:12:02 crc kubenswrapper[4772]: I0127 15:12:02.959406 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5shj" event={"ID":"72dcc284-e96b-4605-a428-176ca549eeb2","Type":"ContainerStarted","Data":"03bb2fbce9122be2ed850d1a52f257693a41c3bbdd07883284ed1f5f69d9a222"} Jan 27 15:12:02 crc kubenswrapper[4772]: I0127 15:12:02.994077 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f5shj" podStartSLOduration=2.38753925 podStartE2EDuration="3.994059738s" podCreationTimestamp="2026-01-27 15:11:59 +0000 UTC" firstStartedPulling="2026-01-27 15:12:00.926748008 +0000 UTC m=+306.907357106" lastFinishedPulling="2026-01-27 15:12:02.533268506 +0000 UTC m=+308.513877594" observedRunningTime="2026-01-27 15:12:02.991297997 +0000 UTC m=+308.971907115" watchObservedRunningTime="2026-01-27 15:12:02.994059738 +0000 UTC m=+308.974668836" Jan 27 15:12:03 crc kubenswrapper[4772]: I0127 15:12:03.965900 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5whzm" event={"ID":"79b85747-dcbc-462d-85d1-3d00801b5106","Type":"ContainerStarted","Data":"a2d56215a8255cd6fe0e2caab61dc965898bd5a40e9690f3cc4fda7a7884bd0c"} Jan 27 15:12:03 crc kubenswrapper[4772]: I0127 15:12:03.988784 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5whzm" podStartSLOduration=2.571040494 podStartE2EDuration="4.988761915s" podCreationTimestamp="2026-01-27 15:11:59 +0000 UTC" firstStartedPulling="2026-01-27 15:12:00.919189605 +0000 UTC m=+306.899798703" lastFinishedPulling="2026-01-27 15:12:03.336911036 +0000 UTC m=+309.317520124" observedRunningTime="2026-01-27 15:12:03.986339294 +0000 UTC m=+309.966948452" watchObservedRunningTime="2026-01-27 15:12:03.988761915 +0000 UTC m=+309.969371013" Jan 27 15:12:07 crc kubenswrapper[4772]: I0127 15:12:07.165693 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4vbvz" Jan 27 15:12:07 crc kubenswrapper[4772]: I0127 15:12:07.166092 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4vbvz" Jan 27 15:12:07 crc kubenswrapper[4772]: I0127 15:12:07.212261 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4vbvz" Jan 27 15:12:07 crc kubenswrapper[4772]: I0127 15:12:07.384538 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dfvcs" Jan 27 15:12:07 crc kubenswrapper[4772]: I0127 15:12:07.384582 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dfvcs" Jan 27 15:12:07 crc kubenswrapper[4772]: I0127 15:12:07.425904 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dfvcs" Jan 27 15:12:08 crc kubenswrapper[4772]: I0127 15:12:08.025196 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4vbvz" Jan 27 15:12:08 crc kubenswrapper[4772]: I0127 15:12:08.025605 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dfvcs" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.563331 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f5shj" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.563386 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f5shj" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.609945 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-2dw59"] Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.610583 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.623388 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-2dw59"] Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.628608 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f5shj" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.751161 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5whzm" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.751247 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5whzm" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.760715 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/566c3aef-cf83-4a7b-a77e-774a9dfb90a6-trusted-ca\") pod \"image-registry-66df7c8f76-2dw59\" (UID: \"566c3aef-cf83-4a7b-a77e-774a9dfb90a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.760791 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/566c3aef-cf83-4a7b-a77e-774a9dfb90a6-bound-sa-token\") pod \"image-registry-66df7c8f76-2dw59\" (UID: \"566c3aef-cf83-4a7b-a77e-774a9dfb90a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.760916 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/566c3aef-cf83-4a7b-a77e-774a9dfb90a6-registry-tls\") pod \"image-registry-66df7c8f76-2dw59\" (UID: \"566c3aef-cf83-4a7b-a77e-774a9dfb90a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.761124 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/566c3aef-cf83-4a7b-a77e-774a9dfb90a6-installation-pull-secrets\") pod \"image-registry-66df7c8f76-2dw59\" (UID: \"566c3aef-cf83-4a7b-a77e-774a9dfb90a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.761227 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/566c3aef-cf83-4a7b-a77e-774a9dfb90a6-registry-certificates\") pod \"image-registry-66df7c8f76-2dw59\" (UID: \"566c3aef-cf83-4a7b-a77e-774a9dfb90a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.761263 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-2dw59\" (UID: \"566c3aef-cf83-4a7b-a77e-774a9dfb90a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.761282 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/566c3aef-cf83-4a7b-a77e-774a9dfb90a6-ca-trust-extracted\") pod \"image-registry-66df7c8f76-2dw59\" (UID: \"566c3aef-cf83-4a7b-a77e-774a9dfb90a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.761314 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm6cr\" (UniqueName: \"kubernetes.io/projected/566c3aef-cf83-4a7b-a77e-774a9dfb90a6-kube-api-access-cm6cr\") pod \"image-registry-66df7c8f76-2dw59\" (UID: \"566c3aef-cf83-4a7b-a77e-774a9dfb90a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.786745 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-2dw59\" (UID: \"566c3aef-cf83-4a7b-a77e-774a9dfb90a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.791663 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5whzm" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.862968 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/566c3aef-cf83-4a7b-a77e-774a9dfb90a6-registry-certificates\") pod \"image-registry-66df7c8f76-2dw59\" (UID: \"566c3aef-cf83-4a7b-a77e-774a9dfb90a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.863464 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/566c3aef-cf83-4a7b-a77e-774a9dfb90a6-ca-trust-extracted\") pod \"image-registry-66df7c8f76-2dw59\" (UID: \"566c3aef-cf83-4a7b-a77e-774a9dfb90a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.863528 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm6cr\" (UniqueName: \"kubernetes.io/projected/566c3aef-cf83-4a7b-a77e-774a9dfb90a6-kube-api-access-cm6cr\") pod \"image-registry-66df7c8f76-2dw59\" (UID: \"566c3aef-cf83-4a7b-a77e-774a9dfb90a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.863573 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/566c3aef-cf83-4a7b-a77e-774a9dfb90a6-trusted-ca\") pod \"image-registry-66df7c8f76-2dw59\" (UID: \"566c3aef-cf83-4a7b-a77e-774a9dfb90a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.863626 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/566c3aef-cf83-4a7b-a77e-774a9dfb90a6-bound-sa-token\") pod \"image-registry-66df7c8f76-2dw59\" (UID: \"566c3aef-cf83-4a7b-a77e-774a9dfb90a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.863669 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/566c3aef-cf83-4a7b-a77e-774a9dfb90a6-registry-tls\") pod \"image-registry-66df7c8f76-2dw59\" (UID: \"566c3aef-cf83-4a7b-a77e-774a9dfb90a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.863701 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/566c3aef-cf83-4a7b-a77e-774a9dfb90a6-installation-pull-secrets\") pod \"image-registry-66df7c8f76-2dw59\" (UID: \"566c3aef-cf83-4a7b-a77e-774a9dfb90a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.864163 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/566c3aef-cf83-4a7b-a77e-774a9dfb90a6-ca-trust-extracted\") pod \"image-registry-66df7c8f76-2dw59\" (UID: \"566c3aef-cf83-4a7b-a77e-774a9dfb90a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.864654 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/566c3aef-cf83-4a7b-a77e-774a9dfb90a6-registry-certificates\") pod \"image-registry-66df7c8f76-2dw59\" (UID: \"566c3aef-cf83-4a7b-a77e-774a9dfb90a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.865318 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/566c3aef-cf83-4a7b-a77e-774a9dfb90a6-trusted-ca\") pod \"image-registry-66df7c8f76-2dw59\" (UID: \"566c3aef-cf83-4a7b-a77e-774a9dfb90a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.869536 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/566c3aef-cf83-4a7b-a77e-774a9dfb90a6-registry-tls\") pod \"image-registry-66df7c8f76-2dw59\" (UID: \"566c3aef-cf83-4a7b-a77e-774a9dfb90a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.869534 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/566c3aef-cf83-4a7b-a77e-774a9dfb90a6-installation-pull-secrets\") pod \"image-registry-66df7c8f76-2dw59\" (UID: \"566c3aef-cf83-4a7b-a77e-774a9dfb90a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.882902 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/566c3aef-cf83-4a7b-a77e-774a9dfb90a6-bound-sa-token\") pod \"image-registry-66df7c8f76-2dw59\" (UID: \"566c3aef-cf83-4a7b-a77e-774a9dfb90a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.883524 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm6cr\" (UniqueName: \"kubernetes.io/projected/566c3aef-cf83-4a7b-a77e-774a9dfb90a6-kube-api-access-cm6cr\") pod \"image-registry-66df7c8f76-2dw59\" (UID: \"566c3aef-cf83-4a7b-a77e-774a9dfb90a6\") " pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:09 crc kubenswrapper[4772]: I0127 15:12:09.944290 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:10 crc kubenswrapper[4772]: I0127 15:12:10.099828 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5whzm" Jan 27 15:12:10 crc kubenswrapper[4772]: I0127 15:12:10.099899 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f5shj" Jan 27 15:12:10 crc kubenswrapper[4772]: I0127 15:12:10.433938 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-2dw59"] Jan 27 15:12:11 crc kubenswrapper[4772]: I0127 15:12:11.008426 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" event={"ID":"566c3aef-cf83-4a7b-a77e-774a9dfb90a6","Type":"ContainerStarted","Data":"bc3cce03415945747e4cc390ca99cce1ebbbb8af445807ab4be29b39e3748ccb"} Jan 27 15:12:12 crc kubenswrapper[4772]: I0127 15:12:12.016041 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" event={"ID":"566c3aef-cf83-4a7b-a77e-774a9dfb90a6","Type":"ContainerStarted","Data":"110f202c0cb81a69a4644066f56a9baff303141c914ee46b5b9318a554daa0e9"} Jan 27 15:12:13 crc kubenswrapper[4772]: I0127 15:12:13.021310 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:13 crc kubenswrapper[4772]: I0127 15:12:13.042145 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" podStartSLOduration=4.042128471 podStartE2EDuration="4.042128471s" podCreationTimestamp="2026-01-27 15:12:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:12:13.040262566 +0000 UTC m=+319.020871684" watchObservedRunningTime="2026-01-27 15:12:13.042128471 +0000 UTC m=+319.022737569" Jan 27 15:12:15 crc kubenswrapper[4772]: I0127 15:12:15.069105 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55985dff9-whh87"] Jan 27 15:12:15 crc kubenswrapper[4772]: I0127 15:12:15.069342 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-55985dff9-whh87" podUID="ce1b4cd5-019c-41b7-a994-e98eb4fd3b11" containerName="route-controller-manager" containerID="cri-o://d424ca2f80c2d25b1c4a792e0cbbcc57ee25bfd64b22c04930189bc1f4e34858" gracePeriod=30 Jan 27 15:12:15 crc kubenswrapper[4772]: I0127 15:12:15.553307 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-55985dff9-whh87" Jan 27 15:12:15 crc kubenswrapper[4772]: I0127 15:12:15.669728 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11-config\") pod \"ce1b4cd5-019c-41b7-a994-e98eb4fd3b11\" (UID: \"ce1b4cd5-019c-41b7-a994-e98eb4fd3b11\") " Jan 27 15:12:15 crc kubenswrapper[4772]: I0127 15:12:15.669798 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-799x8\" (UniqueName: \"kubernetes.io/projected/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11-kube-api-access-799x8\") pod \"ce1b4cd5-019c-41b7-a994-e98eb4fd3b11\" (UID: \"ce1b4cd5-019c-41b7-a994-e98eb4fd3b11\") " Jan 27 15:12:15 crc kubenswrapper[4772]: I0127 15:12:15.669824 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11-serving-cert\") pod \"ce1b4cd5-019c-41b7-a994-e98eb4fd3b11\" (UID: \"ce1b4cd5-019c-41b7-a994-e98eb4fd3b11\") " Jan 27 15:12:15 crc kubenswrapper[4772]: I0127 15:12:15.669888 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11-client-ca\") pod \"ce1b4cd5-019c-41b7-a994-e98eb4fd3b11\" (UID: \"ce1b4cd5-019c-41b7-a994-e98eb4fd3b11\") " Jan 27 15:12:15 crc kubenswrapper[4772]: I0127 15:12:15.670557 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11-client-ca" (OuterVolumeSpecName: "client-ca") pod "ce1b4cd5-019c-41b7-a994-e98eb4fd3b11" (UID: "ce1b4cd5-019c-41b7-a994-e98eb4fd3b11"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:12:15 crc kubenswrapper[4772]: I0127 15:12:15.670580 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11-config" (OuterVolumeSpecName: "config") pod "ce1b4cd5-019c-41b7-a994-e98eb4fd3b11" (UID: "ce1b4cd5-019c-41b7-a994-e98eb4fd3b11"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:12:15 crc kubenswrapper[4772]: I0127 15:12:15.674914 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11-kube-api-access-799x8" (OuterVolumeSpecName: "kube-api-access-799x8") pod "ce1b4cd5-019c-41b7-a994-e98eb4fd3b11" (UID: "ce1b4cd5-019c-41b7-a994-e98eb4fd3b11"). InnerVolumeSpecName "kube-api-access-799x8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:12:15 crc kubenswrapper[4772]: I0127 15:12:15.675048 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ce1b4cd5-019c-41b7-a994-e98eb4fd3b11" (UID: "ce1b4cd5-019c-41b7-a994-e98eb4fd3b11"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:12:15 crc kubenswrapper[4772]: I0127 15:12:15.770943 4772 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11-client-ca\") on node \"crc\" DevicePath \"\"" Jan 27 15:12:15 crc kubenswrapper[4772]: I0127 15:12:15.770980 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:12:15 crc kubenswrapper[4772]: I0127 15:12:15.770990 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-799x8\" (UniqueName: \"kubernetes.io/projected/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11-kube-api-access-799x8\") on node \"crc\" DevicePath \"\"" Jan 27 15:12:15 crc kubenswrapper[4772]: I0127 15:12:15.771001 4772 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.047763 4772 generic.go:334] "Generic (PLEG): container finished" podID="ce1b4cd5-019c-41b7-a994-e98eb4fd3b11" containerID="d424ca2f80c2d25b1c4a792e0cbbcc57ee25bfd64b22c04930189bc1f4e34858" exitCode=0 Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.047821 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-55985dff9-whh87" event={"ID":"ce1b4cd5-019c-41b7-a994-e98eb4fd3b11","Type":"ContainerDied","Data":"d424ca2f80c2d25b1c4a792e0cbbcc57ee25bfd64b22c04930189bc1f4e34858"} Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.047854 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-55985dff9-whh87" event={"ID":"ce1b4cd5-019c-41b7-a994-e98eb4fd3b11","Type":"ContainerDied","Data":"fe97b2d600f6d6cc4091b116b23735caf871c4e2a0580b348ce94443cc624fcf"} Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.047878 4772 scope.go:117] "RemoveContainer" containerID="d424ca2f80c2d25b1c4a792e0cbbcc57ee25bfd64b22c04930189bc1f4e34858" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.048021 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-55985dff9-whh87" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.070387 4772 scope.go:117] "RemoveContainer" containerID="d424ca2f80c2d25b1c4a792e0cbbcc57ee25bfd64b22c04930189bc1f4e34858" Jan 27 15:12:16 crc kubenswrapper[4772]: E0127 15:12:16.071223 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d424ca2f80c2d25b1c4a792e0cbbcc57ee25bfd64b22c04930189bc1f4e34858\": container with ID starting with d424ca2f80c2d25b1c4a792e0cbbcc57ee25bfd64b22c04930189bc1f4e34858 not found: ID does not exist" containerID="d424ca2f80c2d25b1c4a792e0cbbcc57ee25bfd64b22c04930189bc1f4e34858" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.071251 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d424ca2f80c2d25b1c4a792e0cbbcc57ee25bfd64b22c04930189bc1f4e34858"} err="failed to get container status \"d424ca2f80c2d25b1c4a792e0cbbcc57ee25bfd64b22c04930189bc1f4e34858\": rpc error: code = NotFound desc = could not find container \"d424ca2f80c2d25b1c4a792e0cbbcc57ee25bfd64b22c04930189bc1f4e34858\": container with ID starting with d424ca2f80c2d25b1c4a792e0cbbcc57ee25bfd64b22c04930189bc1f4e34858 not found: ID does not exist" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.079699 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55985dff9-whh87"] Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.084285 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55985dff9-whh87"] Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.670593 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce1b4cd5-019c-41b7-a994-e98eb4fd3b11" path="/var/lib/kubelet/pods/ce1b4cd5-019c-41b7-a994-e98eb4fd3b11/volumes" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.724550 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66f698ffcb-hdfc8"] Jan 27 15:12:16 crc kubenswrapper[4772]: E0127 15:12:16.724753 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce1b4cd5-019c-41b7-a994-e98eb4fd3b11" containerName="route-controller-manager" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.724765 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce1b4cd5-019c-41b7-a994-e98eb4fd3b11" containerName="route-controller-manager" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.724851 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce1b4cd5-019c-41b7-a994-e98eb4fd3b11" containerName="route-controller-manager" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.725220 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66f698ffcb-hdfc8" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.728222 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.728872 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.729013 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.728907 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.729451 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.729329 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.736740 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66f698ffcb-hdfc8"] Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.883812 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7ebccc5-7e14-42c0-8324-9acee056f8a7-serving-cert\") pod \"route-controller-manager-66f698ffcb-hdfc8\" (UID: \"f7ebccc5-7e14-42c0-8324-9acee056f8a7\") " pod="openshift-route-controller-manager/route-controller-manager-66f698ffcb-hdfc8" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.883875 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhdcw\" (UniqueName: \"kubernetes.io/projected/f7ebccc5-7e14-42c0-8324-9acee056f8a7-kube-api-access-bhdcw\") pod \"route-controller-manager-66f698ffcb-hdfc8\" (UID: \"f7ebccc5-7e14-42c0-8324-9acee056f8a7\") " pod="openshift-route-controller-manager/route-controller-manager-66f698ffcb-hdfc8" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.883918 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7ebccc5-7e14-42c0-8324-9acee056f8a7-config\") pod \"route-controller-manager-66f698ffcb-hdfc8\" (UID: \"f7ebccc5-7e14-42c0-8324-9acee056f8a7\") " pod="openshift-route-controller-manager/route-controller-manager-66f698ffcb-hdfc8" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.883939 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f7ebccc5-7e14-42c0-8324-9acee056f8a7-client-ca\") pod \"route-controller-manager-66f698ffcb-hdfc8\" (UID: \"f7ebccc5-7e14-42c0-8324-9acee056f8a7\") " pod="openshift-route-controller-manager/route-controller-manager-66f698ffcb-hdfc8" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.985323 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7ebccc5-7e14-42c0-8324-9acee056f8a7-serving-cert\") pod \"route-controller-manager-66f698ffcb-hdfc8\" (UID: \"f7ebccc5-7e14-42c0-8324-9acee056f8a7\") " pod="openshift-route-controller-manager/route-controller-manager-66f698ffcb-hdfc8" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.985373 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhdcw\" (UniqueName: \"kubernetes.io/projected/f7ebccc5-7e14-42c0-8324-9acee056f8a7-kube-api-access-bhdcw\") pod \"route-controller-manager-66f698ffcb-hdfc8\" (UID: \"f7ebccc5-7e14-42c0-8324-9acee056f8a7\") " pod="openshift-route-controller-manager/route-controller-manager-66f698ffcb-hdfc8" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.985413 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7ebccc5-7e14-42c0-8324-9acee056f8a7-config\") pod \"route-controller-manager-66f698ffcb-hdfc8\" (UID: \"f7ebccc5-7e14-42c0-8324-9acee056f8a7\") " pod="openshift-route-controller-manager/route-controller-manager-66f698ffcb-hdfc8" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.985434 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f7ebccc5-7e14-42c0-8324-9acee056f8a7-client-ca\") pod \"route-controller-manager-66f698ffcb-hdfc8\" (UID: \"f7ebccc5-7e14-42c0-8324-9acee056f8a7\") " pod="openshift-route-controller-manager/route-controller-manager-66f698ffcb-hdfc8" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.986328 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f7ebccc5-7e14-42c0-8324-9acee056f8a7-client-ca\") pod \"route-controller-manager-66f698ffcb-hdfc8\" (UID: \"f7ebccc5-7e14-42c0-8324-9acee056f8a7\") " pod="openshift-route-controller-manager/route-controller-manager-66f698ffcb-hdfc8" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.986931 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7ebccc5-7e14-42c0-8324-9acee056f8a7-config\") pod \"route-controller-manager-66f698ffcb-hdfc8\" (UID: \"f7ebccc5-7e14-42c0-8324-9acee056f8a7\") " pod="openshift-route-controller-manager/route-controller-manager-66f698ffcb-hdfc8" Jan 27 15:12:16 crc kubenswrapper[4772]: I0127 15:12:16.989583 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7ebccc5-7e14-42c0-8324-9acee056f8a7-serving-cert\") pod \"route-controller-manager-66f698ffcb-hdfc8\" (UID: \"f7ebccc5-7e14-42c0-8324-9acee056f8a7\") " pod="openshift-route-controller-manager/route-controller-manager-66f698ffcb-hdfc8" Jan 27 15:12:17 crc kubenswrapper[4772]: I0127 15:12:17.007796 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhdcw\" (UniqueName: \"kubernetes.io/projected/f7ebccc5-7e14-42c0-8324-9acee056f8a7-kube-api-access-bhdcw\") pod \"route-controller-manager-66f698ffcb-hdfc8\" (UID: \"f7ebccc5-7e14-42c0-8324-9acee056f8a7\") " pod="openshift-route-controller-manager/route-controller-manager-66f698ffcb-hdfc8" Jan 27 15:12:17 crc kubenswrapper[4772]: I0127 15:12:17.068886 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66f698ffcb-hdfc8" Jan 27 15:12:17 crc kubenswrapper[4772]: I0127 15:12:17.456711 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66f698ffcb-hdfc8"] Jan 27 15:12:17 crc kubenswrapper[4772]: W0127 15:12:17.461786 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7ebccc5_7e14_42c0_8324_9acee056f8a7.slice/crio-1127780f119b177be5380b18f8676f014ff7a27500f8d274643e4193e7cfa2f7 WatchSource:0}: Error finding container 1127780f119b177be5380b18f8676f014ff7a27500f8d274643e4193e7cfa2f7: Status 404 returned error can't find the container with id 1127780f119b177be5380b18f8676f014ff7a27500f8d274643e4193e7cfa2f7 Jan 27 15:12:18 crc kubenswrapper[4772]: I0127 15:12:18.059580 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66f698ffcb-hdfc8" event={"ID":"f7ebccc5-7e14-42c0-8324-9acee056f8a7","Type":"ContainerStarted","Data":"033b0040a797aa93a999c71c60e265daf0395c42e378608e62e4905533abc35e"} Jan 27 15:12:18 crc kubenswrapper[4772]: I0127 15:12:18.059991 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-66f698ffcb-hdfc8" Jan 27 15:12:18 crc kubenswrapper[4772]: I0127 15:12:18.060035 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66f698ffcb-hdfc8" event={"ID":"f7ebccc5-7e14-42c0-8324-9acee056f8a7","Type":"ContainerStarted","Data":"1127780f119b177be5380b18f8676f014ff7a27500f8d274643e4193e7cfa2f7"} Jan 27 15:12:18 crc kubenswrapper[4772]: I0127 15:12:18.081447 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-66f698ffcb-hdfc8" podStartSLOduration=3.0814304359999998 podStartE2EDuration="3.081430436s" podCreationTimestamp="2026-01-27 15:12:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:12:18.080353284 +0000 UTC m=+324.060962412" watchObservedRunningTime="2026-01-27 15:12:18.081430436 +0000 UTC m=+324.062039534" Jan 27 15:12:18 crc kubenswrapper[4772]: I0127 15:12:18.159480 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-66f698ffcb-hdfc8" Jan 27 15:12:29 crc kubenswrapper[4772]: I0127 15:12:29.949072 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-2dw59" Jan 27 15:12:30 crc kubenswrapper[4772]: I0127 15:12:30.001035 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-crlcr"] Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.049296 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" podUID="877de785-bc18-4c1c-970a-1e6533539467" containerName="registry" containerID="cri-o://228e6fd0668bf433c1f6aa09021f79564dfe5e7bb750301de0ab0cbfce9f1ef2" gracePeriod=30 Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.340084 4772 generic.go:334] "Generic (PLEG): container finished" podID="877de785-bc18-4c1c-970a-1e6533539467" containerID="228e6fd0668bf433c1f6aa09021f79564dfe5e7bb750301de0ab0cbfce9f1ef2" exitCode=0 Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.340193 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" event={"ID":"877de785-bc18-4c1c-970a-1e6533539467","Type":"ContainerDied","Data":"228e6fd0668bf433c1f6aa09021f79564dfe5e7bb750301de0ab0cbfce9f1ef2"} Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.402088 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.567125 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/877de785-bc18-4c1c-970a-1e6533539467-trusted-ca\") pod \"877de785-bc18-4c1c-970a-1e6533539467\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.567185 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5h66t\" (UniqueName: \"kubernetes.io/projected/877de785-bc18-4c1c-970a-1e6533539467-kube-api-access-5h66t\") pod \"877de785-bc18-4c1c-970a-1e6533539467\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.567206 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/877de785-bc18-4c1c-970a-1e6533539467-bound-sa-token\") pod \"877de785-bc18-4c1c-970a-1e6533539467\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.567246 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/877de785-bc18-4c1c-970a-1e6533539467-ca-trust-extracted\") pod \"877de785-bc18-4c1c-970a-1e6533539467\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.567322 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/877de785-bc18-4c1c-970a-1e6533539467-registry-certificates\") pod \"877de785-bc18-4c1c-970a-1e6533539467\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.567338 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/877de785-bc18-4c1c-970a-1e6533539467-registry-tls\") pod \"877de785-bc18-4c1c-970a-1e6533539467\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.567539 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"877de785-bc18-4c1c-970a-1e6533539467\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.567578 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/877de785-bc18-4c1c-970a-1e6533539467-installation-pull-secrets\") pod \"877de785-bc18-4c1c-970a-1e6533539467\" (UID: \"877de785-bc18-4c1c-970a-1e6533539467\") " Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.568322 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/877de785-bc18-4c1c-970a-1e6533539467-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "877de785-bc18-4c1c-970a-1e6533539467" (UID: "877de785-bc18-4c1c-970a-1e6533539467"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.568536 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/877de785-bc18-4c1c-970a-1e6533539467-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "877de785-bc18-4c1c-970a-1e6533539467" (UID: "877de785-bc18-4c1c-970a-1e6533539467"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.572774 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/877de785-bc18-4c1c-970a-1e6533539467-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "877de785-bc18-4c1c-970a-1e6533539467" (UID: "877de785-bc18-4c1c-970a-1e6533539467"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.572822 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/877de785-bc18-4c1c-970a-1e6533539467-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "877de785-bc18-4c1c-970a-1e6533539467" (UID: "877de785-bc18-4c1c-970a-1e6533539467"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.572970 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/877de785-bc18-4c1c-970a-1e6533539467-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "877de785-bc18-4c1c-970a-1e6533539467" (UID: "877de785-bc18-4c1c-970a-1e6533539467"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.576460 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "877de785-bc18-4c1c-970a-1e6533539467" (UID: "877de785-bc18-4c1c-970a-1e6533539467"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.577387 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/877de785-bc18-4c1c-970a-1e6533539467-kube-api-access-5h66t" (OuterVolumeSpecName: "kube-api-access-5h66t") pod "877de785-bc18-4c1c-970a-1e6533539467" (UID: "877de785-bc18-4c1c-970a-1e6533539467"). InnerVolumeSpecName "kube-api-access-5h66t". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.582485 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/877de785-bc18-4c1c-970a-1e6533539467-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "877de785-bc18-4c1c-970a-1e6533539467" (UID: "877de785-bc18-4c1c-970a-1e6533539467"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.669672 4772 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/877de785-bc18-4c1c-970a-1e6533539467-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.669726 4772 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/877de785-bc18-4c1c-970a-1e6533539467-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.669751 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5h66t\" (UniqueName: \"kubernetes.io/projected/877de785-bc18-4c1c-970a-1e6533539467-kube-api-access-5h66t\") on node \"crc\" DevicePath \"\"" Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.669769 4772 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/877de785-bc18-4c1c-970a-1e6533539467-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.669786 4772 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/877de785-bc18-4c1c-970a-1e6533539467-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.669803 4772 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/877de785-bc18-4c1c-970a-1e6533539467-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 27 15:12:55 crc kubenswrapper[4772]: I0127 15:12:55.669820 4772 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/877de785-bc18-4c1c-970a-1e6533539467-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 27 15:12:56 crc kubenswrapper[4772]: I0127 15:12:56.347313 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" event={"ID":"877de785-bc18-4c1c-970a-1e6533539467","Type":"ContainerDied","Data":"a3585a039b9cbf60a67ac7ced2eaf947fce2a88abe7705503eb446ef5ad9fc74"} Jan 27 15:12:56 crc kubenswrapper[4772]: I0127 15:12:56.347362 4772 scope.go:117] "RemoveContainer" containerID="228e6fd0668bf433c1f6aa09021f79564dfe5e7bb750301de0ab0cbfce9f1ef2" Jan 27 15:12:56 crc kubenswrapper[4772]: I0127 15:12:56.347398 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-crlcr" Jan 27 15:12:56 crc kubenswrapper[4772]: I0127 15:12:56.373564 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-crlcr"] Jan 27 15:12:56 crc kubenswrapper[4772]: I0127 15:12:56.378382 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-crlcr"] Jan 27 15:12:56 crc kubenswrapper[4772]: I0127 15:12:56.673033 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="877de785-bc18-4c1c-970a-1e6533539467" path="/var/lib/kubelet/pods/877de785-bc18-4c1c-970a-1e6533539467/volumes" Jan 27 15:13:12 crc kubenswrapper[4772]: I0127 15:13:12.058747 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:13:12 crc kubenswrapper[4772]: I0127 15:13:12.059107 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:13:42 crc kubenswrapper[4772]: I0127 15:13:42.058850 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:13:42 crc kubenswrapper[4772]: I0127 15:13:42.059624 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:14:12 crc kubenswrapper[4772]: I0127 15:14:12.058216 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:14:12 crc kubenswrapper[4772]: I0127 15:14:12.058944 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:14:12 crc kubenswrapper[4772]: I0127 15:14:12.058993 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:14:12 crc kubenswrapper[4772]: I0127 15:14:12.059798 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8e72007caa5160368d39dc40b9c7f95a9beba3bef9f9e290eac1d112ef6eeb10"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 15:14:12 crc kubenswrapper[4772]: I0127 15:14:12.059886 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://8e72007caa5160368d39dc40b9c7f95a9beba3bef9f9e290eac1d112ef6eeb10" gracePeriod=600 Jan 27 15:14:12 crc kubenswrapper[4772]: I0127 15:14:12.757429 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="8e72007caa5160368d39dc40b9c7f95a9beba3bef9f9e290eac1d112ef6eeb10" exitCode=0 Jan 27 15:14:12 crc kubenswrapper[4772]: I0127 15:14:12.757481 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"8e72007caa5160368d39dc40b9c7f95a9beba3bef9f9e290eac1d112ef6eeb10"} Jan 27 15:14:12 crc kubenswrapper[4772]: I0127 15:14:12.757886 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"32659ec7f069b0827082828bb6142c20199821498a042e5f263706f6e96e9462"} Jan 27 15:14:12 crc kubenswrapper[4772]: I0127 15:14:12.757935 4772 scope.go:117] "RemoveContainer" containerID="0d95f231ee1013dc5475acac704b796538ef0050cd94e435a3382bd12b7cbf19" Jan 27 15:15:00 crc kubenswrapper[4772]: I0127 15:15:00.158094 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492115-hb89g"] Jan 27 15:15:00 crc kubenswrapper[4772]: E0127 15:15:00.158822 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="877de785-bc18-4c1c-970a-1e6533539467" containerName="registry" Jan 27 15:15:00 crc kubenswrapper[4772]: I0127 15:15:00.158835 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="877de785-bc18-4c1c-970a-1e6533539467" containerName="registry" Jan 27 15:15:00 crc kubenswrapper[4772]: I0127 15:15:00.158929 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="877de785-bc18-4c1c-970a-1e6533539467" containerName="registry" Jan 27 15:15:00 crc kubenswrapper[4772]: I0127 15:15:00.159312 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492115-hb89g" Jan 27 15:15:00 crc kubenswrapper[4772]: I0127 15:15:00.161382 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 27 15:15:00 crc kubenswrapper[4772]: I0127 15:15:00.163274 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 27 15:15:00 crc kubenswrapper[4772]: I0127 15:15:00.174078 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492115-hb89g"] Jan 27 15:15:00 crc kubenswrapper[4772]: I0127 15:15:00.326589 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1-config-volume\") pod \"collect-profiles-29492115-hb89g\" (UID: \"616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492115-hb89g" Jan 27 15:15:00 crc kubenswrapper[4772]: I0127 15:15:00.326661 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r96xf\" (UniqueName: \"kubernetes.io/projected/616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1-kube-api-access-r96xf\") pod \"collect-profiles-29492115-hb89g\" (UID: \"616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492115-hb89g" Jan 27 15:15:00 crc kubenswrapper[4772]: I0127 15:15:00.326703 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1-secret-volume\") pod \"collect-profiles-29492115-hb89g\" (UID: \"616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492115-hb89g" Jan 27 15:15:00 crc kubenswrapper[4772]: I0127 15:15:00.428224 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1-secret-volume\") pod \"collect-profiles-29492115-hb89g\" (UID: \"616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492115-hb89g" Jan 27 15:15:00 crc kubenswrapper[4772]: I0127 15:15:00.428301 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1-config-volume\") pod \"collect-profiles-29492115-hb89g\" (UID: \"616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492115-hb89g" Jan 27 15:15:00 crc kubenswrapper[4772]: I0127 15:15:00.428353 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r96xf\" (UniqueName: \"kubernetes.io/projected/616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1-kube-api-access-r96xf\") pod \"collect-profiles-29492115-hb89g\" (UID: \"616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492115-hb89g" Jan 27 15:15:00 crc kubenswrapper[4772]: I0127 15:15:00.429433 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1-config-volume\") pod \"collect-profiles-29492115-hb89g\" (UID: \"616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492115-hb89g" Jan 27 15:15:00 crc kubenswrapper[4772]: I0127 15:15:00.434676 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1-secret-volume\") pod \"collect-profiles-29492115-hb89g\" (UID: \"616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492115-hb89g" Jan 27 15:15:00 crc kubenswrapper[4772]: I0127 15:15:00.449154 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r96xf\" (UniqueName: \"kubernetes.io/projected/616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1-kube-api-access-r96xf\") pod \"collect-profiles-29492115-hb89g\" (UID: \"616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492115-hb89g" Jan 27 15:15:00 crc kubenswrapper[4772]: I0127 15:15:00.480968 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492115-hb89g" Jan 27 15:15:00 crc kubenswrapper[4772]: I0127 15:15:00.658689 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492115-hb89g"] Jan 27 15:15:01 crc kubenswrapper[4772]: I0127 15:15:01.037121 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492115-hb89g" event={"ID":"616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1","Type":"ContainerStarted","Data":"8931f0cc38dd8c453e687a0b65ac6a9c2d9a0265440b30f5480c6ac0483f9860"} Jan 27 15:15:01 crc kubenswrapper[4772]: I0127 15:15:01.037471 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492115-hb89g" event={"ID":"616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1","Type":"ContainerStarted","Data":"d2a67b192739d3b926b87f8d49b2048266ed3826d7d2d9ed2f68c533ca1f4d5b"} Jan 27 15:15:02 crc kubenswrapper[4772]: I0127 15:15:02.044566 4772 generic.go:334] "Generic (PLEG): container finished" podID="616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1" containerID="8931f0cc38dd8c453e687a0b65ac6a9c2d9a0265440b30f5480c6ac0483f9860" exitCode=0 Jan 27 15:15:02 crc kubenswrapper[4772]: I0127 15:15:02.044609 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492115-hb89g" event={"ID":"616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1","Type":"ContainerDied","Data":"8931f0cc38dd8c453e687a0b65ac6a9c2d9a0265440b30f5480c6ac0483f9860"} Jan 27 15:15:03 crc kubenswrapper[4772]: I0127 15:15:03.261037 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492115-hb89g" Jan 27 15:15:03 crc kubenswrapper[4772]: I0127 15:15:03.363517 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1-config-volume\") pod \"616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1\" (UID: \"616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1\") " Jan 27 15:15:03 crc kubenswrapper[4772]: I0127 15:15:03.363598 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1-secret-volume\") pod \"616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1\" (UID: \"616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1\") " Jan 27 15:15:03 crc kubenswrapper[4772]: I0127 15:15:03.363719 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r96xf\" (UniqueName: \"kubernetes.io/projected/616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1-kube-api-access-r96xf\") pod \"616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1\" (UID: \"616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1\") " Jan 27 15:15:03 crc kubenswrapper[4772]: I0127 15:15:03.364228 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1-config-volume" (OuterVolumeSpecName: "config-volume") pod "616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1" (UID: "616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:15:03 crc kubenswrapper[4772]: I0127 15:15:03.368978 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1" (UID: "616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:15:03 crc kubenswrapper[4772]: I0127 15:15:03.368976 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1-kube-api-access-r96xf" (OuterVolumeSpecName: "kube-api-access-r96xf") pod "616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1" (UID: "616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1"). InnerVolumeSpecName "kube-api-access-r96xf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:15:03 crc kubenswrapper[4772]: I0127 15:15:03.465476 4772 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1-config-volume\") on node \"crc\" DevicePath \"\"" Jan 27 15:15:03 crc kubenswrapper[4772]: I0127 15:15:03.465517 4772 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 27 15:15:03 crc kubenswrapper[4772]: I0127 15:15:03.465529 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r96xf\" (UniqueName: \"kubernetes.io/projected/616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1-kube-api-access-r96xf\") on node \"crc\" DevicePath \"\"" Jan 27 15:15:04 crc kubenswrapper[4772]: I0127 15:15:04.059783 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492115-hb89g" event={"ID":"616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1","Type":"ContainerDied","Data":"d2a67b192739d3b926b87f8d49b2048266ed3826d7d2d9ed2f68c533ca1f4d5b"} Jan 27 15:15:04 crc kubenswrapper[4772]: I0127 15:15:04.059835 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2a67b192739d3b926b87f8d49b2048266ed3826d7d2d9ed2f68c533ca1f4d5b" Jan 27 15:15:04 crc kubenswrapper[4772]: I0127 15:15:04.059902 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492115-hb89g" Jan 27 15:15:54 crc kubenswrapper[4772]: I0127 15:15:54.857366 4772 scope.go:117] "RemoveContainer" containerID="437c578755bfcacf0145c1b3dcede3b1938b4e11e6ad9c7db9d8ac6a8b6df37e" Jan 27 15:16:12 crc kubenswrapper[4772]: I0127 15:16:12.058880 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:16:12 crc kubenswrapper[4772]: I0127 15:16:12.059904 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:16:42 crc kubenswrapper[4772]: I0127 15:16:42.059212 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:16:42 crc kubenswrapper[4772]: I0127 15:16:42.060349 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:17:12 crc kubenswrapper[4772]: I0127 15:17:12.058873 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:17:12 crc kubenswrapper[4772]: I0127 15:17:12.059454 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:17:12 crc kubenswrapper[4772]: I0127 15:17:12.059500 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:17:12 crc kubenswrapper[4772]: I0127 15:17:12.060069 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"32659ec7f069b0827082828bb6142c20199821498a042e5f263706f6e96e9462"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 15:17:12 crc kubenswrapper[4772]: I0127 15:17:12.060120 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://32659ec7f069b0827082828bb6142c20199821498a042e5f263706f6e96e9462" gracePeriod=600 Jan 27 15:17:13 crc kubenswrapper[4772]: I0127 15:17:13.122923 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="32659ec7f069b0827082828bb6142c20199821498a042e5f263706f6e96e9462" exitCode=0 Jan 27 15:17:13 crc kubenswrapper[4772]: I0127 15:17:13.122991 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"32659ec7f069b0827082828bb6142c20199821498a042e5f263706f6e96e9462"} Jan 27 15:17:13 crc kubenswrapper[4772]: I0127 15:17:13.123291 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"60c798dfb542a875b90e857bf6f54352abce005f4bc0c5fd246c1b5d0903e3f3"} Jan 27 15:17:13 crc kubenswrapper[4772]: I0127 15:17:13.123326 4772 scope.go:117] "RemoveContainer" containerID="8e72007caa5160368d39dc40b9c7f95a9beba3bef9f9e290eac1d112ef6eeb10" Jan 27 15:18:48 crc kubenswrapper[4772]: I0127 15:18:48.865063 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-n2khk"] Jan 27 15:18:48 crc kubenswrapper[4772]: I0127 15:18:48.867021 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="ovn-controller" containerID="cri-o://e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e" gracePeriod=30 Jan 27 15:18:48 crc kubenswrapper[4772]: I0127 15:18:48.867108 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="nbdb" containerID="cri-o://0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b" gracePeriod=30 Jan 27 15:18:48 crc kubenswrapper[4772]: I0127 15:18:48.867404 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="sbdb" containerID="cri-o://0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae" gracePeriod=30 Jan 27 15:18:48 crc kubenswrapper[4772]: I0127 15:18:48.867410 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="northd" containerID="cri-o://5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4" gracePeriod=30 Jan 27 15:18:48 crc kubenswrapper[4772]: I0127 15:18:48.867504 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5" gracePeriod=30 Jan 27 15:18:48 crc kubenswrapper[4772]: I0127 15:18:48.867475 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="kube-rbac-proxy-node" containerID="cri-o://321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9" gracePeriod=30 Jan 27 15:18:48 crc kubenswrapper[4772]: I0127 15:18:48.868362 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="ovn-acl-logging" containerID="cri-o://3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854" gracePeriod=30 Jan 27 15:18:48 crc kubenswrapper[4772]: I0127 15:18:48.914255 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="ovnkube-controller" containerID="cri-o://7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823" gracePeriod=30 Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.228342 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n2khk_736264c8-cd18-479a-88ba-e1ec15dbfdae/ovnkube-controller/3.log" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.232084 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n2khk_736264c8-cd18-479a-88ba-e1ec15dbfdae/ovn-acl-logging/0.log" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.232747 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n2khk_736264c8-cd18-479a-88ba-e1ec15dbfdae/ovn-controller/0.log" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.233240 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.285344 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qml47"] Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.285661 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="kubecfg-setup" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.285690 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="kubecfg-setup" Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.285704 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="ovnkube-controller" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.285710 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="ovnkube-controller" Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.285720 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="nbdb" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.285728 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="nbdb" Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.285738 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="ovn-controller" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.285743 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="ovn-controller" Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.285754 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="kube-rbac-proxy-ovn-metrics" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.285761 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="kube-rbac-proxy-ovn-metrics" Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.285773 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="sbdb" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.285783 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="sbdb" Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.285793 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="kube-rbac-proxy-node" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.285800 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="kube-rbac-proxy-node" Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.285807 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="ovnkube-controller" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.285812 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="ovnkube-controller" Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.285819 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="ovnkube-controller" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.285826 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="ovnkube-controller" Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.285834 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="northd" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.285841 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="northd" Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.285855 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="ovnkube-controller" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.285862 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="ovnkube-controller" Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.285871 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1" containerName="collect-profiles" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.285878 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1" containerName="collect-profiles" Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.285888 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="ovn-acl-logging" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.285896 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="ovn-acl-logging" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.286015 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="ovnkube-controller" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.286030 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="ovnkube-controller" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.286039 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="kube-rbac-proxy-ovn-metrics" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.286048 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="ovn-acl-logging" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.286057 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1" containerName="collect-profiles" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.286065 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="kube-rbac-proxy-node" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.286074 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="northd" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.286081 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="sbdb" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.286091 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="nbdb" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.286099 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="ovnkube-controller" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.286106 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="ovn-controller" Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.286307 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="ovnkube-controller" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.286320 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="ovnkube-controller" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.286422 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="ovnkube-controller" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.286433 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerName="ovnkube-controller" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.289430 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.339515 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-run-ovn\") pod \"736264c8-cd18-479a-88ba-e1ec15dbfdae\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.339741 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-var-lib-cni-networks-ovn-kubernetes\") pod \"736264c8-cd18-479a-88ba-e1ec15dbfdae\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.339823 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/736264c8-cd18-479a-88ba-e1ec15dbfdae-env-overrides\") pod \"736264c8-cd18-479a-88ba-e1ec15dbfdae\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.339944 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-log-socket\") pod \"736264c8-cd18-479a-88ba-e1ec15dbfdae\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340031 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-run-netns\") pod \"736264c8-cd18-479a-88ba-e1ec15dbfdae\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340107 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-slash\") pod \"736264c8-cd18-479a-88ba-e1ec15dbfdae\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340211 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-cni-netd\") pod \"736264c8-cd18-479a-88ba-e1ec15dbfdae\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340314 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dt6g\" (UniqueName: \"kubernetes.io/projected/736264c8-cd18-479a-88ba-e1ec15dbfdae-kube-api-access-2dt6g\") pod \"736264c8-cd18-479a-88ba-e1ec15dbfdae\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340408 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-etc-openvswitch\") pod \"736264c8-cd18-479a-88ba-e1ec15dbfdae\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.339642 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "736264c8-cd18-479a-88ba-e1ec15dbfdae" (UID: "736264c8-cd18-479a-88ba-e1ec15dbfdae"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.339786 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "736264c8-cd18-479a-88ba-e1ec15dbfdae" (UID: "736264c8-cd18-479a-88ba-e1ec15dbfdae"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340117 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "736264c8-cd18-479a-88ba-e1ec15dbfdae" (UID: "736264c8-cd18-479a-88ba-e1ec15dbfdae"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340108 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-log-socket" (OuterVolumeSpecName: "log-socket") pod "736264c8-cd18-479a-88ba-e1ec15dbfdae" (UID: "736264c8-cd18-479a-88ba-e1ec15dbfdae"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340134 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-slash" (OuterVolumeSpecName: "host-slash") pod "736264c8-cd18-479a-88ba-e1ec15dbfdae" (UID: "736264c8-cd18-479a-88ba-e1ec15dbfdae"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340310 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "736264c8-cd18-479a-88ba-e1ec15dbfdae" (UID: "736264c8-cd18-479a-88ba-e1ec15dbfdae"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340434 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/736264c8-cd18-479a-88ba-e1ec15dbfdae-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "736264c8-cd18-479a-88ba-e1ec15dbfdae" (UID: "736264c8-cd18-479a-88ba-e1ec15dbfdae"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340471 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "736264c8-cd18-479a-88ba-e1ec15dbfdae" (UID: "736264c8-cd18-479a-88ba-e1ec15dbfdae"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340484 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-run-systemd\") pod \"736264c8-cd18-479a-88ba-e1ec15dbfdae\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340561 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-run-openvswitch\") pod \"736264c8-cd18-479a-88ba-e1ec15dbfdae\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340579 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-cni-bin\") pod \"736264c8-cd18-479a-88ba-e1ec15dbfdae\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340602 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/736264c8-cd18-479a-88ba-e1ec15dbfdae-ovnkube-config\") pod \"736264c8-cd18-479a-88ba-e1ec15dbfdae\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340619 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-node-log\") pod \"736264c8-cd18-479a-88ba-e1ec15dbfdae\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340636 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-var-lib-openvswitch\") pod \"736264c8-cd18-479a-88ba-e1ec15dbfdae\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340655 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-kubelet\") pod \"736264c8-cd18-479a-88ba-e1ec15dbfdae\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340674 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/736264c8-cd18-479a-88ba-e1ec15dbfdae-ovnkube-script-lib\") pod \"736264c8-cd18-479a-88ba-e1ec15dbfdae\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340690 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/736264c8-cd18-479a-88ba-e1ec15dbfdae-ovn-node-metrics-cert\") pod \"736264c8-cd18-479a-88ba-e1ec15dbfdae\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340720 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-systemd-units\") pod \"736264c8-cd18-479a-88ba-e1ec15dbfdae\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340735 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-run-ovn-kubernetes\") pod \"736264c8-cd18-479a-88ba-e1ec15dbfdae\" (UID: \"736264c8-cd18-479a-88ba-e1ec15dbfdae\") " Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340837 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ff07e8b2-325b-4f96-b685-f2068052a960-ovn-node-metrics-cert\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340858 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-run-openvswitch\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340878 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ff07e8b2-325b-4f96-b685-f2068052a960-ovnkube-config\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340895 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-host-run-netns\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340911 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-run-ovn\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340927 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340950 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-var-lib-openvswitch\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340970 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ff07e8b2-325b-4f96-b685-f2068052a960-env-overrides\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.340995 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b7p2\" (UniqueName: \"kubernetes.io/projected/ff07e8b2-325b-4f96-b685-f2068052a960-kube-api-access-8b7p2\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341021 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-log-socket\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341047 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-host-slash\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341071 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-host-cni-bin\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341096 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-host-kubelet\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341122 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-systemd-units\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341145 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-host-cni-netd\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341193 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-node-log\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341216 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-etc-openvswitch\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341236 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ff07e8b2-325b-4f96-b685-f2068052a960-ovnkube-script-lib\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341258 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-run-systemd\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341275 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-host-run-ovn-kubernetes\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341305 4772 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341314 4772 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341325 4772 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/736264c8-cd18-479a-88ba-e1ec15dbfdae-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341334 4772 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-log-socket\") on node \"crc\" DevicePath \"\"" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341342 4772 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-run-netns\") on node \"crc\" DevicePath \"\"" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341349 4772 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-slash\") on node \"crc\" DevicePath \"\"" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341357 4772 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-cni-netd\") on node \"crc\" DevicePath \"\"" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341365 4772 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341401 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "736264c8-cd18-479a-88ba-e1ec15dbfdae" (UID: "736264c8-cd18-479a-88ba-e1ec15dbfdae"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341419 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "736264c8-cd18-479a-88ba-e1ec15dbfdae" (UID: "736264c8-cd18-479a-88ba-e1ec15dbfdae"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341651 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-node-log" (OuterVolumeSpecName: "node-log") pod "736264c8-cd18-479a-88ba-e1ec15dbfdae" (UID: "736264c8-cd18-479a-88ba-e1ec15dbfdae"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341711 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "736264c8-cd18-479a-88ba-e1ec15dbfdae" (UID: "736264c8-cd18-479a-88ba-e1ec15dbfdae"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341741 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "736264c8-cd18-479a-88ba-e1ec15dbfdae" (UID: "736264c8-cd18-479a-88ba-e1ec15dbfdae"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.341793 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/736264c8-cd18-479a-88ba-e1ec15dbfdae-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "736264c8-cd18-479a-88ba-e1ec15dbfdae" (UID: "736264c8-cd18-479a-88ba-e1ec15dbfdae"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.342206 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/736264c8-cd18-479a-88ba-e1ec15dbfdae-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "736264c8-cd18-479a-88ba-e1ec15dbfdae" (UID: "736264c8-cd18-479a-88ba-e1ec15dbfdae"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.342291 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "736264c8-cd18-479a-88ba-e1ec15dbfdae" (UID: "736264c8-cd18-479a-88ba-e1ec15dbfdae"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.342301 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "736264c8-cd18-479a-88ba-e1ec15dbfdae" (UID: "736264c8-cd18-479a-88ba-e1ec15dbfdae"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.348445 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/736264c8-cd18-479a-88ba-e1ec15dbfdae-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "736264c8-cd18-479a-88ba-e1ec15dbfdae" (UID: "736264c8-cd18-479a-88ba-e1ec15dbfdae"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.348561 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/736264c8-cd18-479a-88ba-e1ec15dbfdae-kube-api-access-2dt6g" (OuterVolumeSpecName: "kube-api-access-2dt6g") pod "736264c8-cd18-479a-88ba-e1ec15dbfdae" (UID: "736264c8-cd18-479a-88ba-e1ec15dbfdae"). InnerVolumeSpecName "kube-api-access-2dt6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.359339 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "736264c8-cd18-479a-88ba-e1ec15dbfdae" (UID: "736264c8-cd18-479a-88ba-e1ec15dbfdae"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.442480 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-host-cni-bin\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.442744 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-host-kubelet\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.442771 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-systemd-units\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.442627 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-host-cni-bin\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.442822 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-host-cni-netd\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.442788 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-host-cni-netd\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.442857 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-systemd-units\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.442901 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-host-kubelet\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.442922 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-node-log\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.442990 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-node-log\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443041 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-etc-openvswitch\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443098 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ff07e8b2-325b-4f96-b685-f2068052a960-ovnkube-script-lib\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443111 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-etc-openvswitch\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443159 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-run-systemd\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443260 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-run-systemd\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443271 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-host-run-ovn-kubernetes\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443326 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ff07e8b2-325b-4f96-b685-f2068052a960-ovn-node-metrics-cert\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443376 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-host-run-ovn-kubernetes\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443382 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-run-openvswitch\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443437 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-run-openvswitch\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443461 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ff07e8b2-325b-4f96-b685-f2068052a960-ovnkube-config\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443485 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-host-run-netns\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443506 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-run-ovn\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443530 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443571 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-var-lib-openvswitch\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443595 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ff07e8b2-325b-4f96-b685-f2068052a960-env-overrides\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443623 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b7p2\" (UniqueName: \"kubernetes.io/projected/ff07e8b2-325b-4f96-b685-f2068052a960-kube-api-access-8b7p2\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443633 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-host-run-netns\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443656 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-log-socket\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443684 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-host-slash\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443742 4772 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-kubelet\") on node \"crc\" DevicePath \"\"" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443772 4772 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/736264c8-cd18-479a-88ba-e1ec15dbfdae-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443785 4772 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/736264c8-cd18-479a-88ba-e1ec15dbfdae-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443799 4772 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-systemd-units\") on node \"crc\" DevicePath \"\"" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443814 4772 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443827 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dt6g\" (UniqueName: \"kubernetes.io/projected/736264c8-cd18-479a-88ba-e1ec15dbfdae-kube-api-access-2dt6g\") on node \"crc\" DevicePath \"\"" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443837 4772 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-run-systemd\") on node \"crc\" DevicePath \"\"" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443849 4772 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-run-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443861 4772 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-host-cni-bin\") on node \"crc\" DevicePath \"\"" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443872 4772 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/736264c8-cd18-479a-88ba-e1ec15dbfdae-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443882 4772 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-node-log\") on node \"crc\" DevicePath \"\"" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443893 4772 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/736264c8-cd18-479a-88ba-e1ec15dbfdae-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443926 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-host-slash\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443958 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-run-ovn\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.443987 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.444016 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-var-lib-openvswitch\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.444886 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ff07e8b2-325b-4f96-b685-f2068052a960-log-socket\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.444908 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ff07e8b2-325b-4f96-b685-f2068052a960-ovnkube-script-lib\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.445246 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ff07e8b2-325b-4f96-b685-f2068052a960-env-overrides\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.445422 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ff07e8b2-325b-4f96-b685-f2068052a960-ovnkube-config\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.448340 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ff07e8b2-325b-4f96-b685-f2068052a960-ovn-node-metrics-cert\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.461644 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b7p2\" (UniqueName: \"kubernetes.io/projected/ff07e8b2-325b-4f96-b685-f2068052a960-kube-api-access-8b7p2\") pod \"ovnkube-node-qml47\" (UID: \"ff07e8b2-325b-4f96-b685-f2068052a960\") " pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.605891 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.707541 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n2khk_736264c8-cd18-479a-88ba-e1ec15dbfdae/ovnkube-controller/3.log" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.711459 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n2khk_736264c8-cd18-479a-88ba-e1ec15dbfdae/ovn-acl-logging/0.log" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.711978 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-n2khk_736264c8-cd18-479a-88ba-e1ec15dbfdae/ovn-controller/0.log" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712389 4772 generic.go:334] "Generic (PLEG): container finished" podID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerID="7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823" exitCode=0 Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712436 4772 generic.go:334] "Generic (PLEG): container finished" podID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerID="0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae" exitCode=0 Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712455 4772 generic.go:334] "Generic (PLEG): container finished" podID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerID="0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b" exitCode=0 Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712469 4772 generic.go:334] "Generic (PLEG): container finished" podID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerID="5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4" exitCode=0 Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712486 4772 generic.go:334] "Generic (PLEG): container finished" podID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerID="45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5" exitCode=0 Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712501 4772 generic.go:334] "Generic (PLEG): container finished" podID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerID="321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9" exitCode=0 Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712520 4772 generic.go:334] "Generic (PLEG): container finished" podID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerID="3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854" exitCode=143 Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712535 4772 generic.go:334] "Generic (PLEG): container finished" podID="736264c8-cd18-479a-88ba-e1ec15dbfdae" containerID="e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e" exitCode=143 Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712556 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712621 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerDied","Data":"7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712679 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerDied","Data":"0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712708 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerDied","Data":"0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712730 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerDied","Data":"5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712769 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerDied","Data":"45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712789 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerDied","Data":"321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712812 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712831 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712843 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712856 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712868 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712882 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712895 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712907 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712920 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712937 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerDied","Data":"3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712954 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712970 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712982 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.712995 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713007 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713018 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713031 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713042 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713054 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713066 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713082 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerDied","Data":"e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713099 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713112 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713123 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713135 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713146 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713157 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713204 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713217 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713228 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713239 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713258 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n2khk" event={"ID":"736264c8-cd18-479a-88ba-e1ec15dbfdae","Type":"ContainerDied","Data":"b6a209d8fc4e180971a6f92a0f3c7493472a2095b6c5303a9b0ce0f4e62056a9"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713276 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713289 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713301 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713313 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713324 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713335 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713347 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713358 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713370 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713382 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.713407 4772 scope.go:117] "RemoveContainer" containerID="7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.715709 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7jwx_87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8/kube-multus/2.log" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.724492 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7jwx_87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8/kube-multus/1.log" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.724625 4772 generic.go:334] "Generic (PLEG): container finished" podID="87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8" containerID="a5fee45d3fc79618abfe1fb780f6741fbf20558f07d7edf5c931f442a9c1c7dd" exitCode=2 Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.724849 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x7jwx" event={"ID":"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8","Type":"ContainerDied","Data":"a5fee45d3fc79618abfe1fb780f6741fbf20558f07d7edf5c931f442a9c1c7dd"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.724911 4772 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9f72b451fa77f3fce2c251de546110ab49c7c9e0122759f6ef29a32fde422356"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.725495 4772 scope.go:117] "RemoveContainer" containerID="a5fee45d3fc79618abfe1fb780f6741fbf20558f07d7edf5c931f442a9c1c7dd" Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.725772 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-x7jwx_openshift-multus(87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8)\"" pod="openshift-multus/multus-x7jwx" podUID="87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.731195 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qml47" event={"ID":"ff07e8b2-325b-4f96-b685-f2068052a960","Type":"ContainerStarted","Data":"ac4a5d287e4ae272f3c908bc4bec13137183b7668478f6a7b3fd983a452c66ad"} Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.751646 4772 scope.go:117] "RemoveContainer" containerID="8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.773834 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-n2khk"] Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.778688 4772 scope.go:117] "RemoveContainer" containerID="0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.786988 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-n2khk"] Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.849039 4772 scope.go:117] "RemoveContainer" containerID="0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.860408 4772 scope.go:117] "RemoveContainer" containerID="5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.874055 4772 scope.go:117] "RemoveContainer" containerID="45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.885303 4772 scope.go:117] "RemoveContainer" containerID="321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.896699 4772 scope.go:117] "RemoveContainer" containerID="3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.911506 4772 scope.go:117] "RemoveContainer" containerID="e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.926553 4772 scope.go:117] "RemoveContainer" containerID="c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.937900 4772 scope.go:117] "RemoveContainer" containerID="7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823" Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.938371 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823\": container with ID starting with 7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823 not found: ID does not exist" containerID="7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.938415 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823"} err="failed to get container status \"7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823\": rpc error: code = NotFound desc = could not find container \"7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823\": container with ID starting with 7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823 not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.938434 4772 scope.go:117] "RemoveContainer" containerID="8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a" Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.938686 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a\": container with ID starting with 8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a not found: ID does not exist" containerID="8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.938720 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a"} err="failed to get container status \"8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a\": rpc error: code = NotFound desc = could not find container \"8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a\": container with ID starting with 8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.938743 4772 scope.go:117] "RemoveContainer" containerID="0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae" Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.939083 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\": container with ID starting with 0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae not found: ID does not exist" containerID="0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.939134 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae"} err="failed to get container status \"0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\": rpc error: code = NotFound desc = could not find container \"0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\": container with ID starting with 0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.939154 4772 scope.go:117] "RemoveContainer" containerID="0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b" Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.939502 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\": container with ID starting with 0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b not found: ID does not exist" containerID="0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.939528 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b"} err="failed to get container status \"0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\": rpc error: code = NotFound desc = could not find container \"0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\": container with ID starting with 0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.939544 4772 scope.go:117] "RemoveContainer" containerID="5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4" Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.940155 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\": container with ID starting with 5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4 not found: ID does not exist" containerID="5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.940211 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4"} err="failed to get container status \"5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\": rpc error: code = NotFound desc = could not find container \"5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\": container with ID starting with 5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4 not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.940230 4772 scope.go:117] "RemoveContainer" containerID="45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5" Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.941419 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\": container with ID starting with 45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5 not found: ID does not exist" containerID="45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.941439 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5"} err="failed to get container status \"45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\": rpc error: code = NotFound desc = could not find container \"45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\": container with ID starting with 45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5 not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.941452 4772 scope.go:117] "RemoveContainer" containerID="321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9" Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.941703 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\": container with ID starting with 321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9 not found: ID does not exist" containerID="321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.941728 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9"} err="failed to get container status \"321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\": rpc error: code = NotFound desc = could not find container \"321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\": container with ID starting with 321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9 not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.941742 4772 scope.go:117] "RemoveContainer" containerID="3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854" Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.942109 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\": container with ID starting with 3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854 not found: ID does not exist" containerID="3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.942133 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854"} err="failed to get container status \"3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\": rpc error: code = NotFound desc = could not find container \"3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\": container with ID starting with 3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854 not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.942149 4772 scope.go:117] "RemoveContainer" containerID="e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e" Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.942662 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\": container with ID starting with e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e not found: ID does not exist" containerID="e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.942682 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e"} err="failed to get container status \"e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\": rpc error: code = NotFound desc = could not find container \"e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\": container with ID starting with e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.942695 4772 scope.go:117] "RemoveContainer" containerID="c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d" Jan 27 15:18:49 crc kubenswrapper[4772]: E0127 15:18:49.942888 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\": container with ID starting with c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d not found: ID does not exist" containerID="c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.942910 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d"} err="failed to get container status \"c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\": rpc error: code = NotFound desc = could not find container \"c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\": container with ID starting with c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.942923 4772 scope.go:117] "RemoveContainer" containerID="7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.943102 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823"} err="failed to get container status \"7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823\": rpc error: code = NotFound desc = could not find container \"7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823\": container with ID starting with 7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823 not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.943121 4772 scope.go:117] "RemoveContainer" containerID="8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.943335 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a"} err="failed to get container status \"8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a\": rpc error: code = NotFound desc = could not find container \"8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a\": container with ID starting with 8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.943355 4772 scope.go:117] "RemoveContainer" containerID="0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.943518 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae"} err="failed to get container status \"0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\": rpc error: code = NotFound desc = could not find container \"0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\": container with ID starting with 0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.943540 4772 scope.go:117] "RemoveContainer" containerID="0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.943701 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b"} err="failed to get container status \"0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\": rpc error: code = NotFound desc = could not find container \"0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\": container with ID starting with 0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.943716 4772 scope.go:117] "RemoveContainer" containerID="5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.943849 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4"} err="failed to get container status \"5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\": rpc error: code = NotFound desc = could not find container \"5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\": container with ID starting with 5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4 not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.943863 4772 scope.go:117] "RemoveContainer" containerID="45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.944024 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5"} err="failed to get container status \"45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\": rpc error: code = NotFound desc = could not find container \"45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\": container with ID starting with 45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5 not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.944046 4772 scope.go:117] "RemoveContainer" containerID="321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.944299 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9"} err="failed to get container status \"321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\": rpc error: code = NotFound desc = could not find container \"321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\": container with ID starting with 321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9 not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.944317 4772 scope.go:117] "RemoveContainer" containerID="3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.944482 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854"} err="failed to get container status \"3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\": rpc error: code = NotFound desc = could not find container \"3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\": container with ID starting with 3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854 not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.944505 4772 scope.go:117] "RemoveContainer" containerID="e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.944679 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e"} err="failed to get container status \"e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\": rpc error: code = NotFound desc = could not find container \"e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\": container with ID starting with e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.944702 4772 scope.go:117] "RemoveContainer" containerID="c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.944902 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d"} err="failed to get container status \"c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\": rpc error: code = NotFound desc = could not find container \"c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\": container with ID starting with c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.944933 4772 scope.go:117] "RemoveContainer" containerID="7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.945121 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823"} err="failed to get container status \"7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823\": rpc error: code = NotFound desc = could not find container \"7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823\": container with ID starting with 7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823 not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.945138 4772 scope.go:117] "RemoveContainer" containerID="8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.945376 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a"} err="failed to get container status \"8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a\": rpc error: code = NotFound desc = could not find container \"8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a\": container with ID starting with 8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.945395 4772 scope.go:117] "RemoveContainer" containerID="0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.945684 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae"} err="failed to get container status \"0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\": rpc error: code = NotFound desc = could not find container \"0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\": container with ID starting with 0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.945706 4772 scope.go:117] "RemoveContainer" containerID="0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.945930 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b"} err="failed to get container status \"0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\": rpc error: code = NotFound desc = could not find container \"0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\": container with ID starting with 0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.945954 4772 scope.go:117] "RemoveContainer" containerID="5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.946330 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4"} err="failed to get container status \"5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\": rpc error: code = NotFound desc = could not find container \"5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\": container with ID starting with 5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4 not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.946379 4772 scope.go:117] "RemoveContainer" containerID="45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.947089 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5"} err="failed to get container status \"45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\": rpc error: code = NotFound desc = could not find container \"45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\": container with ID starting with 45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5 not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.947110 4772 scope.go:117] "RemoveContainer" containerID="321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.947437 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9"} err="failed to get container status \"321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\": rpc error: code = NotFound desc = could not find container \"321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\": container with ID starting with 321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9 not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.947483 4772 scope.go:117] "RemoveContainer" containerID="3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.947725 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854"} err="failed to get container status \"3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\": rpc error: code = NotFound desc = could not find container \"3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\": container with ID starting with 3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854 not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.947746 4772 scope.go:117] "RemoveContainer" containerID="e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.948073 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e"} err="failed to get container status \"e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\": rpc error: code = NotFound desc = could not find container \"e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\": container with ID starting with e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.948096 4772 scope.go:117] "RemoveContainer" containerID="c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.948411 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d"} err="failed to get container status \"c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\": rpc error: code = NotFound desc = could not find container \"c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\": container with ID starting with c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.948432 4772 scope.go:117] "RemoveContainer" containerID="7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.948602 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823"} err="failed to get container status \"7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823\": rpc error: code = NotFound desc = could not find container \"7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823\": container with ID starting with 7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823 not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.948621 4772 scope.go:117] "RemoveContainer" containerID="8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.948796 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a"} err="failed to get container status \"8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a\": rpc error: code = NotFound desc = could not find container \"8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a\": container with ID starting with 8724eea2ef6df0dd65b48200e79bf0dd04d8e8a658ef59e0d5dbbc706d4cf68a not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.948832 4772 scope.go:117] "RemoveContainer" containerID="0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.949027 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae"} err="failed to get container status \"0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\": rpc error: code = NotFound desc = could not find container \"0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae\": container with ID starting with 0b8a45e6cc2ed92af13f48ae2d40ef8ba713fda7f78a0f0df8728375aa1326ae not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.949051 4772 scope.go:117] "RemoveContainer" containerID="0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.949230 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b"} err="failed to get container status \"0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\": rpc error: code = NotFound desc = could not find container \"0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b\": container with ID starting with 0672785b54a52ada7c5ecb3813df46790ca221c9234910436fb48c6c16bdbb3b not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.949253 4772 scope.go:117] "RemoveContainer" containerID="5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.949436 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4"} err="failed to get container status \"5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\": rpc error: code = NotFound desc = could not find container \"5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4\": container with ID starting with 5589864ab0df6b8ca9f810dea1168b5b16ac2b158531aac819758f0281dadba4 not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.949455 4772 scope.go:117] "RemoveContainer" containerID="45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.949635 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5"} err="failed to get container status \"45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\": rpc error: code = NotFound desc = could not find container \"45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5\": container with ID starting with 45a46f78c67486ce0034eefead74ba09c52e5ced21e7914963a219fef85efcc5 not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.949655 4772 scope.go:117] "RemoveContainer" containerID="321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.949864 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9"} err="failed to get container status \"321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\": rpc error: code = NotFound desc = could not find container \"321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9\": container with ID starting with 321c3da626a22e29558283e4efb292c6e039764e36318c3686f783e48ad876b9 not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.949884 4772 scope.go:117] "RemoveContainer" containerID="3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.950376 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854"} err="failed to get container status \"3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\": rpc error: code = NotFound desc = could not find container \"3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854\": container with ID starting with 3de2db48c4b82a23cce25c1be5a9c9e66439c283f7651d322e94015b94dc7854 not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.950415 4772 scope.go:117] "RemoveContainer" containerID="e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.950666 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e"} err="failed to get container status \"e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\": rpc error: code = NotFound desc = could not find container \"e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e\": container with ID starting with e68c9acebcb335ee630582ff98a6406849766800d848c6f4a7d87e22a65e1e1e not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.950684 4772 scope.go:117] "RemoveContainer" containerID="c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.950891 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d"} err="failed to get container status \"c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\": rpc error: code = NotFound desc = could not find container \"c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d\": container with ID starting with c55ceb84c37125c4dc988d6e8c3ea65d854c89190c2ac8f32c191d6d55e2982d not found: ID does not exist" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.950928 4772 scope.go:117] "RemoveContainer" containerID="7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823" Jan 27 15:18:49 crc kubenswrapper[4772]: I0127 15:18:49.951192 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823"} err="failed to get container status \"7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823\": rpc error: code = NotFound desc = could not find container \"7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823\": container with ID starting with 7bcf11983997321ee81682fb1ab65d69810342d1d15ef9f8da9f8d1344cdc823 not found: ID does not exist" Jan 27 15:18:50 crc kubenswrapper[4772]: I0127 15:18:50.670672 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="736264c8-cd18-479a-88ba-e1ec15dbfdae" path="/var/lib/kubelet/pods/736264c8-cd18-479a-88ba-e1ec15dbfdae/volumes" Jan 27 15:18:50 crc kubenswrapper[4772]: I0127 15:18:50.739765 4772 generic.go:334] "Generic (PLEG): container finished" podID="ff07e8b2-325b-4f96-b685-f2068052a960" containerID="93cf0cb8c7e8fc64f2417d90ba2ce26fdbca27a4bbf34b0ab6539feb8b83bf96" exitCode=0 Jan 27 15:18:50 crc kubenswrapper[4772]: I0127 15:18:50.739845 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qml47" event={"ID":"ff07e8b2-325b-4f96-b685-f2068052a960","Type":"ContainerDied","Data":"93cf0cb8c7e8fc64f2417d90ba2ce26fdbca27a4bbf34b0ab6539feb8b83bf96"} Jan 27 15:18:51 crc kubenswrapper[4772]: I0127 15:18:51.752690 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qml47" event={"ID":"ff07e8b2-325b-4f96-b685-f2068052a960","Type":"ContainerStarted","Data":"f75eb4ed67c4fe2b28f19628b2f7e25812cf89c51eb13a3fc08dbf049eddcaa5"} Jan 27 15:18:51 crc kubenswrapper[4772]: I0127 15:18:51.753140 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qml47" event={"ID":"ff07e8b2-325b-4f96-b685-f2068052a960","Type":"ContainerStarted","Data":"aaca2c527bc95b49c9ba9bae28800d2aae6784e5558342ef024c11be4d341405"} Jan 27 15:18:51 crc kubenswrapper[4772]: I0127 15:18:51.753155 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qml47" event={"ID":"ff07e8b2-325b-4f96-b685-f2068052a960","Type":"ContainerStarted","Data":"537c0ae0bfa35738f9cff8a98c8948a3a4160a7c65ec7f2a8b377177e94338d4"} Jan 27 15:18:51 crc kubenswrapper[4772]: I0127 15:18:51.753186 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qml47" event={"ID":"ff07e8b2-325b-4f96-b685-f2068052a960","Type":"ContainerStarted","Data":"500953b46c83a06e86a04b002f891b91017d531482a6d4c341f1b652687440e7"} Jan 27 15:18:51 crc kubenswrapper[4772]: I0127 15:18:51.753199 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qml47" event={"ID":"ff07e8b2-325b-4f96-b685-f2068052a960","Type":"ContainerStarted","Data":"d9c68bc8eb48037e17d05979d4605cba35a4f6b495e3950a2b99610cd7b14c31"} Jan 27 15:18:51 crc kubenswrapper[4772]: I0127 15:18:51.753210 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qml47" event={"ID":"ff07e8b2-325b-4f96-b685-f2068052a960","Type":"ContainerStarted","Data":"58042ebd592c2e1a476cd7c933626dc515b1acd1a401b13d469c688dacd59b3f"} Jan 27 15:18:53 crc kubenswrapper[4772]: I0127 15:18:53.774620 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qml47" event={"ID":"ff07e8b2-325b-4f96-b685-f2068052a960","Type":"ContainerStarted","Data":"2562ba36a42df9c7d6943315e35e2a1bf5af741384e77eb33177ba315e7fbc56"} Jan 27 15:18:54 crc kubenswrapper[4772]: I0127 15:18:54.917809 4772 scope.go:117] "RemoveContainer" containerID="9f72b451fa77f3fce2c251de546110ab49c7c9e0122759f6ef29a32fde422356" Jan 27 15:18:55 crc kubenswrapper[4772]: I0127 15:18:55.787858 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7jwx_87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8/kube-multus/2.log" Jan 27 15:18:56 crc kubenswrapper[4772]: I0127 15:18:56.801624 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qml47" event={"ID":"ff07e8b2-325b-4f96-b685-f2068052a960","Type":"ContainerStarted","Data":"22cf781b1425a3e566222d387ab7f24fc9a07b9dde7c46d5838e46a939cb4c7a"} Jan 27 15:18:56 crc kubenswrapper[4772]: I0127 15:18:56.801960 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:56 crc kubenswrapper[4772]: I0127 15:18:56.801974 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:56 crc kubenswrapper[4772]: I0127 15:18:56.801984 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:56 crc kubenswrapper[4772]: I0127 15:18:56.836474 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:56 crc kubenswrapper[4772]: I0127 15:18:56.836922 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:18:56 crc kubenswrapper[4772]: I0127 15:18:56.840968 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-qml47" podStartSLOduration=7.840952459 podStartE2EDuration="7.840952459s" podCreationTimestamp="2026-01-27 15:18:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:18:56.839968571 +0000 UTC m=+722.820577699" watchObservedRunningTime="2026-01-27 15:18:56.840952459 +0000 UTC m=+722.821561557" Jan 27 15:18:56 crc kubenswrapper[4772]: I0127 15:18:56.974012 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-pmrs5"] Jan 27 15:18:56 crc kubenswrapper[4772]: I0127 15:18:56.974727 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-pmrs5" Jan 27 15:18:56 crc kubenswrapper[4772]: I0127 15:18:56.976627 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Jan 27 15:18:56 crc kubenswrapper[4772]: I0127 15:18:56.976757 4772 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-r9nmx" Jan 27 15:18:56 crc kubenswrapper[4772]: I0127 15:18:56.977084 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Jan 27 15:18:56 crc kubenswrapper[4772]: I0127 15:18:56.977373 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Jan 27 15:18:57 crc kubenswrapper[4772]: I0127 15:18:57.018594 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-pmrs5"] Jan 27 15:18:57 crc kubenswrapper[4772]: I0127 15:18:57.040743 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79mhd\" (UniqueName: \"kubernetes.io/projected/32a5cafc-0519-4e90-9456-acb182176c41-kube-api-access-79mhd\") pod \"crc-storage-crc-pmrs5\" (UID: \"32a5cafc-0519-4e90-9456-acb182176c41\") " pod="crc-storage/crc-storage-crc-pmrs5" Jan 27 15:18:57 crc kubenswrapper[4772]: I0127 15:18:57.040813 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/32a5cafc-0519-4e90-9456-acb182176c41-crc-storage\") pod \"crc-storage-crc-pmrs5\" (UID: \"32a5cafc-0519-4e90-9456-acb182176c41\") " pod="crc-storage/crc-storage-crc-pmrs5" Jan 27 15:18:57 crc kubenswrapper[4772]: I0127 15:18:57.040840 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/32a5cafc-0519-4e90-9456-acb182176c41-node-mnt\") pod \"crc-storage-crc-pmrs5\" (UID: \"32a5cafc-0519-4e90-9456-acb182176c41\") " pod="crc-storage/crc-storage-crc-pmrs5" Jan 27 15:18:57 crc kubenswrapper[4772]: I0127 15:18:57.141458 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79mhd\" (UniqueName: \"kubernetes.io/projected/32a5cafc-0519-4e90-9456-acb182176c41-kube-api-access-79mhd\") pod \"crc-storage-crc-pmrs5\" (UID: \"32a5cafc-0519-4e90-9456-acb182176c41\") " pod="crc-storage/crc-storage-crc-pmrs5" Jan 27 15:18:57 crc kubenswrapper[4772]: I0127 15:18:57.141526 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/32a5cafc-0519-4e90-9456-acb182176c41-crc-storage\") pod \"crc-storage-crc-pmrs5\" (UID: \"32a5cafc-0519-4e90-9456-acb182176c41\") " pod="crc-storage/crc-storage-crc-pmrs5" Jan 27 15:18:57 crc kubenswrapper[4772]: I0127 15:18:57.141553 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/32a5cafc-0519-4e90-9456-acb182176c41-node-mnt\") pod \"crc-storage-crc-pmrs5\" (UID: \"32a5cafc-0519-4e90-9456-acb182176c41\") " pod="crc-storage/crc-storage-crc-pmrs5" Jan 27 15:18:57 crc kubenswrapper[4772]: I0127 15:18:57.141867 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/32a5cafc-0519-4e90-9456-acb182176c41-node-mnt\") pod \"crc-storage-crc-pmrs5\" (UID: \"32a5cafc-0519-4e90-9456-acb182176c41\") " pod="crc-storage/crc-storage-crc-pmrs5" Jan 27 15:18:57 crc kubenswrapper[4772]: I0127 15:18:57.143284 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/32a5cafc-0519-4e90-9456-acb182176c41-crc-storage\") pod \"crc-storage-crc-pmrs5\" (UID: \"32a5cafc-0519-4e90-9456-acb182176c41\") " pod="crc-storage/crc-storage-crc-pmrs5" Jan 27 15:18:57 crc kubenswrapper[4772]: I0127 15:18:57.161007 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79mhd\" (UniqueName: \"kubernetes.io/projected/32a5cafc-0519-4e90-9456-acb182176c41-kube-api-access-79mhd\") pod \"crc-storage-crc-pmrs5\" (UID: \"32a5cafc-0519-4e90-9456-acb182176c41\") " pod="crc-storage/crc-storage-crc-pmrs5" Jan 27 15:18:57 crc kubenswrapper[4772]: I0127 15:18:57.289364 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-pmrs5" Jan 27 15:18:57 crc kubenswrapper[4772]: E0127 15:18:57.313534 4772 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-pmrs5_crc-storage_32a5cafc-0519-4e90-9456-acb182176c41_0(e72254d7c4699290e0bb88e53cbb2017bdf725db6e241c5fea9269ea5eac8473): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 27 15:18:57 crc kubenswrapper[4772]: E0127 15:18:57.313724 4772 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-pmrs5_crc-storage_32a5cafc-0519-4e90-9456-acb182176c41_0(e72254d7c4699290e0bb88e53cbb2017bdf725db6e241c5fea9269ea5eac8473): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-pmrs5" Jan 27 15:18:57 crc kubenswrapper[4772]: E0127 15:18:57.313798 4772 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-pmrs5_crc-storage_32a5cafc-0519-4e90-9456-acb182176c41_0(e72254d7c4699290e0bb88e53cbb2017bdf725db6e241c5fea9269ea5eac8473): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-pmrs5" Jan 27 15:18:57 crc kubenswrapper[4772]: E0127 15:18:57.313895 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-pmrs5_crc-storage(32a5cafc-0519-4e90-9456-acb182176c41)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-pmrs5_crc-storage(32a5cafc-0519-4e90-9456-acb182176c41)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-pmrs5_crc-storage_32a5cafc-0519-4e90-9456-acb182176c41_0(e72254d7c4699290e0bb88e53cbb2017bdf725db6e241c5fea9269ea5eac8473): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-pmrs5" podUID="32a5cafc-0519-4e90-9456-acb182176c41" Jan 27 15:18:57 crc kubenswrapper[4772]: I0127 15:18:57.807082 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-pmrs5" Jan 27 15:18:57 crc kubenswrapper[4772]: I0127 15:18:57.807582 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-pmrs5" Jan 27 15:18:57 crc kubenswrapper[4772]: E0127 15:18:57.835432 4772 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-pmrs5_crc-storage_32a5cafc-0519-4e90-9456-acb182176c41_0(0b81e4bb3394a9db9e768032e5b4b727da59aed8fa29834672c7ceadbc239a0e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 27 15:18:57 crc kubenswrapper[4772]: E0127 15:18:57.835503 4772 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-pmrs5_crc-storage_32a5cafc-0519-4e90-9456-acb182176c41_0(0b81e4bb3394a9db9e768032e5b4b727da59aed8fa29834672c7ceadbc239a0e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-pmrs5" Jan 27 15:18:57 crc kubenswrapper[4772]: E0127 15:18:57.835528 4772 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-pmrs5_crc-storage_32a5cafc-0519-4e90-9456-acb182176c41_0(0b81e4bb3394a9db9e768032e5b4b727da59aed8fa29834672c7ceadbc239a0e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-pmrs5" Jan 27 15:18:57 crc kubenswrapper[4772]: E0127 15:18:57.835580 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-pmrs5_crc-storage(32a5cafc-0519-4e90-9456-acb182176c41)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-pmrs5_crc-storage(32a5cafc-0519-4e90-9456-acb182176c41)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-pmrs5_crc-storage_32a5cafc-0519-4e90-9456-acb182176c41_0(0b81e4bb3394a9db9e768032e5b4b727da59aed8fa29834672c7ceadbc239a0e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-pmrs5" podUID="32a5cafc-0519-4e90-9456-acb182176c41" Jan 27 15:19:03 crc kubenswrapper[4772]: I0127 15:19:03.663886 4772 scope.go:117] "RemoveContainer" containerID="a5fee45d3fc79618abfe1fb780f6741fbf20558f07d7edf5c931f442a9c1c7dd" Jan 27 15:19:03 crc kubenswrapper[4772]: I0127 15:19:03.847302 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7jwx_87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8/kube-multus/2.log" Jan 27 15:19:04 crc kubenswrapper[4772]: I0127 15:19:04.856996 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x7jwx_87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8/kube-multus/2.log" Jan 27 15:19:04 crc kubenswrapper[4772]: I0127 15:19:04.858038 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x7jwx" event={"ID":"87cb2a5b-099e-4a3b-a0bc-cba76a1a00a8","Type":"ContainerStarted","Data":"2f8786cb2f7e8a784f7db02cc01cff531e7a0af9267bee88f64e672f700cc88f"} Jan 27 15:19:12 crc kubenswrapper[4772]: I0127 15:19:12.058266 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:19:12 crc kubenswrapper[4772]: I0127 15:19:12.058817 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:19:12 crc kubenswrapper[4772]: I0127 15:19:12.663644 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-pmrs5" Jan 27 15:19:12 crc kubenswrapper[4772]: I0127 15:19:12.664454 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-pmrs5" Jan 27 15:19:12 crc kubenswrapper[4772]: I0127 15:19:12.865480 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-pmrs5"] Jan 27 15:19:12 crc kubenswrapper[4772]: W0127 15:19:12.870199 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32a5cafc_0519_4e90_9456_acb182176c41.slice/crio-cc832d30aabf335cefcf2aff9b9f4129cc8d1a2b7ef5bc41178d9a657d238dcc WatchSource:0}: Error finding container cc832d30aabf335cefcf2aff9b9f4129cc8d1a2b7ef5bc41178d9a657d238dcc: Status 404 returned error can't find the container with id cc832d30aabf335cefcf2aff9b9f4129cc8d1a2b7ef5bc41178d9a657d238dcc Jan 27 15:19:12 crc kubenswrapper[4772]: I0127 15:19:12.872748 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 15:19:12 crc kubenswrapper[4772]: I0127 15:19:12.914226 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-pmrs5" event={"ID":"32a5cafc-0519-4e90-9456-acb182176c41","Type":"ContainerStarted","Data":"cc832d30aabf335cefcf2aff9b9f4129cc8d1a2b7ef5bc41178d9a657d238dcc"} Jan 27 15:19:14 crc kubenswrapper[4772]: I0127 15:19:14.925951 4772 generic.go:334] "Generic (PLEG): container finished" podID="32a5cafc-0519-4e90-9456-acb182176c41" containerID="f4e8f8b6e9c9139e4588eb373fc8616c60521a1a5d0cbfb79f4f8c9d4dc676b9" exitCode=0 Jan 27 15:19:14 crc kubenswrapper[4772]: I0127 15:19:14.926049 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-pmrs5" event={"ID":"32a5cafc-0519-4e90-9456-acb182176c41","Type":"ContainerDied","Data":"f4e8f8b6e9c9139e4588eb373fc8616c60521a1a5d0cbfb79f4f8c9d4dc676b9"} Jan 27 15:19:16 crc kubenswrapper[4772]: I0127 15:19:16.207656 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-pmrs5" Jan 27 15:19:16 crc kubenswrapper[4772]: I0127 15:19:16.327047 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79mhd\" (UniqueName: \"kubernetes.io/projected/32a5cafc-0519-4e90-9456-acb182176c41-kube-api-access-79mhd\") pod \"32a5cafc-0519-4e90-9456-acb182176c41\" (UID: \"32a5cafc-0519-4e90-9456-acb182176c41\") " Jan 27 15:19:16 crc kubenswrapper[4772]: I0127 15:19:16.327212 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/32a5cafc-0519-4e90-9456-acb182176c41-node-mnt\") pod \"32a5cafc-0519-4e90-9456-acb182176c41\" (UID: \"32a5cafc-0519-4e90-9456-acb182176c41\") " Jan 27 15:19:16 crc kubenswrapper[4772]: I0127 15:19:16.327284 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/32a5cafc-0519-4e90-9456-acb182176c41-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "32a5cafc-0519-4e90-9456-acb182176c41" (UID: "32a5cafc-0519-4e90-9456-acb182176c41"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:19:16 crc kubenswrapper[4772]: I0127 15:19:16.327963 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/32a5cafc-0519-4e90-9456-acb182176c41-crc-storage\") pod \"32a5cafc-0519-4e90-9456-acb182176c41\" (UID: \"32a5cafc-0519-4e90-9456-acb182176c41\") " Jan 27 15:19:16 crc kubenswrapper[4772]: I0127 15:19:16.328189 4772 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/32a5cafc-0519-4e90-9456-acb182176c41-node-mnt\") on node \"crc\" DevicePath \"\"" Jan 27 15:19:16 crc kubenswrapper[4772]: I0127 15:19:16.333727 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32a5cafc-0519-4e90-9456-acb182176c41-kube-api-access-79mhd" (OuterVolumeSpecName: "kube-api-access-79mhd") pod "32a5cafc-0519-4e90-9456-acb182176c41" (UID: "32a5cafc-0519-4e90-9456-acb182176c41"). InnerVolumeSpecName "kube-api-access-79mhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:19:16 crc kubenswrapper[4772]: I0127 15:19:16.347332 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32a5cafc-0519-4e90-9456-acb182176c41-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "32a5cafc-0519-4e90-9456-acb182176c41" (UID: "32a5cafc-0519-4e90-9456-acb182176c41"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:19:16 crc kubenswrapper[4772]: I0127 15:19:16.429088 4772 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/32a5cafc-0519-4e90-9456-acb182176c41-crc-storage\") on node \"crc\" DevicePath \"\"" Jan 27 15:19:16 crc kubenswrapper[4772]: I0127 15:19:16.429134 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79mhd\" (UniqueName: \"kubernetes.io/projected/32a5cafc-0519-4e90-9456-acb182176c41-kube-api-access-79mhd\") on node \"crc\" DevicePath \"\"" Jan 27 15:19:16 crc kubenswrapper[4772]: I0127 15:19:16.945825 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-pmrs5" event={"ID":"32a5cafc-0519-4e90-9456-acb182176c41","Type":"ContainerDied","Data":"cc832d30aabf335cefcf2aff9b9f4129cc8d1a2b7ef5bc41178d9a657d238dcc"} Jan 27 15:19:16 crc kubenswrapper[4772]: I0127 15:19:16.945879 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc832d30aabf335cefcf2aff9b9f4129cc8d1a2b7ef5bc41178d9a657d238dcc" Jan 27 15:19:16 crc kubenswrapper[4772]: I0127 15:19:16.945947 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-pmrs5" Jan 27 15:19:19 crc kubenswrapper[4772]: I0127 15:19:19.634584 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qml47" Jan 27 15:19:23 crc kubenswrapper[4772]: I0127 15:19:23.124718 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf"] Jan 27 15:19:23 crc kubenswrapper[4772]: E0127 15:19:23.125256 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32a5cafc-0519-4e90-9456-acb182176c41" containerName="storage" Jan 27 15:19:23 crc kubenswrapper[4772]: I0127 15:19:23.125269 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="32a5cafc-0519-4e90-9456-acb182176c41" containerName="storage" Jan 27 15:19:23 crc kubenswrapper[4772]: I0127 15:19:23.125373 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="32a5cafc-0519-4e90-9456-acb182176c41" containerName="storage" Jan 27 15:19:23 crc kubenswrapper[4772]: I0127 15:19:23.126021 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf" Jan 27 15:19:23 crc kubenswrapper[4772]: I0127 15:19:23.128321 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 27 15:19:23 crc kubenswrapper[4772]: I0127 15:19:23.133456 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf"] Jan 27 15:19:23 crc kubenswrapper[4772]: I0127 15:19:23.322750 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d583171b-99cd-49da-9a9f-48931806cb45-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf\" (UID: \"d583171b-99cd-49da-9a9f-48931806cb45\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf" Jan 27 15:19:23 crc kubenswrapper[4772]: I0127 15:19:23.322843 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwcjv\" (UniqueName: \"kubernetes.io/projected/d583171b-99cd-49da-9a9f-48931806cb45-kube-api-access-lwcjv\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf\" (UID: \"d583171b-99cd-49da-9a9f-48931806cb45\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf" Jan 27 15:19:23 crc kubenswrapper[4772]: I0127 15:19:23.322968 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d583171b-99cd-49da-9a9f-48931806cb45-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf\" (UID: \"d583171b-99cd-49da-9a9f-48931806cb45\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf" Jan 27 15:19:23 crc kubenswrapper[4772]: I0127 15:19:23.423983 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d583171b-99cd-49da-9a9f-48931806cb45-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf\" (UID: \"d583171b-99cd-49da-9a9f-48931806cb45\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf" Jan 27 15:19:23 crc kubenswrapper[4772]: I0127 15:19:23.424070 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwcjv\" (UniqueName: \"kubernetes.io/projected/d583171b-99cd-49da-9a9f-48931806cb45-kube-api-access-lwcjv\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf\" (UID: \"d583171b-99cd-49da-9a9f-48931806cb45\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf" Jan 27 15:19:23 crc kubenswrapper[4772]: I0127 15:19:23.424601 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d583171b-99cd-49da-9a9f-48931806cb45-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf\" (UID: \"d583171b-99cd-49da-9a9f-48931806cb45\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf" Jan 27 15:19:23 crc kubenswrapper[4772]: I0127 15:19:23.425300 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d583171b-99cd-49da-9a9f-48931806cb45-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf\" (UID: \"d583171b-99cd-49da-9a9f-48931806cb45\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf" Jan 27 15:19:23 crc kubenswrapper[4772]: I0127 15:19:23.425591 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d583171b-99cd-49da-9a9f-48931806cb45-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf\" (UID: \"d583171b-99cd-49da-9a9f-48931806cb45\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf" Jan 27 15:19:23 crc kubenswrapper[4772]: I0127 15:19:23.444467 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwcjv\" (UniqueName: \"kubernetes.io/projected/d583171b-99cd-49da-9a9f-48931806cb45-kube-api-access-lwcjv\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf\" (UID: \"d583171b-99cd-49da-9a9f-48931806cb45\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf" Jan 27 15:19:23 crc kubenswrapper[4772]: I0127 15:19:23.743628 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf" Jan 27 15:19:23 crc kubenswrapper[4772]: I0127 15:19:23.971346 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf"] Jan 27 15:19:24 crc kubenswrapper[4772]: I0127 15:19:24.994919 4772 generic.go:334] "Generic (PLEG): container finished" podID="d583171b-99cd-49da-9a9f-48931806cb45" containerID="809f06b005772e28360eaf3f49b070a296c1f8e64d6ac832b3fd457afe7571b1" exitCode=0 Jan 27 15:19:24 crc kubenswrapper[4772]: I0127 15:19:24.995010 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf" event={"ID":"d583171b-99cd-49da-9a9f-48931806cb45","Type":"ContainerDied","Data":"809f06b005772e28360eaf3f49b070a296c1f8e64d6ac832b3fd457afe7571b1"} Jan 27 15:19:24 crc kubenswrapper[4772]: I0127 15:19:24.995304 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf" event={"ID":"d583171b-99cd-49da-9a9f-48931806cb45","Type":"ContainerStarted","Data":"aaf6d95824b5eed2d7c5b70fa582cfb3a7afd152705a6bd3badccbd05f100bf9"} Jan 27 15:19:27 crc kubenswrapper[4772]: I0127 15:19:27.009947 4772 generic.go:334] "Generic (PLEG): container finished" podID="d583171b-99cd-49da-9a9f-48931806cb45" containerID="3edc59963bb2eb5ccee335b1bb8a54cddb9975464ce2916da2d9b900d384092c" exitCode=0 Jan 27 15:19:27 crc kubenswrapper[4772]: I0127 15:19:27.010085 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf" event={"ID":"d583171b-99cd-49da-9a9f-48931806cb45","Type":"ContainerDied","Data":"3edc59963bb2eb5ccee335b1bb8a54cddb9975464ce2916da2d9b900d384092c"} Jan 27 15:19:28 crc kubenswrapper[4772]: I0127 15:19:28.016357 4772 generic.go:334] "Generic (PLEG): container finished" podID="d583171b-99cd-49da-9a9f-48931806cb45" containerID="9a8a4fd5cb4bc936b5cac14dfa14122988a0b6fbe2b78903ad0bf46c28a9b6d8" exitCode=0 Jan 27 15:19:28 crc kubenswrapper[4772]: I0127 15:19:28.016395 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf" event={"ID":"d583171b-99cd-49da-9a9f-48931806cb45","Type":"ContainerDied","Data":"9a8a4fd5cb4bc936b5cac14dfa14122988a0b6fbe2b78903ad0bf46c28a9b6d8"} Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.274772 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf" Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.406135 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d583171b-99cd-49da-9a9f-48931806cb45-util\") pod \"d583171b-99cd-49da-9a9f-48931806cb45\" (UID: \"d583171b-99cd-49da-9a9f-48931806cb45\") " Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.406257 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d583171b-99cd-49da-9a9f-48931806cb45-bundle\") pod \"d583171b-99cd-49da-9a9f-48931806cb45\" (UID: \"d583171b-99cd-49da-9a9f-48931806cb45\") " Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.406290 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwcjv\" (UniqueName: \"kubernetes.io/projected/d583171b-99cd-49da-9a9f-48931806cb45-kube-api-access-lwcjv\") pod \"d583171b-99cd-49da-9a9f-48931806cb45\" (UID: \"d583171b-99cd-49da-9a9f-48931806cb45\") " Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.406867 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d583171b-99cd-49da-9a9f-48931806cb45-bundle" (OuterVolumeSpecName: "bundle") pod "d583171b-99cd-49da-9a9f-48931806cb45" (UID: "d583171b-99cd-49da-9a9f-48931806cb45"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.412320 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d583171b-99cd-49da-9a9f-48931806cb45-kube-api-access-lwcjv" (OuterVolumeSpecName: "kube-api-access-lwcjv") pod "d583171b-99cd-49da-9a9f-48931806cb45" (UID: "d583171b-99cd-49da-9a9f-48931806cb45"). InnerVolumeSpecName "kube-api-access-lwcjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.420663 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d583171b-99cd-49da-9a9f-48931806cb45-util" (OuterVolumeSpecName: "util") pod "d583171b-99cd-49da-9a9f-48931806cb45" (UID: "d583171b-99cd-49da-9a9f-48931806cb45"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.497773 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wjx8r"] Jan 27 15:19:29 crc kubenswrapper[4772]: E0127 15:19:29.498032 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d583171b-99cd-49da-9a9f-48931806cb45" containerName="pull" Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.498045 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="d583171b-99cd-49da-9a9f-48931806cb45" containerName="pull" Jan 27 15:19:29 crc kubenswrapper[4772]: E0127 15:19:29.498059 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d583171b-99cd-49da-9a9f-48931806cb45" containerName="extract" Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.498065 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="d583171b-99cd-49da-9a9f-48931806cb45" containerName="extract" Jan 27 15:19:29 crc kubenswrapper[4772]: E0127 15:19:29.498083 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d583171b-99cd-49da-9a9f-48931806cb45" containerName="util" Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.498091 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="d583171b-99cd-49da-9a9f-48931806cb45" containerName="util" Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.498213 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="d583171b-99cd-49da-9a9f-48931806cb45" containerName="extract" Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.498959 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wjx8r" Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.504916 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wjx8r"] Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.507886 4772 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d583171b-99cd-49da-9a9f-48931806cb45-util\") on node \"crc\" DevicePath \"\"" Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.508033 4772 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d583171b-99cd-49da-9a9f-48931806cb45-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.508137 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwcjv\" (UniqueName: \"kubernetes.io/projected/d583171b-99cd-49da-9a9f-48931806cb45-kube-api-access-lwcjv\") on node \"crc\" DevicePath \"\"" Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.610005 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09a099f0-cf75-4584-9369-e146e1898ee5-catalog-content\") pod \"redhat-operators-wjx8r\" (UID: \"09a099f0-cf75-4584-9369-e146e1898ee5\") " pod="openshift-marketplace/redhat-operators-wjx8r" Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.610080 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09a099f0-cf75-4584-9369-e146e1898ee5-utilities\") pod \"redhat-operators-wjx8r\" (UID: \"09a099f0-cf75-4584-9369-e146e1898ee5\") " pod="openshift-marketplace/redhat-operators-wjx8r" Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.610132 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbc5f\" (UniqueName: \"kubernetes.io/projected/09a099f0-cf75-4584-9369-e146e1898ee5-kube-api-access-pbc5f\") pod \"redhat-operators-wjx8r\" (UID: \"09a099f0-cf75-4584-9369-e146e1898ee5\") " pod="openshift-marketplace/redhat-operators-wjx8r" Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.710995 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbc5f\" (UniqueName: \"kubernetes.io/projected/09a099f0-cf75-4584-9369-e146e1898ee5-kube-api-access-pbc5f\") pod \"redhat-operators-wjx8r\" (UID: \"09a099f0-cf75-4584-9369-e146e1898ee5\") " pod="openshift-marketplace/redhat-operators-wjx8r" Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.711069 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09a099f0-cf75-4584-9369-e146e1898ee5-catalog-content\") pod \"redhat-operators-wjx8r\" (UID: \"09a099f0-cf75-4584-9369-e146e1898ee5\") " pod="openshift-marketplace/redhat-operators-wjx8r" Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.711133 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09a099f0-cf75-4584-9369-e146e1898ee5-utilities\") pod \"redhat-operators-wjx8r\" (UID: \"09a099f0-cf75-4584-9369-e146e1898ee5\") " pod="openshift-marketplace/redhat-operators-wjx8r" Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.711856 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09a099f0-cf75-4584-9369-e146e1898ee5-utilities\") pod \"redhat-operators-wjx8r\" (UID: \"09a099f0-cf75-4584-9369-e146e1898ee5\") " pod="openshift-marketplace/redhat-operators-wjx8r" Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.712070 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09a099f0-cf75-4584-9369-e146e1898ee5-catalog-content\") pod \"redhat-operators-wjx8r\" (UID: \"09a099f0-cf75-4584-9369-e146e1898ee5\") " pod="openshift-marketplace/redhat-operators-wjx8r" Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.731969 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbc5f\" (UniqueName: \"kubernetes.io/projected/09a099f0-cf75-4584-9369-e146e1898ee5-kube-api-access-pbc5f\") pod \"redhat-operators-wjx8r\" (UID: \"09a099f0-cf75-4584-9369-e146e1898ee5\") " pod="openshift-marketplace/redhat-operators-wjx8r" Jan 27 15:19:29 crc kubenswrapper[4772]: I0127 15:19:29.829655 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wjx8r" Jan 27 15:19:30 crc kubenswrapper[4772]: I0127 15:19:30.043479 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wjx8r"] Jan 27 15:19:30 crc kubenswrapper[4772]: I0127 15:19:30.043855 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf" event={"ID":"d583171b-99cd-49da-9a9f-48931806cb45","Type":"ContainerDied","Data":"aaf6d95824b5eed2d7c5b70fa582cfb3a7afd152705a6bd3badccbd05f100bf9"} Jan 27 15:19:30 crc kubenswrapper[4772]: I0127 15:19:30.043883 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aaf6d95824b5eed2d7c5b70fa582cfb3a7afd152705a6bd3badccbd05f100bf9" Jan 27 15:19:30 crc kubenswrapper[4772]: I0127 15:19:30.043652 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf" Jan 27 15:19:30 crc kubenswrapper[4772]: W0127 15:19:30.043916 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09a099f0_cf75_4584_9369_e146e1898ee5.slice/crio-5400ef2d30491c889dad91bc65f72f745b56aba86dca10ba2c7a00a785a6ae51 WatchSource:0}: Error finding container 5400ef2d30491c889dad91bc65f72f745b56aba86dca10ba2c7a00a785a6ae51: Status 404 returned error can't find the container with id 5400ef2d30491c889dad91bc65f72f745b56aba86dca10ba2c7a00a785a6ae51 Jan 27 15:19:30 crc kubenswrapper[4772]: I0127 15:19:30.203937 4772 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 27 15:19:31 crc kubenswrapper[4772]: I0127 15:19:31.050485 4772 generic.go:334] "Generic (PLEG): container finished" podID="09a099f0-cf75-4584-9369-e146e1898ee5" containerID="db8ffbbcaeec3abdc8460ac13f60a94b11906e285d02efbfc2441a057fa123e1" exitCode=0 Jan 27 15:19:31 crc kubenswrapper[4772]: I0127 15:19:31.050542 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjx8r" event={"ID":"09a099f0-cf75-4584-9369-e146e1898ee5","Type":"ContainerDied","Data":"db8ffbbcaeec3abdc8460ac13f60a94b11906e285d02efbfc2441a057fa123e1"} Jan 27 15:19:31 crc kubenswrapper[4772]: I0127 15:19:31.050810 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjx8r" event={"ID":"09a099f0-cf75-4584-9369-e146e1898ee5","Type":"ContainerStarted","Data":"5400ef2d30491c889dad91bc65f72f745b56aba86dca10ba2c7a00a785a6ae51"} Jan 27 15:19:31 crc kubenswrapper[4772]: I0127 15:19:31.379444 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-mdjph"] Jan 27 15:19:31 crc kubenswrapper[4772]: I0127 15:19:31.380257 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-mdjph" Jan 27 15:19:31 crc kubenswrapper[4772]: I0127 15:19:31.382925 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Jan 27 15:19:31 crc kubenswrapper[4772]: I0127 15:19:31.383028 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-bcpxd" Jan 27 15:19:31 crc kubenswrapper[4772]: I0127 15:19:31.385568 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Jan 27 15:19:31 crc kubenswrapper[4772]: I0127 15:19:31.397669 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-mdjph"] Jan 27 15:19:31 crc kubenswrapper[4772]: I0127 15:19:31.433285 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5qg8\" (UniqueName: \"kubernetes.io/projected/2914eab0-19c8-464b-a774-d30a492f6763-kube-api-access-n5qg8\") pod \"nmstate-operator-646758c888-mdjph\" (UID: \"2914eab0-19c8-464b-a774-d30a492f6763\") " pod="openshift-nmstate/nmstate-operator-646758c888-mdjph" Jan 27 15:19:31 crc kubenswrapper[4772]: I0127 15:19:31.534339 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5qg8\" (UniqueName: \"kubernetes.io/projected/2914eab0-19c8-464b-a774-d30a492f6763-kube-api-access-n5qg8\") pod \"nmstate-operator-646758c888-mdjph\" (UID: \"2914eab0-19c8-464b-a774-d30a492f6763\") " pod="openshift-nmstate/nmstate-operator-646758c888-mdjph" Jan 27 15:19:31 crc kubenswrapper[4772]: I0127 15:19:31.559663 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5qg8\" (UniqueName: \"kubernetes.io/projected/2914eab0-19c8-464b-a774-d30a492f6763-kube-api-access-n5qg8\") pod \"nmstate-operator-646758c888-mdjph\" (UID: \"2914eab0-19c8-464b-a774-d30a492f6763\") " pod="openshift-nmstate/nmstate-operator-646758c888-mdjph" Jan 27 15:19:31 crc kubenswrapper[4772]: I0127 15:19:31.697676 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-mdjph" Jan 27 15:19:31 crc kubenswrapper[4772]: I0127 15:19:31.960418 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-mdjph"] Jan 27 15:19:31 crc kubenswrapper[4772]: W0127 15:19:31.971600 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2914eab0_19c8_464b_a774_d30a492f6763.slice/crio-6c6fdd92e5a1aee735a193b46e66d29f340458e8944dfdf216d0e03d42c2b335 WatchSource:0}: Error finding container 6c6fdd92e5a1aee735a193b46e66d29f340458e8944dfdf216d0e03d42c2b335: Status 404 returned error can't find the container with id 6c6fdd92e5a1aee735a193b46e66d29f340458e8944dfdf216d0e03d42c2b335 Jan 27 15:19:32 crc kubenswrapper[4772]: I0127 15:19:32.057608 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjx8r" event={"ID":"09a099f0-cf75-4584-9369-e146e1898ee5","Type":"ContainerStarted","Data":"a779b499675d02b76b862fa155c06070ae6ad9b590cec73ad50ae9e5d0cd44a9"} Jan 27 15:19:32 crc kubenswrapper[4772]: I0127 15:19:32.059016 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-mdjph" event={"ID":"2914eab0-19c8-464b-a774-d30a492f6763","Type":"ContainerStarted","Data":"6c6fdd92e5a1aee735a193b46e66d29f340458e8944dfdf216d0e03d42c2b335"} Jan 27 15:19:33 crc kubenswrapper[4772]: I0127 15:19:33.066640 4772 generic.go:334] "Generic (PLEG): container finished" podID="09a099f0-cf75-4584-9369-e146e1898ee5" containerID="a779b499675d02b76b862fa155c06070ae6ad9b590cec73ad50ae9e5d0cd44a9" exitCode=0 Jan 27 15:19:33 crc kubenswrapper[4772]: I0127 15:19:33.066681 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjx8r" event={"ID":"09a099f0-cf75-4584-9369-e146e1898ee5","Type":"ContainerDied","Data":"a779b499675d02b76b862fa155c06070ae6ad9b590cec73ad50ae9e5d0cd44a9"} Jan 27 15:19:35 crc kubenswrapper[4772]: I0127 15:19:35.076567 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-mdjph" event={"ID":"2914eab0-19c8-464b-a774-d30a492f6763","Type":"ContainerStarted","Data":"15e2a155bf300de194c72a842f30d8c0505b2430fc699c4937e38716d89daf3e"} Jan 27 15:19:35 crc kubenswrapper[4772]: I0127 15:19:35.079561 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjx8r" event={"ID":"09a099f0-cf75-4584-9369-e146e1898ee5","Type":"ContainerStarted","Data":"c89368d0949f4bd3cbe349eab2729e718b9b735415848548f5754e07c04a3cef"} Jan 27 15:19:35 crc kubenswrapper[4772]: I0127 15:19:35.099184 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-646758c888-mdjph" podStartSLOduration=2.004480249 podStartE2EDuration="4.09915238s" podCreationTimestamp="2026-01-27 15:19:31 +0000 UTC" firstStartedPulling="2026-01-27 15:19:31.97434937 +0000 UTC m=+757.954958468" lastFinishedPulling="2026-01-27 15:19:34.069021501 +0000 UTC m=+760.049630599" observedRunningTime="2026-01-27 15:19:35.095484415 +0000 UTC m=+761.076093513" watchObservedRunningTime="2026-01-27 15:19:35.09915238 +0000 UTC m=+761.079761478" Jan 27 15:19:35 crc kubenswrapper[4772]: I0127 15:19:35.121237 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wjx8r" podStartSLOduration=3.106218951 podStartE2EDuration="6.121208746s" podCreationTimestamp="2026-01-27 15:19:29 +0000 UTC" firstStartedPulling="2026-01-27 15:19:31.052306666 +0000 UTC m=+757.032915764" lastFinishedPulling="2026-01-27 15:19:34.067296461 +0000 UTC m=+760.047905559" observedRunningTime="2026-01-27 15:19:35.120345321 +0000 UTC m=+761.100954479" watchObservedRunningTime="2026-01-27 15:19:35.121208746 +0000 UTC m=+761.101817884" Jan 27 15:19:35 crc kubenswrapper[4772]: I0127 15:19:35.951260 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-g7d66"] Jan 27 15:19:35 crc kubenswrapper[4772]: I0127 15:19:35.952798 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-g7d66" Jan 27 15:19:35 crc kubenswrapper[4772]: I0127 15:19:35.954981 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-9xfrz" Jan 27 15:19:35 crc kubenswrapper[4772]: I0127 15:19:35.956479 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-gpfqc"] Jan 27 15:19:35 crc kubenswrapper[4772]: I0127 15:19:35.957332 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gpfqc" Jan 27 15:19:35 crc kubenswrapper[4772]: I0127 15:19:35.961414 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Jan 27 15:19:35 crc kubenswrapper[4772]: I0127 15:19:35.964275 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-g7d66"] Jan 27 15:19:35 crc kubenswrapper[4772]: I0127 15:19:35.982703 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-gpfqc"] Jan 27 15:19:35 crc kubenswrapper[4772]: I0127 15:19:35.986504 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-mmdc5"] Jan 27 15:19:35 crc kubenswrapper[4772]: I0127 15:19:35.987247 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-mmdc5" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.000809 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/cf21b49c-f01b-4c7c-bdb9-57e115b364d9-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-gpfqc\" (UID: \"cf21b49c-f01b-4c7c-bdb9-57e115b364d9\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gpfqc" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.000879 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp8kd\" (UniqueName: \"kubernetes.io/projected/d834ccf6-9b3a-4a3e-8980-7f0a102babd0-kube-api-access-hp8kd\") pod \"nmstate-handler-mmdc5\" (UID: \"d834ccf6-9b3a-4a3e-8980-7f0a102babd0\") " pod="openshift-nmstate/nmstate-handler-mmdc5" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.000924 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d834ccf6-9b3a-4a3e-8980-7f0a102babd0-nmstate-lock\") pod \"nmstate-handler-mmdc5\" (UID: \"d834ccf6-9b3a-4a3e-8980-7f0a102babd0\") " pod="openshift-nmstate/nmstate-handler-mmdc5" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.001029 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d834ccf6-9b3a-4a3e-8980-7f0a102babd0-dbus-socket\") pod \"nmstate-handler-mmdc5\" (UID: \"d834ccf6-9b3a-4a3e-8980-7f0a102babd0\") " pod="openshift-nmstate/nmstate-handler-mmdc5" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.001061 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rjm9\" (UniqueName: \"kubernetes.io/projected/cf21b49c-f01b-4c7c-bdb9-57e115b364d9-kube-api-access-8rjm9\") pod \"nmstate-webhook-8474b5b9d8-gpfqc\" (UID: \"cf21b49c-f01b-4c7c-bdb9-57e115b364d9\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gpfqc" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.001110 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d834ccf6-9b3a-4a3e-8980-7f0a102babd0-ovs-socket\") pod \"nmstate-handler-mmdc5\" (UID: \"d834ccf6-9b3a-4a3e-8980-7f0a102babd0\") " pod="openshift-nmstate/nmstate-handler-mmdc5" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.001142 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2skgs\" (UniqueName: \"kubernetes.io/projected/004d59b7-1d3b-41af-8c3d-c6562dd9716a-kube-api-access-2skgs\") pod \"nmstate-metrics-54757c584b-g7d66\" (UID: \"004d59b7-1d3b-41af-8c3d-c6562dd9716a\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-g7d66" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.092225 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-jlq99"] Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.093009 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-jlq99" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.095836 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-cc7j4" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.096044 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.096188 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.105379 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3706a5f9-4370-4cca-abb9-b23e8b9c828f-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-jlq99\" (UID: \"3706a5f9-4370-4cca-abb9-b23e8b9c828f\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-jlq99" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.105444 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d834ccf6-9b3a-4a3e-8980-7f0a102babd0-ovs-socket\") pod \"nmstate-handler-mmdc5\" (UID: \"d834ccf6-9b3a-4a3e-8980-7f0a102babd0\") " pod="openshift-nmstate/nmstate-handler-mmdc5" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.105480 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2skgs\" (UniqueName: \"kubernetes.io/projected/004d59b7-1d3b-41af-8c3d-c6562dd9716a-kube-api-access-2skgs\") pod \"nmstate-metrics-54757c584b-g7d66\" (UID: \"004d59b7-1d3b-41af-8c3d-c6562dd9716a\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-g7d66" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.105517 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3706a5f9-4370-4cca-abb9-b23e8b9c828f-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-jlq99\" (UID: \"3706a5f9-4370-4cca-abb9-b23e8b9c828f\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-jlq99" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.105541 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/cf21b49c-f01b-4c7c-bdb9-57e115b364d9-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-gpfqc\" (UID: \"cf21b49c-f01b-4c7c-bdb9-57e115b364d9\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gpfqc" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.105578 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp8kd\" (UniqueName: \"kubernetes.io/projected/d834ccf6-9b3a-4a3e-8980-7f0a102babd0-kube-api-access-hp8kd\") pod \"nmstate-handler-mmdc5\" (UID: \"d834ccf6-9b3a-4a3e-8980-7f0a102babd0\") " pod="openshift-nmstate/nmstate-handler-mmdc5" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.105621 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d834ccf6-9b3a-4a3e-8980-7f0a102babd0-nmstate-lock\") pod \"nmstate-handler-mmdc5\" (UID: \"d834ccf6-9b3a-4a3e-8980-7f0a102babd0\") " pod="openshift-nmstate/nmstate-handler-mmdc5" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.105667 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlw6p\" (UniqueName: \"kubernetes.io/projected/3706a5f9-4370-4cca-abb9-b23e8b9c828f-kube-api-access-mlw6p\") pod \"nmstate-console-plugin-7754f76f8b-jlq99\" (UID: \"3706a5f9-4370-4cca-abb9-b23e8b9c828f\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-jlq99" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.105701 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d834ccf6-9b3a-4a3e-8980-7f0a102babd0-dbus-socket\") pod \"nmstate-handler-mmdc5\" (UID: \"d834ccf6-9b3a-4a3e-8980-7f0a102babd0\") " pod="openshift-nmstate/nmstate-handler-mmdc5" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.105898 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rjm9\" (UniqueName: \"kubernetes.io/projected/cf21b49c-f01b-4c7c-bdb9-57e115b364d9-kube-api-access-8rjm9\") pod \"nmstate-webhook-8474b5b9d8-gpfqc\" (UID: \"cf21b49c-f01b-4c7c-bdb9-57e115b364d9\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gpfqc" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.105958 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d834ccf6-9b3a-4a3e-8980-7f0a102babd0-ovs-socket\") pod \"nmstate-handler-mmdc5\" (UID: \"d834ccf6-9b3a-4a3e-8980-7f0a102babd0\") " pod="openshift-nmstate/nmstate-handler-mmdc5" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.106008 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d834ccf6-9b3a-4a3e-8980-7f0a102babd0-dbus-socket\") pod \"nmstate-handler-mmdc5\" (UID: \"d834ccf6-9b3a-4a3e-8980-7f0a102babd0\") " pod="openshift-nmstate/nmstate-handler-mmdc5" Jan 27 15:19:36 crc kubenswrapper[4772]: E0127 15:19:36.106106 4772 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Jan 27 15:19:36 crc kubenswrapper[4772]: E0127 15:19:36.106145 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cf21b49c-f01b-4c7c-bdb9-57e115b364d9-tls-key-pair podName:cf21b49c-f01b-4c7c-bdb9-57e115b364d9 nodeName:}" failed. No retries permitted until 2026-01-27 15:19:36.606130313 +0000 UTC m=+762.586739411 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/cf21b49c-f01b-4c7c-bdb9-57e115b364d9-tls-key-pair") pod "nmstate-webhook-8474b5b9d8-gpfqc" (UID: "cf21b49c-f01b-4c7c-bdb9-57e115b364d9") : secret "openshift-nmstate-webhook" not found Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.106479 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d834ccf6-9b3a-4a3e-8980-7f0a102babd0-nmstate-lock\") pod \"nmstate-handler-mmdc5\" (UID: \"d834ccf6-9b3a-4a3e-8980-7f0a102babd0\") " pod="openshift-nmstate/nmstate-handler-mmdc5" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.108356 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-jlq99"] Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.137881 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp8kd\" (UniqueName: \"kubernetes.io/projected/d834ccf6-9b3a-4a3e-8980-7f0a102babd0-kube-api-access-hp8kd\") pod \"nmstate-handler-mmdc5\" (UID: \"d834ccf6-9b3a-4a3e-8980-7f0a102babd0\") " pod="openshift-nmstate/nmstate-handler-mmdc5" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.140138 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2skgs\" (UniqueName: \"kubernetes.io/projected/004d59b7-1d3b-41af-8c3d-c6562dd9716a-kube-api-access-2skgs\") pod \"nmstate-metrics-54757c584b-g7d66\" (UID: \"004d59b7-1d3b-41af-8c3d-c6562dd9716a\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-g7d66" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.145932 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rjm9\" (UniqueName: \"kubernetes.io/projected/cf21b49c-f01b-4c7c-bdb9-57e115b364d9-kube-api-access-8rjm9\") pod \"nmstate-webhook-8474b5b9d8-gpfqc\" (UID: \"cf21b49c-f01b-4c7c-bdb9-57e115b364d9\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gpfqc" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.207355 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3706a5f9-4370-4cca-abb9-b23e8b9c828f-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-jlq99\" (UID: \"3706a5f9-4370-4cca-abb9-b23e8b9c828f\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-jlq99" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.207492 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlw6p\" (UniqueName: \"kubernetes.io/projected/3706a5f9-4370-4cca-abb9-b23e8b9c828f-kube-api-access-mlw6p\") pod \"nmstate-console-plugin-7754f76f8b-jlq99\" (UID: \"3706a5f9-4370-4cca-abb9-b23e8b9c828f\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-jlq99" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.207550 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3706a5f9-4370-4cca-abb9-b23e8b9c828f-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-jlq99\" (UID: \"3706a5f9-4370-4cca-abb9-b23e8b9c828f\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-jlq99" Jan 27 15:19:36 crc kubenswrapper[4772]: E0127 15:19:36.208158 4772 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Jan 27 15:19:36 crc kubenswrapper[4772]: E0127 15:19:36.208258 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3706a5f9-4370-4cca-abb9-b23e8b9c828f-plugin-serving-cert podName:3706a5f9-4370-4cca-abb9-b23e8b9c828f nodeName:}" failed. No retries permitted until 2026-01-27 15:19:36.708236866 +0000 UTC m=+762.688845974 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/3706a5f9-4370-4cca-abb9-b23e8b9c828f-plugin-serving-cert") pod "nmstate-console-plugin-7754f76f8b-jlq99" (UID: "3706a5f9-4370-4cca-abb9-b23e8b9c828f") : secret "plugin-serving-cert" not found Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.210548 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3706a5f9-4370-4cca-abb9-b23e8b9c828f-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-jlq99\" (UID: \"3706a5f9-4370-4cca-abb9-b23e8b9c828f\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-jlq99" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.232657 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlw6p\" (UniqueName: \"kubernetes.io/projected/3706a5f9-4370-4cca-abb9-b23e8b9c828f-kube-api-access-mlw6p\") pod \"nmstate-console-plugin-7754f76f8b-jlq99\" (UID: \"3706a5f9-4370-4cca-abb9-b23e8b9c828f\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-jlq99" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.274679 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-g7d66" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.290425 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7587874955-cvhbn"] Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.291073 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.316450 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b88eb902-4d85-4dc3-9336-daec648107cf-trusted-ca-bundle\") pod \"console-7587874955-cvhbn\" (UID: \"b88eb902-4d85-4dc3-9336-daec648107cf\") " pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.316489 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b88eb902-4d85-4dc3-9336-daec648107cf-oauth-serving-cert\") pod \"console-7587874955-cvhbn\" (UID: \"b88eb902-4d85-4dc3-9336-daec648107cf\") " pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.316516 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh7rn\" (UniqueName: \"kubernetes.io/projected/b88eb902-4d85-4dc3-9336-daec648107cf-kube-api-access-sh7rn\") pod \"console-7587874955-cvhbn\" (UID: \"b88eb902-4d85-4dc3-9336-daec648107cf\") " pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.316639 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b88eb902-4d85-4dc3-9336-daec648107cf-service-ca\") pod \"console-7587874955-cvhbn\" (UID: \"b88eb902-4d85-4dc3-9336-daec648107cf\") " pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.316683 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b88eb902-4d85-4dc3-9336-daec648107cf-console-oauth-config\") pod \"console-7587874955-cvhbn\" (UID: \"b88eb902-4d85-4dc3-9336-daec648107cf\") " pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.316761 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b88eb902-4d85-4dc3-9336-daec648107cf-console-config\") pod \"console-7587874955-cvhbn\" (UID: \"b88eb902-4d85-4dc3-9336-daec648107cf\") " pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.316831 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b88eb902-4d85-4dc3-9336-daec648107cf-console-serving-cert\") pod \"console-7587874955-cvhbn\" (UID: \"b88eb902-4d85-4dc3-9336-daec648107cf\") " pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.318032 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-mmdc5" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.320065 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7587874955-cvhbn"] Jan 27 15:19:36 crc kubenswrapper[4772]: W0127 15:19:36.345543 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd834ccf6_9b3a_4a3e_8980_7f0a102babd0.slice/crio-74cd6762872bf6b32e6299908edcfd6a6e011bf2e45c20e81a5fc62f98313049 WatchSource:0}: Error finding container 74cd6762872bf6b32e6299908edcfd6a6e011bf2e45c20e81a5fc62f98313049: Status 404 returned error can't find the container with id 74cd6762872bf6b32e6299908edcfd6a6e011bf2e45c20e81a5fc62f98313049 Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.417801 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b88eb902-4d85-4dc3-9336-daec648107cf-console-serving-cert\") pod \"console-7587874955-cvhbn\" (UID: \"b88eb902-4d85-4dc3-9336-daec648107cf\") " pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.418277 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b88eb902-4d85-4dc3-9336-daec648107cf-trusted-ca-bundle\") pod \"console-7587874955-cvhbn\" (UID: \"b88eb902-4d85-4dc3-9336-daec648107cf\") " pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.418300 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b88eb902-4d85-4dc3-9336-daec648107cf-oauth-serving-cert\") pod \"console-7587874955-cvhbn\" (UID: \"b88eb902-4d85-4dc3-9336-daec648107cf\") " pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.418326 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh7rn\" (UniqueName: \"kubernetes.io/projected/b88eb902-4d85-4dc3-9336-daec648107cf-kube-api-access-sh7rn\") pod \"console-7587874955-cvhbn\" (UID: \"b88eb902-4d85-4dc3-9336-daec648107cf\") " pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.418364 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b88eb902-4d85-4dc3-9336-daec648107cf-service-ca\") pod \"console-7587874955-cvhbn\" (UID: \"b88eb902-4d85-4dc3-9336-daec648107cf\") " pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.418391 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b88eb902-4d85-4dc3-9336-daec648107cf-console-oauth-config\") pod \"console-7587874955-cvhbn\" (UID: \"b88eb902-4d85-4dc3-9336-daec648107cf\") " pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.418440 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b88eb902-4d85-4dc3-9336-daec648107cf-console-config\") pod \"console-7587874955-cvhbn\" (UID: \"b88eb902-4d85-4dc3-9336-daec648107cf\") " pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.419933 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b88eb902-4d85-4dc3-9336-daec648107cf-oauth-serving-cert\") pod \"console-7587874955-cvhbn\" (UID: \"b88eb902-4d85-4dc3-9336-daec648107cf\") " pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.420027 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b88eb902-4d85-4dc3-9336-daec648107cf-service-ca\") pod \"console-7587874955-cvhbn\" (UID: \"b88eb902-4d85-4dc3-9336-daec648107cf\") " pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.422026 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b88eb902-4d85-4dc3-9336-daec648107cf-trusted-ca-bundle\") pod \"console-7587874955-cvhbn\" (UID: \"b88eb902-4d85-4dc3-9336-daec648107cf\") " pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.423915 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b88eb902-4d85-4dc3-9336-daec648107cf-console-config\") pod \"console-7587874955-cvhbn\" (UID: \"b88eb902-4d85-4dc3-9336-daec648107cf\") " pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.425138 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b88eb902-4d85-4dc3-9336-daec648107cf-console-oauth-config\") pod \"console-7587874955-cvhbn\" (UID: \"b88eb902-4d85-4dc3-9336-daec648107cf\") " pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.425182 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b88eb902-4d85-4dc3-9336-daec648107cf-console-serving-cert\") pod \"console-7587874955-cvhbn\" (UID: \"b88eb902-4d85-4dc3-9336-daec648107cf\") " pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.442718 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh7rn\" (UniqueName: \"kubernetes.io/projected/b88eb902-4d85-4dc3-9336-daec648107cf-kube-api-access-sh7rn\") pod \"console-7587874955-cvhbn\" (UID: \"b88eb902-4d85-4dc3-9336-daec648107cf\") " pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.533458 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-g7d66"] Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.620994 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/cf21b49c-f01b-4c7c-bdb9-57e115b364d9-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-gpfqc\" (UID: \"cf21b49c-f01b-4c7c-bdb9-57e115b364d9\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gpfqc" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.624950 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/cf21b49c-f01b-4c7c-bdb9-57e115b364d9-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-gpfqc\" (UID: \"cf21b49c-f01b-4c7c-bdb9-57e115b364d9\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gpfqc" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.649128 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.721868 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3706a5f9-4370-4cca-abb9-b23e8b9c828f-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-jlq99\" (UID: \"3706a5f9-4370-4cca-abb9-b23e8b9c828f\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-jlq99" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.725432 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3706a5f9-4370-4cca-abb9-b23e8b9c828f-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-jlq99\" (UID: \"3706a5f9-4370-4cca-abb9-b23e8b9c828f\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-jlq99" Jan 27 15:19:36 crc kubenswrapper[4772]: I0127 15:19:36.883921 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gpfqc" Jan 27 15:19:37 crc kubenswrapper[4772]: I0127 15:19:37.020343 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-jlq99" Jan 27 15:19:37 crc kubenswrapper[4772]: I0127 15:19:37.049507 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-gpfqc"] Jan 27 15:19:37 crc kubenswrapper[4772]: W0127 15:19:37.056757 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf21b49c_f01b_4c7c_bdb9_57e115b364d9.slice/crio-cb4a0bda422e0b5b92ec8b7453b03549dbfb27214199a98cdb713ea2cb4962a5 WatchSource:0}: Error finding container cb4a0bda422e0b5b92ec8b7453b03549dbfb27214199a98cdb713ea2cb4962a5: Status 404 returned error can't find the container with id cb4a0bda422e0b5b92ec8b7453b03549dbfb27214199a98cdb713ea2cb4962a5 Jan 27 15:19:37 crc kubenswrapper[4772]: I0127 15:19:37.096151 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-g7d66" event={"ID":"004d59b7-1d3b-41af-8c3d-c6562dd9716a","Type":"ContainerStarted","Data":"bbc361de727bd2369e0573f411a5acac824bb4d34535d084ae81adcfc98bf23f"} Jan 27 15:19:37 crc kubenswrapper[4772]: I0127 15:19:37.097236 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-mmdc5" event={"ID":"d834ccf6-9b3a-4a3e-8980-7f0a102babd0","Type":"ContainerStarted","Data":"74cd6762872bf6b32e6299908edcfd6a6e011bf2e45c20e81a5fc62f98313049"} Jan 27 15:19:37 crc kubenswrapper[4772]: I0127 15:19:37.098351 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gpfqc" event={"ID":"cf21b49c-f01b-4c7c-bdb9-57e115b364d9","Type":"ContainerStarted","Data":"cb4a0bda422e0b5b92ec8b7453b03549dbfb27214199a98cdb713ea2cb4962a5"} Jan 27 15:19:37 crc kubenswrapper[4772]: I0127 15:19:37.101835 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7587874955-cvhbn"] Jan 27 15:19:37 crc kubenswrapper[4772]: W0127 15:19:37.124713 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb88eb902_4d85_4dc3_9336_daec648107cf.slice/crio-8f0d2778bdef050f5fb3d11428ef1cb145879b491c8be1450efa14ff19e4b0f5 WatchSource:0}: Error finding container 8f0d2778bdef050f5fb3d11428ef1cb145879b491c8be1450efa14ff19e4b0f5: Status 404 returned error can't find the container with id 8f0d2778bdef050f5fb3d11428ef1cb145879b491c8be1450efa14ff19e4b0f5 Jan 27 15:19:37 crc kubenswrapper[4772]: I0127 15:19:37.256568 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-jlq99"] Jan 27 15:19:37 crc kubenswrapper[4772]: W0127 15:19:37.263431 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3706a5f9_4370_4cca_abb9_b23e8b9c828f.slice/crio-b16dd94e3004bea0e6593538807931500b18201385283e58cf176558aa7a8a91 WatchSource:0}: Error finding container b16dd94e3004bea0e6593538807931500b18201385283e58cf176558aa7a8a91: Status 404 returned error can't find the container with id b16dd94e3004bea0e6593538807931500b18201385283e58cf176558aa7a8a91 Jan 27 15:19:38 crc kubenswrapper[4772]: I0127 15:19:38.107673 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-jlq99" event={"ID":"3706a5f9-4370-4cca-abb9-b23e8b9c828f","Type":"ContainerStarted","Data":"b16dd94e3004bea0e6593538807931500b18201385283e58cf176558aa7a8a91"} Jan 27 15:19:38 crc kubenswrapper[4772]: I0127 15:19:38.110353 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7587874955-cvhbn" event={"ID":"b88eb902-4d85-4dc3-9336-daec648107cf","Type":"ContainerStarted","Data":"c2d567fb0b2f6dbdef0f4485f0fba6e81e898002431a48fc7faa6d20c6dde697"} Jan 27 15:19:38 crc kubenswrapper[4772]: I0127 15:19:38.110383 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7587874955-cvhbn" event={"ID":"b88eb902-4d85-4dc3-9336-daec648107cf","Type":"ContainerStarted","Data":"8f0d2778bdef050f5fb3d11428ef1cb145879b491c8be1450efa14ff19e4b0f5"} Jan 27 15:19:38 crc kubenswrapper[4772]: I0127 15:19:38.150341 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7587874955-cvhbn" podStartSLOduration=2.150321828 podStartE2EDuration="2.150321828s" podCreationTimestamp="2026-01-27 15:19:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:19:38.1466067 +0000 UTC m=+764.127215798" watchObservedRunningTime="2026-01-27 15:19:38.150321828 +0000 UTC m=+764.130930946" Jan 27 15:19:39 crc kubenswrapper[4772]: I0127 15:19:39.830320 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wjx8r" Jan 27 15:19:39 crc kubenswrapper[4772]: I0127 15:19:39.830637 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wjx8r" Jan 27 15:19:40 crc kubenswrapper[4772]: I0127 15:19:40.865272 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wjx8r" podUID="09a099f0-cf75-4584-9369-e146e1898ee5" containerName="registry-server" probeResult="failure" output=< Jan 27 15:19:40 crc kubenswrapper[4772]: timeout: failed to connect service ":50051" within 1s Jan 27 15:19:40 crc kubenswrapper[4772]: > Jan 27 15:19:41 crc kubenswrapper[4772]: I0127 15:19:41.144026 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gpfqc" event={"ID":"cf21b49c-f01b-4c7c-bdb9-57e115b364d9","Type":"ContainerStarted","Data":"936192f8fc1de4210fb2ca33015efcb6990e6ba9a61d80f6df411fc8be3e1edc"} Jan 27 15:19:41 crc kubenswrapper[4772]: I0127 15:19:41.144220 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gpfqc" Jan 27 15:19:41 crc kubenswrapper[4772]: I0127 15:19:41.147957 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-g7d66" event={"ID":"004d59b7-1d3b-41af-8c3d-c6562dd9716a","Type":"ContainerStarted","Data":"c90797f7a014cf412ef32210553a472dee22c58f527a9768fc7f51ad8be45f87"} Jan 27 15:19:41 crc kubenswrapper[4772]: I0127 15:19:41.150765 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-mmdc5" event={"ID":"d834ccf6-9b3a-4a3e-8980-7f0a102babd0","Type":"ContainerStarted","Data":"a907a3c154ee056d510026b823e6420e785bf97e6a98b4c06bb43b132e2ed452"} Jan 27 15:19:41 crc kubenswrapper[4772]: I0127 15:19:41.150989 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-mmdc5" Jan 27 15:19:41 crc kubenswrapper[4772]: I0127 15:19:41.161289 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gpfqc" podStartSLOduration=3.071497556 podStartE2EDuration="6.161269216s" podCreationTimestamp="2026-01-27 15:19:35 +0000 UTC" firstStartedPulling="2026-01-27 15:19:37.060159229 +0000 UTC m=+763.040768327" lastFinishedPulling="2026-01-27 15:19:40.149930889 +0000 UTC m=+766.130539987" observedRunningTime="2026-01-27 15:19:41.160203775 +0000 UTC m=+767.140812893" watchObservedRunningTime="2026-01-27 15:19:41.161269216 +0000 UTC m=+767.141878314" Jan 27 15:19:41 crc kubenswrapper[4772]: I0127 15:19:41.183307 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-mmdc5" podStartSLOduration=2.353455832 podStartE2EDuration="6.183286851s" podCreationTimestamp="2026-01-27 15:19:35 +0000 UTC" firstStartedPulling="2026-01-27 15:19:36.350790294 +0000 UTC m=+762.331399392" lastFinishedPulling="2026-01-27 15:19:40.180621313 +0000 UTC m=+766.161230411" observedRunningTime="2026-01-27 15:19:41.182137817 +0000 UTC m=+767.162746935" watchObservedRunningTime="2026-01-27 15:19:41.183286851 +0000 UTC m=+767.163895949" Jan 27 15:19:42 crc kubenswrapper[4772]: I0127 15:19:42.059021 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:19:42 crc kubenswrapper[4772]: I0127 15:19:42.059091 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:19:43 crc kubenswrapper[4772]: I0127 15:19:43.165696 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-g7d66" event={"ID":"004d59b7-1d3b-41af-8c3d-c6562dd9716a","Type":"ContainerStarted","Data":"3e8f259bfabcaf4122367dc58f7e2c3cab40fdbf3006a7ed27c8bd4cfc4c525c"} Jan 27 15:19:43 crc kubenswrapper[4772]: I0127 15:19:43.189059 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-54757c584b-g7d66" podStartSLOduration=2.200072662 podStartE2EDuration="8.189039939s" podCreationTimestamp="2026-01-27 15:19:35 +0000 UTC" firstStartedPulling="2026-01-27 15:19:36.557567694 +0000 UTC m=+762.538176792" lastFinishedPulling="2026-01-27 15:19:42.546534931 +0000 UTC m=+768.527144069" observedRunningTime="2026-01-27 15:19:43.183092768 +0000 UTC m=+769.163701876" watchObservedRunningTime="2026-01-27 15:19:43.189039939 +0000 UTC m=+769.169649037" Jan 27 15:19:46 crc kubenswrapper[4772]: I0127 15:19:46.352534 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-mmdc5" Jan 27 15:19:46 crc kubenswrapper[4772]: I0127 15:19:46.649304 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:46 crc kubenswrapper[4772]: I0127 15:19:46.649367 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:46 crc kubenswrapper[4772]: I0127 15:19:46.659374 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:47 crc kubenswrapper[4772]: I0127 15:19:47.198032 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7587874955-cvhbn" Jan 27 15:19:47 crc kubenswrapper[4772]: I0127 15:19:47.251264 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-7qfrl"] Jan 27 15:19:49 crc kubenswrapper[4772]: I0127 15:19:49.869368 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wjx8r" Jan 27 15:19:49 crc kubenswrapper[4772]: I0127 15:19:49.959428 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wjx8r" Jan 27 15:19:50 crc kubenswrapper[4772]: I0127 15:19:50.121976 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wjx8r"] Jan 27 15:19:51 crc kubenswrapper[4772]: I0127 15:19:51.220586 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-jlq99" event={"ID":"3706a5f9-4370-4cca-abb9-b23e8b9c828f","Type":"ContainerStarted","Data":"1bf82255d91ce78638063827979b83642dcb63bc78c82d7f9b9cce2badea6e57"} Jan 27 15:19:51 crc kubenswrapper[4772]: I0127 15:19:51.220761 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wjx8r" podUID="09a099f0-cf75-4584-9369-e146e1898ee5" containerName="registry-server" containerID="cri-o://c89368d0949f4bd3cbe349eab2729e718b9b735415848548f5754e07c04a3cef" gracePeriod=2 Jan 27 15:19:51 crc kubenswrapper[4772]: I0127 15:19:51.260007 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-jlq99" podStartSLOduration=2.230594602 podStartE2EDuration="15.259971589s" podCreationTimestamp="2026-01-27 15:19:36 +0000 UTC" firstStartedPulling="2026-01-27 15:19:37.265919129 +0000 UTC m=+763.246528217" lastFinishedPulling="2026-01-27 15:19:50.295296116 +0000 UTC m=+776.275905204" observedRunningTime="2026-01-27 15:19:51.246341066 +0000 UTC m=+777.226950204" watchObservedRunningTime="2026-01-27 15:19:51.259971589 +0000 UTC m=+777.240580727" Jan 27 15:19:51 crc kubenswrapper[4772]: I0127 15:19:51.554484 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wjx8r" Jan 27 15:19:51 crc kubenswrapper[4772]: I0127 15:19:51.575273 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09a099f0-cf75-4584-9369-e146e1898ee5-utilities\") pod \"09a099f0-cf75-4584-9369-e146e1898ee5\" (UID: \"09a099f0-cf75-4584-9369-e146e1898ee5\") " Jan 27 15:19:51 crc kubenswrapper[4772]: I0127 15:19:51.575364 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbc5f\" (UniqueName: \"kubernetes.io/projected/09a099f0-cf75-4584-9369-e146e1898ee5-kube-api-access-pbc5f\") pod \"09a099f0-cf75-4584-9369-e146e1898ee5\" (UID: \"09a099f0-cf75-4584-9369-e146e1898ee5\") " Jan 27 15:19:51 crc kubenswrapper[4772]: I0127 15:19:51.575490 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09a099f0-cf75-4584-9369-e146e1898ee5-catalog-content\") pod \"09a099f0-cf75-4584-9369-e146e1898ee5\" (UID: \"09a099f0-cf75-4584-9369-e146e1898ee5\") " Jan 27 15:19:51 crc kubenswrapper[4772]: I0127 15:19:51.580449 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09a099f0-cf75-4584-9369-e146e1898ee5-utilities" (OuterVolumeSpecName: "utilities") pod "09a099f0-cf75-4584-9369-e146e1898ee5" (UID: "09a099f0-cf75-4584-9369-e146e1898ee5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:19:51 crc kubenswrapper[4772]: I0127 15:19:51.580887 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09a099f0-cf75-4584-9369-e146e1898ee5-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 15:19:51 crc kubenswrapper[4772]: I0127 15:19:51.581155 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09a099f0-cf75-4584-9369-e146e1898ee5-kube-api-access-pbc5f" (OuterVolumeSpecName: "kube-api-access-pbc5f") pod "09a099f0-cf75-4584-9369-e146e1898ee5" (UID: "09a099f0-cf75-4584-9369-e146e1898ee5"). InnerVolumeSpecName "kube-api-access-pbc5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:19:51 crc kubenswrapper[4772]: I0127 15:19:51.682331 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbc5f\" (UniqueName: \"kubernetes.io/projected/09a099f0-cf75-4584-9369-e146e1898ee5-kube-api-access-pbc5f\") on node \"crc\" DevicePath \"\"" Jan 27 15:19:51 crc kubenswrapper[4772]: I0127 15:19:51.731677 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09a099f0-cf75-4584-9369-e146e1898ee5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09a099f0-cf75-4584-9369-e146e1898ee5" (UID: "09a099f0-cf75-4584-9369-e146e1898ee5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:19:51 crc kubenswrapper[4772]: I0127 15:19:51.783913 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09a099f0-cf75-4584-9369-e146e1898ee5-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 15:19:52 crc kubenswrapper[4772]: I0127 15:19:52.236944 4772 generic.go:334] "Generic (PLEG): container finished" podID="09a099f0-cf75-4584-9369-e146e1898ee5" containerID="c89368d0949f4bd3cbe349eab2729e718b9b735415848548f5754e07c04a3cef" exitCode=0 Jan 27 15:19:52 crc kubenswrapper[4772]: I0127 15:19:52.237013 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wjx8r" Jan 27 15:19:52 crc kubenswrapper[4772]: I0127 15:19:52.237061 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjx8r" event={"ID":"09a099f0-cf75-4584-9369-e146e1898ee5","Type":"ContainerDied","Data":"c89368d0949f4bd3cbe349eab2729e718b9b735415848548f5754e07c04a3cef"} Jan 27 15:19:52 crc kubenswrapper[4772]: I0127 15:19:52.237123 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wjx8r" event={"ID":"09a099f0-cf75-4584-9369-e146e1898ee5","Type":"ContainerDied","Data":"5400ef2d30491c889dad91bc65f72f745b56aba86dca10ba2c7a00a785a6ae51"} Jan 27 15:19:52 crc kubenswrapper[4772]: I0127 15:19:52.237152 4772 scope.go:117] "RemoveContainer" containerID="c89368d0949f4bd3cbe349eab2729e718b9b735415848548f5754e07c04a3cef" Jan 27 15:19:52 crc kubenswrapper[4772]: I0127 15:19:52.261886 4772 scope.go:117] "RemoveContainer" containerID="a779b499675d02b76b862fa155c06070ae6ad9b590cec73ad50ae9e5d0cd44a9" Jan 27 15:19:52 crc kubenswrapper[4772]: I0127 15:19:52.276332 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wjx8r"] Jan 27 15:19:52 crc kubenswrapper[4772]: I0127 15:19:52.282148 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wjx8r"] Jan 27 15:19:52 crc kubenswrapper[4772]: I0127 15:19:52.298963 4772 scope.go:117] "RemoveContainer" containerID="db8ffbbcaeec3abdc8460ac13f60a94b11906e285d02efbfc2441a057fa123e1" Jan 27 15:19:52 crc kubenswrapper[4772]: I0127 15:19:52.313431 4772 scope.go:117] "RemoveContainer" containerID="c89368d0949f4bd3cbe349eab2729e718b9b735415848548f5754e07c04a3cef" Jan 27 15:19:52 crc kubenswrapper[4772]: E0127 15:19:52.314109 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c89368d0949f4bd3cbe349eab2729e718b9b735415848548f5754e07c04a3cef\": container with ID starting with c89368d0949f4bd3cbe349eab2729e718b9b735415848548f5754e07c04a3cef not found: ID does not exist" containerID="c89368d0949f4bd3cbe349eab2729e718b9b735415848548f5754e07c04a3cef" Jan 27 15:19:52 crc kubenswrapper[4772]: I0127 15:19:52.314159 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c89368d0949f4bd3cbe349eab2729e718b9b735415848548f5754e07c04a3cef"} err="failed to get container status \"c89368d0949f4bd3cbe349eab2729e718b9b735415848548f5754e07c04a3cef\": rpc error: code = NotFound desc = could not find container \"c89368d0949f4bd3cbe349eab2729e718b9b735415848548f5754e07c04a3cef\": container with ID starting with c89368d0949f4bd3cbe349eab2729e718b9b735415848548f5754e07c04a3cef not found: ID does not exist" Jan 27 15:19:52 crc kubenswrapper[4772]: I0127 15:19:52.314203 4772 scope.go:117] "RemoveContainer" containerID="a779b499675d02b76b862fa155c06070ae6ad9b590cec73ad50ae9e5d0cd44a9" Jan 27 15:19:52 crc kubenswrapper[4772]: E0127 15:19:52.314613 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a779b499675d02b76b862fa155c06070ae6ad9b590cec73ad50ae9e5d0cd44a9\": container with ID starting with a779b499675d02b76b862fa155c06070ae6ad9b590cec73ad50ae9e5d0cd44a9 not found: ID does not exist" containerID="a779b499675d02b76b862fa155c06070ae6ad9b590cec73ad50ae9e5d0cd44a9" Jan 27 15:19:52 crc kubenswrapper[4772]: I0127 15:19:52.314668 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a779b499675d02b76b862fa155c06070ae6ad9b590cec73ad50ae9e5d0cd44a9"} err="failed to get container status \"a779b499675d02b76b862fa155c06070ae6ad9b590cec73ad50ae9e5d0cd44a9\": rpc error: code = NotFound desc = could not find container \"a779b499675d02b76b862fa155c06070ae6ad9b590cec73ad50ae9e5d0cd44a9\": container with ID starting with a779b499675d02b76b862fa155c06070ae6ad9b590cec73ad50ae9e5d0cd44a9 not found: ID does not exist" Jan 27 15:19:52 crc kubenswrapper[4772]: I0127 15:19:52.314705 4772 scope.go:117] "RemoveContainer" containerID="db8ffbbcaeec3abdc8460ac13f60a94b11906e285d02efbfc2441a057fa123e1" Jan 27 15:19:52 crc kubenswrapper[4772]: E0127 15:19:52.315094 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db8ffbbcaeec3abdc8460ac13f60a94b11906e285d02efbfc2441a057fa123e1\": container with ID starting with db8ffbbcaeec3abdc8460ac13f60a94b11906e285d02efbfc2441a057fa123e1 not found: ID does not exist" containerID="db8ffbbcaeec3abdc8460ac13f60a94b11906e285d02efbfc2441a057fa123e1" Jan 27 15:19:52 crc kubenswrapper[4772]: I0127 15:19:52.315123 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db8ffbbcaeec3abdc8460ac13f60a94b11906e285d02efbfc2441a057fa123e1"} err="failed to get container status \"db8ffbbcaeec3abdc8460ac13f60a94b11906e285d02efbfc2441a057fa123e1\": rpc error: code = NotFound desc = could not find container \"db8ffbbcaeec3abdc8460ac13f60a94b11906e285d02efbfc2441a057fa123e1\": container with ID starting with db8ffbbcaeec3abdc8460ac13f60a94b11906e285d02efbfc2441a057fa123e1 not found: ID does not exist" Jan 27 15:19:52 crc kubenswrapper[4772]: I0127 15:19:52.669153 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09a099f0-cf75-4584-9369-e146e1898ee5" path="/var/lib/kubelet/pods/09a099f0-cf75-4584-9369-e146e1898ee5/volumes" Jan 27 15:19:56 crc kubenswrapper[4772]: I0127 15:19:56.893423 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-gpfqc" Jan 27 15:20:08 crc kubenswrapper[4772]: I0127 15:20:08.693106 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk"] Jan 27 15:20:08 crc kubenswrapper[4772]: E0127 15:20:08.694487 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09a099f0-cf75-4584-9369-e146e1898ee5" containerName="registry-server" Jan 27 15:20:08 crc kubenswrapper[4772]: I0127 15:20:08.694506 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="09a099f0-cf75-4584-9369-e146e1898ee5" containerName="registry-server" Jan 27 15:20:08 crc kubenswrapper[4772]: E0127 15:20:08.694523 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09a099f0-cf75-4584-9369-e146e1898ee5" containerName="extract-content" Jan 27 15:20:08 crc kubenswrapper[4772]: I0127 15:20:08.694532 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="09a099f0-cf75-4584-9369-e146e1898ee5" containerName="extract-content" Jan 27 15:20:08 crc kubenswrapper[4772]: E0127 15:20:08.694555 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09a099f0-cf75-4584-9369-e146e1898ee5" containerName="extract-utilities" Jan 27 15:20:08 crc kubenswrapper[4772]: I0127 15:20:08.694565 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="09a099f0-cf75-4584-9369-e146e1898ee5" containerName="extract-utilities" Jan 27 15:20:08 crc kubenswrapper[4772]: I0127 15:20:08.694763 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="09a099f0-cf75-4584-9369-e146e1898ee5" containerName="registry-server" Jan 27 15:20:08 crc kubenswrapper[4772]: I0127 15:20:08.697829 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk" Jan 27 15:20:08 crc kubenswrapper[4772]: I0127 15:20:08.699866 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 27 15:20:08 crc kubenswrapper[4772]: I0127 15:20:08.715670 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk"] Jan 27 15:20:08 crc kubenswrapper[4772]: I0127 15:20:08.881570 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/09090577-fdfa-4f36-badf-f32c6ee2ab7d-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk\" (UID: \"09090577-fdfa-4f36-badf-f32c6ee2ab7d\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk" Jan 27 15:20:08 crc kubenswrapper[4772]: I0127 15:20:08.881646 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/09090577-fdfa-4f36-badf-f32c6ee2ab7d-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk\" (UID: \"09090577-fdfa-4f36-badf-f32c6ee2ab7d\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk" Jan 27 15:20:08 crc kubenswrapper[4772]: I0127 15:20:08.881689 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jxtg\" (UniqueName: \"kubernetes.io/projected/09090577-fdfa-4f36-badf-f32c6ee2ab7d-kube-api-access-9jxtg\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk\" (UID: \"09090577-fdfa-4f36-badf-f32c6ee2ab7d\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk" Jan 27 15:20:08 crc kubenswrapper[4772]: I0127 15:20:08.983646 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/09090577-fdfa-4f36-badf-f32c6ee2ab7d-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk\" (UID: \"09090577-fdfa-4f36-badf-f32c6ee2ab7d\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk" Jan 27 15:20:08 crc kubenswrapper[4772]: I0127 15:20:08.983733 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/09090577-fdfa-4f36-badf-f32c6ee2ab7d-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk\" (UID: \"09090577-fdfa-4f36-badf-f32c6ee2ab7d\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk" Jan 27 15:20:08 crc kubenswrapper[4772]: I0127 15:20:08.983797 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jxtg\" (UniqueName: \"kubernetes.io/projected/09090577-fdfa-4f36-badf-f32c6ee2ab7d-kube-api-access-9jxtg\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk\" (UID: \"09090577-fdfa-4f36-badf-f32c6ee2ab7d\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk" Jan 27 15:20:08 crc kubenswrapper[4772]: I0127 15:20:08.984299 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/09090577-fdfa-4f36-badf-f32c6ee2ab7d-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk\" (UID: \"09090577-fdfa-4f36-badf-f32c6ee2ab7d\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk" Jan 27 15:20:08 crc kubenswrapper[4772]: I0127 15:20:08.984402 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/09090577-fdfa-4f36-badf-f32c6ee2ab7d-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk\" (UID: \"09090577-fdfa-4f36-badf-f32c6ee2ab7d\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk" Jan 27 15:20:09 crc kubenswrapper[4772]: I0127 15:20:09.008711 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jxtg\" (UniqueName: \"kubernetes.io/projected/09090577-fdfa-4f36-badf-f32c6ee2ab7d-kube-api-access-9jxtg\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk\" (UID: \"09090577-fdfa-4f36-badf-f32c6ee2ab7d\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk" Jan 27 15:20:09 crc kubenswrapper[4772]: I0127 15:20:09.017128 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk" Jan 27 15:20:09 crc kubenswrapper[4772]: I0127 15:20:09.458737 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk"] Jan 27 15:20:09 crc kubenswrapper[4772]: W0127 15:20:09.470788 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09090577_fdfa_4f36_badf_f32c6ee2ab7d.slice/crio-5eace4e07ae1ddfd9ca33204554ccf4c3b1f352fe916de49d02bb84876aa31e2 WatchSource:0}: Error finding container 5eace4e07ae1ddfd9ca33204554ccf4c3b1f352fe916de49d02bb84876aa31e2: Status 404 returned error can't find the container with id 5eace4e07ae1ddfd9ca33204554ccf4c3b1f352fe916de49d02bb84876aa31e2 Jan 27 15:20:10 crc kubenswrapper[4772]: I0127 15:20:10.355002 4772 generic.go:334] "Generic (PLEG): container finished" podID="09090577-fdfa-4f36-badf-f32c6ee2ab7d" containerID="439c3bfaa41668060ae94dd39d5966e30d1a6826569c7e63cae066908ef685a6" exitCode=0 Jan 27 15:20:10 crc kubenswrapper[4772]: I0127 15:20:10.355085 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk" event={"ID":"09090577-fdfa-4f36-badf-f32c6ee2ab7d","Type":"ContainerDied","Data":"439c3bfaa41668060ae94dd39d5966e30d1a6826569c7e63cae066908ef685a6"} Jan 27 15:20:10 crc kubenswrapper[4772]: I0127 15:20:10.355143 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk" event={"ID":"09090577-fdfa-4f36-badf-f32c6ee2ab7d","Type":"ContainerStarted","Data":"5eace4e07ae1ddfd9ca33204554ccf4c3b1f352fe916de49d02bb84876aa31e2"} Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.058335 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.058863 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.058919 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.059590 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"60c798dfb542a875b90e857bf6f54352abce005f4bc0c5fd246c1b5d0903e3f3"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.059664 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://60c798dfb542a875b90e857bf6f54352abce005f4bc0c5fd246c1b5d0903e3f3" gracePeriod=600 Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.300681 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-7qfrl" podUID="e2e31e5f-3a41-42f5-90b0-99c05a8033a6" containerName="console" containerID="cri-o://5fa14544d0f474c1dab5359f3cbcee7247c22e26d183454cd30d9fa3ab064652" gracePeriod=15 Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.370303 4772 generic.go:334] "Generic (PLEG): container finished" podID="09090577-fdfa-4f36-badf-f32c6ee2ab7d" containerID="85f37a805cce12fdeb9e9c9c28438e3e6b7638e7b3599585f99fb6660d5b1264" exitCode=0 Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.370387 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk" event={"ID":"09090577-fdfa-4f36-badf-f32c6ee2ab7d","Type":"ContainerDied","Data":"85f37a805cce12fdeb9e9c9c28438e3e6b7638e7b3599585f99fb6660d5b1264"} Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.372892 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="60c798dfb542a875b90e857bf6f54352abce005f4bc0c5fd246c1b5d0903e3f3" exitCode=0 Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.372940 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"60c798dfb542a875b90e857bf6f54352abce005f4bc0c5fd246c1b5d0903e3f3"} Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.373042 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"c8213e4fa74445d3800c2dbcb45efc3fb34a6f40c3d5ed5845b811a51d3d8497"} Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.373073 4772 scope.go:117] "RemoveContainer" containerID="32659ec7f069b0827082828bb6142c20199821498a042e5f263706f6e96e9462" Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.646309 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-7qfrl_e2e31e5f-3a41-42f5-90b0-99c05a8033a6/console/0.log" Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.646622 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.844554 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-oauth-serving-cert\") pod \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.845074 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-trusted-ca-bundle\") pod \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.845205 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-console-oauth-config\") pod \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.845297 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "e2e31e5f-3a41-42f5-90b0-99c05a8033a6" (UID: "e2e31e5f-3a41-42f5-90b0-99c05a8033a6"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.845345 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmgkc\" (UniqueName: \"kubernetes.io/projected/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-kube-api-access-vmgkc\") pod \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.845463 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "e2e31e5f-3a41-42f5-90b0-99c05a8033a6" (UID: "e2e31e5f-3a41-42f5-90b0-99c05a8033a6"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.845474 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-service-ca\") pod \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.845546 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-console-serving-cert\") pod \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.845577 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-console-config\") pod \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\" (UID: \"e2e31e5f-3a41-42f5-90b0-99c05a8033a6\") " Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.845937 4772 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.845952 4772 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.846060 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-service-ca" (OuterVolumeSpecName: "service-ca") pod "e2e31e5f-3a41-42f5-90b0-99c05a8033a6" (UID: "e2e31e5f-3a41-42f5-90b0-99c05a8033a6"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.846374 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-console-config" (OuterVolumeSpecName: "console-config") pod "e2e31e5f-3a41-42f5-90b0-99c05a8033a6" (UID: "e2e31e5f-3a41-42f5-90b0-99c05a8033a6"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.850210 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "e2e31e5f-3a41-42f5-90b0-99c05a8033a6" (UID: "e2e31e5f-3a41-42f5-90b0-99c05a8033a6"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.850290 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "e2e31e5f-3a41-42f5-90b0-99c05a8033a6" (UID: "e2e31e5f-3a41-42f5-90b0-99c05a8033a6"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.850363 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-kube-api-access-vmgkc" (OuterVolumeSpecName: "kube-api-access-vmgkc") pod "e2e31e5f-3a41-42f5-90b0-99c05a8033a6" (UID: "e2e31e5f-3a41-42f5-90b0-99c05a8033a6"). InnerVolumeSpecName "kube-api-access-vmgkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.947082 4772 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.947130 4772 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-console-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.947159 4772 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.947203 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmgkc\" (UniqueName: \"kubernetes.io/projected/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-kube-api-access-vmgkc\") on node \"crc\" DevicePath \"\"" Jan 27 15:20:12 crc kubenswrapper[4772]: I0127 15:20:12.947222 4772 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e2e31e5f-3a41-42f5-90b0-99c05a8033a6-service-ca\") on node \"crc\" DevicePath \"\"" Jan 27 15:20:13 crc kubenswrapper[4772]: I0127 15:20:13.385360 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-7qfrl_e2e31e5f-3a41-42f5-90b0-99c05a8033a6/console/0.log" Jan 27 15:20:13 crc kubenswrapper[4772]: I0127 15:20:13.386430 4772 generic.go:334] "Generic (PLEG): container finished" podID="e2e31e5f-3a41-42f5-90b0-99c05a8033a6" containerID="5fa14544d0f474c1dab5359f3cbcee7247c22e26d183454cd30d9fa3ab064652" exitCode=2 Jan 27 15:20:13 crc kubenswrapper[4772]: I0127 15:20:13.386506 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7qfrl" Jan 27 15:20:13 crc kubenswrapper[4772]: I0127 15:20:13.386523 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7qfrl" event={"ID":"e2e31e5f-3a41-42f5-90b0-99c05a8033a6","Type":"ContainerDied","Data":"5fa14544d0f474c1dab5359f3cbcee7247c22e26d183454cd30d9fa3ab064652"} Jan 27 15:20:13 crc kubenswrapper[4772]: I0127 15:20:13.386561 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7qfrl" event={"ID":"e2e31e5f-3a41-42f5-90b0-99c05a8033a6","Type":"ContainerDied","Data":"689105dc82b6dcc122fad60678c44aee714f4e2b250e67f0c76903dd34d0b5c3"} Jan 27 15:20:13 crc kubenswrapper[4772]: I0127 15:20:13.386582 4772 scope.go:117] "RemoveContainer" containerID="5fa14544d0f474c1dab5359f3cbcee7247c22e26d183454cd30d9fa3ab064652" Jan 27 15:20:13 crc kubenswrapper[4772]: I0127 15:20:13.389793 4772 generic.go:334] "Generic (PLEG): container finished" podID="09090577-fdfa-4f36-badf-f32c6ee2ab7d" containerID="1700e0008ac67d7585a3770cc2c3ff72683a1c89a65d5123d41fbd195bc151f9" exitCode=0 Jan 27 15:20:13 crc kubenswrapper[4772]: I0127 15:20:13.389835 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk" event={"ID":"09090577-fdfa-4f36-badf-f32c6ee2ab7d","Type":"ContainerDied","Data":"1700e0008ac67d7585a3770cc2c3ff72683a1c89a65d5123d41fbd195bc151f9"} Jan 27 15:20:13 crc kubenswrapper[4772]: I0127 15:20:13.407463 4772 scope.go:117] "RemoveContainer" containerID="5fa14544d0f474c1dab5359f3cbcee7247c22e26d183454cd30d9fa3ab064652" Jan 27 15:20:13 crc kubenswrapper[4772]: E0127 15:20:13.408130 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fa14544d0f474c1dab5359f3cbcee7247c22e26d183454cd30d9fa3ab064652\": container with ID starting with 5fa14544d0f474c1dab5359f3cbcee7247c22e26d183454cd30d9fa3ab064652 not found: ID does not exist" containerID="5fa14544d0f474c1dab5359f3cbcee7247c22e26d183454cd30d9fa3ab064652" Jan 27 15:20:13 crc kubenswrapper[4772]: I0127 15:20:13.408318 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fa14544d0f474c1dab5359f3cbcee7247c22e26d183454cd30d9fa3ab064652"} err="failed to get container status \"5fa14544d0f474c1dab5359f3cbcee7247c22e26d183454cd30d9fa3ab064652\": rpc error: code = NotFound desc = could not find container \"5fa14544d0f474c1dab5359f3cbcee7247c22e26d183454cd30d9fa3ab064652\": container with ID starting with 5fa14544d0f474c1dab5359f3cbcee7247c22e26d183454cd30d9fa3ab064652 not found: ID does not exist" Jan 27 15:20:13 crc kubenswrapper[4772]: I0127 15:20:13.426882 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-7qfrl"] Jan 27 15:20:13 crc kubenswrapper[4772]: I0127 15:20:13.434120 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-7qfrl"] Jan 27 15:20:14 crc kubenswrapper[4772]: I0127 15:20:14.677498 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2e31e5f-3a41-42f5-90b0-99c05a8033a6" path="/var/lib/kubelet/pods/e2e31e5f-3a41-42f5-90b0-99c05a8033a6/volumes" Jan 27 15:20:14 crc kubenswrapper[4772]: I0127 15:20:14.688028 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk" Jan 27 15:20:14 crc kubenswrapper[4772]: I0127 15:20:14.875400 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/09090577-fdfa-4f36-badf-f32c6ee2ab7d-util\") pod \"09090577-fdfa-4f36-badf-f32c6ee2ab7d\" (UID: \"09090577-fdfa-4f36-badf-f32c6ee2ab7d\") " Jan 27 15:20:14 crc kubenswrapper[4772]: I0127 15:20:14.875564 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/09090577-fdfa-4f36-badf-f32c6ee2ab7d-bundle\") pod \"09090577-fdfa-4f36-badf-f32c6ee2ab7d\" (UID: \"09090577-fdfa-4f36-badf-f32c6ee2ab7d\") " Jan 27 15:20:14 crc kubenswrapper[4772]: I0127 15:20:14.875621 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jxtg\" (UniqueName: \"kubernetes.io/projected/09090577-fdfa-4f36-badf-f32c6ee2ab7d-kube-api-access-9jxtg\") pod \"09090577-fdfa-4f36-badf-f32c6ee2ab7d\" (UID: \"09090577-fdfa-4f36-badf-f32c6ee2ab7d\") " Jan 27 15:20:14 crc kubenswrapper[4772]: I0127 15:20:14.877504 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09090577-fdfa-4f36-badf-f32c6ee2ab7d-bundle" (OuterVolumeSpecName: "bundle") pod "09090577-fdfa-4f36-badf-f32c6ee2ab7d" (UID: "09090577-fdfa-4f36-badf-f32c6ee2ab7d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:20:14 crc kubenswrapper[4772]: I0127 15:20:14.884579 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09090577-fdfa-4f36-badf-f32c6ee2ab7d-kube-api-access-9jxtg" (OuterVolumeSpecName: "kube-api-access-9jxtg") pod "09090577-fdfa-4f36-badf-f32c6ee2ab7d" (UID: "09090577-fdfa-4f36-badf-f32c6ee2ab7d"). InnerVolumeSpecName "kube-api-access-9jxtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:20:14 crc kubenswrapper[4772]: I0127 15:20:14.909557 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09090577-fdfa-4f36-badf-f32c6ee2ab7d-util" (OuterVolumeSpecName: "util") pod "09090577-fdfa-4f36-badf-f32c6ee2ab7d" (UID: "09090577-fdfa-4f36-badf-f32c6ee2ab7d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:20:14 crc kubenswrapper[4772]: I0127 15:20:14.978883 4772 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/09090577-fdfa-4f36-badf-f32c6ee2ab7d-util\") on node \"crc\" DevicePath \"\"" Jan 27 15:20:14 crc kubenswrapper[4772]: I0127 15:20:14.978934 4772 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/09090577-fdfa-4f36-badf-f32c6ee2ab7d-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:20:14 crc kubenswrapper[4772]: I0127 15:20:14.978950 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jxtg\" (UniqueName: \"kubernetes.io/projected/09090577-fdfa-4f36-badf-f32c6ee2ab7d-kube-api-access-9jxtg\") on node \"crc\" DevicePath \"\"" Jan 27 15:20:15 crc kubenswrapper[4772]: I0127 15:20:15.410155 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk" event={"ID":"09090577-fdfa-4f36-badf-f32c6ee2ab7d","Type":"ContainerDied","Data":"5eace4e07ae1ddfd9ca33204554ccf4c3b1f352fe916de49d02bb84876aa31e2"} Jan 27 15:20:15 crc kubenswrapper[4772]: I0127 15:20:15.410869 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5eace4e07ae1ddfd9ca33204554ccf4c3b1f352fe916de49d02bb84876aa31e2" Jan 27 15:20:15 crc kubenswrapper[4772]: I0127 15:20:15.410291 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.392647 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5c6dd9c74b-84qz4"] Jan 27 15:20:24 crc kubenswrapper[4772]: E0127 15:20:24.393686 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09090577-fdfa-4f36-badf-f32c6ee2ab7d" containerName="util" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.393702 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="09090577-fdfa-4f36-badf-f32c6ee2ab7d" containerName="util" Jan 27 15:20:24 crc kubenswrapper[4772]: E0127 15:20:24.393717 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09090577-fdfa-4f36-badf-f32c6ee2ab7d" containerName="extract" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.393725 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="09090577-fdfa-4f36-badf-f32c6ee2ab7d" containerName="extract" Jan 27 15:20:24 crc kubenswrapper[4772]: E0127 15:20:24.393735 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2e31e5f-3a41-42f5-90b0-99c05a8033a6" containerName="console" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.393743 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2e31e5f-3a41-42f5-90b0-99c05a8033a6" containerName="console" Jan 27 15:20:24 crc kubenswrapper[4772]: E0127 15:20:24.393761 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09090577-fdfa-4f36-badf-f32c6ee2ab7d" containerName="pull" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.393767 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="09090577-fdfa-4f36-badf-f32c6ee2ab7d" containerName="pull" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.393881 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2e31e5f-3a41-42f5-90b0-99c05a8033a6" containerName="console" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.393896 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="09090577-fdfa-4f36-badf-f32c6ee2ab7d" containerName="extract" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.394438 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5c6dd9c74b-84qz4" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.401396 4772 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.401580 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.401739 4772 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.401935 4772 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-tpk9v" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.406279 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.412594 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5c6dd9c74b-84qz4"] Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.493722 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbnhq\" (UniqueName: \"kubernetes.io/projected/f72c611d-60d8-4649-a410-38434d01d8e2-kube-api-access-sbnhq\") pod \"metallb-operator-controller-manager-5c6dd9c74b-84qz4\" (UID: \"f72c611d-60d8-4649-a410-38434d01d8e2\") " pod="metallb-system/metallb-operator-controller-manager-5c6dd9c74b-84qz4" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.493778 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f72c611d-60d8-4649-a410-38434d01d8e2-webhook-cert\") pod \"metallb-operator-controller-manager-5c6dd9c74b-84qz4\" (UID: \"f72c611d-60d8-4649-a410-38434d01d8e2\") " pod="metallb-system/metallb-operator-controller-manager-5c6dd9c74b-84qz4" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.493816 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f72c611d-60d8-4649-a410-38434d01d8e2-apiservice-cert\") pod \"metallb-operator-controller-manager-5c6dd9c74b-84qz4\" (UID: \"f72c611d-60d8-4649-a410-38434d01d8e2\") " pod="metallb-system/metallb-operator-controller-manager-5c6dd9c74b-84qz4" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.595599 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbnhq\" (UniqueName: \"kubernetes.io/projected/f72c611d-60d8-4649-a410-38434d01d8e2-kube-api-access-sbnhq\") pod \"metallb-operator-controller-manager-5c6dd9c74b-84qz4\" (UID: \"f72c611d-60d8-4649-a410-38434d01d8e2\") " pod="metallb-system/metallb-operator-controller-manager-5c6dd9c74b-84qz4" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.595910 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f72c611d-60d8-4649-a410-38434d01d8e2-webhook-cert\") pod \"metallb-operator-controller-manager-5c6dd9c74b-84qz4\" (UID: \"f72c611d-60d8-4649-a410-38434d01d8e2\") " pod="metallb-system/metallb-operator-controller-manager-5c6dd9c74b-84qz4" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.596045 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f72c611d-60d8-4649-a410-38434d01d8e2-apiservice-cert\") pod \"metallb-operator-controller-manager-5c6dd9c74b-84qz4\" (UID: \"f72c611d-60d8-4649-a410-38434d01d8e2\") " pod="metallb-system/metallb-operator-controller-manager-5c6dd9c74b-84qz4" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.606289 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f72c611d-60d8-4649-a410-38434d01d8e2-apiservice-cert\") pod \"metallb-operator-controller-manager-5c6dd9c74b-84qz4\" (UID: \"f72c611d-60d8-4649-a410-38434d01d8e2\") " pod="metallb-system/metallb-operator-controller-manager-5c6dd9c74b-84qz4" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.607002 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f72c611d-60d8-4649-a410-38434d01d8e2-webhook-cert\") pod \"metallb-operator-controller-manager-5c6dd9c74b-84qz4\" (UID: \"f72c611d-60d8-4649-a410-38434d01d8e2\") " pod="metallb-system/metallb-operator-controller-manager-5c6dd9c74b-84qz4" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.625975 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbnhq\" (UniqueName: \"kubernetes.io/projected/f72c611d-60d8-4649-a410-38434d01d8e2-kube-api-access-sbnhq\") pod \"metallb-operator-controller-manager-5c6dd9c74b-84qz4\" (UID: \"f72c611d-60d8-4649-a410-38434d01d8e2\") " pod="metallb-system/metallb-operator-controller-manager-5c6dd9c74b-84qz4" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.698632 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-66986f9f9f-bmvms"] Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.699724 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-66986f9f9f-bmvms" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.707007 4772 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-zhlwq" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.707177 4772 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.707361 4772 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.712648 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-66986f9f9f-bmvms"] Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.731544 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5c6dd9c74b-84qz4" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.899882 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c5ee8d7f-0160-4526-8ae0-45a50a450725-apiservice-cert\") pod \"metallb-operator-webhook-server-66986f9f9f-bmvms\" (UID: \"c5ee8d7f-0160-4526-8ae0-45a50a450725\") " pod="metallb-system/metallb-operator-webhook-server-66986f9f9f-bmvms" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.900100 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c5ee8d7f-0160-4526-8ae0-45a50a450725-webhook-cert\") pod \"metallb-operator-webhook-server-66986f9f9f-bmvms\" (UID: \"c5ee8d7f-0160-4526-8ae0-45a50a450725\") " pod="metallb-system/metallb-operator-webhook-server-66986f9f9f-bmvms" Jan 27 15:20:24 crc kubenswrapper[4772]: I0127 15:20:24.900151 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4kqk\" (UniqueName: \"kubernetes.io/projected/c5ee8d7f-0160-4526-8ae0-45a50a450725-kube-api-access-w4kqk\") pod \"metallb-operator-webhook-server-66986f9f9f-bmvms\" (UID: \"c5ee8d7f-0160-4526-8ae0-45a50a450725\") " pod="metallb-system/metallb-operator-webhook-server-66986f9f9f-bmvms" Jan 27 15:20:25 crc kubenswrapper[4772]: I0127 15:20:25.001115 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c5ee8d7f-0160-4526-8ae0-45a50a450725-apiservice-cert\") pod \"metallb-operator-webhook-server-66986f9f9f-bmvms\" (UID: \"c5ee8d7f-0160-4526-8ae0-45a50a450725\") " pod="metallb-system/metallb-operator-webhook-server-66986f9f9f-bmvms" Jan 27 15:20:25 crc kubenswrapper[4772]: I0127 15:20:25.001194 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c5ee8d7f-0160-4526-8ae0-45a50a450725-webhook-cert\") pod \"metallb-operator-webhook-server-66986f9f9f-bmvms\" (UID: \"c5ee8d7f-0160-4526-8ae0-45a50a450725\") " pod="metallb-system/metallb-operator-webhook-server-66986f9f9f-bmvms" Jan 27 15:20:25 crc kubenswrapper[4772]: I0127 15:20:25.001228 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4kqk\" (UniqueName: \"kubernetes.io/projected/c5ee8d7f-0160-4526-8ae0-45a50a450725-kube-api-access-w4kqk\") pod \"metallb-operator-webhook-server-66986f9f9f-bmvms\" (UID: \"c5ee8d7f-0160-4526-8ae0-45a50a450725\") " pod="metallb-system/metallb-operator-webhook-server-66986f9f9f-bmvms" Jan 27 15:20:25 crc kubenswrapper[4772]: I0127 15:20:25.007819 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c5ee8d7f-0160-4526-8ae0-45a50a450725-apiservice-cert\") pod \"metallb-operator-webhook-server-66986f9f9f-bmvms\" (UID: \"c5ee8d7f-0160-4526-8ae0-45a50a450725\") " pod="metallb-system/metallb-operator-webhook-server-66986f9f9f-bmvms" Jan 27 15:20:25 crc kubenswrapper[4772]: I0127 15:20:25.007888 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c5ee8d7f-0160-4526-8ae0-45a50a450725-webhook-cert\") pod \"metallb-operator-webhook-server-66986f9f9f-bmvms\" (UID: \"c5ee8d7f-0160-4526-8ae0-45a50a450725\") " pod="metallb-system/metallb-operator-webhook-server-66986f9f9f-bmvms" Jan 27 15:20:25 crc kubenswrapper[4772]: I0127 15:20:25.024753 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4kqk\" (UniqueName: \"kubernetes.io/projected/c5ee8d7f-0160-4526-8ae0-45a50a450725-kube-api-access-w4kqk\") pod \"metallb-operator-webhook-server-66986f9f9f-bmvms\" (UID: \"c5ee8d7f-0160-4526-8ae0-45a50a450725\") " pod="metallb-system/metallb-operator-webhook-server-66986f9f9f-bmvms" Jan 27 15:20:25 crc kubenswrapper[4772]: I0127 15:20:25.025751 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-66986f9f9f-bmvms" Jan 27 15:20:25 crc kubenswrapper[4772]: I0127 15:20:25.215129 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5c6dd9c74b-84qz4"] Jan 27 15:20:25 crc kubenswrapper[4772]: W0127 15:20:25.229685 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf72c611d_60d8_4649_a410_38434d01d8e2.slice/crio-4caf69be229218f9b04edf0ad07b0614ac6cb4226b5bfdaff0acb1e3267678b2 WatchSource:0}: Error finding container 4caf69be229218f9b04edf0ad07b0614ac6cb4226b5bfdaff0acb1e3267678b2: Status 404 returned error can't find the container with id 4caf69be229218f9b04edf0ad07b0614ac6cb4226b5bfdaff0acb1e3267678b2 Jan 27 15:20:25 crc kubenswrapper[4772]: I0127 15:20:25.229824 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-66986f9f9f-bmvms"] Jan 27 15:20:25 crc kubenswrapper[4772]: I0127 15:20:25.475057 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-66986f9f9f-bmvms" event={"ID":"c5ee8d7f-0160-4526-8ae0-45a50a450725","Type":"ContainerStarted","Data":"7a92d7b3bea78e77245e067e3091d09809cb9ab9290860ca7457eadb8a23f48f"} Jan 27 15:20:25 crc kubenswrapper[4772]: I0127 15:20:25.477897 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5c6dd9c74b-84qz4" event={"ID":"f72c611d-60d8-4649-a410-38434d01d8e2","Type":"ContainerStarted","Data":"4caf69be229218f9b04edf0ad07b0614ac6cb4226b5bfdaff0acb1e3267678b2"} Jan 27 15:20:28 crc kubenswrapper[4772]: I0127 15:20:28.498149 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5c6dd9c74b-84qz4" event={"ID":"f72c611d-60d8-4649-a410-38434d01d8e2","Type":"ContainerStarted","Data":"2a97c2748ab3c37fc26e8f99921f757adb8b041f0526a0b3adbe03801cfd948f"} Jan 27 15:20:28 crc kubenswrapper[4772]: I0127 15:20:28.498966 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5c6dd9c74b-84qz4" Jan 27 15:20:34 crc kubenswrapper[4772]: I0127 15:20:34.551619 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-66986f9f9f-bmvms" event={"ID":"c5ee8d7f-0160-4526-8ae0-45a50a450725","Type":"ContainerStarted","Data":"dca38d4ff748fb7a714a98e080648bb0e1e2a8e4d7decd87038933d75280339d"} Jan 27 15:20:34 crc kubenswrapper[4772]: I0127 15:20:34.552663 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-66986f9f9f-bmvms" Jan 27 15:20:34 crc kubenswrapper[4772]: I0127 15:20:34.579130 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5c6dd9c74b-84qz4" podStartSLOduration=7.516448158 podStartE2EDuration="10.579110537s" podCreationTimestamp="2026-01-27 15:20:24 +0000 UTC" firstStartedPulling="2026-01-27 15:20:25.233055544 +0000 UTC m=+811.213664642" lastFinishedPulling="2026-01-27 15:20:28.295717923 +0000 UTC m=+814.276327021" observedRunningTime="2026-01-27 15:20:28.525072423 +0000 UTC m=+814.505681541" watchObservedRunningTime="2026-01-27 15:20:34.579110537 +0000 UTC m=+820.559719635" Jan 27 15:20:34 crc kubenswrapper[4772]: I0127 15:20:34.579601 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-66986f9f9f-bmvms" podStartSLOduration=1.689105377 podStartE2EDuration="10.579594371s" podCreationTimestamp="2026-01-27 15:20:24 +0000 UTC" firstStartedPulling="2026-01-27 15:20:25.24403269 +0000 UTC m=+811.224641798" lastFinishedPulling="2026-01-27 15:20:34.134521694 +0000 UTC m=+820.115130792" observedRunningTime="2026-01-27 15:20:34.570979973 +0000 UTC m=+820.551589131" watchObservedRunningTime="2026-01-27 15:20:34.579594371 +0000 UTC m=+820.560203469" Jan 27 15:20:45 crc kubenswrapper[4772]: I0127 15:20:45.032371 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-66986f9f9f-bmvms" Jan 27 15:21:04 crc kubenswrapper[4772]: I0127 15:21:04.734654 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5c6dd9c74b-84qz4" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.490793 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-jhpnb"] Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.493693 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.495332 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-qpxrs"] Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.496145 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qpxrs" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.496814 4772 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.497687 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.497804 4772 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-l97nh" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.498140 4772 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.511014 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-qpxrs"] Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.594443 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-cl54q"] Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.595508 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-cl54q" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.597524 4772 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.597744 4772 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.598032 4772 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-jz9xz" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.598235 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.622513 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6968d8fdc4-vdg69"] Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.624541 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-vdg69" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.626654 4772 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.633079 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-vdg69"] Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.667209 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6ghc\" (UniqueName: \"kubernetes.io/projected/28ed9da3-cd29-4d80-9703-472bdbb3c64b-kube-api-access-z6ghc\") pod \"frr-k8s-webhook-server-7df86c4f6c-qpxrs\" (UID: \"28ed9da3-cd29-4d80-9703-472bdbb3c64b\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qpxrs" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.667260 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxg4j\" (UniqueName: \"kubernetes.io/projected/5d41fcac-7044-4f36-b9f8-0b656bb3bcca-kube-api-access-hxg4j\") pod \"frr-k8s-jhpnb\" (UID: \"5d41fcac-7044-4f36-b9f8-0b656bb3bcca\") " pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.667299 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/5d41fcac-7044-4f36-b9f8-0b656bb3bcca-reloader\") pod \"frr-k8s-jhpnb\" (UID: \"5d41fcac-7044-4f36-b9f8-0b656bb3bcca\") " pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.667339 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/5d41fcac-7044-4f36-b9f8-0b656bb3bcca-metrics\") pod \"frr-k8s-jhpnb\" (UID: \"5d41fcac-7044-4f36-b9f8-0b656bb3bcca\") " pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.667359 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/5d41fcac-7044-4f36-b9f8-0b656bb3bcca-frr-startup\") pod \"frr-k8s-jhpnb\" (UID: \"5d41fcac-7044-4f36-b9f8-0b656bb3bcca\") " pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.667381 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28ed9da3-cd29-4d80-9703-472bdbb3c64b-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-qpxrs\" (UID: \"28ed9da3-cd29-4d80-9703-472bdbb3c64b\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qpxrs" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.667414 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/5d41fcac-7044-4f36-b9f8-0b656bb3bcca-frr-conf\") pod \"frr-k8s-jhpnb\" (UID: \"5d41fcac-7044-4f36-b9f8-0b656bb3bcca\") " pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.667441 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/5d41fcac-7044-4f36-b9f8-0b656bb3bcca-frr-sockets\") pod \"frr-k8s-jhpnb\" (UID: \"5d41fcac-7044-4f36-b9f8-0b656bb3bcca\") " pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.667487 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5d41fcac-7044-4f36-b9f8-0b656bb3bcca-metrics-certs\") pod \"frr-k8s-jhpnb\" (UID: \"5d41fcac-7044-4f36-b9f8-0b656bb3bcca\") " pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.768111 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gljb\" (UniqueName: \"kubernetes.io/projected/24577bed-b34e-4419-9e9e-7068155ba0d1-kube-api-access-4gljb\") pod \"speaker-cl54q\" (UID: \"24577bed-b34e-4419-9e9e-7068155ba0d1\") " pod="metallb-system/speaker-cl54q" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.768165 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/5d41fcac-7044-4f36-b9f8-0b656bb3bcca-frr-sockets\") pod \"frr-k8s-jhpnb\" (UID: \"5d41fcac-7044-4f36-b9f8-0b656bb3bcca\") " pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.768228 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5d41fcac-7044-4f36-b9f8-0b656bb3bcca-metrics-certs\") pod \"frr-k8s-jhpnb\" (UID: \"5d41fcac-7044-4f36-b9f8-0b656bb3bcca\") " pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.768260 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6ghc\" (UniqueName: \"kubernetes.io/projected/28ed9da3-cd29-4d80-9703-472bdbb3c64b-kube-api-access-z6ghc\") pod \"frr-k8s-webhook-server-7df86c4f6c-qpxrs\" (UID: \"28ed9da3-cd29-4d80-9703-472bdbb3c64b\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qpxrs" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.768282 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxg4j\" (UniqueName: \"kubernetes.io/projected/5d41fcac-7044-4f36-b9f8-0b656bb3bcca-kube-api-access-hxg4j\") pod \"frr-k8s-jhpnb\" (UID: \"5d41fcac-7044-4f36-b9f8-0b656bb3bcca\") " pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.768321 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/5d41fcac-7044-4f36-b9f8-0b656bb3bcca-reloader\") pod \"frr-k8s-jhpnb\" (UID: \"5d41fcac-7044-4f36-b9f8-0b656bb3bcca\") " pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.768341 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/24577bed-b34e-4419-9e9e-7068155ba0d1-memberlist\") pod \"speaker-cl54q\" (UID: \"24577bed-b34e-4419-9e9e-7068155ba0d1\") " pod="metallb-system/speaker-cl54q" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.768360 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/24577bed-b34e-4419-9e9e-7068155ba0d1-metallb-excludel2\") pod \"speaker-cl54q\" (UID: \"24577bed-b34e-4419-9e9e-7068155ba0d1\") " pod="metallb-system/speaker-cl54q" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.768389 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gs9dw\" (UniqueName: \"kubernetes.io/projected/d2282b46-452e-402e-b929-23875b572727-kube-api-access-gs9dw\") pod \"controller-6968d8fdc4-vdg69\" (UID: \"d2282b46-452e-402e-b929-23875b572727\") " pod="metallb-system/controller-6968d8fdc4-vdg69" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.768411 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d2282b46-452e-402e-b929-23875b572727-metrics-certs\") pod \"controller-6968d8fdc4-vdg69\" (UID: \"d2282b46-452e-402e-b929-23875b572727\") " pod="metallb-system/controller-6968d8fdc4-vdg69" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.768437 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/5d41fcac-7044-4f36-b9f8-0b656bb3bcca-metrics\") pod \"frr-k8s-jhpnb\" (UID: \"5d41fcac-7044-4f36-b9f8-0b656bb3bcca\") " pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.768456 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/24577bed-b34e-4419-9e9e-7068155ba0d1-metrics-certs\") pod \"speaker-cl54q\" (UID: \"24577bed-b34e-4419-9e9e-7068155ba0d1\") " pod="metallb-system/speaker-cl54q" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.768477 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/5d41fcac-7044-4f36-b9f8-0b656bb3bcca-frr-startup\") pod \"frr-k8s-jhpnb\" (UID: \"5d41fcac-7044-4f36-b9f8-0b656bb3bcca\") " pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.768499 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28ed9da3-cd29-4d80-9703-472bdbb3c64b-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-qpxrs\" (UID: \"28ed9da3-cd29-4d80-9703-472bdbb3c64b\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qpxrs" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.768525 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d2282b46-452e-402e-b929-23875b572727-cert\") pod \"controller-6968d8fdc4-vdg69\" (UID: \"d2282b46-452e-402e-b929-23875b572727\") " pod="metallb-system/controller-6968d8fdc4-vdg69" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.768553 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/5d41fcac-7044-4f36-b9f8-0b656bb3bcca-frr-conf\") pod \"frr-k8s-jhpnb\" (UID: \"5d41fcac-7044-4f36-b9f8-0b656bb3bcca\") " pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.768943 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/5d41fcac-7044-4f36-b9f8-0b656bb3bcca-frr-conf\") pod \"frr-k8s-jhpnb\" (UID: \"5d41fcac-7044-4f36-b9f8-0b656bb3bcca\") " pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.769161 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/5d41fcac-7044-4f36-b9f8-0b656bb3bcca-frr-sockets\") pod \"frr-k8s-jhpnb\" (UID: \"5d41fcac-7044-4f36-b9f8-0b656bb3bcca\") " pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.770566 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/5d41fcac-7044-4f36-b9f8-0b656bb3bcca-reloader\") pod \"frr-k8s-jhpnb\" (UID: \"5d41fcac-7044-4f36-b9f8-0b656bb3bcca\") " pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.771245 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/5d41fcac-7044-4f36-b9f8-0b656bb3bcca-metrics\") pod \"frr-k8s-jhpnb\" (UID: \"5d41fcac-7044-4f36-b9f8-0b656bb3bcca\") " pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.771532 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/5d41fcac-7044-4f36-b9f8-0b656bb3bcca-frr-startup\") pod \"frr-k8s-jhpnb\" (UID: \"5d41fcac-7044-4f36-b9f8-0b656bb3bcca\") " pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.774551 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5d41fcac-7044-4f36-b9f8-0b656bb3bcca-metrics-certs\") pod \"frr-k8s-jhpnb\" (UID: \"5d41fcac-7044-4f36-b9f8-0b656bb3bcca\") " pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.775226 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28ed9da3-cd29-4d80-9703-472bdbb3c64b-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-qpxrs\" (UID: \"28ed9da3-cd29-4d80-9703-472bdbb3c64b\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qpxrs" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.788987 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6ghc\" (UniqueName: \"kubernetes.io/projected/28ed9da3-cd29-4d80-9703-472bdbb3c64b-kube-api-access-z6ghc\") pod \"frr-k8s-webhook-server-7df86c4f6c-qpxrs\" (UID: \"28ed9da3-cd29-4d80-9703-472bdbb3c64b\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qpxrs" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.798764 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxg4j\" (UniqueName: \"kubernetes.io/projected/5d41fcac-7044-4f36-b9f8-0b656bb3bcca-kube-api-access-hxg4j\") pod \"frr-k8s-jhpnb\" (UID: \"5d41fcac-7044-4f36-b9f8-0b656bb3bcca\") " pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.816353 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.829121 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qpxrs" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.869755 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d2282b46-452e-402e-b929-23875b572727-cert\") pod \"controller-6968d8fdc4-vdg69\" (UID: \"d2282b46-452e-402e-b929-23875b572727\") " pod="metallb-system/controller-6968d8fdc4-vdg69" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.869813 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gljb\" (UniqueName: \"kubernetes.io/projected/24577bed-b34e-4419-9e9e-7068155ba0d1-kube-api-access-4gljb\") pod \"speaker-cl54q\" (UID: \"24577bed-b34e-4419-9e9e-7068155ba0d1\") " pod="metallb-system/speaker-cl54q" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.869872 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/24577bed-b34e-4419-9e9e-7068155ba0d1-memberlist\") pod \"speaker-cl54q\" (UID: \"24577bed-b34e-4419-9e9e-7068155ba0d1\") " pod="metallb-system/speaker-cl54q" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.869887 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/24577bed-b34e-4419-9e9e-7068155ba0d1-metallb-excludel2\") pod \"speaker-cl54q\" (UID: \"24577bed-b34e-4419-9e9e-7068155ba0d1\") " pod="metallb-system/speaker-cl54q" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.869908 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gs9dw\" (UniqueName: \"kubernetes.io/projected/d2282b46-452e-402e-b929-23875b572727-kube-api-access-gs9dw\") pod \"controller-6968d8fdc4-vdg69\" (UID: \"d2282b46-452e-402e-b929-23875b572727\") " pod="metallb-system/controller-6968d8fdc4-vdg69" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.869924 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d2282b46-452e-402e-b929-23875b572727-metrics-certs\") pod \"controller-6968d8fdc4-vdg69\" (UID: \"d2282b46-452e-402e-b929-23875b572727\") " pod="metallb-system/controller-6968d8fdc4-vdg69" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.869943 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/24577bed-b34e-4419-9e9e-7068155ba0d1-metrics-certs\") pod \"speaker-cl54q\" (UID: \"24577bed-b34e-4419-9e9e-7068155ba0d1\") " pod="metallb-system/speaker-cl54q" Jan 27 15:21:05 crc kubenswrapper[4772]: E0127 15:21:05.870325 4772 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 27 15:21:05 crc kubenswrapper[4772]: E0127 15:21:05.870400 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/24577bed-b34e-4419-9e9e-7068155ba0d1-memberlist podName:24577bed-b34e-4419-9e9e-7068155ba0d1 nodeName:}" failed. No retries permitted until 2026-01-27 15:21:06.370382282 +0000 UTC m=+852.350991380 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/24577bed-b34e-4419-9e9e-7068155ba0d1-memberlist") pod "speaker-cl54q" (UID: "24577bed-b34e-4419-9e9e-7068155ba0d1") : secret "metallb-memberlist" not found Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.872944 4772 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.874046 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/24577bed-b34e-4419-9e9e-7068155ba0d1-metrics-certs\") pod \"speaker-cl54q\" (UID: \"24577bed-b34e-4419-9e9e-7068155ba0d1\") " pod="metallb-system/speaker-cl54q" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.880912 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d2282b46-452e-402e-b929-23875b572727-metrics-certs\") pod \"controller-6968d8fdc4-vdg69\" (UID: \"d2282b46-452e-402e-b929-23875b572727\") " pod="metallb-system/controller-6968d8fdc4-vdg69" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.883492 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d2282b46-452e-402e-b929-23875b572727-cert\") pod \"controller-6968d8fdc4-vdg69\" (UID: \"d2282b46-452e-402e-b929-23875b572727\") " pod="metallb-system/controller-6968d8fdc4-vdg69" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.875424 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/24577bed-b34e-4419-9e9e-7068155ba0d1-metallb-excludel2\") pod \"speaker-cl54q\" (UID: \"24577bed-b34e-4419-9e9e-7068155ba0d1\") " pod="metallb-system/speaker-cl54q" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.889135 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gs9dw\" (UniqueName: \"kubernetes.io/projected/d2282b46-452e-402e-b929-23875b572727-kube-api-access-gs9dw\") pod \"controller-6968d8fdc4-vdg69\" (UID: \"d2282b46-452e-402e-b929-23875b572727\") " pod="metallb-system/controller-6968d8fdc4-vdg69" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.896702 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gljb\" (UniqueName: \"kubernetes.io/projected/24577bed-b34e-4419-9e9e-7068155ba0d1-kube-api-access-4gljb\") pod \"speaker-cl54q\" (UID: \"24577bed-b34e-4419-9e9e-7068155ba0d1\") " pod="metallb-system/speaker-cl54q" Jan 27 15:21:05 crc kubenswrapper[4772]: I0127 15:21:05.952538 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-vdg69" Jan 27 15:21:06 crc kubenswrapper[4772]: I0127 15:21:06.040621 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-qpxrs"] Jan 27 15:21:06 crc kubenswrapper[4772]: I0127 15:21:06.232414 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-vdg69"] Jan 27 15:21:06 crc kubenswrapper[4772]: I0127 15:21:06.376147 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/24577bed-b34e-4419-9e9e-7068155ba0d1-memberlist\") pod \"speaker-cl54q\" (UID: \"24577bed-b34e-4419-9e9e-7068155ba0d1\") " pod="metallb-system/speaker-cl54q" Jan 27 15:21:06 crc kubenswrapper[4772]: E0127 15:21:06.376294 4772 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 27 15:21:06 crc kubenswrapper[4772]: E0127 15:21:06.376590 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/24577bed-b34e-4419-9e9e-7068155ba0d1-memberlist podName:24577bed-b34e-4419-9e9e-7068155ba0d1 nodeName:}" failed. No retries permitted until 2026-01-27 15:21:07.376567138 +0000 UTC m=+853.357176236 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/24577bed-b34e-4419-9e9e-7068155ba0d1-memberlist") pod "speaker-cl54q" (UID: "24577bed-b34e-4419-9e9e-7068155ba0d1") : secret "metallb-memberlist" not found Jan 27 15:21:06 crc kubenswrapper[4772]: I0127 15:21:06.778609 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qpxrs" event={"ID":"28ed9da3-cd29-4d80-9703-472bdbb3c64b","Type":"ContainerStarted","Data":"780aca966d9b36d525c38f5fb2440499e489ad5d95ae0544152e425165926e70"} Jan 27 15:21:06 crc kubenswrapper[4772]: I0127 15:21:06.780044 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jhpnb" event={"ID":"5d41fcac-7044-4f36-b9f8-0b656bb3bcca","Type":"ContainerStarted","Data":"bca95dea96b7d9c84b4d9aac1543290086fcf9e1cd46a039c27562c97260ed25"} Jan 27 15:21:06 crc kubenswrapper[4772]: I0127 15:21:06.782078 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-vdg69" event={"ID":"d2282b46-452e-402e-b929-23875b572727","Type":"ContainerStarted","Data":"341afb24b9025101bf1fb56bdbe3737452cb4cc86d59fff9042e46c9401b3d15"} Jan 27 15:21:06 crc kubenswrapper[4772]: I0127 15:21:06.782145 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-vdg69" event={"ID":"d2282b46-452e-402e-b929-23875b572727","Type":"ContainerStarted","Data":"2c9ba4d1eaa3351a582a108e113dc0dc908f1c4c3bd2e5a07a9d796df90bb0fb"} Jan 27 15:21:06 crc kubenswrapper[4772]: I0127 15:21:06.782161 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-vdg69" event={"ID":"d2282b46-452e-402e-b929-23875b572727","Type":"ContainerStarted","Data":"ceb56c6378c104b948c9d6b068a254ebf7b9f9fe3f2d109e87f59bc0984f28a7"} Jan 27 15:21:06 crc kubenswrapper[4772]: I0127 15:21:06.807294 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6968d8fdc4-vdg69" podStartSLOduration=1.807269733 podStartE2EDuration="1.807269733s" podCreationTimestamp="2026-01-27 15:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:21:06.803078132 +0000 UTC m=+852.783687250" watchObservedRunningTime="2026-01-27 15:21:06.807269733 +0000 UTC m=+852.787878851" Jan 27 15:21:07 crc kubenswrapper[4772]: I0127 15:21:07.390065 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/24577bed-b34e-4419-9e9e-7068155ba0d1-memberlist\") pod \"speaker-cl54q\" (UID: \"24577bed-b34e-4419-9e9e-7068155ba0d1\") " pod="metallb-system/speaker-cl54q" Jan 27 15:21:07 crc kubenswrapper[4772]: I0127 15:21:07.399839 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/24577bed-b34e-4419-9e9e-7068155ba0d1-memberlist\") pod \"speaker-cl54q\" (UID: \"24577bed-b34e-4419-9e9e-7068155ba0d1\") " pod="metallb-system/speaker-cl54q" Jan 27 15:21:07 crc kubenswrapper[4772]: I0127 15:21:07.409291 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-cl54q" Jan 27 15:21:07 crc kubenswrapper[4772]: I0127 15:21:07.814289 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-cl54q" event={"ID":"24577bed-b34e-4419-9e9e-7068155ba0d1","Type":"ContainerStarted","Data":"73276f07d5615044b17ec0fef1b7f7d6010cd89f4ec2ae0dd0ddf1ddd8568bde"} Jan 27 15:21:07 crc kubenswrapper[4772]: I0127 15:21:07.814331 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6968d8fdc4-vdg69" Jan 27 15:21:07 crc kubenswrapper[4772]: I0127 15:21:07.814347 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-cl54q" event={"ID":"24577bed-b34e-4419-9e9e-7068155ba0d1","Type":"ContainerStarted","Data":"2240600dba6c2d78ad94c28d4ef41d4fc16fd68d2e6744503d736eb88ce1582e"} Jan 27 15:21:08 crc kubenswrapper[4772]: I0127 15:21:08.822643 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-cl54q" event={"ID":"24577bed-b34e-4419-9e9e-7068155ba0d1","Type":"ContainerStarted","Data":"2b5b3a1b39aae66538d36485921a8a98cabe7ffbece55252ae7e9dfb5706bbd5"} Jan 27 15:21:09 crc kubenswrapper[4772]: I0127 15:21:09.849660 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-cl54q" Jan 27 15:21:14 crc kubenswrapper[4772]: I0127 15:21:14.686914 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-cl54q" podStartSLOduration=9.686898099 podStartE2EDuration="9.686898099s" podCreationTimestamp="2026-01-27 15:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:21:08.846138009 +0000 UTC m=+854.826747127" watchObservedRunningTime="2026-01-27 15:21:14.686898099 +0000 UTC m=+860.667507197" Jan 27 15:21:14 crc kubenswrapper[4772]: I0127 15:21:14.889600 4772 generic.go:334] "Generic (PLEG): container finished" podID="5d41fcac-7044-4f36-b9f8-0b656bb3bcca" containerID="4152684fe40ccfc62765e2d701064a0eddfcc9efca765746312af1435e2e0743" exitCode=0 Jan 27 15:21:14 crc kubenswrapper[4772]: I0127 15:21:14.889713 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jhpnb" event={"ID":"5d41fcac-7044-4f36-b9f8-0b656bb3bcca","Type":"ContainerDied","Data":"4152684fe40ccfc62765e2d701064a0eddfcc9efca765746312af1435e2e0743"} Jan 27 15:21:14 crc kubenswrapper[4772]: I0127 15:21:14.893526 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qpxrs" event={"ID":"28ed9da3-cd29-4d80-9703-472bdbb3c64b","Type":"ContainerStarted","Data":"e6f53ee01d2b56f4345ee3366629355e1ed57bd397dfeb870e6198fbb52190c3"} Jan 27 15:21:14 crc kubenswrapper[4772]: I0127 15:21:14.894138 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qpxrs" Jan 27 15:21:14 crc kubenswrapper[4772]: I0127 15:21:14.958815 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qpxrs" podStartSLOduration=2.282162238 podStartE2EDuration="9.958792538s" podCreationTimestamp="2026-01-27 15:21:05 +0000 UTC" firstStartedPulling="2026-01-27 15:21:06.046780155 +0000 UTC m=+852.027389253" lastFinishedPulling="2026-01-27 15:21:13.723410455 +0000 UTC m=+859.704019553" observedRunningTime="2026-01-27 15:21:14.952325732 +0000 UTC m=+860.932934840" watchObservedRunningTime="2026-01-27 15:21:14.958792538 +0000 UTC m=+860.939401646" Jan 27 15:21:15 crc kubenswrapper[4772]: I0127 15:21:15.915843 4772 generic.go:334] "Generic (PLEG): container finished" podID="5d41fcac-7044-4f36-b9f8-0b656bb3bcca" containerID="9967ee6e447cf31d2e31e46708ac245669b58f5587a46e4d3ff44b9263ca1a06" exitCode=0 Jan 27 15:21:15 crc kubenswrapper[4772]: I0127 15:21:15.917547 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jhpnb" event={"ID":"5d41fcac-7044-4f36-b9f8-0b656bb3bcca","Type":"ContainerDied","Data":"9967ee6e447cf31d2e31e46708ac245669b58f5587a46e4d3ff44b9263ca1a06"} Jan 27 15:21:16 crc kubenswrapper[4772]: I0127 15:21:16.926539 4772 generic.go:334] "Generic (PLEG): container finished" podID="5d41fcac-7044-4f36-b9f8-0b656bb3bcca" containerID="e945c2714806c1f70092a08fa71ca1b875f7d14aaaea2b52d560d0eba8daf73f" exitCode=0 Jan 27 15:21:16 crc kubenswrapper[4772]: I0127 15:21:16.926616 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jhpnb" event={"ID":"5d41fcac-7044-4f36-b9f8-0b656bb3bcca","Type":"ContainerDied","Data":"e945c2714806c1f70092a08fa71ca1b875f7d14aaaea2b52d560d0eba8daf73f"} Jan 27 15:21:17 crc kubenswrapper[4772]: I0127 15:21:17.413289 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-cl54q" Jan 27 15:21:17 crc kubenswrapper[4772]: I0127 15:21:17.937128 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jhpnb" event={"ID":"5d41fcac-7044-4f36-b9f8-0b656bb3bcca","Type":"ContainerStarted","Data":"bfed979fcfe5bd812af2efe31e19fb0d984b1fd6aaa19f7833a20a8f64b840b0"} Jan 27 15:21:17 crc kubenswrapper[4772]: I0127 15:21:17.937228 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jhpnb" event={"ID":"5d41fcac-7044-4f36-b9f8-0b656bb3bcca","Type":"ContainerStarted","Data":"4436e7005084c425e3a9d35a46fa4e7c4f85b3ed742ac9d293c8d22b9b849826"} Jan 27 15:21:17 crc kubenswrapper[4772]: I0127 15:21:17.937249 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jhpnb" event={"ID":"5d41fcac-7044-4f36-b9f8-0b656bb3bcca","Type":"ContainerStarted","Data":"e923f314d195a301af3bc586b97562896538b03ab9e799d82ee601b7992f2baa"} Jan 27 15:21:17 crc kubenswrapper[4772]: I0127 15:21:17.937267 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jhpnb" event={"ID":"5d41fcac-7044-4f36-b9f8-0b656bb3bcca","Type":"ContainerStarted","Data":"0b029840d4b2e66b85e73893e9bb791afc17bfb135aa1f86951376b56341b53e"} Jan 27 15:21:17 crc kubenswrapper[4772]: I0127 15:21:17.937284 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jhpnb" event={"ID":"5d41fcac-7044-4f36-b9f8-0b656bb3bcca","Type":"ContainerStarted","Data":"2789caab35376e23e9fc736f8a3a208ae2e5b419ee54bf1ea2c19c82e715d467"} Jan 27 15:21:17 crc kubenswrapper[4772]: I0127 15:21:17.937310 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jhpnb" event={"ID":"5d41fcac-7044-4f36-b9f8-0b656bb3bcca","Type":"ContainerStarted","Data":"e4c82b762cf9effb2b151661ce0a466f1b96207b155c3c9f65d48e7046a558a3"} Jan 27 15:21:17 crc kubenswrapper[4772]: I0127 15:21:17.937340 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:18 crc kubenswrapper[4772]: I0127 15:21:18.800081 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-jhpnb" podStartSLOduration=6.066893337 podStartE2EDuration="13.800062963s" podCreationTimestamp="2026-01-27 15:21:05 +0000 UTC" firstStartedPulling="2026-01-27 15:21:06.003906252 +0000 UTC m=+851.984515360" lastFinishedPulling="2026-01-27 15:21:13.737075888 +0000 UTC m=+859.717684986" observedRunningTime="2026-01-27 15:21:17.989767543 +0000 UTC m=+863.970376751" watchObservedRunningTime="2026-01-27 15:21:18.800062963 +0000 UTC m=+864.780672061" Jan 27 15:21:18 crc kubenswrapper[4772]: I0127 15:21:18.804105 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg"] Jan 27 15:21:18 crc kubenswrapper[4772]: I0127 15:21:18.805150 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg" Jan 27 15:21:18 crc kubenswrapper[4772]: I0127 15:21:18.807539 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 27 15:21:18 crc kubenswrapper[4772]: I0127 15:21:18.814299 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg"] Jan 27 15:21:18 crc kubenswrapper[4772]: I0127 15:21:18.953667 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl89w\" (UniqueName: \"kubernetes.io/projected/44b239be-466d-4995-9c33-38d68a00550d-kube-api-access-rl89w\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg\" (UID: \"44b239be-466d-4995-9c33-38d68a00550d\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg" Jan 27 15:21:18 crc kubenswrapper[4772]: I0127 15:21:18.953754 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44b239be-466d-4995-9c33-38d68a00550d-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg\" (UID: \"44b239be-466d-4995-9c33-38d68a00550d\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg" Jan 27 15:21:18 crc kubenswrapper[4772]: I0127 15:21:18.953782 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44b239be-466d-4995-9c33-38d68a00550d-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg\" (UID: \"44b239be-466d-4995-9c33-38d68a00550d\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg" Jan 27 15:21:19 crc kubenswrapper[4772]: I0127 15:21:19.055545 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44b239be-466d-4995-9c33-38d68a00550d-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg\" (UID: \"44b239be-466d-4995-9c33-38d68a00550d\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg" Jan 27 15:21:19 crc kubenswrapper[4772]: I0127 15:21:19.055620 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44b239be-466d-4995-9c33-38d68a00550d-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg\" (UID: \"44b239be-466d-4995-9c33-38d68a00550d\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg" Jan 27 15:21:19 crc kubenswrapper[4772]: I0127 15:21:19.055715 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rl89w\" (UniqueName: \"kubernetes.io/projected/44b239be-466d-4995-9c33-38d68a00550d-kube-api-access-rl89w\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg\" (UID: \"44b239be-466d-4995-9c33-38d68a00550d\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg" Jan 27 15:21:19 crc kubenswrapper[4772]: I0127 15:21:19.056137 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44b239be-466d-4995-9c33-38d68a00550d-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg\" (UID: \"44b239be-466d-4995-9c33-38d68a00550d\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg" Jan 27 15:21:19 crc kubenswrapper[4772]: I0127 15:21:19.056257 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44b239be-466d-4995-9c33-38d68a00550d-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg\" (UID: \"44b239be-466d-4995-9c33-38d68a00550d\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg" Jan 27 15:21:19 crc kubenswrapper[4772]: I0127 15:21:19.089067 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rl89w\" (UniqueName: \"kubernetes.io/projected/44b239be-466d-4995-9c33-38d68a00550d-kube-api-access-rl89w\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg\" (UID: \"44b239be-466d-4995-9c33-38d68a00550d\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg" Jan 27 15:21:19 crc kubenswrapper[4772]: I0127 15:21:19.120809 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg" Jan 27 15:21:19 crc kubenswrapper[4772]: I0127 15:21:19.363365 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg"] Jan 27 15:21:19 crc kubenswrapper[4772]: W0127 15:21:19.371018 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44b239be_466d_4995_9c33_38d68a00550d.slice/crio-ef16ee5c9b0f6269a183df8b440e38235f63a12a2af4b9d1a81c3a7c37222be9 WatchSource:0}: Error finding container ef16ee5c9b0f6269a183df8b440e38235f63a12a2af4b9d1a81c3a7c37222be9: Status 404 returned error can't find the container with id ef16ee5c9b0f6269a183df8b440e38235f63a12a2af4b9d1a81c3a7c37222be9 Jan 27 15:21:19 crc kubenswrapper[4772]: I0127 15:21:19.952863 4772 generic.go:334] "Generic (PLEG): container finished" podID="44b239be-466d-4995-9c33-38d68a00550d" containerID="b21c4d3041de620d506243172af9adcd979dc0a44622b73db1c752a3490807be" exitCode=0 Jan 27 15:21:19 crc kubenswrapper[4772]: I0127 15:21:19.952991 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg" event={"ID":"44b239be-466d-4995-9c33-38d68a00550d","Type":"ContainerDied","Data":"b21c4d3041de620d506243172af9adcd979dc0a44622b73db1c752a3490807be"} Jan 27 15:21:19 crc kubenswrapper[4772]: I0127 15:21:19.953214 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg" event={"ID":"44b239be-466d-4995-9c33-38d68a00550d","Type":"ContainerStarted","Data":"ef16ee5c9b0f6269a183df8b440e38235f63a12a2af4b9d1a81c3a7c37222be9"} Jan 27 15:21:20 crc kubenswrapper[4772]: I0127 15:21:20.816712 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:20 crc kubenswrapper[4772]: I0127 15:21:20.879362 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:23 crc kubenswrapper[4772]: I0127 15:21:23.995303 4772 generic.go:334] "Generic (PLEG): container finished" podID="44b239be-466d-4995-9c33-38d68a00550d" containerID="cf0e54cd4817a89906c4991b53edd466f93be54e83126f84100f6a7ca6530b41" exitCode=0 Jan 27 15:21:23 crc kubenswrapper[4772]: I0127 15:21:23.995371 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg" event={"ID":"44b239be-466d-4995-9c33-38d68a00550d","Type":"ContainerDied","Data":"cf0e54cd4817a89906c4991b53edd466f93be54e83126f84100f6a7ca6530b41"} Jan 27 15:21:25 crc kubenswrapper[4772]: I0127 15:21:25.014155 4772 generic.go:334] "Generic (PLEG): container finished" podID="44b239be-466d-4995-9c33-38d68a00550d" containerID="9c82381351805aef1a46dfd050b839a3c77763cc224fe4c1bf4a211c797b81c0" exitCode=0 Jan 27 15:21:25 crc kubenswrapper[4772]: I0127 15:21:25.014311 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg" event={"ID":"44b239be-466d-4995-9c33-38d68a00550d","Type":"ContainerDied","Data":"9c82381351805aef1a46dfd050b839a3c77763cc224fe4c1bf4a211c797b81c0"} Jan 27 15:21:25 crc kubenswrapper[4772]: I0127 15:21:25.838375 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qpxrs" Jan 27 15:21:25 crc kubenswrapper[4772]: I0127 15:21:25.961714 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6968d8fdc4-vdg69" Jan 27 15:21:26 crc kubenswrapper[4772]: I0127 15:21:26.305292 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg" Jan 27 15:21:26 crc kubenswrapper[4772]: I0127 15:21:26.476987 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44b239be-466d-4995-9c33-38d68a00550d-util\") pod \"44b239be-466d-4995-9c33-38d68a00550d\" (UID: \"44b239be-466d-4995-9c33-38d68a00550d\") " Jan 27 15:21:26 crc kubenswrapper[4772]: I0127 15:21:26.477092 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44b239be-466d-4995-9c33-38d68a00550d-bundle\") pod \"44b239be-466d-4995-9c33-38d68a00550d\" (UID: \"44b239be-466d-4995-9c33-38d68a00550d\") " Jan 27 15:21:26 crc kubenswrapper[4772]: I0127 15:21:26.477119 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rl89w\" (UniqueName: \"kubernetes.io/projected/44b239be-466d-4995-9c33-38d68a00550d-kube-api-access-rl89w\") pod \"44b239be-466d-4995-9c33-38d68a00550d\" (UID: \"44b239be-466d-4995-9c33-38d68a00550d\") " Jan 27 15:21:26 crc kubenswrapper[4772]: I0127 15:21:26.477987 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44b239be-466d-4995-9c33-38d68a00550d-bundle" (OuterVolumeSpecName: "bundle") pod "44b239be-466d-4995-9c33-38d68a00550d" (UID: "44b239be-466d-4995-9c33-38d68a00550d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:21:26 crc kubenswrapper[4772]: I0127 15:21:26.482938 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44b239be-466d-4995-9c33-38d68a00550d-kube-api-access-rl89w" (OuterVolumeSpecName: "kube-api-access-rl89w") pod "44b239be-466d-4995-9c33-38d68a00550d" (UID: "44b239be-466d-4995-9c33-38d68a00550d"). InnerVolumeSpecName "kube-api-access-rl89w". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:21:26 crc kubenswrapper[4772]: I0127 15:21:26.493591 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44b239be-466d-4995-9c33-38d68a00550d-util" (OuterVolumeSpecName: "util") pod "44b239be-466d-4995-9c33-38d68a00550d" (UID: "44b239be-466d-4995-9c33-38d68a00550d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:21:26 crc kubenswrapper[4772]: I0127 15:21:26.578434 4772 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/44b239be-466d-4995-9c33-38d68a00550d-util\") on node \"crc\" DevicePath \"\"" Jan 27 15:21:26 crc kubenswrapper[4772]: I0127 15:21:26.578475 4772 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/44b239be-466d-4995-9c33-38d68a00550d-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:21:26 crc kubenswrapper[4772]: I0127 15:21:26.578494 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rl89w\" (UniqueName: \"kubernetes.io/projected/44b239be-466d-4995-9c33-38d68a00550d-kube-api-access-rl89w\") on node \"crc\" DevicePath \"\"" Jan 27 15:21:27 crc kubenswrapper[4772]: I0127 15:21:27.032987 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg" event={"ID":"44b239be-466d-4995-9c33-38d68a00550d","Type":"ContainerDied","Data":"ef16ee5c9b0f6269a183df8b440e38235f63a12a2af4b9d1a81c3a7c37222be9"} Jan 27 15:21:27 crc kubenswrapper[4772]: I0127 15:21:27.033043 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef16ee5c9b0f6269a183df8b440e38235f63a12a2af4b9d1a81c3a7c37222be9" Jan 27 15:21:27 crc kubenswrapper[4772]: I0127 15:21:27.033128 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg" Jan 27 15:21:32 crc kubenswrapper[4772]: I0127 15:21:32.065355 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fh8c9"] Jan 27 15:21:32 crc kubenswrapper[4772]: E0127 15:21:32.065920 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44b239be-466d-4995-9c33-38d68a00550d" containerName="extract" Jan 27 15:21:32 crc kubenswrapper[4772]: I0127 15:21:32.065934 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="44b239be-466d-4995-9c33-38d68a00550d" containerName="extract" Jan 27 15:21:32 crc kubenswrapper[4772]: E0127 15:21:32.065958 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44b239be-466d-4995-9c33-38d68a00550d" containerName="pull" Jan 27 15:21:32 crc kubenswrapper[4772]: I0127 15:21:32.065965 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="44b239be-466d-4995-9c33-38d68a00550d" containerName="pull" Jan 27 15:21:32 crc kubenswrapper[4772]: E0127 15:21:32.065978 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44b239be-466d-4995-9c33-38d68a00550d" containerName="util" Jan 27 15:21:32 crc kubenswrapper[4772]: I0127 15:21:32.065987 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="44b239be-466d-4995-9c33-38d68a00550d" containerName="util" Jan 27 15:21:32 crc kubenswrapper[4772]: I0127 15:21:32.066109 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="44b239be-466d-4995-9c33-38d68a00550d" containerName="extract" Jan 27 15:21:32 crc kubenswrapper[4772]: I0127 15:21:32.066594 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fh8c9" Jan 27 15:21:32 crc kubenswrapper[4772]: I0127 15:21:32.071226 4772 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-52slv" Jan 27 15:21:32 crc kubenswrapper[4772]: I0127 15:21:32.073037 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Jan 27 15:21:32 crc kubenswrapper[4772]: I0127 15:21:32.073519 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Jan 27 15:21:32 crc kubenswrapper[4772]: I0127 15:21:32.079045 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zgq5\" (UniqueName: \"kubernetes.io/projected/649069f7-0947-4089-9e65-ae192e952f8e-kube-api-access-8zgq5\") pod \"cert-manager-operator-controller-manager-64cf6dff88-fh8c9\" (UID: \"649069f7-0947-4089-9e65-ae192e952f8e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fh8c9" Jan 27 15:21:32 crc kubenswrapper[4772]: I0127 15:21:32.079118 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/649069f7-0947-4089-9e65-ae192e952f8e-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-fh8c9\" (UID: \"649069f7-0947-4089-9e65-ae192e952f8e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fh8c9" Jan 27 15:21:32 crc kubenswrapper[4772]: I0127 15:21:32.095873 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fh8c9"] Jan 27 15:21:32 crc kubenswrapper[4772]: I0127 15:21:32.180528 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/649069f7-0947-4089-9e65-ae192e952f8e-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-fh8c9\" (UID: \"649069f7-0947-4089-9e65-ae192e952f8e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fh8c9" Jan 27 15:21:32 crc kubenswrapper[4772]: I0127 15:21:32.180640 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zgq5\" (UniqueName: \"kubernetes.io/projected/649069f7-0947-4089-9e65-ae192e952f8e-kube-api-access-8zgq5\") pod \"cert-manager-operator-controller-manager-64cf6dff88-fh8c9\" (UID: \"649069f7-0947-4089-9e65-ae192e952f8e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fh8c9" Jan 27 15:21:32 crc kubenswrapper[4772]: I0127 15:21:32.181138 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/649069f7-0947-4089-9e65-ae192e952f8e-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-fh8c9\" (UID: \"649069f7-0947-4089-9e65-ae192e952f8e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fh8c9" Jan 27 15:21:32 crc kubenswrapper[4772]: I0127 15:21:32.221925 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zgq5\" (UniqueName: \"kubernetes.io/projected/649069f7-0947-4089-9e65-ae192e952f8e-kube-api-access-8zgq5\") pod \"cert-manager-operator-controller-manager-64cf6dff88-fh8c9\" (UID: \"649069f7-0947-4089-9e65-ae192e952f8e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fh8c9" Jan 27 15:21:32 crc kubenswrapper[4772]: I0127 15:21:32.381696 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fh8c9" Jan 27 15:21:32 crc kubenswrapper[4772]: I0127 15:21:32.637585 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fh8c9"] Jan 27 15:21:32 crc kubenswrapper[4772]: W0127 15:21:32.643380 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod649069f7_0947_4089_9e65_ae192e952f8e.slice/crio-bf7f6f71230081901c74a00d96446a7b7c53a966a8f273c8c4a01df45ebc987a WatchSource:0}: Error finding container bf7f6f71230081901c74a00d96446a7b7c53a966a8f273c8c4a01df45ebc987a: Status 404 returned error can't find the container with id bf7f6f71230081901c74a00d96446a7b7c53a966a8f273c8c4a01df45ebc987a Jan 27 15:21:33 crc kubenswrapper[4772]: I0127 15:21:33.069160 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fh8c9" event={"ID":"649069f7-0947-4089-9e65-ae192e952f8e","Type":"ContainerStarted","Data":"bf7f6f71230081901c74a00d96446a7b7c53a966a8f273c8c4a01df45ebc987a"} Jan 27 15:21:35 crc kubenswrapper[4772]: I0127 15:21:35.821106 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-jhpnb" Jan 27 15:21:40 crc kubenswrapper[4772]: I0127 15:21:40.124267 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fh8c9" event={"ID":"649069f7-0947-4089-9e65-ae192e952f8e","Type":"ContainerStarted","Data":"8e9ab94d2d2b624ead64b2485da2e8caa695e68f615d94bc2982b5a1ceebe2ed"} Jan 27 15:21:40 crc kubenswrapper[4772]: I0127 15:21:40.142723 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-fh8c9" podStartSLOduration=1.22977818 podStartE2EDuration="8.14269739s" podCreationTimestamp="2026-01-27 15:21:32 +0000 UTC" firstStartedPulling="2026-01-27 15:21:32.648100354 +0000 UTC m=+878.628709452" lastFinishedPulling="2026-01-27 15:21:39.561019564 +0000 UTC m=+885.541628662" observedRunningTime="2026-01-27 15:21:40.141122274 +0000 UTC m=+886.121731392" watchObservedRunningTime="2026-01-27 15:21:40.14269739 +0000 UTC m=+886.123306518" Jan 27 15:21:43 crc kubenswrapper[4772]: I0127 15:21:43.554650 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-4886f"] Jan 27 15:21:43 crc kubenswrapper[4772]: I0127 15:21:43.556318 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-4886f" Jan 27 15:21:43 crc kubenswrapper[4772]: I0127 15:21:43.561050 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Jan 27 15:21:43 crc kubenswrapper[4772]: I0127 15:21:43.561087 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Jan 27 15:21:43 crc kubenswrapper[4772]: I0127 15:21:43.565510 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-4886f"] Jan 27 15:21:43 crc kubenswrapper[4772]: I0127 15:21:43.566818 4772 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-8r264" Jan 27 15:21:43 crc kubenswrapper[4772]: I0127 15:21:43.745565 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqlbf\" (UniqueName: \"kubernetes.io/projected/ab67a3dd-5a79-400f-9b27-294ef256823d-kube-api-access-hqlbf\") pod \"cert-manager-webhook-f4fb5df64-4886f\" (UID: \"ab67a3dd-5a79-400f-9b27-294ef256823d\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-4886f" Jan 27 15:21:43 crc kubenswrapper[4772]: I0127 15:21:43.745679 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ab67a3dd-5a79-400f-9b27-294ef256823d-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-4886f\" (UID: \"ab67a3dd-5a79-400f-9b27-294ef256823d\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-4886f" Jan 27 15:21:43 crc kubenswrapper[4772]: I0127 15:21:43.848394 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ab67a3dd-5a79-400f-9b27-294ef256823d-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-4886f\" (UID: \"ab67a3dd-5a79-400f-9b27-294ef256823d\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-4886f" Jan 27 15:21:43 crc kubenswrapper[4772]: I0127 15:21:43.848530 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqlbf\" (UniqueName: \"kubernetes.io/projected/ab67a3dd-5a79-400f-9b27-294ef256823d-kube-api-access-hqlbf\") pod \"cert-manager-webhook-f4fb5df64-4886f\" (UID: \"ab67a3dd-5a79-400f-9b27-294ef256823d\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-4886f" Jan 27 15:21:43 crc kubenswrapper[4772]: I0127 15:21:43.872895 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ab67a3dd-5a79-400f-9b27-294ef256823d-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-4886f\" (UID: \"ab67a3dd-5a79-400f-9b27-294ef256823d\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-4886f" Jan 27 15:21:43 crc kubenswrapper[4772]: I0127 15:21:43.874324 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqlbf\" (UniqueName: \"kubernetes.io/projected/ab67a3dd-5a79-400f-9b27-294ef256823d-kube-api-access-hqlbf\") pod \"cert-manager-webhook-f4fb5df64-4886f\" (UID: \"ab67a3dd-5a79-400f-9b27-294ef256823d\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-4886f" Jan 27 15:21:43 crc kubenswrapper[4772]: I0127 15:21:43.899216 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-4886f" Jan 27 15:21:44 crc kubenswrapper[4772]: I0127 15:21:44.119100 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-4886f"] Jan 27 15:21:44 crc kubenswrapper[4772]: I0127 15:21:44.153962 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-4886f" event={"ID":"ab67a3dd-5a79-400f-9b27-294ef256823d","Type":"ContainerStarted","Data":"064f31700f599bf883e42ed7b95cbf043907cbd1f50c0ddff52a145423ce43af"} Jan 27 15:21:46 crc kubenswrapper[4772]: I0127 15:21:46.194250 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-s7ksj"] Jan 27 15:21:46 crc kubenswrapper[4772]: I0127 15:21:46.195260 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-s7ksj" Jan 27 15:21:46 crc kubenswrapper[4772]: I0127 15:21:46.197379 4772 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-8pmkj" Jan 27 15:21:46 crc kubenswrapper[4772]: I0127 15:21:46.201626 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-s7ksj"] Jan 27 15:21:46 crc kubenswrapper[4772]: I0127 15:21:46.276856 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fk7dp\" (UniqueName: \"kubernetes.io/projected/8d401dfc-33d3-416f-abba-cad4a1e173bd-kube-api-access-fk7dp\") pod \"cert-manager-cainjector-855d9ccff4-s7ksj\" (UID: \"8d401dfc-33d3-416f-abba-cad4a1e173bd\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-s7ksj" Jan 27 15:21:46 crc kubenswrapper[4772]: I0127 15:21:46.276957 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8d401dfc-33d3-416f-abba-cad4a1e173bd-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-s7ksj\" (UID: \"8d401dfc-33d3-416f-abba-cad4a1e173bd\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-s7ksj" Jan 27 15:21:46 crc kubenswrapper[4772]: I0127 15:21:46.377803 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fk7dp\" (UniqueName: \"kubernetes.io/projected/8d401dfc-33d3-416f-abba-cad4a1e173bd-kube-api-access-fk7dp\") pod \"cert-manager-cainjector-855d9ccff4-s7ksj\" (UID: \"8d401dfc-33d3-416f-abba-cad4a1e173bd\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-s7ksj" Jan 27 15:21:46 crc kubenswrapper[4772]: I0127 15:21:46.378108 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8d401dfc-33d3-416f-abba-cad4a1e173bd-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-s7ksj\" (UID: \"8d401dfc-33d3-416f-abba-cad4a1e173bd\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-s7ksj" Jan 27 15:21:46 crc kubenswrapper[4772]: I0127 15:21:46.397107 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8d401dfc-33d3-416f-abba-cad4a1e173bd-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-s7ksj\" (UID: \"8d401dfc-33d3-416f-abba-cad4a1e173bd\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-s7ksj" Jan 27 15:21:46 crc kubenswrapper[4772]: I0127 15:21:46.398304 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fk7dp\" (UniqueName: \"kubernetes.io/projected/8d401dfc-33d3-416f-abba-cad4a1e173bd-kube-api-access-fk7dp\") pod \"cert-manager-cainjector-855d9ccff4-s7ksj\" (UID: \"8d401dfc-33d3-416f-abba-cad4a1e173bd\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-s7ksj" Jan 27 15:21:46 crc kubenswrapper[4772]: I0127 15:21:46.517848 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-s7ksj" Jan 27 15:21:46 crc kubenswrapper[4772]: I0127 15:21:46.738087 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-s7ksj"] Jan 27 15:21:47 crc kubenswrapper[4772]: I0127 15:21:47.176822 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-s7ksj" event={"ID":"8d401dfc-33d3-416f-abba-cad4a1e173bd","Type":"ContainerStarted","Data":"1fe539967b131aa9595b66edc664c29e3406effab2e49a53e7c776bc52a08faa"} Jan 27 15:21:53 crc kubenswrapper[4772]: I0127 15:21:53.207699 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-s7ksj" event={"ID":"8d401dfc-33d3-416f-abba-cad4a1e173bd","Type":"ContainerStarted","Data":"2af0c1e9f079137b54a488e65ed83dd57de1eb3f2452bab5e7bd7083aefa0e52"} Jan 27 15:21:53 crc kubenswrapper[4772]: I0127 15:21:53.210129 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-4886f" event={"ID":"ab67a3dd-5a79-400f-9b27-294ef256823d","Type":"ContainerStarted","Data":"cd9cd0ee0baebdf60fd4c2c607e947c53e17c28f3d2ced16c697a54ad535f73b"} Jan 27 15:21:53 crc kubenswrapper[4772]: I0127 15:21:53.210252 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-4886f" Jan 27 15:21:53 crc kubenswrapper[4772]: I0127 15:21:53.251592 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-s7ksj" podStartSLOduration=1.980454093 podStartE2EDuration="7.251572894s" podCreationTimestamp="2026-01-27 15:21:46 +0000 UTC" firstStartedPulling="2026-01-27 15:21:46.754332726 +0000 UTC m=+892.734941824" lastFinishedPulling="2026-01-27 15:21:52.025451517 +0000 UTC m=+898.006060625" observedRunningTime="2026-01-27 15:21:52.220357672 +0000 UTC m=+898.200966770" watchObservedRunningTime="2026-01-27 15:21:53.251572894 +0000 UTC m=+899.232181992" Jan 27 15:21:53 crc kubenswrapper[4772]: I0127 15:21:53.253243 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-4886f" podStartSLOduration=2.328066595 podStartE2EDuration="10.253235712s" podCreationTimestamp="2026-01-27 15:21:43 +0000 UTC" firstStartedPulling="2026-01-27 15:21:44.13853726 +0000 UTC m=+890.119146348" lastFinishedPulling="2026-01-27 15:21:52.063706367 +0000 UTC m=+898.044315465" observedRunningTime="2026-01-27 15:21:53.249893816 +0000 UTC m=+899.230502934" watchObservedRunningTime="2026-01-27 15:21:53.253235712 +0000 UTC m=+899.233844810" Jan 27 15:21:54 crc kubenswrapper[4772]: I0127 15:21:54.309495 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-bmz8p"] Jan 27 15:21:54 crc kubenswrapper[4772]: I0127 15:21:54.310731 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-bmz8p" Jan 27 15:21:54 crc kubenswrapper[4772]: I0127 15:21:54.313995 4772 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-wrc5s" Jan 27 15:21:54 crc kubenswrapper[4772]: I0127 15:21:54.322389 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-bmz8p"] Jan 27 15:21:54 crc kubenswrapper[4772]: I0127 15:21:54.347862 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7b51171e-5d65-415e-8052-3cc8991f5de4-bound-sa-token\") pod \"cert-manager-86cb77c54b-bmz8p\" (UID: \"7b51171e-5d65-415e-8052-3cc8991f5de4\") " pod="cert-manager/cert-manager-86cb77c54b-bmz8p" Jan 27 15:21:54 crc kubenswrapper[4772]: I0127 15:21:54.347925 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9rcj\" (UniqueName: \"kubernetes.io/projected/7b51171e-5d65-415e-8052-3cc8991f5de4-kube-api-access-x9rcj\") pod \"cert-manager-86cb77c54b-bmz8p\" (UID: \"7b51171e-5d65-415e-8052-3cc8991f5de4\") " pod="cert-manager/cert-manager-86cb77c54b-bmz8p" Jan 27 15:21:54 crc kubenswrapper[4772]: I0127 15:21:54.448785 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7b51171e-5d65-415e-8052-3cc8991f5de4-bound-sa-token\") pod \"cert-manager-86cb77c54b-bmz8p\" (UID: \"7b51171e-5d65-415e-8052-3cc8991f5de4\") " pod="cert-manager/cert-manager-86cb77c54b-bmz8p" Jan 27 15:21:54 crc kubenswrapper[4772]: I0127 15:21:54.448863 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9rcj\" (UniqueName: \"kubernetes.io/projected/7b51171e-5d65-415e-8052-3cc8991f5de4-kube-api-access-x9rcj\") pod \"cert-manager-86cb77c54b-bmz8p\" (UID: \"7b51171e-5d65-415e-8052-3cc8991f5de4\") " pod="cert-manager/cert-manager-86cb77c54b-bmz8p" Jan 27 15:21:54 crc kubenswrapper[4772]: I0127 15:21:54.467064 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9rcj\" (UniqueName: \"kubernetes.io/projected/7b51171e-5d65-415e-8052-3cc8991f5de4-kube-api-access-x9rcj\") pod \"cert-manager-86cb77c54b-bmz8p\" (UID: \"7b51171e-5d65-415e-8052-3cc8991f5de4\") " pod="cert-manager/cert-manager-86cb77c54b-bmz8p" Jan 27 15:21:54 crc kubenswrapper[4772]: I0127 15:21:54.470960 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7b51171e-5d65-415e-8052-3cc8991f5de4-bound-sa-token\") pod \"cert-manager-86cb77c54b-bmz8p\" (UID: \"7b51171e-5d65-415e-8052-3cc8991f5de4\") " pod="cert-manager/cert-manager-86cb77c54b-bmz8p" Jan 27 15:21:54 crc kubenswrapper[4772]: I0127 15:21:54.634945 4772 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-wrc5s" Jan 27 15:21:54 crc kubenswrapper[4772]: I0127 15:21:54.643832 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-bmz8p" Jan 27 15:21:55 crc kubenswrapper[4772]: I0127 15:21:55.046015 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-bmz8p"] Jan 27 15:21:55 crc kubenswrapper[4772]: W0127 15:21:55.050303 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b51171e_5d65_415e_8052_3cc8991f5de4.slice/crio-9f7057cb198afc00edbb3717c98144058c5b710f4507c7ac0b1afce1b6e73def WatchSource:0}: Error finding container 9f7057cb198afc00edbb3717c98144058c5b710f4507c7ac0b1afce1b6e73def: Status 404 returned error can't find the container with id 9f7057cb198afc00edbb3717c98144058c5b710f4507c7ac0b1afce1b6e73def Jan 27 15:21:55 crc kubenswrapper[4772]: I0127 15:21:55.225256 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-bmz8p" event={"ID":"7b51171e-5d65-415e-8052-3cc8991f5de4","Type":"ContainerStarted","Data":"b9cc0505c7e78010c61f93057c4ae5e6a5ca8b108d2e0825f7101acb70d5db1d"} Jan 27 15:21:55 crc kubenswrapper[4772]: I0127 15:21:55.225657 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-bmz8p" event={"ID":"7b51171e-5d65-415e-8052-3cc8991f5de4","Type":"ContainerStarted","Data":"9f7057cb198afc00edbb3717c98144058c5b710f4507c7ac0b1afce1b6e73def"} Jan 27 15:21:55 crc kubenswrapper[4772]: I0127 15:21:55.246450 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-bmz8p" podStartSLOduration=1.246429646 podStartE2EDuration="1.246429646s" podCreationTimestamp="2026-01-27 15:21:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:21:55.244637535 +0000 UTC m=+901.225246643" watchObservedRunningTime="2026-01-27 15:21:55.246429646 +0000 UTC m=+901.227038744" Jan 27 15:21:57 crc kubenswrapper[4772]: I0127 15:21:57.933783 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6pm7m"] Jan 27 15:21:57 crc kubenswrapper[4772]: I0127 15:21:57.935392 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6pm7m" Jan 27 15:21:57 crc kubenswrapper[4772]: I0127 15:21:57.955059 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6pm7m"] Jan 27 15:21:58 crc kubenswrapper[4772]: I0127 15:21:58.002917 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxgxq\" (UniqueName: \"kubernetes.io/projected/bdc687da-0857-4070-a27a-90b08ca108c9-kube-api-access-gxgxq\") pod \"redhat-marketplace-6pm7m\" (UID: \"bdc687da-0857-4070-a27a-90b08ca108c9\") " pod="openshift-marketplace/redhat-marketplace-6pm7m" Jan 27 15:21:58 crc kubenswrapper[4772]: I0127 15:21:58.003006 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdc687da-0857-4070-a27a-90b08ca108c9-utilities\") pod \"redhat-marketplace-6pm7m\" (UID: \"bdc687da-0857-4070-a27a-90b08ca108c9\") " pod="openshift-marketplace/redhat-marketplace-6pm7m" Jan 27 15:21:58 crc kubenswrapper[4772]: I0127 15:21:58.003093 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdc687da-0857-4070-a27a-90b08ca108c9-catalog-content\") pod \"redhat-marketplace-6pm7m\" (UID: \"bdc687da-0857-4070-a27a-90b08ca108c9\") " pod="openshift-marketplace/redhat-marketplace-6pm7m" Jan 27 15:21:58 crc kubenswrapper[4772]: I0127 15:21:58.104821 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxgxq\" (UniqueName: \"kubernetes.io/projected/bdc687da-0857-4070-a27a-90b08ca108c9-kube-api-access-gxgxq\") pod \"redhat-marketplace-6pm7m\" (UID: \"bdc687da-0857-4070-a27a-90b08ca108c9\") " pod="openshift-marketplace/redhat-marketplace-6pm7m" Jan 27 15:21:58 crc kubenswrapper[4772]: I0127 15:21:58.104927 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdc687da-0857-4070-a27a-90b08ca108c9-utilities\") pod \"redhat-marketplace-6pm7m\" (UID: \"bdc687da-0857-4070-a27a-90b08ca108c9\") " pod="openshift-marketplace/redhat-marketplace-6pm7m" Jan 27 15:21:58 crc kubenswrapper[4772]: I0127 15:21:58.104988 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdc687da-0857-4070-a27a-90b08ca108c9-catalog-content\") pod \"redhat-marketplace-6pm7m\" (UID: \"bdc687da-0857-4070-a27a-90b08ca108c9\") " pod="openshift-marketplace/redhat-marketplace-6pm7m" Jan 27 15:21:58 crc kubenswrapper[4772]: I0127 15:21:58.105453 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdc687da-0857-4070-a27a-90b08ca108c9-utilities\") pod \"redhat-marketplace-6pm7m\" (UID: \"bdc687da-0857-4070-a27a-90b08ca108c9\") " pod="openshift-marketplace/redhat-marketplace-6pm7m" Jan 27 15:21:58 crc kubenswrapper[4772]: I0127 15:21:58.105534 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdc687da-0857-4070-a27a-90b08ca108c9-catalog-content\") pod \"redhat-marketplace-6pm7m\" (UID: \"bdc687da-0857-4070-a27a-90b08ca108c9\") " pod="openshift-marketplace/redhat-marketplace-6pm7m" Jan 27 15:21:58 crc kubenswrapper[4772]: I0127 15:21:58.128364 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxgxq\" (UniqueName: \"kubernetes.io/projected/bdc687da-0857-4070-a27a-90b08ca108c9-kube-api-access-gxgxq\") pod \"redhat-marketplace-6pm7m\" (UID: \"bdc687da-0857-4070-a27a-90b08ca108c9\") " pod="openshift-marketplace/redhat-marketplace-6pm7m" Jan 27 15:21:58 crc kubenswrapper[4772]: I0127 15:21:58.261139 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6pm7m" Jan 27 15:21:58 crc kubenswrapper[4772]: I0127 15:21:58.507236 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6pm7m"] Jan 27 15:21:58 crc kubenswrapper[4772]: I0127 15:21:58.901399 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-4886f" Jan 27 15:21:59 crc kubenswrapper[4772]: I0127 15:21:59.252488 4772 generic.go:334] "Generic (PLEG): container finished" podID="bdc687da-0857-4070-a27a-90b08ca108c9" containerID="94686f897895f9959663a28bfa07c2aaa2acb444714ed141333b20d65eed7fc2" exitCode=0 Jan 27 15:21:59 crc kubenswrapper[4772]: I0127 15:21:59.252549 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6pm7m" event={"ID":"bdc687da-0857-4070-a27a-90b08ca108c9","Type":"ContainerDied","Data":"94686f897895f9959663a28bfa07c2aaa2acb444714ed141333b20d65eed7fc2"} Jan 27 15:21:59 crc kubenswrapper[4772]: I0127 15:21:59.252585 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6pm7m" event={"ID":"bdc687da-0857-4070-a27a-90b08ca108c9","Type":"ContainerStarted","Data":"41d3f481a49e568f6c8ae964bb8f25577c85d9700b3c5df1b8ce8fff898b4f29"} Jan 27 15:22:01 crc kubenswrapper[4772]: I0127 15:22:01.272294 4772 generic.go:334] "Generic (PLEG): container finished" podID="bdc687da-0857-4070-a27a-90b08ca108c9" containerID="6a60cf692ab3a3eae06e2802f1931e40f49d38619fda7b12604a36a8f58a1dbe" exitCode=0 Jan 27 15:22:01 crc kubenswrapper[4772]: I0127 15:22:01.272442 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6pm7m" event={"ID":"bdc687da-0857-4070-a27a-90b08ca108c9","Type":"ContainerDied","Data":"6a60cf692ab3a3eae06e2802f1931e40f49d38619fda7b12604a36a8f58a1dbe"} Jan 27 15:22:02 crc kubenswrapper[4772]: I0127 15:22:02.280962 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6pm7m" event={"ID":"bdc687da-0857-4070-a27a-90b08ca108c9","Type":"ContainerStarted","Data":"5b7851c728141b28bc16da228e27af7b01b686d419421b07643b0ccbbf7acd5a"} Jan 27 15:22:02 crc kubenswrapper[4772]: I0127 15:22:02.314107 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6pm7m" podStartSLOduration=2.8753379199999998 podStartE2EDuration="5.314081786s" podCreationTimestamp="2026-01-27 15:21:57 +0000 UTC" firstStartedPulling="2026-01-27 15:21:59.254256461 +0000 UTC m=+905.234865599" lastFinishedPulling="2026-01-27 15:22:01.693000327 +0000 UTC m=+907.673609465" observedRunningTime="2026-01-27 15:22:02.306692724 +0000 UTC m=+908.287301842" watchObservedRunningTime="2026-01-27 15:22:02.314081786 +0000 UTC m=+908.294690924" Jan 27 15:22:05 crc kubenswrapper[4772]: I0127 15:22:05.324946 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-7lq2b"] Jan 27 15:22:05 crc kubenswrapper[4772]: I0127 15:22:05.327094 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7lq2b" Jan 27 15:22:05 crc kubenswrapper[4772]: I0127 15:22:05.334997 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Jan 27 15:22:05 crc kubenswrapper[4772]: I0127 15:22:05.335196 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Jan 27 15:22:05 crc kubenswrapper[4772]: I0127 15:22:05.335484 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-4m6p5" Jan 27 15:22:05 crc kubenswrapper[4772]: I0127 15:22:05.336503 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-7lq2b"] Jan 27 15:22:05 crc kubenswrapper[4772]: I0127 15:22:05.411822 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5hnl\" (UniqueName: \"kubernetes.io/projected/13489498-3c32-4ef1-baf5-99f6907d07e4-kube-api-access-j5hnl\") pod \"openstack-operator-index-7lq2b\" (UID: \"13489498-3c32-4ef1-baf5-99f6907d07e4\") " pod="openstack-operators/openstack-operator-index-7lq2b" Jan 27 15:22:05 crc kubenswrapper[4772]: I0127 15:22:05.513815 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5hnl\" (UniqueName: \"kubernetes.io/projected/13489498-3c32-4ef1-baf5-99f6907d07e4-kube-api-access-j5hnl\") pod \"openstack-operator-index-7lq2b\" (UID: \"13489498-3c32-4ef1-baf5-99f6907d07e4\") " pod="openstack-operators/openstack-operator-index-7lq2b" Jan 27 15:22:05 crc kubenswrapper[4772]: I0127 15:22:05.547943 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5hnl\" (UniqueName: \"kubernetes.io/projected/13489498-3c32-4ef1-baf5-99f6907d07e4-kube-api-access-j5hnl\") pod \"openstack-operator-index-7lq2b\" (UID: \"13489498-3c32-4ef1-baf5-99f6907d07e4\") " pod="openstack-operators/openstack-operator-index-7lq2b" Jan 27 15:22:05 crc kubenswrapper[4772]: I0127 15:22:05.663949 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7lq2b" Jan 27 15:22:06 crc kubenswrapper[4772]: I0127 15:22:06.140478 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-7lq2b"] Jan 27 15:22:06 crc kubenswrapper[4772]: I0127 15:22:06.307247 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7lq2b" event={"ID":"13489498-3c32-4ef1-baf5-99f6907d07e4","Type":"ContainerStarted","Data":"055263e39566289af029844504c4a8375ca8aa346f646c3a569e5ad465cedaa7"} Jan 27 15:22:08 crc kubenswrapper[4772]: I0127 15:22:08.262081 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6pm7m" Jan 27 15:22:08 crc kubenswrapper[4772]: I0127 15:22:08.262497 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6pm7m" Jan 27 15:22:08 crc kubenswrapper[4772]: I0127 15:22:08.302506 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6pm7m" Jan 27 15:22:08 crc kubenswrapper[4772]: I0127 15:22:08.367468 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6pm7m" Jan 27 15:22:09 crc kubenswrapper[4772]: I0127 15:22:09.326455 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7lq2b" event={"ID":"13489498-3c32-4ef1-baf5-99f6907d07e4","Type":"ContainerStarted","Data":"d289b63c1084f4442f5f1a2e1c20d7abe82ebe1562d47d96d12194c6053992a4"} Jan 27 15:22:09 crc kubenswrapper[4772]: I0127 15:22:09.342964 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-7lq2b" podStartSLOduration=1.7184273129999998 podStartE2EDuration="4.34293812s" podCreationTimestamp="2026-01-27 15:22:05 +0000 UTC" firstStartedPulling="2026-01-27 15:22:06.208593123 +0000 UTC m=+912.189202221" lastFinishedPulling="2026-01-27 15:22:08.83310389 +0000 UTC m=+914.813713028" observedRunningTime="2026-01-27 15:22:09.341386925 +0000 UTC m=+915.321996063" watchObservedRunningTime="2026-01-27 15:22:09.34293812 +0000 UTC m=+915.323547238" Jan 27 15:22:09 crc kubenswrapper[4772]: I0127 15:22:09.902822 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-7lq2b"] Jan 27 15:22:10 crc kubenswrapper[4772]: I0127 15:22:10.512119 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-vs9rk"] Jan 27 15:22:10 crc kubenswrapper[4772]: I0127 15:22:10.513037 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vs9rk" Jan 27 15:22:10 crc kubenswrapper[4772]: I0127 15:22:10.522695 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-vs9rk"] Jan 27 15:22:10 crc kubenswrapper[4772]: I0127 15:22:10.585806 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq9lb\" (UniqueName: \"kubernetes.io/projected/96fcf3a5-2584-4590-8057-9c18a9866bd4-kube-api-access-mq9lb\") pod \"openstack-operator-index-vs9rk\" (UID: \"96fcf3a5-2584-4590-8057-9c18a9866bd4\") " pod="openstack-operators/openstack-operator-index-vs9rk" Jan 27 15:22:10 crc kubenswrapper[4772]: I0127 15:22:10.687418 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq9lb\" (UniqueName: \"kubernetes.io/projected/96fcf3a5-2584-4590-8057-9c18a9866bd4-kube-api-access-mq9lb\") pod \"openstack-operator-index-vs9rk\" (UID: \"96fcf3a5-2584-4590-8057-9c18a9866bd4\") " pod="openstack-operators/openstack-operator-index-vs9rk" Jan 27 15:22:10 crc kubenswrapper[4772]: I0127 15:22:10.722262 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq9lb\" (UniqueName: \"kubernetes.io/projected/96fcf3a5-2584-4590-8057-9c18a9866bd4-kube-api-access-mq9lb\") pod \"openstack-operator-index-vs9rk\" (UID: \"96fcf3a5-2584-4590-8057-9c18a9866bd4\") " pod="openstack-operators/openstack-operator-index-vs9rk" Jan 27 15:22:10 crc kubenswrapper[4772]: I0127 15:22:10.843271 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vs9rk" Jan 27 15:22:11 crc kubenswrapper[4772]: W0127 15:22:11.287248 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96fcf3a5_2584_4590_8057_9c18a9866bd4.slice/crio-51e818974842a23b8b8c4b017ed47491446211a1c89d4658b8224801df3d4faf WatchSource:0}: Error finding container 51e818974842a23b8b8c4b017ed47491446211a1c89d4658b8224801df3d4faf: Status 404 returned error can't find the container with id 51e818974842a23b8b8c4b017ed47491446211a1c89d4658b8224801df3d4faf Jan 27 15:22:11 crc kubenswrapper[4772]: I0127 15:22:11.287736 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-vs9rk"] Jan 27 15:22:11 crc kubenswrapper[4772]: I0127 15:22:11.341430 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vs9rk" event={"ID":"96fcf3a5-2584-4590-8057-9c18a9866bd4","Type":"ContainerStarted","Data":"51e818974842a23b8b8c4b017ed47491446211a1c89d4658b8224801df3d4faf"} Jan 27 15:22:11 crc kubenswrapper[4772]: I0127 15:22:11.341664 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-7lq2b" podUID="13489498-3c32-4ef1-baf5-99f6907d07e4" containerName="registry-server" containerID="cri-o://d289b63c1084f4442f5f1a2e1c20d7abe82ebe1562d47d96d12194c6053992a4" gracePeriod=2 Jan 27 15:22:11 crc kubenswrapper[4772]: I0127 15:22:11.692534 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7lq2b" Jan 27 15:22:11 crc kubenswrapper[4772]: I0127 15:22:11.804580 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5hnl\" (UniqueName: \"kubernetes.io/projected/13489498-3c32-4ef1-baf5-99f6907d07e4-kube-api-access-j5hnl\") pod \"13489498-3c32-4ef1-baf5-99f6907d07e4\" (UID: \"13489498-3c32-4ef1-baf5-99f6907d07e4\") " Jan 27 15:22:11 crc kubenswrapper[4772]: I0127 15:22:11.809641 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13489498-3c32-4ef1-baf5-99f6907d07e4-kube-api-access-j5hnl" (OuterVolumeSpecName: "kube-api-access-j5hnl") pod "13489498-3c32-4ef1-baf5-99f6907d07e4" (UID: "13489498-3c32-4ef1-baf5-99f6907d07e4"). InnerVolumeSpecName "kube-api-access-j5hnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:22:11 crc kubenswrapper[4772]: I0127 15:22:11.906447 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5hnl\" (UniqueName: \"kubernetes.io/projected/13489498-3c32-4ef1-baf5-99f6907d07e4-kube-api-access-j5hnl\") on node \"crc\" DevicePath \"\"" Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.058746 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.059257 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.107645 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6pm7m"] Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.107983 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6pm7m" podUID="bdc687da-0857-4070-a27a-90b08ca108c9" containerName="registry-server" containerID="cri-o://5b7851c728141b28bc16da228e27af7b01b686d419421b07643b0ccbbf7acd5a" gracePeriod=2 Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.350951 4772 generic.go:334] "Generic (PLEG): container finished" podID="bdc687da-0857-4070-a27a-90b08ca108c9" containerID="5b7851c728141b28bc16da228e27af7b01b686d419421b07643b0ccbbf7acd5a" exitCode=0 Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.351044 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6pm7m" event={"ID":"bdc687da-0857-4070-a27a-90b08ca108c9","Type":"ContainerDied","Data":"5b7851c728141b28bc16da228e27af7b01b686d419421b07643b0ccbbf7acd5a"} Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.352538 4772 generic.go:334] "Generic (PLEG): container finished" podID="13489498-3c32-4ef1-baf5-99f6907d07e4" containerID="d289b63c1084f4442f5f1a2e1c20d7abe82ebe1562d47d96d12194c6053992a4" exitCode=0 Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.352614 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7lq2b" Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.352625 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7lq2b" event={"ID":"13489498-3c32-4ef1-baf5-99f6907d07e4","Type":"ContainerDied","Data":"d289b63c1084f4442f5f1a2e1c20d7abe82ebe1562d47d96d12194c6053992a4"} Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.352666 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7lq2b" event={"ID":"13489498-3c32-4ef1-baf5-99f6907d07e4","Type":"ContainerDied","Data":"055263e39566289af029844504c4a8375ca8aa346f646c3a569e5ad465cedaa7"} Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.352697 4772 scope.go:117] "RemoveContainer" containerID="d289b63c1084f4442f5f1a2e1c20d7abe82ebe1562d47d96d12194c6053992a4" Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.354299 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vs9rk" event={"ID":"96fcf3a5-2584-4590-8057-9c18a9866bd4","Type":"ContainerStarted","Data":"38b35858a30e11e3719ec119912ad142df8868545490ca4c217b5258f55412a0"} Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.381953 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-vs9rk" podStartSLOduration=2.305204858 podStartE2EDuration="2.381929635s" podCreationTimestamp="2026-01-27 15:22:10 +0000 UTC" firstStartedPulling="2026-01-27 15:22:11.295899087 +0000 UTC m=+917.276508205" lastFinishedPulling="2026-01-27 15:22:11.372623884 +0000 UTC m=+917.353232982" observedRunningTime="2026-01-27 15:22:12.378403784 +0000 UTC m=+918.359012912" watchObservedRunningTime="2026-01-27 15:22:12.381929635 +0000 UTC m=+918.362538753" Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.398816 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-7lq2b"] Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.399353 4772 scope.go:117] "RemoveContainer" containerID="d289b63c1084f4442f5f1a2e1c20d7abe82ebe1562d47d96d12194c6053992a4" Jan 27 15:22:12 crc kubenswrapper[4772]: E0127 15:22:12.399794 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d289b63c1084f4442f5f1a2e1c20d7abe82ebe1562d47d96d12194c6053992a4\": container with ID starting with d289b63c1084f4442f5f1a2e1c20d7abe82ebe1562d47d96d12194c6053992a4 not found: ID does not exist" containerID="d289b63c1084f4442f5f1a2e1c20d7abe82ebe1562d47d96d12194c6053992a4" Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.399831 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d289b63c1084f4442f5f1a2e1c20d7abe82ebe1562d47d96d12194c6053992a4"} err="failed to get container status \"d289b63c1084f4442f5f1a2e1c20d7abe82ebe1562d47d96d12194c6053992a4\": rpc error: code = NotFound desc = could not find container \"d289b63c1084f4442f5f1a2e1c20d7abe82ebe1562d47d96d12194c6053992a4\": container with ID starting with d289b63c1084f4442f5f1a2e1c20d7abe82ebe1562d47d96d12194c6053992a4 not found: ID does not exist" Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.410465 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-7lq2b"] Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.510679 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6pm7m" Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.621471 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxgxq\" (UniqueName: \"kubernetes.io/projected/bdc687da-0857-4070-a27a-90b08ca108c9-kube-api-access-gxgxq\") pod \"bdc687da-0857-4070-a27a-90b08ca108c9\" (UID: \"bdc687da-0857-4070-a27a-90b08ca108c9\") " Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.621541 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdc687da-0857-4070-a27a-90b08ca108c9-utilities\") pod \"bdc687da-0857-4070-a27a-90b08ca108c9\" (UID: \"bdc687da-0857-4070-a27a-90b08ca108c9\") " Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.621579 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdc687da-0857-4070-a27a-90b08ca108c9-catalog-content\") pod \"bdc687da-0857-4070-a27a-90b08ca108c9\" (UID: \"bdc687da-0857-4070-a27a-90b08ca108c9\") " Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.622351 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdc687da-0857-4070-a27a-90b08ca108c9-utilities" (OuterVolumeSpecName: "utilities") pod "bdc687da-0857-4070-a27a-90b08ca108c9" (UID: "bdc687da-0857-4070-a27a-90b08ca108c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.625037 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdc687da-0857-4070-a27a-90b08ca108c9-kube-api-access-gxgxq" (OuterVolumeSpecName: "kube-api-access-gxgxq") pod "bdc687da-0857-4070-a27a-90b08ca108c9" (UID: "bdc687da-0857-4070-a27a-90b08ca108c9"). InnerVolumeSpecName "kube-api-access-gxgxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.666041 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdc687da-0857-4070-a27a-90b08ca108c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bdc687da-0857-4070-a27a-90b08ca108c9" (UID: "bdc687da-0857-4070-a27a-90b08ca108c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.677410 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13489498-3c32-4ef1-baf5-99f6907d07e4" path="/var/lib/kubelet/pods/13489498-3c32-4ef1-baf5-99f6907d07e4/volumes" Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.722736 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdc687da-0857-4070-a27a-90b08ca108c9-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.722760 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxgxq\" (UniqueName: \"kubernetes.io/projected/bdc687da-0857-4070-a27a-90b08ca108c9-kube-api-access-gxgxq\") on node \"crc\" DevicePath \"\"" Jan 27 15:22:12 crc kubenswrapper[4772]: I0127 15:22:12.722772 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdc687da-0857-4070-a27a-90b08ca108c9-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 15:22:13 crc kubenswrapper[4772]: I0127 15:22:13.365540 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6pm7m" event={"ID":"bdc687da-0857-4070-a27a-90b08ca108c9","Type":"ContainerDied","Data":"41d3f481a49e568f6c8ae964bb8f25577c85d9700b3c5df1b8ce8fff898b4f29"} Jan 27 15:22:13 crc kubenswrapper[4772]: I0127 15:22:13.365612 4772 scope.go:117] "RemoveContainer" containerID="5b7851c728141b28bc16da228e27af7b01b686d419421b07643b0ccbbf7acd5a" Jan 27 15:22:13 crc kubenswrapper[4772]: I0127 15:22:13.365684 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6pm7m" Jan 27 15:22:13 crc kubenswrapper[4772]: I0127 15:22:13.396210 4772 scope.go:117] "RemoveContainer" containerID="6a60cf692ab3a3eae06e2802f1931e40f49d38619fda7b12604a36a8f58a1dbe" Jan 27 15:22:13 crc kubenswrapper[4772]: I0127 15:22:13.404430 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6pm7m"] Jan 27 15:22:13 crc kubenswrapper[4772]: I0127 15:22:13.410467 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6pm7m"] Jan 27 15:22:13 crc kubenswrapper[4772]: I0127 15:22:13.420867 4772 scope.go:117] "RemoveContainer" containerID="94686f897895f9959663a28bfa07c2aaa2acb444714ed141333b20d65eed7fc2" Jan 27 15:22:14 crc kubenswrapper[4772]: I0127 15:22:14.677623 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdc687da-0857-4070-a27a-90b08ca108c9" path="/var/lib/kubelet/pods/bdc687da-0857-4070-a27a-90b08ca108c9/volumes" Jan 27 15:22:20 crc kubenswrapper[4772]: I0127 15:22:20.844443 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-vs9rk" Jan 27 15:22:20 crc kubenswrapper[4772]: I0127 15:22:20.845021 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-vs9rk" Jan 27 15:22:20 crc kubenswrapper[4772]: I0127 15:22:20.881925 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-vs9rk" Jan 27 15:22:21 crc kubenswrapper[4772]: I0127 15:22:21.462600 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-vs9rk" Jan 27 15:22:22 crc kubenswrapper[4772]: I0127 15:22:22.952273 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4"] Jan 27 15:22:22 crc kubenswrapper[4772]: E0127 15:22:22.952967 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdc687da-0857-4070-a27a-90b08ca108c9" containerName="registry-server" Jan 27 15:22:22 crc kubenswrapper[4772]: I0127 15:22:22.952982 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdc687da-0857-4070-a27a-90b08ca108c9" containerName="registry-server" Jan 27 15:22:22 crc kubenswrapper[4772]: E0127 15:22:22.952996 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdc687da-0857-4070-a27a-90b08ca108c9" containerName="extract-content" Jan 27 15:22:22 crc kubenswrapper[4772]: I0127 15:22:22.953004 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdc687da-0857-4070-a27a-90b08ca108c9" containerName="extract-content" Jan 27 15:22:22 crc kubenswrapper[4772]: E0127 15:22:22.953030 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13489498-3c32-4ef1-baf5-99f6907d07e4" containerName="registry-server" Jan 27 15:22:22 crc kubenswrapper[4772]: I0127 15:22:22.953038 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="13489498-3c32-4ef1-baf5-99f6907d07e4" containerName="registry-server" Jan 27 15:22:22 crc kubenswrapper[4772]: E0127 15:22:22.953047 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdc687da-0857-4070-a27a-90b08ca108c9" containerName="extract-utilities" Jan 27 15:22:22 crc kubenswrapper[4772]: I0127 15:22:22.953053 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdc687da-0857-4070-a27a-90b08ca108c9" containerName="extract-utilities" Jan 27 15:22:22 crc kubenswrapper[4772]: I0127 15:22:22.953197 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdc687da-0857-4070-a27a-90b08ca108c9" containerName="registry-server" Jan 27 15:22:22 crc kubenswrapper[4772]: I0127 15:22:22.953212 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="13489498-3c32-4ef1-baf5-99f6907d07e4" containerName="registry-server" Jan 27 15:22:22 crc kubenswrapper[4772]: I0127 15:22:22.953965 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4" Jan 27 15:22:22 crc kubenswrapper[4772]: I0127 15:22:22.957028 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-nwbtg" Jan 27 15:22:22 crc kubenswrapper[4772]: I0127 15:22:22.967771 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4"] Jan 27 15:22:23 crc kubenswrapper[4772]: I0127 15:22:23.062541 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff7tn\" (UniqueName: \"kubernetes.io/projected/0f29ea34-f593-4806-b5f6-2f9976c46a12-kube-api-access-ff7tn\") pod \"6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4\" (UID: \"0f29ea34-f593-4806-b5f6-2f9976c46a12\") " pod="openstack-operators/6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4" Jan 27 15:22:23 crc kubenswrapper[4772]: I0127 15:22:23.062581 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0f29ea34-f593-4806-b5f6-2f9976c46a12-util\") pod \"6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4\" (UID: \"0f29ea34-f593-4806-b5f6-2f9976c46a12\") " pod="openstack-operators/6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4" Jan 27 15:22:23 crc kubenswrapper[4772]: I0127 15:22:23.062611 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0f29ea34-f593-4806-b5f6-2f9976c46a12-bundle\") pod \"6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4\" (UID: \"0f29ea34-f593-4806-b5f6-2f9976c46a12\") " pod="openstack-operators/6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4" Jan 27 15:22:23 crc kubenswrapper[4772]: I0127 15:22:23.163683 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0f29ea34-f593-4806-b5f6-2f9976c46a12-util\") pod \"6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4\" (UID: \"0f29ea34-f593-4806-b5f6-2f9976c46a12\") " pod="openstack-operators/6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4" Jan 27 15:22:23 crc kubenswrapper[4772]: I0127 15:22:23.163759 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0f29ea34-f593-4806-b5f6-2f9976c46a12-bundle\") pod \"6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4\" (UID: \"0f29ea34-f593-4806-b5f6-2f9976c46a12\") " pod="openstack-operators/6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4" Jan 27 15:22:23 crc kubenswrapper[4772]: I0127 15:22:23.163877 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff7tn\" (UniqueName: \"kubernetes.io/projected/0f29ea34-f593-4806-b5f6-2f9976c46a12-kube-api-access-ff7tn\") pod \"6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4\" (UID: \"0f29ea34-f593-4806-b5f6-2f9976c46a12\") " pod="openstack-operators/6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4" Jan 27 15:22:23 crc kubenswrapper[4772]: I0127 15:22:23.164410 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0f29ea34-f593-4806-b5f6-2f9976c46a12-util\") pod \"6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4\" (UID: \"0f29ea34-f593-4806-b5f6-2f9976c46a12\") " pod="openstack-operators/6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4" Jan 27 15:22:23 crc kubenswrapper[4772]: I0127 15:22:23.164412 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0f29ea34-f593-4806-b5f6-2f9976c46a12-bundle\") pod \"6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4\" (UID: \"0f29ea34-f593-4806-b5f6-2f9976c46a12\") " pod="openstack-operators/6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4" Jan 27 15:22:23 crc kubenswrapper[4772]: I0127 15:22:23.183125 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff7tn\" (UniqueName: \"kubernetes.io/projected/0f29ea34-f593-4806-b5f6-2f9976c46a12-kube-api-access-ff7tn\") pod \"6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4\" (UID: \"0f29ea34-f593-4806-b5f6-2f9976c46a12\") " pod="openstack-operators/6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4" Jan 27 15:22:23 crc kubenswrapper[4772]: I0127 15:22:23.269229 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4" Jan 27 15:22:23 crc kubenswrapper[4772]: I0127 15:22:23.565707 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4"] Jan 27 15:22:24 crc kubenswrapper[4772]: I0127 15:22:24.447690 4772 generic.go:334] "Generic (PLEG): container finished" podID="0f29ea34-f593-4806-b5f6-2f9976c46a12" containerID="3eed0491dec9fe3e452816050dcb0d79415f19c85bbaf20cf8830f09120d7753" exitCode=0 Jan 27 15:22:24 crc kubenswrapper[4772]: I0127 15:22:24.447744 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4" event={"ID":"0f29ea34-f593-4806-b5f6-2f9976c46a12","Type":"ContainerDied","Data":"3eed0491dec9fe3e452816050dcb0d79415f19c85bbaf20cf8830f09120d7753"} Jan 27 15:22:24 crc kubenswrapper[4772]: I0127 15:22:24.447788 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4" event={"ID":"0f29ea34-f593-4806-b5f6-2f9976c46a12","Type":"ContainerStarted","Data":"c56568fce76e8df1f0305837d648767268ec5d470bc759ec6509cc21e8e2cbc4"} Jan 27 15:22:27 crc kubenswrapper[4772]: I0127 15:22:27.478279 4772 generic.go:334] "Generic (PLEG): container finished" podID="0f29ea34-f593-4806-b5f6-2f9976c46a12" containerID="79203eaf770d19a8deea1ab6fa237bbc2f2bf50b4932961ae73b8351e90c5716" exitCode=0 Jan 27 15:22:27 crc kubenswrapper[4772]: I0127 15:22:27.478394 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4" event={"ID":"0f29ea34-f593-4806-b5f6-2f9976c46a12","Type":"ContainerDied","Data":"79203eaf770d19a8deea1ab6fa237bbc2f2bf50b4932961ae73b8351e90c5716"} Jan 27 15:22:28 crc kubenswrapper[4772]: I0127 15:22:28.488303 4772 generic.go:334] "Generic (PLEG): container finished" podID="0f29ea34-f593-4806-b5f6-2f9976c46a12" containerID="e5cb8ba5058b870db0b2b7ce8c9625be9b47bfc608e7bdd04c686bcba481c1b6" exitCode=0 Jan 27 15:22:28 crc kubenswrapper[4772]: I0127 15:22:28.488375 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4" event={"ID":"0f29ea34-f593-4806-b5f6-2f9976c46a12","Type":"ContainerDied","Data":"e5cb8ba5058b870db0b2b7ce8c9625be9b47bfc608e7bdd04c686bcba481c1b6"} Jan 27 15:22:29 crc kubenswrapper[4772]: I0127 15:22:29.824462 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4" Jan 27 15:22:29 crc kubenswrapper[4772]: I0127 15:22:29.877098 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0f29ea34-f593-4806-b5f6-2f9976c46a12-bundle\") pod \"0f29ea34-f593-4806-b5f6-2f9976c46a12\" (UID: \"0f29ea34-f593-4806-b5f6-2f9976c46a12\") " Jan 27 15:22:29 crc kubenswrapper[4772]: I0127 15:22:29.877221 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0f29ea34-f593-4806-b5f6-2f9976c46a12-util\") pod \"0f29ea34-f593-4806-b5f6-2f9976c46a12\" (UID: \"0f29ea34-f593-4806-b5f6-2f9976c46a12\") " Jan 27 15:22:29 crc kubenswrapper[4772]: I0127 15:22:29.877248 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ff7tn\" (UniqueName: \"kubernetes.io/projected/0f29ea34-f593-4806-b5f6-2f9976c46a12-kube-api-access-ff7tn\") pod \"0f29ea34-f593-4806-b5f6-2f9976c46a12\" (UID: \"0f29ea34-f593-4806-b5f6-2f9976c46a12\") " Jan 27 15:22:29 crc kubenswrapper[4772]: I0127 15:22:29.877887 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f29ea34-f593-4806-b5f6-2f9976c46a12-bundle" (OuterVolumeSpecName: "bundle") pod "0f29ea34-f593-4806-b5f6-2f9976c46a12" (UID: "0f29ea34-f593-4806-b5f6-2f9976c46a12"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:22:29 crc kubenswrapper[4772]: I0127 15:22:29.883437 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f29ea34-f593-4806-b5f6-2f9976c46a12-kube-api-access-ff7tn" (OuterVolumeSpecName: "kube-api-access-ff7tn") pod "0f29ea34-f593-4806-b5f6-2f9976c46a12" (UID: "0f29ea34-f593-4806-b5f6-2f9976c46a12"). InnerVolumeSpecName "kube-api-access-ff7tn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:22:29 crc kubenswrapper[4772]: I0127 15:22:29.887699 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f29ea34-f593-4806-b5f6-2f9976c46a12-util" (OuterVolumeSpecName: "util") pod "0f29ea34-f593-4806-b5f6-2f9976c46a12" (UID: "0f29ea34-f593-4806-b5f6-2f9976c46a12"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:22:29 crc kubenswrapper[4772]: I0127 15:22:29.979329 4772 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0f29ea34-f593-4806-b5f6-2f9976c46a12-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:22:29 crc kubenswrapper[4772]: I0127 15:22:29.979365 4772 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0f29ea34-f593-4806-b5f6-2f9976c46a12-util\") on node \"crc\" DevicePath \"\"" Jan 27 15:22:29 crc kubenswrapper[4772]: I0127 15:22:29.979379 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ff7tn\" (UniqueName: \"kubernetes.io/projected/0f29ea34-f593-4806-b5f6-2f9976c46a12-kube-api-access-ff7tn\") on node \"crc\" DevicePath \"\"" Jan 27 15:22:30 crc kubenswrapper[4772]: I0127 15:22:30.508399 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4" event={"ID":"0f29ea34-f593-4806-b5f6-2f9976c46a12","Type":"ContainerDied","Data":"c56568fce76e8df1f0305837d648767268ec5d470bc759ec6509cc21e8e2cbc4"} Jan 27 15:22:30 crc kubenswrapper[4772]: I0127 15:22:30.508470 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c56568fce76e8df1f0305837d648767268ec5d470bc759ec6509cc21e8e2cbc4" Jan 27 15:22:30 crc kubenswrapper[4772]: I0127 15:22:30.508741 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4" Jan 27 15:22:35 crc kubenswrapper[4772]: I0127 15:22:35.294300 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-6fb647f7d4-gkjgt"] Jan 27 15:22:35 crc kubenswrapper[4772]: E0127 15:22:35.295136 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f29ea34-f593-4806-b5f6-2f9976c46a12" containerName="util" Jan 27 15:22:35 crc kubenswrapper[4772]: I0127 15:22:35.295151 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f29ea34-f593-4806-b5f6-2f9976c46a12" containerName="util" Jan 27 15:22:35 crc kubenswrapper[4772]: E0127 15:22:35.295182 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f29ea34-f593-4806-b5f6-2f9976c46a12" containerName="pull" Jan 27 15:22:35 crc kubenswrapper[4772]: I0127 15:22:35.295190 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f29ea34-f593-4806-b5f6-2f9976c46a12" containerName="pull" Jan 27 15:22:35 crc kubenswrapper[4772]: E0127 15:22:35.295204 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f29ea34-f593-4806-b5f6-2f9976c46a12" containerName="extract" Jan 27 15:22:35 crc kubenswrapper[4772]: I0127 15:22:35.295213 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f29ea34-f593-4806-b5f6-2f9976c46a12" containerName="extract" Jan 27 15:22:35 crc kubenswrapper[4772]: I0127 15:22:35.295356 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f29ea34-f593-4806-b5f6-2f9976c46a12" containerName="extract" Jan 27 15:22:35 crc kubenswrapper[4772]: I0127 15:22:35.295860 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-6fb647f7d4-gkjgt" Jan 27 15:22:35 crc kubenswrapper[4772]: I0127 15:22:35.298544 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-rhmmw" Jan 27 15:22:35 crc kubenswrapper[4772]: I0127 15:22:35.326112 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-6fb647f7d4-gkjgt"] Jan 27 15:22:35 crc kubenswrapper[4772]: I0127 15:22:35.361512 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2z4nv\" (UniqueName: \"kubernetes.io/projected/939a692e-65d1-4be8-b78a-22ae83072d51-kube-api-access-2z4nv\") pod \"openstack-operator-controller-init-6fb647f7d4-gkjgt\" (UID: \"939a692e-65d1-4be8-b78a-22ae83072d51\") " pod="openstack-operators/openstack-operator-controller-init-6fb647f7d4-gkjgt" Jan 27 15:22:35 crc kubenswrapper[4772]: I0127 15:22:35.462553 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2z4nv\" (UniqueName: \"kubernetes.io/projected/939a692e-65d1-4be8-b78a-22ae83072d51-kube-api-access-2z4nv\") pod \"openstack-operator-controller-init-6fb647f7d4-gkjgt\" (UID: \"939a692e-65d1-4be8-b78a-22ae83072d51\") " pod="openstack-operators/openstack-operator-controller-init-6fb647f7d4-gkjgt" Jan 27 15:22:35 crc kubenswrapper[4772]: I0127 15:22:35.485409 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2z4nv\" (UniqueName: \"kubernetes.io/projected/939a692e-65d1-4be8-b78a-22ae83072d51-kube-api-access-2z4nv\") pod \"openstack-operator-controller-init-6fb647f7d4-gkjgt\" (UID: \"939a692e-65d1-4be8-b78a-22ae83072d51\") " pod="openstack-operators/openstack-operator-controller-init-6fb647f7d4-gkjgt" Jan 27 15:22:35 crc kubenswrapper[4772]: I0127 15:22:35.614593 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-6fb647f7d4-gkjgt" Jan 27 15:22:36 crc kubenswrapper[4772]: I0127 15:22:36.152912 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-6fb647f7d4-gkjgt"] Jan 27 15:22:36 crc kubenswrapper[4772]: I0127 15:22:36.556809 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-6fb647f7d4-gkjgt" event={"ID":"939a692e-65d1-4be8-b78a-22ae83072d51","Type":"ContainerStarted","Data":"285430d1d2da8bf5c3321cad317f7ec39aefceedaed1c75e38d18002294544e1"} Jan 27 15:22:40 crc kubenswrapper[4772]: I0127 15:22:40.581769 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-6fb647f7d4-gkjgt" event={"ID":"939a692e-65d1-4be8-b78a-22ae83072d51","Type":"ContainerStarted","Data":"0bb32ca286b4bb63a4601ea17d681fa5e57fc6fd9be09d28d3803335c16b8c89"} Jan 27 15:22:40 crc kubenswrapper[4772]: I0127 15:22:40.582285 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-6fb647f7d4-gkjgt" Jan 27 15:22:40 crc kubenswrapper[4772]: I0127 15:22:40.606238 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-6fb647f7d4-gkjgt" podStartSLOduration=2.124067436 podStartE2EDuration="5.606223645s" podCreationTimestamp="2026-01-27 15:22:35 +0000 UTC" firstStartedPulling="2026-01-27 15:22:36.158154182 +0000 UTC m=+942.138763280" lastFinishedPulling="2026-01-27 15:22:39.640310391 +0000 UTC m=+945.620919489" observedRunningTime="2026-01-27 15:22:40.602734035 +0000 UTC m=+946.583343163" watchObservedRunningTime="2026-01-27 15:22:40.606223645 +0000 UTC m=+946.586832753" Jan 27 15:22:42 crc kubenswrapper[4772]: I0127 15:22:42.059179 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:22:42 crc kubenswrapper[4772]: I0127 15:22:42.059550 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:22:45 crc kubenswrapper[4772]: I0127 15:22:45.617996 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-6fb647f7d4-gkjgt" Jan 27 15:22:54 crc kubenswrapper[4772]: I0127 15:22:54.016562 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wbmd9"] Jan 27 15:22:54 crc kubenswrapper[4772]: I0127 15:22:54.018189 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wbmd9" Jan 27 15:22:54 crc kubenswrapper[4772]: I0127 15:22:54.030954 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e682da4-be9e-4318-8f65-cd879f9a826a-catalog-content\") pod \"certified-operators-wbmd9\" (UID: \"8e682da4-be9e-4318-8f65-cd879f9a826a\") " pod="openshift-marketplace/certified-operators-wbmd9" Jan 27 15:22:54 crc kubenswrapper[4772]: I0127 15:22:54.031037 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e682da4-be9e-4318-8f65-cd879f9a826a-utilities\") pod \"certified-operators-wbmd9\" (UID: \"8e682da4-be9e-4318-8f65-cd879f9a826a\") " pod="openshift-marketplace/certified-operators-wbmd9" Jan 27 15:22:54 crc kubenswrapper[4772]: I0127 15:22:54.031089 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv5fc\" (UniqueName: \"kubernetes.io/projected/8e682da4-be9e-4318-8f65-cd879f9a826a-kube-api-access-tv5fc\") pod \"certified-operators-wbmd9\" (UID: \"8e682da4-be9e-4318-8f65-cd879f9a826a\") " pod="openshift-marketplace/certified-operators-wbmd9" Jan 27 15:22:54 crc kubenswrapper[4772]: I0127 15:22:54.036794 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wbmd9"] Jan 27 15:22:54 crc kubenswrapper[4772]: I0127 15:22:54.132722 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tv5fc\" (UniqueName: \"kubernetes.io/projected/8e682da4-be9e-4318-8f65-cd879f9a826a-kube-api-access-tv5fc\") pod \"certified-operators-wbmd9\" (UID: \"8e682da4-be9e-4318-8f65-cd879f9a826a\") " pod="openshift-marketplace/certified-operators-wbmd9" Jan 27 15:22:54 crc kubenswrapper[4772]: I0127 15:22:54.132786 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e682da4-be9e-4318-8f65-cd879f9a826a-catalog-content\") pod \"certified-operators-wbmd9\" (UID: \"8e682da4-be9e-4318-8f65-cd879f9a826a\") " pod="openshift-marketplace/certified-operators-wbmd9" Jan 27 15:22:54 crc kubenswrapper[4772]: I0127 15:22:54.132842 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e682da4-be9e-4318-8f65-cd879f9a826a-utilities\") pod \"certified-operators-wbmd9\" (UID: \"8e682da4-be9e-4318-8f65-cd879f9a826a\") " pod="openshift-marketplace/certified-operators-wbmd9" Jan 27 15:22:54 crc kubenswrapper[4772]: I0127 15:22:54.133317 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e682da4-be9e-4318-8f65-cd879f9a826a-utilities\") pod \"certified-operators-wbmd9\" (UID: \"8e682da4-be9e-4318-8f65-cd879f9a826a\") " pod="openshift-marketplace/certified-operators-wbmd9" Jan 27 15:22:54 crc kubenswrapper[4772]: I0127 15:22:54.133407 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e682da4-be9e-4318-8f65-cd879f9a826a-catalog-content\") pod \"certified-operators-wbmd9\" (UID: \"8e682da4-be9e-4318-8f65-cd879f9a826a\") " pod="openshift-marketplace/certified-operators-wbmd9" Jan 27 15:22:54 crc kubenswrapper[4772]: I0127 15:22:54.163204 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv5fc\" (UniqueName: \"kubernetes.io/projected/8e682da4-be9e-4318-8f65-cd879f9a826a-kube-api-access-tv5fc\") pod \"certified-operators-wbmd9\" (UID: \"8e682da4-be9e-4318-8f65-cd879f9a826a\") " pod="openshift-marketplace/certified-operators-wbmd9" Jan 27 15:22:54 crc kubenswrapper[4772]: I0127 15:22:54.381511 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wbmd9" Jan 27 15:22:54 crc kubenswrapper[4772]: I0127 15:22:54.815177 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wbmd9"] Jan 27 15:22:54 crc kubenswrapper[4772]: I0127 15:22:54.872084 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbmd9" event={"ID":"8e682da4-be9e-4318-8f65-cd879f9a826a","Type":"ContainerStarted","Data":"ae5e43e320906455ee157d2ed44f17f1214d92e37e6149b34c7998975892c8dd"} Jan 27 15:22:56 crc kubenswrapper[4772]: I0127 15:22:56.893778 4772 generic.go:334] "Generic (PLEG): container finished" podID="8e682da4-be9e-4318-8f65-cd879f9a826a" containerID="2f8832235d7bf2dd65342391017e414aa8e6b691bf56ffa84bb5000edb279643" exitCode=0 Jan 27 15:22:56 crc kubenswrapper[4772]: I0127 15:22:56.893868 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbmd9" event={"ID":"8e682da4-be9e-4318-8f65-cd879f9a826a","Type":"ContainerDied","Data":"2f8832235d7bf2dd65342391017e414aa8e6b691bf56ffa84bb5000edb279643"} Jan 27 15:22:59 crc kubenswrapper[4772]: I0127 15:22:59.920534 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbmd9" event={"ID":"8e682da4-be9e-4318-8f65-cd879f9a826a","Type":"ContainerStarted","Data":"bee09a387c7651d30762718fdf9481148ff7173f222fd6351786d33b9f772e6b"} Jan 27 15:23:00 crc kubenswrapper[4772]: I0127 15:23:00.929004 4772 generic.go:334] "Generic (PLEG): container finished" podID="8e682da4-be9e-4318-8f65-cd879f9a826a" containerID="bee09a387c7651d30762718fdf9481148ff7173f222fd6351786d33b9f772e6b" exitCode=0 Jan 27 15:23:00 crc kubenswrapper[4772]: I0127 15:23:00.929063 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbmd9" event={"ID":"8e682da4-be9e-4318-8f65-cd879f9a826a","Type":"ContainerDied","Data":"bee09a387c7651d30762718fdf9481148ff7173f222fd6351786d33b9f772e6b"} Jan 27 15:23:01 crc kubenswrapper[4772]: I0127 15:23:01.939253 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbmd9" event={"ID":"8e682da4-be9e-4318-8f65-cd879f9a826a","Type":"ContainerStarted","Data":"370eac75d4ce32b16b99194d5e4a9c1cd98d3301cc51e5eca10f39a5246044cb"} Jan 27 15:23:01 crc kubenswrapper[4772]: I0127 15:23:01.962406 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wbmd9" podStartSLOduration=5.380610437 podStartE2EDuration="8.96238828s" podCreationTimestamp="2026-01-27 15:22:53 +0000 UTC" firstStartedPulling="2026-01-27 15:22:57.903635592 +0000 UTC m=+963.884244690" lastFinishedPulling="2026-01-27 15:23:01.485413435 +0000 UTC m=+967.466022533" observedRunningTime="2026-01-27 15:23:01.956535432 +0000 UTC m=+967.937144530" watchObservedRunningTime="2026-01-27 15:23:01.96238828 +0000 UTC m=+967.942997378" Jan 27 15:23:04 crc kubenswrapper[4772]: I0127 15:23:04.382696 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wbmd9" Jan 27 15:23:04 crc kubenswrapper[4772]: I0127 15:23:04.383475 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wbmd9" Jan 27 15:23:04 crc kubenswrapper[4772]: I0127 15:23:04.436829 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wbmd9" Jan 27 15:23:12 crc kubenswrapper[4772]: I0127 15:23:12.059031 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:23:12 crc kubenswrapper[4772]: I0127 15:23:12.059799 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:23:12 crc kubenswrapper[4772]: I0127 15:23:12.059850 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:23:12 crc kubenswrapper[4772]: I0127 15:23:12.060566 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c8213e4fa74445d3800c2dbcb45efc3fb34a6f40c3d5ed5845b811a51d3d8497"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 15:23:12 crc kubenswrapper[4772]: I0127 15:23:12.060636 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://c8213e4fa74445d3800c2dbcb45efc3fb34a6f40c3d5ed5845b811a51d3d8497" gracePeriod=600 Jan 27 15:23:13 crc kubenswrapper[4772]: I0127 15:23:13.010444 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="c8213e4fa74445d3800c2dbcb45efc3fb34a6f40c3d5ed5845b811a51d3d8497" exitCode=0 Jan 27 15:23:13 crc kubenswrapper[4772]: I0127 15:23:13.010492 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"c8213e4fa74445d3800c2dbcb45efc3fb34a6f40c3d5ed5845b811a51d3d8497"} Jan 27 15:23:13 crc kubenswrapper[4772]: I0127 15:23:13.010800 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"ed9bc8d4920540552bc96f7af996996e69c893224418d74c897e7298ed107163"} Jan 27 15:23:13 crc kubenswrapper[4772]: I0127 15:23:13.010823 4772 scope.go:117] "RemoveContainer" containerID="60c798dfb542a875b90e857bf6f54352abce005f4bc0c5fd246c1b5d0903e3f3" Jan 27 15:23:14 crc kubenswrapper[4772]: I0127 15:23:14.443666 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wbmd9" Jan 27 15:23:14 crc kubenswrapper[4772]: I0127 15:23:14.502573 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wbmd9"] Jan 27 15:23:15 crc kubenswrapper[4772]: I0127 15:23:15.026051 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wbmd9" podUID="8e682da4-be9e-4318-8f65-cd879f9a826a" containerName="registry-server" containerID="cri-o://370eac75d4ce32b16b99194d5e4a9c1cd98d3301cc51e5eca10f39a5246044cb" gracePeriod=2 Jan 27 15:23:15 crc kubenswrapper[4772]: I0127 15:23:15.375030 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wbmd9" Jan 27 15:23:15 crc kubenswrapper[4772]: I0127 15:23:15.550741 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e682da4-be9e-4318-8f65-cd879f9a826a-catalog-content\") pod \"8e682da4-be9e-4318-8f65-cd879f9a826a\" (UID: \"8e682da4-be9e-4318-8f65-cd879f9a826a\") " Jan 27 15:23:15 crc kubenswrapper[4772]: I0127 15:23:15.550855 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e682da4-be9e-4318-8f65-cd879f9a826a-utilities\") pod \"8e682da4-be9e-4318-8f65-cd879f9a826a\" (UID: \"8e682da4-be9e-4318-8f65-cd879f9a826a\") " Jan 27 15:23:15 crc kubenswrapper[4772]: I0127 15:23:15.550897 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tv5fc\" (UniqueName: \"kubernetes.io/projected/8e682da4-be9e-4318-8f65-cd879f9a826a-kube-api-access-tv5fc\") pod \"8e682da4-be9e-4318-8f65-cd879f9a826a\" (UID: \"8e682da4-be9e-4318-8f65-cd879f9a826a\") " Jan 27 15:23:15 crc kubenswrapper[4772]: I0127 15:23:15.551808 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e682da4-be9e-4318-8f65-cd879f9a826a-utilities" (OuterVolumeSpecName: "utilities") pod "8e682da4-be9e-4318-8f65-cd879f9a826a" (UID: "8e682da4-be9e-4318-8f65-cd879f9a826a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:23:15 crc kubenswrapper[4772]: I0127 15:23:15.556459 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e682da4-be9e-4318-8f65-cd879f9a826a-kube-api-access-tv5fc" (OuterVolumeSpecName: "kube-api-access-tv5fc") pod "8e682da4-be9e-4318-8f65-cd879f9a826a" (UID: "8e682da4-be9e-4318-8f65-cd879f9a826a"). InnerVolumeSpecName "kube-api-access-tv5fc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:23:15 crc kubenswrapper[4772]: I0127 15:23:15.602858 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e682da4-be9e-4318-8f65-cd879f9a826a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e682da4-be9e-4318-8f65-cd879f9a826a" (UID: "8e682da4-be9e-4318-8f65-cd879f9a826a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:23:15 crc kubenswrapper[4772]: I0127 15:23:15.652468 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e682da4-be9e-4318-8f65-cd879f9a826a-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 15:23:15 crc kubenswrapper[4772]: I0127 15:23:15.652504 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tv5fc\" (UniqueName: \"kubernetes.io/projected/8e682da4-be9e-4318-8f65-cd879f9a826a-kube-api-access-tv5fc\") on node \"crc\" DevicePath \"\"" Jan 27 15:23:15 crc kubenswrapper[4772]: I0127 15:23:15.652515 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e682da4-be9e-4318-8f65-cd879f9a826a-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 15:23:16 crc kubenswrapper[4772]: I0127 15:23:16.043437 4772 generic.go:334] "Generic (PLEG): container finished" podID="8e682da4-be9e-4318-8f65-cd879f9a826a" containerID="370eac75d4ce32b16b99194d5e4a9c1cd98d3301cc51e5eca10f39a5246044cb" exitCode=0 Jan 27 15:23:16 crc kubenswrapper[4772]: I0127 15:23:16.043489 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wbmd9" Jan 27 15:23:16 crc kubenswrapper[4772]: I0127 15:23:16.043487 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbmd9" event={"ID":"8e682da4-be9e-4318-8f65-cd879f9a826a","Type":"ContainerDied","Data":"370eac75d4ce32b16b99194d5e4a9c1cd98d3301cc51e5eca10f39a5246044cb"} Jan 27 15:23:16 crc kubenswrapper[4772]: I0127 15:23:16.043598 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wbmd9" event={"ID":"8e682da4-be9e-4318-8f65-cd879f9a826a","Type":"ContainerDied","Data":"ae5e43e320906455ee157d2ed44f17f1214d92e37e6149b34c7998975892c8dd"} Jan 27 15:23:16 crc kubenswrapper[4772]: I0127 15:23:16.043620 4772 scope.go:117] "RemoveContainer" containerID="370eac75d4ce32b16b99194d5e4a9c1cd98d3301cc51e5eca10f39a5246044cb" Jan 27 15:23:16 crc kubenswrapper[4772]: I0127 15:23:16.058815 4772 scope.go:117] "RemoveContainer" containerID="bee09a387c7651d30762718fdf9481148ff7173f222fd6351786d33b9f772e6b" Jan 27 15:23:16 crc kubenswrapper[4772]: I0127 15:23:16.078571 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wbmd9"] Jan 27 15:23:16 crc kubenswrapper[4772]: I0127 15:23:16.082987 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wbmd9"] Jan 27 15:23:16 crc kubenswrapper[4772]: I0127 15:23:16.104503 4772 scope.go:117] "RemoveContainer" containerID="2f8832235d7bf2dd65342391017e414aa8e6b691bf56ffa84bb5000edb279643" Jan 27 15:23:16 crc kubenswrapper[4772]: I0127 15:23:16.125907 4772 scope.go:117] "RemoveContainer" containerID="370eac75d4ce32b16b99194d5e4a9c1cd98d3301cc51e5eca10f39a5246044cb" Jan 27 15:23:16 crc kubenswrapper[4772]: E0127 15:23:16.126345 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"370eac75d4ce32b16b99194d5e4a9c1cd98d3301cc51e5eca10f39a5246044cb\": container with ID starting with 370eac75d4ce32b16b99194d5e4a9c1cd98d3301cc51e5eca10f39a5246044cb not found: ID does not exist" containerID="370eac75d4ce32b16b99194d5e4a9c1cd98d3301cc51e5eca10f39a5246044cb" Jan 27 15:23:16 crc kubenswrapper[4772]: I0127 15:23:16.126387 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"370eac75d4ce32b16b99194d5e4a9c1cd98d3301cc51e5eca10f39a5246044cb"} err="failed to get container status \"370eac75d4ce32b16b99194d5e4a9c1cd98d3301cc51e5eca10f39a5246044cb\": rpc error: code = NotFound desc = could not find container \"370eac75d4ce32b16b99194d5e4a9c1cd98d3301cc51e5eca10f39a5246044cb\": container with ID starting with 370eac75d4ce32b16b99194d5e4a9c1cd98d3301cc51e5eca10f39a5246044cb not found: ID does not exist" Jan 27 15:23:16 crc kubenswrapper[4772]: I0127 15:23:16.126415 4772 scope.go:117] "RemoveContainer" containerID="bee09a387c7651d30762718fdf9481148ff7173f222fd6351786d33b9f772e6b" Jan 27 15:23:16 crc kubenswrapper[4772]: E0127 15:23:16.126795 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bee09a387c7651d30762718fdf9481148ff7173f222fd6351786d33b9f772e6b\": container with ID starting with bee09a387c7651d30762718fdf9481148ff7173f222fd6351786d33b9f772e6b not found: ID does not exist" containerID="bee09a387c7651d30762718fdf9481148ff7173f222fd6351786d33b9f772e6b" Jan 27 15:23:16 crc kubenswrapper[4772]: I0127 15:23:16.126815 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bee09a387c7651d30762718fdf9481148ff7173f222fd6351786d33b9f772e6b"} err="failed to get container status \"bee09a387c7651d30762718fdf9481148ff7173f222fd6351786d33b9f772e6b\": rpc error: code = NotFound desc = could not find container \"bee09a387c7651d30762718fdf9481148ff7173f222fd6351786d33b9f772e6b\": container with ID starting with bee09a387c7651d30762718fdf9481148ff7173f222fd6351786d33b9f772e6b not found: ID does not exist" Jan 27 15:23:16 crc kubenswrapper[4772]: I0127 15:23:16.126827 4772 scope.go:117] "RemoveContainer" containerID="2f8832235d7bf2dd65342391017e414aa8e6b691bf56ffa84bb5000edb279643" Jan 27 15:23:16 crc kubenswrapper[4772]: E0127 15:23:16.128509 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f8832235d7bf2dd65342391017e414aa8e6b691bf56ffa84bb5000edb279643\": container with ID starting with 2f8832235d7bf2dd65342391017e414aa8e6b691bf56ffa84bb5000edb279643 not found: ID does not exist" containerID="2f8832235d7bf2dd65342391017e414aa8e6b691bf56ffa84bb5000edb279643" Jan 27 15:23:16 crc kubenswrapper[4772]: I0127 15:23:16.128546 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f8832235d7bf2dd65342391017e414aa8e6b691bf56ffa84bb5000edb279643"} err="failed to get container status \"2f8832235d7bf2dd65342391017e414aa8e6b691bf56ffa84bb5000edb279643\": rpc error: code = NotFound desc = could not find container \"2f8832235d7bf2dd65342391017e414aa8e6b691bf56ffa84bb5000edb279643\": container with ID starting with 2f8832235d7bf2dd65342391017e414aa8e6b691bf56ffa84bb5000edb279643 not found: ID does not exist" Jan 27 15:23:16 crc kubenswrapper[4772]: I0127 15:23:16.670229 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e682da4-be9e-4318-8f65-cd879f9a826a" path="/var/lib/kubelet/pods/8e682da4-be9e-4318-8f65-cd879f9a826a/volumes" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.710848 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-65ff799cfd-t42n9"] Jan 27 15:23:21 crc kubenswrapper[4772]: E0127 15:23:21.711752 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e682da4-be9e-4318-8f65-cd879f9a826a" containerName="extract-content" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.711774 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e682da4-be9e-4318-8f65-cd879f9a826a" containerName="extract-content" Jan 27 15:23:21 crc kubenswrapper[4772]: E0127 15:23:21.711812 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e682da4-be9e-4318-8f65-cd879f9a826a" containerName="extract-utilities" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.711823 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e682da4-be9e-4318-8f65-cd879f9a826a" containerName="extract-utilities" Jan 27 15:23:21 crc kubenswrapper[4772]: E0127 15:23:21.711848 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e682da4-be9e-4318-8f65-cd879f9a826a" containerName="registry-server" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.711856 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e682da4-be9e-4318-8f65-cd879f9a826a" containerName="registry-server" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.712005 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e682da4-be9e-4318-8f65-cd879f9a826a" containerName="registry-server" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.712580 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-65ff799cfd-t42n9" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.716770 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-lmxhh" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.718093 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-655bf9cfbb-cgh7j"] Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.719125 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-655bf9cfbb-cgh7j" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.721441 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-qm6dd" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.725794 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-65ff799cfd-t42n9"] Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.734294 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-77554cdc5c-tkr6j"] Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.735004 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-77554cdc5c-tkr6j" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.737084 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-7pnmk" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.746714 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-67dd55ff59-hgscb"] Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.747601 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-67dd55ff59-hgscb" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.760419 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-xtchh" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.767351 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-655bf9cfbb-cgh7j"] Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.767396 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-77554cdc5c-tkr6j"] Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.777679 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-67dd55ff59-hgscb"] Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.843207 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-575ffb885b-mtd9d"] Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.844133 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-575ffb885b-mtd9d" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.849411 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-ffgpg" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.872136 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-575ffb885b-mtd9d"] Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.884081 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xv2c\" (UniqueName: \"kubernetes.io/projected/d395f105-54f0-4497-a119-57802be313a3-kube-api-access-4xv2c\") pod \"designate-operator-controller-manager-77554cdc5c-tkr6j\" (UID: \"d395f105-54f0-4497-a119-57802be313a3\") " pod="openstack-operators/designate-operator-controller-manager-77554cdc5c-tkr6j" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.884124 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b7hp\" (UniqueName: \"kubernetes.io/projected/4c63a702-50b9-42f3-858e-7e27da0a8d8f-kube-api-access-9b7hp\") pod \"heat-operator-controller-manager-575ffb885b-mtd9d\" (UID: \"4c63a702-50b9-42f3-858e-7e27da0a8d8f\") " pod="openstack-operators/heat-operator-controller-manager-575ffb885b-mtd9d" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.884145 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snjkj\" (UniqueName: \"kubernetes.io/projected/fb300814-fca7-4419-ac6e-c08b33edd4be-kube-api-access-snjkj\") pod \"glance-operator-controller-manager-67dd55ff59-hgscb\" (UID: \"fb300814-fca7-4419-ac6e-c08b33edd4be\") " pod="openstack-operators/glance-operator-controller-manager-67dd55ff59-hgscb" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.884203 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brwq5\" (UniqueName: \"kubernetes.io/projected/674f4da6-f50d-4bab-808d-56ab3b9e2cb4-kube-api-access-brwq5\") pod \"barbican-operator-controller-manager-65ff799cfd-t42n9\" (UID: \"674f4da6-f50d-4bab-808d-56ab3b9e2cb4\") " pod="openstack-operators/barbican-operator-controller-manager-65ff799cfd-t42n9" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.884229 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4fn8\" (UniqueName: \"kubernetes.io/projected/fde95124-892b-411a-ba05-fa70927c8838-kube-api-access-z4fn8\") pod \"cinder-operator-controller-manager-655bf9cfbb-cgh7j\" (UID: \"fde95124-892b-411a-ba05-fa70927c8838\") " pod="openstack-operators/cinder-operator-controller-manager-655bf9cfbb-cgh7j" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.897232 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-77d5c5b54f-jcb4p"] Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.898025 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-jcb4p" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.908833 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-97xd2" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.943302 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-77d5c5b54f-jcb4p"] Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.948240 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-768b776ffb-sxbjn"] Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.949047 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-768b776ffb-sxbjn" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.954526 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-5ldvt" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.958226 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-7d75bc88d5-t54fr"] Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.959032 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7d75bc88d5-t54fr" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.960806 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.961288 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-kdbs8" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.984913 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4fn8\" (UniqueName: \"kubernetes.io/projected/fde95124-892b-411a-ba05-fa70927c8838-kube-api-access-z4fn8\") pod \"cinder-operator-controller-manager-655bf9cfbb-cgh7j\" (UID: \"fde95124-892b-411a-ba05-fa70927c8838\") " pod="openstack-operators/cinder-operator-controller-manager-655bf9cfbb-cgh7j" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.985008 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xv2c\" (UniqueName: \"kubernetes.io/projected/d395f105-54f0-4497-a119-57802be313a3-kube-api-access-4xv2c\") pod \"designate-operator-controller-manager-77554cdc5c-tkr6j\" (UID: \"d395f105-54f0-4497-a119-57802be313a3\") " pod="openstack-operators/designate-operator-controller-manager-77554cdc5c-tkr6j" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.985042 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9b7hp\" (UniqueName: \"kubernetes.io/projected/4c63a702-50b9-42f3-858e-7e27da0a8d8f-kube-api-access-9b7hp\") pod \"heat-operator-controller-manager-575ffb885b-mtd9d\" (UID: \"4c63a702-50b9-42f3-858e-7e27da0a8d8f\") " pod="openstack-operators/heat-operator-controller-manager-575ffb885b-mtd9d" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.985068 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snjkj\" (UniqueName: \"kubernetes.io/projected/fb300814-fca7-4419-ac6e-c08b33edd4be-kube-api-access-snjkj\") pod \"glance-operator-controller-manager-67dd55ff59-hgscb\" (UID: \"fb300814-fca7-4419-ac6e-c08b33edd4be\") " pod="openstack-operators/glance-operator-controller-manager-67dd55ff59-hgscb" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.985125 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brwq5\" (UniqueName: \"kubernetes.io/projected/674f4da6-f50d-4bab-808d-56ab3b9e2cb4-kube-api-access-brwq5\") pod \"barbican-operator-controller-manager-65ff799cfd-t42n9\" (UID: \"674f4da6-f50d-4bab-808d-56ab3b9e2cb4\") " pod="openstack-operators/barbican-operator-controller-manager-65ff799cfd-t42n9" Jan 27 15:23:21 crc kubenswrapper[4772]: I0127 15:23:21.986600 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7d75bc88d5-t54fr"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.003236 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-768b776ffb-sxbjn"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.016061 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4fn8\" (UniqueName: \"kubernetes.io/projected/fde95124-892b-411a-ba05-fa70927c8838-kube-api-access-z4fn8\") pod \"cinder-operator-controller-manager-655bf9cfbb-cgh7j\" (UID: \"fde95124-892b-411a-ba05-fa70927c8838\") " pod="openstack-operators/cinder-operator-controller-manager-655bf9cfbb-cgh7j" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.024401 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snjkj\" (UniqueName: \"kubernetes.io/projected/fb300814-fca7-4419-ac6e-c08b33edd4be-kube-api-access-snjkj\") pod \"glance-operator-controller-manager-67dd55ff59-hgscb\" (UID: \"fb300814-fca7-4419-ac6e-c08b33edd4be\") " pod="openstack-operators/glance-operator-controller-manager-67dd55ff59-hgscb" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.029461 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xv2c\" (UniqueName: \"kubernetes.io/projected/d395f105-54f0-4497-a119-57802be313a3-kube-api-access-4xv2c\") pod \"designate-operator-controller-manager-77554cdc5c-tkr6j\" (UID: \"d395f105-54f0-4497-a119-57802be313a3\") " pod="openstack-operators/designate-operator-controller-manager-77554cdc5c-tkr6j" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.039777 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9b7hp\" (UniqueName: \"kubernetes.io/projected/4c63a702-50b9-42f3-858e-7e27da0a8d8f-kube-api-access-9b7hp\") pod \"heat-operator-controller-manager-575ffb885b-mtd9d\" (UID: \"4c63a702-50b9-42f3-858e-7e27da0a8d8f\") " pod="openstack-operators/heat-operator-controller-manager-575ffb885b-mtd9d" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.046868 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brwq5\" (UniqueName: \"kubernetes.io/projected/674f4da6-f50d-4bab-808d-56ab3b9e2cb4-kube-api-access-brwq5\") pod \"barbican-operator-controller-manager-65ff799cfd-t42n9\" (UID: \"674f4da6-f50d-4bab-808d-56ab3b9e2cb4\") " pod="openstack-operators/barbican-operator-controller-manager-65ff799cfd-t42n9" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.075231 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55f684fd56-wzjrz"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.076026 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55f684fd56-wzjrz" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.079463 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-4mmkm" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.085120 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-849fcfbb6b-tvrx9"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.085911 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-849fcfbb6b-tvrx9" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.088187 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-65ff799cfd-t42n9" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.088526 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2qbl\" (UniqueName: \"kubernetes.io/projected/27ec5082-c170-465b-b3a3-1f27a545fd71-kube-api-access-x2qbl\") pod \"manila-operator-controller-manager-849fcfbb6b-tvrx9\" (UID: \"27ec5082-c170-465b-b3a3-1f27a545fd71\") " pod="openstack-operators/manila-operator-controller-manager-849fcfbb6b-tvrx9" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.088596 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnnt7\" (UniqueName: \"kubernetes.io/projected/783d8159-e67a-4796-83d8-4eff27d79505-kube-api-access-dnnt7\") pod \"keystone-operator-controller-manager-55f684fd56-wzjrz\" (UID: \"783d8159-e67a-4796-83d8-4eff27d79505\") " pod="openstack-operators/keystone-operator-controller-manager-55f684fd56-wzjrz" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.088626 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsmq4\" (UniqueName: \"kubernetes.io/projected/e85aef3a-e235-473c-94cc-1f6237798b3e-kube-api-access-fsmq4\") pod \"horizon-operator-controller-manager-77d5c5b54f-jcb4p\" (UID: \"e85aef3a-e235-473c-94cc-1f6237798b3e\") " pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-jcb4p" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.088643 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z74pn\" (UniqueName: \"kubernetes.io/projected/e7465bd0-3b6e-4199-9ee6-28b512198847-kube-api-access-z74pn\") pod \"infra-operator-controller-manager-7d75bc88d5-t54fr\" (UID: \"e7465bd0-3b6e-4199-9ee6-28b512198847\") " pod="openstack-operators/infra-operator-controller-manager-7d75bc88d5-t54fr" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.088662 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e7465bd0-3b6e-4199-9ee6-28b512198847-cert\") pod \"infra-operator-controller-manager-7d75bc88d5-t54fr\" (UID: \"e7465bd0-3b6e-4199-9ee6-28b512198847\") " pod="openstack-operators/infra-operator-controller-manager-7d75bc88d5-t54fr" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.088689 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs77k\" (UniqueName: \"kubernetes.io/projected/2df85221-33ed-49be-949c-516810279e4d-kube-api-access-zs77k\") pod \"ironic-operator-controller-manager-768b776ffb-sxbjn\" (UID: \"2df85221-33ed-49be-949c-516810279e4d\") " pod="openstack-operators/ironic-operator-controller-manager-768b776ffb-sxbjn" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.101310 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-849fcfbb6b-tvrx9"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.101423 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-hqfvl" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.116954 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-655bf9cfbb-cgh7j" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.126346 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55f684fd56-wzjrz"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.131734 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-77554cdc5c-tkr6j" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.151011 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-67dd55ff59-hgscb" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.186769 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-575ffb885b-mtd9d" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.191602 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7ffd8d76d4-gcpj4"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.212574 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7ffd8d76d4-gcpj4" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.221091 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2qbl\" (UniqueName: \"kubernetes.io/projected/27ec5082-c170-465b-b3a3-1f27a545fd71-kube-api-access-x2qbl\") pod \"manila-operator-controller-manager-849fcfbb6b-tvrx9\" (UID: \"27ec5082-c170-465b-b3a3-1f27a545fd71\") " pod="openstack-operators/manila-operator-controller-manager-849fcfbb6b-tvrx9" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.224057 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnnt7\" (UniqueName: \"kubernetes.io/projected/783d8159-e67a-4796-83d8-4eff27d79505-kube-api-access-dnnt7\") pod \"keystone-operator-controller-manager-55f684fd56-wzjrz\" (UID: \"783d8159-e67a-4796-83d8-4eff27d79505\") " pod="openstack-operators/keystone-operator-controller-manager-55f684fd56-wzjrz" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.224150 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsmq4\" (UniqueName: \"kubernetes.io/projected/e85aef3a-e235-473c-94cc-1f6237798b3e-kube-api-access-fsmq4\") pod \"horizon-operator-controller-manager-77d5c5b54f-jcb4p\" (UID: \"e85aef3a-e235-473c-94cc-1f6237798b3e\") " pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-jcb4p" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.224206 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z74pn\" (UniqueName: \"kubernetes.io/projected/e7465bd0-3b6e-4199-9ee6-28b512198847-kube-api-access-z74pn\") pod \"infra-operator-controller-manager-7d75bc88d5-t54fr\" (UID: \"e7465bd0-3b6e-4199-9ee6-28b512198847\") " pod="openstack-operators/infra-operator-controller-manager-7d75bc88d5-t54fr" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.224241 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e7465bd0-3b6e-4199-9ee6-28b512198847-cert\") pod \"infra-operator-controller-manager-7d75bc88d5-t54fr\" (UID: \"e7465bd0-3b6e-4199-9ee6-28b512198847\") " pod="openstack-operators/infra-operator-controller-manager-7d75bc88d5-t54fr" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.224283 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs77k\" (UniqueName: \"kubernetes.io/projected/2df85221-33ed-49be-949c-516810279e4d-kube-api-access-zs77k\") pod \"ironic-operator-controller-manager-768b776ffb-sxbjn\" (UID: \"2df85221-33ed-49be-949c-516810279e4d\") " pod="openstack-operators/ironic-operator-controller-manager-768b776ffb-sxbjn" Jan 27 15:23:22 crc kubenswrapper[4772]: E0127 15:23:22.225470 4772 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 27 15:23:22 crc kubenswrapper[4772]: E0127 15:23:22.225518 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e7465bd0-3b6e-4199-9ee6-28b512198847-cert podName:e7465bd0-3b6e-4199-9ee6-28b512198847 nodeName:}" failed. No retries permitted until 2026-01-27 15:23:22.725502245 +0000 UTC m=+988.706111343 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e7465bd0-3b6e-4199-9ee6-28b512198847-cert") pod "infra-operator-controller-manager-7d75bc88d5-t54fr" (UID: "e7465bd0-3b6e-4199-9ee6-28b512198847") : secret "infra-operator-webhook-server-cert" not found Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.228619 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-qxqv9" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.242263 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-ddcbfd695-6wltn"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.250810 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-ddcbfd695-6wltn" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.261875 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2qbl\" (UniqueName: \"kubernetes.io/projected/27ec5082-c170-465b-b3a3-1f27a545fd71-kube-api-access-x2qbl\") pod \"manila-operator-controller-manager-849fcfbb6b-tvrx9\" (UID: \"27ec5082-c170-465b-b3a3-1f27a545fd71\") " pod="openstack-operators/manila-operator-controller-manager-849fcfbb6b-tvrx9" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.282487 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-8wz4d" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.283401 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsmq4\" (UniqueName: \"kubernetes.io/projected/e85aef3a-e235-473c-94cc-1f6237798b3e-kube-api-access-fsmq4\") pod \"horizon-operator-controller-manager-77d5c5b54f-jcb4p\" (UID: \"e85aef3a-e235-473c-94cc-1f6237798b3e\") " pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-jcb4p" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.287971 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs77k\" (UniqueName: \"kubernetes.io/projected/2df85221-33ed-49be-949c-516810279e4d-kube-api-access-zs77k\") pod \"ironic-operator-controller-manager-768b776ffb-sxbjn\" (UID: \"2df85221-33ed-49be-949c-516810279e4d\") " pod="openstack-operators/ironic-operator-controller-manager-768b776ffb-sxbjn" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.298117 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z74pn\" (UniqueName: \"kubernetes.io/projected/e7465bd0-3b6e-4199-9ee6-28b512198847-kube-api-access-z74pn\") pod \"infra-operator-controller-manager-7d75bc88d5-t54fr\" (UID: \"e7465bd0-3b6e-4199-9ee6-28b512198847\") " pod="openstack-operators/infra-operator-controller-manager-7d75bc88d5-t54fr" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.298708 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-zhd82"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.299641 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-zhd82" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.302228 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnnt7\" (UniqueName: \"kubernetes.io/projected/783d8159-e67a-4796-83d8-4eff27d79505-kube-api-access-dnnt7\") pod \"keystone-operator-controller-manager-55f684fd56-wzjrz\" (UID: \"783d8159-e67a-4796-83d8-4eff27d79505\") " pod="openstack-operators/keystone-operator-controller-manager-55f684fd56-wzjrz" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.313457 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-768b776ffb-sxbjn" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.316726 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-kh8qn" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.324842 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-849fcfbb6b-tvrx9" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.329654 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kqwk\" (UniqueName: \"kubernetes.io/projected/0a88aa66-b634-44ee-8e5b-bfeacb765e57-kube-api-access-6kqwk\") pod \"neutron-operator-controller-manager-7ffd8d76d4-gcpj4\" (UID: \"0a88aa66-b634-44ee-8e5b-bfeacb765e57\") " pod="openstack-operators/neutron-operator-controller-manager-7ffd8d76d4-gcpj4" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.343218 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7875d7675-ktfbt"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.344069 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7875d7675-ktfbt" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.345962 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-rlp2h" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.369900 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7ffd8d76d4-gcpj4"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.395350 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-ddcbfd695-6wltn"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.410714 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-zhd82"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.430939 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7875d7675-ktfbt"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.433448 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zmgp\" (UniqueName: \"kubernetes.io/projected/80584c24-3c75-4624-802f-e608f640eeaa-kube-api-access-2zmgp\") pod \"octavia-operator-controller-manager-7875d7675-ktfbt\" (UID: \"80584c24-3c75-4624-802f-e608f640eeaa\") " pod="openstack-operators/octavia-operator-controller-manager-7875d7675-ktfbt" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.433489 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbb4j\" (UniqueName: \"kubernetes.io/projected/e7fc5297-101a-496e-a7c6-e7296e08a5af-kube-api-access-gbb4j\") pod \"nova-operator-controller-manager-ddcbfd695-6wltn\" (UID: \"e7fc5297-101a-496e-a7c6-e7296e08a5af\") " pod="openstack-operators/nova-operator-controller-manager-ddcbfd695-6wltn" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.433524 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kqwk\" (UniqueName: \"kubernetes.io/projected/0a88aa66-b634-44ee-8e5b-bfeacb765e57-kube-api-access-6kqwk\") pod \"neutron-operator-controller-manager-7ffd8d76d4-gcpj4\" (UID: \"0a88aa66-b634-44ee-8e5b-bfeacb765e57\") " pod="openstack-operators/neutron-operator-controller-manager-7ffd8d76d4-gcpj4" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.433570 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8n6m\" (UniqueName: \"kubernetes.io/projected/b73c175a-e89e-434f-996a-65c1140bb8dd-kube-api-access-t8n6m\") pod \"mariadb-operator-controller-manager-6b9fb5fdcb-zhd82\" (UID: \"b73c175a-e89e-434f-996a-65c1140bb8dd\") " pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-zhd82" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.440305 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854q994c"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.441138 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854q994c" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.443302 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.443588 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-tbw52" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.452701 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-6f75f45d54-ww79v"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.454825 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-ww79v" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.470925 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-ljmlt" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.484010 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-6f75f45d54-ww79v"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.496706 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-79d5ccc684-vwnwk"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.497577 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-vwnwk" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.504480 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kqwk\" (UniqueName: \"kubernetes.io/projected/0a88aa66-b634-44ee-8e5b-bfeacb765e57-kube-api-access-6kqwk\") pod \"neutron-operator-controller-manager-7ffd8d76d4-gcpj4\" (UID: \"0a88aa66-b634-44ee-8e5b-bfeacb765e57\") " pod="openstack-operators/neutron-operator-controller-manager-7ffd8d76d4-gcpj4" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.513718 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-hf92f" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.518939 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-547cbdb99f-l8d48"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.520796 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-jcb4p" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.521544 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-l8d48" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.528658 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-64w92" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.538372 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1389813b-42ea-433f-820c-e5b8b41713d7-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854q994c\" (UID: \"1389813b-42ea-433f-820c-e5b8b41713d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854q994c" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.541373 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zmgp\" (UniqueName: \"kubernetes.io/projected/80584c24-3c75-4624-802f-e608f640eeaa-kube-api-access-2zmgp\") pod \"octavia-operator-controller-manager-7875d7675-ktfbt\" (UID: \"80584c24-3c75-4624-802f-e608f640eeaa\") " pod="openstack-operators/octavia-operator-controller-manager-7875d7675-ktfbt" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.541411 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmnhg\" (UniqueName: \"kubernetes.io/projected/1389813b-42ea-433f-820c-e5b8b41713d7-kube-api-access-gmnhg\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854q994c\" (UID: \"1389813b-42ea-433f-820c-e5b8b41713d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854q994c" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.541443 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmzpp\" (UniqueName: \"kubernetes.io/projected/e76712a7-ebf6-4f04-a52c-c8d2bacb87f7-kube-api-access-lmzpp\") pod \"ovn-operator-controller-manager-6f75f45d54-ww79v\" (UID: \"e76712a7-ebf6-4f04-a52c-c8d2bacb87f7\") " pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-ww79v" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.541471 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbb4j\" (UniqueName: \"kubernetes.io/projected/e7fc5297-101a-496e-a7c6-e7296e08a5af-kube-api-access-gbb4j\") pod \"nova-operator-controller-manager-ddcbfd695-6wltn\" (UID: \"e7fc5297-101a-496e-a7c6-e7296e08a5af\") " pod="openstack-operators/nova-operator-controller-manager-ddcbfd695-6wltn" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.541569 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8n6m\" (UniqueName: \"kubernetes.io/projected/b73c175a-e89e-434f-996a-65c1140bb8dd-kube-api-access-t8n6m\") pod \"mariadb-operator-controller-manager-6b9fb5fdcb-zhd82\" (UID: \"b73c175a-e89e-434f-996a-65c1140bb8dd\") " pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-zhd82" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.564899 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-79d5ccc684-vwnwk"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.564951 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-547cbdb99f-l8d48"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.569721 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854q994c"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.577507 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-69797bbcbd-ln7xf"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.578301 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-ln7xf" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.580340 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8n6m\" (UniqueName: \"kubernetes.io/projected/b73c175a-e89e-434f-996a-65c1140bb8dd-kube-api-access-t8n6m\") pod \"mariadb-operator-controller-manager-6b9fb5fdcb-zhd82\" (UID: \"b73c175a-e89e-434f-996a-65c1140bb8dd\") " pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-zhd82" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.580616 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55f684fd56-wzjrz" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.580613 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-q2hmw" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.599305 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbb4j\" (UniqueName: \"kubernetes.io/projected/e7fc5297-101a-496e-a7c6-e7296e08a5af-kube-api-access-gbb4j\") pod \"nova-operator-controller-manager-ddcbfd695-6wltn\" (UID: \"e7fc5297-101a-496e-a7c6-e7296e08a5af\") " pod="openstack-operators/nova-operator-controller-manager-ddcbfd695-6wltn" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.599520 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zmgp\" (UniqueName: \"kubernetes.io/projected/80584c24-3c75-4624-802f-e608f640eeaa-kube-api-access-2zmgp\") pod \"octavia-operator-controller-manager-7875d7675-ktfbt\" (UID: \"80584c24-3c75-4624-802f-e608f640eeaa\") " pod="openstack-operators/octavia-operator-controller-manager-7875d7675-ktfbt" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.599585 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-799bc87c89-k2l8k"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.600493 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-799bc87c89-k2l8k" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.602078 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-6hhp6" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.612505 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-69797bbcbd-ln7xf"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.618253 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-799bc87c89-k2l8k"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.627897 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c9bb4b66c-ws2mh"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.629049 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c9bb4b66c-ws2mh" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.633272 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-2w5kx" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.642834 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmnhg\" (UniqueName: \"kubernetes.io/projected/1389813b-42ea-433f-820c-e5b8b41713d7-kube-api-access-gmnhg\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854q994c\" (UID: \"1389813b-42ea-433f-820c-e5b8b41713d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854q994c" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.642871 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmzpp\" (UniqueName: \"kubernetes.io/projected/e76712a7-ebf6-4f04-a52c-c8d2bacb87f7-kube-api-access-lmzpp\") pod \"ovn-operator-controller-manager-6f75f45d54-ww79v\" (UID: \"e76712a7-ebf6-4f04-a52c-c8d2bacb87f7\") " pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-ww79v" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.642914 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6k4pg\" (UniqueName: \"kubernetes.io/projected/c3d2f06d-4dd3-49a8-a0a0-54a83cc3f4e8-kube-api-access-6k4pg\") pod \"placement-operator-controller-manager-79d5ccc684-vwnwk\" (UID: \"c3d2f06d-4dd3-49a8-a0a0-54a83cc3f4e8\") " pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-vwnwk" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.642943 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7qgw\" (UniqueName: \"kubernetes.io/projected/c09741c3-6bae-487a-9b4c-7c9f01d8c5bf-kube-api-access-l7qgw\") pod \"swift-operator-controller-manager-547cbdb99f-l8d48\" (UID: \"c09741c3-6bae-487a-9b4c-7c9f01d8c5bf\") " pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-l8d48" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.642978 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1389813b-42ea-433f-820c-e5b8b41713d7-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854q994c\" (UID: \"1389813b-42ea-433f-820c-e5b8b41713d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854q994c" Jan 27 15:23:22 crc kubenswrapper[4772]: E0127 15:23:22.643079 4772 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 27 15:23:22 crc kubenswrapper[4772]: E0127 15:23:22.643115 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1389813b-42ea-433f-820c-e5b8b41713d7-cert podName:1389813b-42ea-433f-820c-e5b8b41713d7 nodeName:}" failed. No retries permitted until 2026-01-27 15:23:23.143102743 +0000 UTC m=+989.123711841 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1389813b-42ea-433f-820c-e5b8b41713d7-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854q994c" (UID: "1389813b-42ea-433f-820c-e5b8b41713d7") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.647907 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c9bb4b66c-ws2mh"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.655429 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.656275 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7ffd8d76d4-gcpj4" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.657128 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.658452 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.659015 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-xvq6z" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.663665 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.664374 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmzpp\" (UniqueName: \"kubernetes.io/projected/e76712a7-ebf6-4f04-a52c-c8d2bacb87f7-kube-api-access-lmzpp\") pod \"ovn-operator-controller-manager-6f75f45d54-ww79v\" (UID: \"e76712a7-ebf6-4f04-a52c-c8d2bacb87f7\") " pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-ww79v" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.664477 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmnhg\" (UniqueName: \"kubernetes.io/projected/1389813b-42ea-433f-820c-e5b8b41713d7-kube-api-access-gmnhg\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854q994c\" (UID: \"1389813b-42ea-433f-820c-e5b8b41713d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854q994c" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.701810 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.713468 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-ddcbfd695-6wltn" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.732583 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h9297"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.741389 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h9297" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.753034 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-sn68d" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.761889 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrxgx\" (UniqueName: \"kubernetes.io/projected/e4a99865-64a7-49e5-bdce-ff929105fc0d-kube-api-access-lrxgx\") pod \"telemetry-operator-controller-manager-799bc87c89-k2l8k\" (UID: \"e4a99865-64a7-49e5-bdce-ff929105fc0d\") " pod="openstack-operators/telemetry-operator-controller-manager-799bc87c89-k2l8k" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.767548 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-zhd82" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.770310 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-webhook-certs\") pod \"openstack-operator-controller-manager-ff554fc88-clt4p\" (UID: \"8087d8d3-d2f6-4bca-abec-f5b5335f26fa\") " pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.770925 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8vnj\" (UniqueName: \"kubernetes.io/projected/6242683c-24ad-4e22-a7b3-8463e07388c2-kube-api-access-s8vnj\") pod \"test-operator-controller-manager-69797bbcbd-ln7xf\" (UID: \"6242683c-24ad-4e22-a7b3-8463e07388c2\") " pod="openstack-operators/test-operator-controller-manager-69797bbcbd-ln7xf" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.771090 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h9297"] Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.771116 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6k4pg\" (UniqueName: \"kubernetes.io/projected/c3d2f06d-4dd3-49a8-a0a0-54a83cc3f4e8-kube-api-access-6k4pg\") pod \"placement-operator-controller-manager-79d5ccc684-vwnwk\" (UID: \"c3d2f06d-4dd3-49a8-a0a0-54a83cc3f4e8\") " pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-vwnwk" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.771574 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7qgw\" (UniqueName: \"kubernetes.io/projected/c09741c3-6bae-487a-9b4c-7c9f01d8c5bf-kube-api-access-l7qgw\") pod \"swift-operator-controller-manager-547cbdb99f-l8d48\" (UID: \"c09741c3-6bae-487a-9b4c-7c9f01d8c5bf\") " pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-l8d48" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.771631 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-metrics-certs\") pod \"openstack-operator-controller-manager-ff554fc88-clt4p\" (UID: \"8087d8d3-d2f6-4bca-abec-f5b5335f26fa\") " pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.771715 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcm8k\" (UniqueName: \"kubernetes.io/projected/783285f4-2e9d-4af5-b017-32676e7d1b01-kube-api-access-jcm8k\") pod \"watcher-operator-controller-manager-6c9bb4b66c-ws2mh\" (UID: \"783285f4-2e9d-4af5-b017-32676e7d1b01\") " pod="openstack-operators/watcher-operator-controller-manager-6c9bb4b66c-ws2mh" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.771800 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e7465bd0-3b6e-4199-9ee6-28b512198847-cert\") pod \"infra-operator-controller-manager-7d75bc88d5-t54fr\" (UID: \"e7465bd0-3b6e-4199-9ee6-28b512198847\") " pod="openstack-operators/infra-operator-controller-manager-7d75bc88d5-t54fr" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.771856 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7phdt\" (UniqueName: \"kubernetes.io/projected/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-kube-api-access-7phdt\") pod \"openstack-operator-controller-manager-ff554fc88-clt4p\" (UID: \"8087d8d3-d2f6-4bca-abec-f5b5335f26fa\") " pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:23:22 crc kubenswrapper[4772]: E0127 15:23:22.772352 4772 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 27 15:23:22 crc kubenswrapper[4772]: E0127 15:23:22.772424 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e7465bd0-3b6e-4199-9ee6-28b512198847-cert podName:e7465bd0-3b6e-4199-9ee6-28b512198847 nodeName:}" failed. No retries permitted until 2026-01-27 15:23:23.772390461 +0000 UTC m=+989.752999559 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e7465bd0-3b6e-4199-9ee6-28b512198847-cert") pod "infra-operator-controller-manager-7d75bc88d5-t54fr" (UID: "e7465bd0-3b6e-4199-9ee6-28b512198847") : secret "infra-operator-webhook-server-cert" not found Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.805785 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6k4pg\" (UniqueName: \"kubernetes.io/projected/c3d2f06d-4dd3-49a8-a0a0-54a83cc3f4e8-kube-api-access-6k4pg\") pod \"placement-operator-controller-manager-79d5ccc684-vwnwk\" (UID: \"c3d2f06d-4dd3-49a8-a0a0-54a83cc3f4e8\") " pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-vwnwk" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.806940 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7qgw\" (UniqueName: \"kubernetes.io/projected/c09741c3-6bae-487a-9b4c-7c9f01d8c5bf-kube-api-access-l7qgw\") pod \"swift-operator-controller-manager-547cbdb99f-l8d48\" (UID: \"c09741c3-6bae-487a-9b4c-7c9f01d8c5bf\") " pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-l8d48" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.815866 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7875d7675-ktfbt" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.828393 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-65ff799cfd-t42n9"] Jan 27 15:23:22 crc kubenswrapper[4772]: W0127 15:23:22.852522 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod674f4da6_f50d_4bab_808d_56ab3b9e2cb4.slice/crio-327b15a8084a569369ed015449a1ec6b61f653b1f47c81d392243142feb80680 WatchSource:0}: Error finding container 327b15a8084a569369ed015449a1ec6b61f653b1f47c81d392243142feb80680: Status 404 returned error can't find the container with id 327b15a8084a569369ed015449a1ec6b61f653b1f47c81d392243142feb80680 Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.874361 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8vnj\" (UniqueName: \"kubernetes.io/projected/6242683c-24ad-4e22-a7b3-8463e07388c2-kube-api-access-s8vnj\") pod \"test-operator-controller-manager-69797bbcbd-ln7xf\" (UID: \"6242683c-24ad-4e22-a7b3-8463e07388c2\") " pod="openstack-operators/test-operator-controller-manager-69797bbcbd-ln7xf" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.874443 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-metrics-certs\") pod \"openstack-operator-controller-manager-ff554fc88-clt4p\" (UID: \"8087d8d3-d2f6-4bca-abec-f5b5335f26fa\") " pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.874475 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t57ck\" (UniqueName: \"kubernetes.io/projected/abaf1142-1b7c-4987-8a9d-c91e6456c4a5-kube-api-access-t57ck\") pod \"rabbitmq-cluster-operator-manager-668c99d594-h9297\" (UID: \"abaf1142-1b7c-4987-8a9d-c91e6456c4a5\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h9297" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.874524 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcm8k\" (UniqueName: \"kubernetes.io/projected/783285f4-2e9d-4af5-b017-32676e7d1b01-kube-api-access-jcm8k\") pod \"watcher-operator-controller-manager-6c9bb4b66c-ws2mh\" (UID: \"783285f4-2e9d-4af5-b017-32676e7d1b01\") " pod="openstack-operators/watcher-operator-controller-manager-6c9bb4b66c-ws2mh" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.874565 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7phdt\" (UniqueName: \"kubernetes.io/projected/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-kube-api-access-7phdt\") pod \"openstack-operator-controller-manager-ff554fc88-clt4p\" (UID: \"8087d8d3-d2f6-4bca-abec-f5b5335f26fa\") " pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.874585 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrxgx\" (UniqueName: \"kubernetes.io/projected/e4a99865-64a7-49e5-bdce-ff929105fc0d-kube-api-access-lrxgx\") pod \"telemetry-operator-controller-manager-799bc87c89-k2l8k\" (UID: \"e4a99865-64a7-49e5-bdce-ff929105fc0d\") " pod="openstack-operators/telemetry-operator-controller-manager-799bc87c89-k2l8k" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.874619 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-webhook-certs\") pod \"openstack-operator-controller-manager-ff554fc88-clt4p\" (UID: \"8087d8d3-d2f6-4bca-abec-f5b5335f26fa\") " pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:23:22 crc kubenswrapper[4772]: E0127 15:23:22.874764 4772 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 27 15:23:22 crc kubenswrapper[4772]: E0127 15:23:22.874813 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-webhook-certs podName:8087d8d3-d2f6-4bca-abec-f5b5335f26fa nodeName:}" failed. No retries permitted until 2026-01-27 15:23:23.374797836 +0000 UTC m=+989.355406934 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-webhook-certs") pod "openstack-operator-controller-manager-ff554fc88-clt4p" (UID: "8087d8d3-d2f6-4bca-abec-f5b5335f26fa") : secret "webhook-server-cert" not found Jan 27 15:23:22 crc kubenswrapper[4772]: E0127 15:23:22.875197 4772 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 27 15:23:22 crc kubenswrapper[4772]: E0127 15:23:22.875221 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-metrics-certs podName:8087d8d3-d2f6-4bca-abec-f5b5335f26fa nodeName:}" failed. No retries permitted until 2026-01-27 15:23:23.375214388 +0000 UTC m=+989.355823486 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-metrics-certs") pod "openstack-operator-controller-manager-ff554fc88-clt4p" (UID: "8087d8d3-d2f6-4bca-abec-f5b5335f26fa") : secret "metrics-server-cert" not found Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.905106 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcm8k\" (UniqueName: \"kubernetes.io/projected/783285f4-2e9d-4af5-b017-32676e7d1b01-kube-api-access-jcm8k\") pod \"watcher-operator-controller-manager-6c9bb4b66c-ws2mh\" (UID: \"783285f4-2e9d-4af5-b017-32676e7d1b01\") " pod="openstack-operators/watcher-operator-controller-manager-6c9bb4b66c-ws2mh" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.907684 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrxgx\" (UniqueName: \"kubernetes.io/projected/e4a99865-64a7-49e5-bdce-ff929105fc0d-kube-api-access-lrxgx\") pod \"telemetry-operator-controller-manager-799bc87c89-k2l8k\" (UID: \"e4a99865-64a7-49e5-bdce-ff929105fc0d\") " pod="openstack-operators/telemetry-operator-controller-manager-799bc87c89-k2l8k" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.910978 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7phdt\" (UniqueName: \"kubernetes.io/projected/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-kube-api-access-7phdt\") pod \"openstack-operator-controller-manager-ff554fc88-clt4p\" (UID: \"8087d8d3-d2f6-4bca-abec-f5b5335f26fa\") " pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.915543 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8vnj\" (UniqueName: \"kubernetes.io/projected/6242683c-24ad-4e22-a7b3-8463e07388c2-kube-api-access-s8vnj\") pod \"test-operator-controller-manager-69797bbcbd-ln7xf\" (UID: \"6242683c-24ad-4e22-a7b3-8463e07388c2\") " pod="openstack-operators/test-operator-controller-manager-69797bbcbd-ln7xf" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.959273 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-ww79v" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.976032 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t57ck\" (UniqueName: \"kubernetes.io/projected/abaf1142-1b7c-4987-8a9d-c91e6456c4a5-kube-api-access-t57ck\") pod \"rabbitmq-cluster-operator-manager-668c99d594-h9297\" (UID: \"abaf1142-1b7c-4987-8a9d-c91e6456c4a5\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h9297" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.985798 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-vwnwk" Jan 27 15:23:22 crc kubenswrapper[4772]: I0127 15:23:22.996767 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t57ck\" (UniqueName: \"kubernetes.io/projected/abaf1142-1b7c-4987-8a9d-c91e6456c4a5-kube-api-access-t57ck\") pod \"rabbitmq-cluster-operator-manager-668c99d594-h9297\" (UID: \"abaf1142-1b7c-4987-8a9d-c91e6456c4a5\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h9297" Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.021519 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-l8d48" Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.038092 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-ln7xf" Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.107205 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-799bc87c89-k2l8k" Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.122796 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-65ff799cfd-t42n9" event={"ID":"674f4da6-f50d-4bab-808d-56ab3b9e2cb4","Type":"ContainerStarted","Data":"327b15a8084a569369ed015449a1ec6b61f653b1f47c81d392243142feb80680"} Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.161265 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c9bb4b66c-ws2mh" Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.181035 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1389813b-42ea-433f-820c-e5b8b41713d7-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854q994c\" (UID: \"1389813b-42ea-433f-820c-e5b8b41713d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854q994c" Jan 27 15:23:23 crc kubenswrapper[4772]: E0127 15:23:23.181210 4772 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 27 15:23:23 crc kubenswrapper[4772]: E0127 15:23:23.181258 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1389813b-42ea-433f-820c-e5b8b41713d7-cert podName:1389813b-42ea-433f-820c-e5b8b41713d7 nodeName:}" failed. No retries permitted until 2026-01-27 15:23:24.181242117 +0000 UTC m=+990.161851215 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1389813b-42ea-433f-820c-e5b8b41713d7-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854q994c" (UID: "1389813b-42ea-433f-820c-e5b8b41713d7") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.224111 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-849fcfbb6b-tvrx9"] Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.236310 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h9297" Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.254861 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-655bf9cfbb-cgh7j"] Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.265938 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-77554cdc5c-tkr6j"] Jan 27 15:23:23 crc kubenswrapper[4772]: W0127 15:23:23.315940 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd395f105_54f0_4497_a119_57802be313a3.slice/crio-173210d17fd463ed8fef2550b0ce546bd294ef92d7429c1055eeb446216d0cb0 WatchSource:0}: Error finding container 173210d17fd463ed8fef2550b0ce546bd294ef92d7429c1055eeb446216d0cb0: Status 404 returned error can't find the container with id 173210d17fd463ed8fef2550b0ce546bd294ef92d7429c1055eeb446216d0cb0 Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.384714 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-metrics-certs\") pod \"openstack-operator-controller-manager-ff554fc88-clt4p\" (UID: \"8087d8d3-d2f6-4bca-abec-f5b5335f26fa\") " pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.384859 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-webhook-certs\") pod \"openstack-operator-controller-manager-ff554fc88-clt4p\" (UID: \"8087d8d3-d2f6-4bca-abec-f5b5335f26fa\") " pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:23:23 crc kubenswrapper[4772]: E0127 15:23:23.385014 4772 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 27 15:23:23 crc kubenswrapper[4772]: E0127 15:23:23.385071 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-webhook-certs podName:8087d8d3-d2f6-4bca-abec-f5b5335f26fa nodeName:}" failed. No retries permitted until 2026-01-27 15:23:24.385054097 +0000 UTC m=+990.365663195 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-webhook-certs") pod "openstack-operator-controller-manager-ff554fc88-clt4p" (UID: "8087d8d3-d2f6-4bca-abec-f5b5335f26fa") : secret "webhook-server-cert" not found Jan 27 15:23:23 crc kubenswrapper[4772]: E0127 15:23:23.385135 4772 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 27 15:23:23 crc kubenswrapper[4772]: E0127 15:23:23.385161 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-metrics-certs podName:8087d8d3-d2f6-4bca-abec-f5b5335f26fa nodeName:}" failed. No retries permitted until 2026-01-27 15:23:24.38515289 +0000 UTC m=+990.365761998 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-metrics-certs") pod "openstack-operator-controller-manager-ff554fc88-clt4p" (UID: "8087d8d3-d2f6-4bca-abec-f5b5335f26fa") : secret "metrics-server-cert" not found Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.404100 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-768b776ffb-sxbjn"] Jan 27 15:23:23 crc kubenswrapper[4772]: W0127 15:23:23.423926 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2df85221_33ed_49be_949c_516810279e4d.slice/crio-923cea042cc049643b9f18ac2932392946acb25bc60a1395d9554055be88e0fd WatchSource:0}: Error finding container 923cea042cc049643b9f18ac2932392946acb25bc60a1395d9554055be88e0fd: Status 404 returned error can't find the container with id 923cea042cc049643b9f18ac2932392946acb25bc60a1395d9554055be88e0fd Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.441654 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-575ffb885b-mtd9d"] Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.457367 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-67dd55ff59-hgscb"] Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.469112 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55f684fd56-wzjrz"] Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.493985 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-77d5c5b54f-jcb4p"] Jan 27 15:23:23 crc kubenswrapper[4772]: W0127 15:23:23.501104 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode85aef3a_e235_473c_94cc_1f6237798b3e.slice/crio-56bf3334bca8a317a9c02e89bea12c6ad7b7a61a941db3290a4791db7563a506 WatchSource:0}: Error finding container 56bf3334bca8a317a9c02e89bea12c6ad7b7a61a941db3290a4791db7563a506: Status 404 returned error can't find the container with id 56bf3334bca8a317a9c02e89bea12c6ad7b7a61a941db3290a4791db7563a506 Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.592028 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7875d7675-ktfbt"] Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.606714 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7ffd8d76d4-gcpj4"] Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.626464 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-zhd82"] Jan 27 15:23:23 crc kubenswrapper[4772]: W0127 15:23:23.630990 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a88aa66_b634_44ee_8e5b_bfeacb765e57.slice/crio-4bb52078231b82897e96da9e073eb926d06bf3f15389686c1533aa0597ef25b0 WatchSource:0}: Error finding container 4bb52078231b82897e96da9e073eb926d06bf3f15389686c1533aa0597ef25b0: Status 404 returned error can't find the container with id 4bb52078231b82897e96da9e073eb926d06bf3f15389686c1533aa0597ef25b0 Jan 27 15:23:23 crc kubenswrapper[4772]: W0127 15:23:23.631594 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80584c24_3c75_4624_802f_e608f640eeaa.slice/crio-75a5b9256fe3030df211e53616d6aaacfa661c5623aee3562a86d05a5b0c0ee1 WatchSource:0}: Error finding container 75a5b9256fe3030df211e53616d6aaacfa661c5623aee3562a86d05a5b0c0ee1: Status 404 returned error can't find the container with id 75a5b9256fe3030df211e53616d6aaacfa661c5623aee3562a86d05a5b0c0ee1 Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.635547 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-ddcbfd695-6wltn"] Jan 27 15:23:23 crc kubenswrapper[4772]: W0127 15:23:23.641345 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7fc5297_101a_496e_a7c6_e7296e08a5af.slice/crio-f223d3b2deea88b7b47d9c6cfd2561cd263d36b3f2656a548273fd94975efdad WatchSource:0}: Error finding container f223d3b2deea88b7b47d9c6cfd2561cd263d36b3f2656a548273fd94975efdad: Status 404 returned error can't find the container with id f223d3b2deea88b7b47d9c6cfd2561cd263d36b3f2656a548273fd94975efdad Jan 27 15:23:23 crc kubenswrapper[4772]: W0127 15:23:23.641681 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb73c175a_e89e_434f_996a_65c1140bb8dd.slice/crio-217e47d1ea1e577f9f3d26d127deeb429193e347283ee4598ecaa45a8a6c4104 WatchSource:0}: Error finding container 217e47d1ea1e577f9f3d26d127deeb429193e347283ee4598ecaa45a8a6c4104: Status 404 returned error can't find the container with id 217e47d1ea1e577f9f3d26d127deeb429193e347283ee4598ecaa45a8a6c4104 Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.743379 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-79d5ccc684-vwnwk"] Jan 27 15:23:23 crc kubenswrapper[4772]: E0127 15:23:23.754190 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:013c0ad82d21a21c7eece5cd4b5d5c4b8eb410b6671ac33a6f3fb78c8510811d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6k4pg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-79d5ccc684-vwnwk_openstack-operators(c3d2f06d-4dd3-49a8-a0a0-54a83cc3f4e8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 27 15:23:23 crc kubenswrapper[4772]: E0127 15:23:23.755398 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-vwnwk" podUID="c3d2f06d-4dd3-49a8-a0a0-54a83cc3f4e8" Jan 27 15:23:23 crc kubenswrapper[4772]: W0127 15:23:23.755865 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode76712a7_ebf6_4f04_a52c_c8d2bacb87f7.slice/crio-f752e4541b769e913ac3b0d8af7e559c30963fa17cfa0958a239f359bea1fb22 WatchSource:0}: Error finding container f752e4541b769e913ac3b0d8af7e559c30963fa17cfa0958a239f359bea1fb22: Status 404 returned error can't find the container with id f752e4541b769e913ac3b0d8af7e559c30963fa17cfa0958a239f359bea1fb22 Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.758261 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-547cbdb99f-l8d48"] Jan 27 15:23:23 crc kubenswrapper[4772]: E0127 15:23:23.761536 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:fa46fc14710961e6b4a76a3522dca3aa3cfa71436c7cf7ade533d3712822f327,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lmzpp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-6f75f45d54-ww79v_openstack-operators(e76712a7-ebf6-4f04-a52c-c8d2bacb87f7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 27 15:23:23 crc kubenswrapper[4772]: E0127 15:23:23.762672 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-ww79v" podUID="e76712a7-ebf6-4f04-a52c-c8d2bacb87f7" Jan 27 15:23:23 crc kubenswrapper[4772]: W0127 15:23:23.762949 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc09741c3_6bae_487a_9b4c_7c9f01d8c5bf.slice/crio-ccbef40d188785ee8381011d390820b0ef2381f4740dc83e4e0fcb6544f9a8fb WatchSource:0}: Error finding container ccbef40d188785ee8381011d390820b0ef2381f4740dc83e4e0fcb6544f9a8fb: Status 404 returned error can't find the container with id ccbef40d188785ee8381011d390820b0ef2381f4740dc83e4e0fcb6544f9a8fb Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.763248 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-6f75f45d54-ww79v"] Jan 27 15:23:23 crc kubenswrapper[4772]: E0127 15:23:23.768527 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:445e951df2f21df6d33a466f75917e0f6103052ae751ae11887136e8ab165922,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l7qgw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-547cbdb99f-l8d48_openstack-operators(c09741c3-6bae-487a-9b4c-7c9f01d8c5bf): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 27 15:23:23 crc kubenswrapper[4772]: E0127 15:23:23.769727 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-l8d48" podUID="c09741c3-6bae-487a-9b4c-7c9f01d8c5bf" Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.791578 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e7465bd0-3b6e-4199-9ee6-28b512198847-cert\") pod \"infra-operator-controller-manager-7d75bc88d5-t54fr\" (UID: \"e7465bd0-3b6e-4199-9ee6-28b512198847\") " pod="openstack-operators/infra-operator-controller-manager-7d75bc88d5-t54fr" Jan 27 15:23:23 crc kubenswrapper[4772]: E0127 15:23:23.791812 4772 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 27 15:23:23 crc kubenswrapper[4772]: E0127 15:23:23.791873 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e7465bd0-3b6e-4199-9ee6-28b512198847-cert podName:e7465bd0-3b6e-4199-9ee6-28b512198847 nodeName:}" failed. No retries permitted until 2026-01-27 15:23:25.791854795 +0000 UTC m=+991.772463893 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e7465bd0-3b6e-4199-9ee6-28b512198847-cert") pod "infra-operator-controller-manager-7d75bc88d5-t54fr" (UID: "e7465bd0-3b6e-4199-9ee6-28b512198847") : secret "infra-operator-webhook-server-cert" not found Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.813853 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c9bb4b66c-ws2mh"] Jan 27 15:23:23 crc kubenswrapper[4772]: W0127 15:23:23.825524 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod783285f4_2e9d_4af5_b017_32676e7d1b01.slice/crio-ed6acc5f50bf0655622492cac0b3f0b4b220be2099e1a5d911dcf85528df3745 WatchSource:0}: Error finding container ed6acc5f50bf0655622492cac0b3f0b4b220be2099e1a5d911dcf85528df3745: Status 404 returned error can't find the container with id ed6acc5f50bf0655622492cac0b3f0b4b220be2099e1a5d911dcf85528df3745 Jan 27 15:23:23 crc kubenswrapper[4772]: E0127 15:23:23.828188 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/lmiccini/watcher-operator@sha256:162fb83ed76cbf5d44ba057fbeee02a9182fdf02346afadb3e16b2e3627e1940,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jcm8k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6c9bb4b66c-ws2mh_openstack-operators(783285f4-2e9d-4af5-b017-32676e7d1b01): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 27 15:23:23 crc kubenswrapper[4772]: E0127 15:23:23.830318 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-6c9bb4b66c-ws2mh" podUID="783285f4-2e9d-4af5-b017-32676e7d1b01" Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.831460 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h9297"] Jan 27 15:23:23 crc kubenswrapper[4772]: W0127 15:23:23.834946 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podabaf1142_1b7c_4987_8a9d_c91e6456c4a5.slice/crio-5d0e7b142093081addf02d817aa0ec9a28effd35fcb9f32597195a91e4afc707 WatchSource:0}: Error finding container 5d0e7b142093081addf02d817aa0ec9a28effd35fcb9f32597195a91e4afc707: Status 404 returned error can't find the container with id 5d0e7b142093081addf02d817aa0ec9a28effd35fcb9f32597195a91e4afc707 Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.851193 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-69797bbcbd-ln7xf"] Jan 27 15:23:23 crc kubenswrapper[4772]: I0127 15:23:23.859556 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-799bc87c89-k2l8k"] Jan 27 15:23:23 crc kubenswrapper[4772]: W0127 15:23:23.865332 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6242683c_24ad_4e22_a7b3_8463e07388c2.slice/crio-f2ece18acf3e1a9e84f9affb02aa64941da9edaa5466989d492565763b3e141f WatchSource:0}: Error finding container f2ece18acf3e1a9e84f9affb02aa64941da9edaa5466989d492565763b3e141f: Status 404 returned error can't find the container with id f2ece18acf3e1a9e84f9affb02aa64941da9edaa5466989d492565763b3e141f Jan 27 15:23:23 crc kubenswrapper[4772]: E0127 15:23:23.867570 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:c8dde42dafd41026ed2e4cfc26efc0fff63c4ba9d31326ae7dc644ccceaafa9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s8vnj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-69797bbcbd-ln7xf_openstack-operators(6242683c-24ad-4e22-a7b3-8463e07388c2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 27 15:23:23 crc kubenswrapper[4772]: W0127 15:23:23.868378 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4a99865_64a7_49e5_bdce_ff929105fc0d.slice/crio-d0b20d02410e5c548ba2aadd079cbe4efef894fa30ffeedf5ad011a05a057ca7 WatchSource:0}: Error finding container d0b20d02410e5c548ba2aadd079cbe4efef894fa30ffeedf5ad011a05a057ca7: Status 404 returned error can't find the container with id d0b20d02410e5c548ba2aadd079cbe4efef894fa30ffeedf5ad011a05a057ca7 Jan 27 15:23:23 crc kubenswrapper[4772]: E0127 15:23:23.869087 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-ln7xf" podUID="6242683c-24ad-4e22-a7b3-8463e07388c2" Jan 27 15:23:23 crc kubenswrapper[4772]: E0127 15:23:23.870654 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/lmiccini/telemetry-operator@sha256:1f1fea3b7df89b81756eab8e6f4c9bed01ab7e949a6ce2d7692c260f41dfbc20,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lrxgx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-799bc87c89-k2l8k_openstack-operators(e4a99865-64a7-49e5-bdce-ff929105fc0d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 27 15:23:23 crc kubenswrapper[4772]: E0127 15:23:23.872121 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-799bc87c89-k2l8k" podUID="e4a99865-64a7-49e5-bdce-ff929105fc0d" Jan 27 15:23:24 crc kubenswrapper[4772]: I0127 15:23:24.128330 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-799bc87c89-k2l8k" event={"ID":"e4a99865-64a7-49e5-bdce-ff929105fc0d","Type":"ContainerStarted","Data":"d0b20d02410e5c548ba2aadd079cbe4efef894fa30ffeedf5ad011a05a057ca7"} Jan 27 15:23:24 crc kubenswrapper[4772]: E0127 15:23:24.129836 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/telemetry-operator@sha256:1f1fea3b7df89b81756eab8e6f4c9bed01ab7e949a6ce2d7692c260f41dfbc20\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-799bc87c89-k2l8k" podUID="e4a99865-64a7-49e5-bdce-ff929105fc0d" Jan 27 15:23:24 crc kubenswrapper[4772]: I0127 15:23:24.130433 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-849fcfbb6b-tvrx9" event={"ID":"27ec5082-c170-465b-b3a3-1f27a545fd71","Type":"ContainerStarted","Data":"2f5dcda5d379d1885b362cc12de1a01612d50caf320579c3c8a372c1ff596cec"} Jan 27 15:23:24 crc kubenswrapper[4772]: I0127 15:23:24.131097 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-77554cdc5c-tkr6j" event={"ID":"d395f105-54f0-4497-a119-57802be313a3","Type":"ContainerStarted","Data":"173210d17fd463ed8fef2550b0ce546bd294ef92d7429c1055eeb446216d0cb0"} Jan 27 15:23:24 crc kubenswrapper[4772]: I0127 15:23:24.132934 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-jcb4p" event={"ID":"e85aef3a-e235-473c-94cc-1f6237798b3e","Type":"ContainerStarted","Data":"56bf3334bca8a317a9c02e89bea12c6ad7b7a61a941db3290a4791db7563a506"} Jan 27 15:23:24 crc kubenswrapper[4772]: I0127 15:23:24.133712 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-ww79v" event={"ID":"e76712a7-ebf6-4f04-a52c-c8d2bacb87f7","Type":"ContainerStarted","Data":"f752e4541b769e913ac3b0d8af7e559c30963fa17cfa0958a239f359bea1fb22"} Jan 27 15:23:24 crc kubenswrapper[4772]: E0127 15:23:24.134632 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:fa46fc14710961e6b4a76a3522dca3aa3cfa71436c7cf7ade533d3712822f327\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-ww79v" podUID="e76712a7-ebf6-4f04-a52c-c8d2bacb87f7" Jan 27 15:23:24 crc kubenswrapper[4772]: I0127 15:23:24.134963 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c9bb4b66c-ws2mh" event={"ID":"783285f4-2e9d-4af5-b017-32676e7d1b01","Type":"ContainerStarted","Data":"ed6acc5f50bf0655622492cac0b3f0b4b220be2099e1a5d911dcf85528df3745"} Jan 27 15:23:24 crc kubenswrapper[4772]: E0127 15:23:24.135665 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/watcher-operator@sha256:162fb83ed76cbf5d44ba057fbeee02a9182fdf02346afadb3e16b2e3627e1940\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6c9bb4b66c-ws2mh" podUID="783285f4-2e9d-4af5-b017-32676e7d1b01" Jan 27 15:23:24 crc kubenswrapper[4772]: I0127 15:23:24.135959 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-655bf9cfbb-cgh7j" event={"ID":"fde95124-892b-411a-ba05-fa70927c8838","Type":"ContainerStarted","Data":"779c536711b064a2c67dc5eaa590675460d3c315e4c4be0ba5fe5fb942f89869"} Jan 27 15:23:24 crc kubenswrapper[4772]: I0127 15:23:24.136590 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-ddcbfd695-6wltn" event={"ID":"e7fc5297-101a-496e-a7c6-e7296e08a5af","Type":"ContainerStarted","Data":"f223d3b2deea88b7b47d9c6cfd2561cd263d36b3f2656a548273fd94975efdad"} Jan 27 15:23:24 crc kubenswrapper[4772]: I0127 15:23:24.137493 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-zhd82" event={"ID":"b73c175a-e89e-434f-996a-65c1140bb8dd","Type":"ContainerStarted","Data":"217e47d1ea1e577f9f3d26d127deeb429193e347283ee4598ecaa45a8a6c4104"} Jan 27 15:23:24 crc kubenswrapper[4772]: I0127 15:23:24.138130 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-768b776ffb-sxbjn" event={"ID":"2df85221-33ed-49be-949c-516810279e4d","Type":"ContainerStarted","Data":"923cea042cc049643b9f18ac2932392946acb25bc60a1395d9554055be88e0fd"} Jan 27 15:23:24 crc kubenswrapper[4772]: I0127 15:23:24.139149 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-575ffb885b-mtd9d" event={"ID":"4c63a702-50b9-42f3-858e-7e27da0a8d8f","Type":"ContainerStarted","Data":"d39c45eba42b53622727a2983e3648d45ebd67d17e99c62446159a41354d110a"} Jan 27 15:23:24 crc kubenswrapper[4772]: I0127 15:23:24.142503 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h9297" event={"ID":"abaf1142-1b7c-4987-8a9d-c91e6456c4a5","Type":"ContainerStarted","Data":"5d0e7b142093081addf02d817aa0ec9a28effd35fcb9f32597195a91e4afc707"} Jan 27 15:23:24 crc kubenswrapper[4772]: I0127 15:23:24.144382 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-l8d48" event={"ID":"c09741c3-6bae-487a-9b4c-7c9f01d8c5bf","Type":"ContainerStarted","Data":"ccbef40d188785ee8381011d390820b0ef2381f4740dc83e4e0fcb6544f9a8fb"} Jan 27 15:23:24 crc kubenswrapper[4772]: E0127 15:23:24.146224 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:445e951df2f21df6d33a466f75917e0f6103052ae751ae11887136e8ab165922\\\"\"" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-l8d48" podUID="c09741c3-6bae-487a-9b4c-7c9f01d8c5bf" Jan 27 15:23:24 crc kubenswrapper[4772]: I0127 15:23:24.148545 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55f684fd56-wzjrz" event={"ID":"783d8159-e67a-4796-83d8-4eff27d79505","Type":"ContainerStarted","Data":"f39c14e2a877fc5caf6155f90f1f56d6b7d577ac42e65ad3f28efe83d3254529"} Jan 27 15:23:24 crc kubenswrapper[4772]: I0127 15:23:24.149635 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-vwnwk" event={"ID":"c3d2f06d-4dd3-49a8-a0a0-54a83cc3f4e8","Type":"ContainerStarted","Data":"2e3eb37d74672104ba96a984370bae529b30991e71efba38da9c8804708dfe62"} Jan 27 15:23:24 crc kubenswrapper[4772]: E0127 15:23:24.152540 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:013c0ad82d21a21c7eece5cd4b5d5c4b8eb410b6671ac33a6f3fb78c8510811d\\\"\"" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-vwnwk" podUID="c3d2f06d-4dd3-49a8-a0a0-54a83cc3f4e8" Jan 27 15:23:24 crc kubenswrapper[4772]: I0127 15:23:24.161529 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-67dd55ff59-hgscb" event={"ID":"fb300814-fca7-4419-ac6e-c08b33edd4be","Type":"ContainerStarted","Data":"b1121b1aaf2c2d1be47c65ced15bfe248060e0013f876ad5bec34c29761f43c5"} Jan 27 15:23:24 crc kubenswrapper[4772]: I0127 15:23:24.165403 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7ffd8d76d4-gcpj4" event={"ID":"0a88aa66-b634-44ee-8e5b-bfeacb765e57","Type":"ContainerStarted","Data":"4bb52078231b82897e96da9e073eb926d06bf3f15389686c1533aa0597ef25b0"} Jan 27 15:23:24 crc kubenswrapper[4772]: I0127 15:23:24.166777 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-ln7xf" event={"ID":"6242683c-24ad-4e22-a7b3-8463e07388c2","Type":"ContainerStarted","Data":"f2ece18acf3e1a9e84f9affb02aa64941da9edaa5466989d492565763b3e141f"} Jan 27 15:23:24 crc kubenswrapper[4772]: I0127 15:23:24.167715 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7875d7675-ktfbt" event={"ID":"80584c24-3c75-4624-802f-e608f640eeaa","Type":"ContainerStarted","Data":"75a5b9256fe3030df211e53616d6aaacfa661c5623aee3562a86d05a5b0c0ee1"} Jan 27 15:23:24 crc kubenswrapper[4772]: E0127 15:23:24.168130 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:c8dde42dafd41026ed2e4cfc26efc0fff63c4ba9d31326ae7dc644ccceaafa9d\\\"\"" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-ln7xf" podUID="6242683c-24ad-4e22-a7b3-8463e07388c2" Jan 27 15:23:24 crc kubenswrapper[4772]: I0127 15:23:24.202880 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1389813b-42ea-433f-820c-e5b8b41713d7-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854q994c\" (UID: \"1389813b-42ea-433f-820c-e5b8b41713d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854q994c" Jan 27 15:23:24 crc kubenswrapper[4772]: E0127 15:23:24.203252 4772 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 27 15:23:24 crc kubenswrapper[4772]: E0127 15:23:24.203351 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1389813b-42ea-433f-820c-e5b8b41713d7-cert podName:1389813b-42ea-433f-820c-e5b8b41713d7 nodeName:}" failed. No retries permitted until 2026-01-27 15:23:26.203327087 +0000 UTC m=+992.183936185 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1389813b-42ea-433f-820c-e5b8b41713d7-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854q994c" (UID: "1389813b-42ea-433f-820c-e5b8b41713d7") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 27 15:23:24 crc kubenswrapper[4772]: I0127 15:23:24.407820 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-metrics-certs\") pod \"openstack-operator-controller-manager-ff554fc88-clt4p\" (UID: \"8087d8d3-d2f6-4bca-abec-f5b5335f26fa\") " pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:23:24 crc kubenswrapper[4772]: I0127 15:23:24.407942 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-webhook-certs\") pod \"openstack-operator-controller-manager-ff554fc88-clt4p\" (UID: \"8087d8d3-d2f6-4bca-abec-f5b5335f26fa\") " pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:23:24 crc kubenswrapper[4772]: E0127 15:23:24.408108 4772 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 27 15:23:24 crc kubenswrapper[4772]: E0127 15:23:24.408245 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-webhook-certs podName:8087d8d3-d2f6-4bca-abec-f5b5335f26fa nodeName:}" failed. No retries permitted until 2026-01-27 15:23:26.408151246 +0000 UTC m=+992.388760334 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-webhook-certs") pod "openstack-operator-controller-manager-ff554fc88-clt4p" (UID: "8087d8d3-d2f6-4bca-abec-f5b5335f26fa") : secret "webhook-server-cert" not found Jan 27 15:23:24 crc kubenswrapper[4772]: E0127 15:23:24.408228 4772 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 27 15:23:24 crc kubenswrapper[4772]: E0127 15:23:24.408549 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-metrics-certs podName:8087d8d3-d2f6-4bca-abec-f5b5335f26fa nodeName:}" failed. No retries permitted until 2026-01-27 15:23:26.408539478 +0000 UTC m=+992.389148576 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-metrics-certs") pod "openstack-operator-controller-manager-ff554fc88-clt4p" (UID: "8087d8d3-d2f6-4bca-abec-f5b5335f26fa") : secret "metrics-server-cert" not found Jan 27 15:23:25 crc kubenswrapper[4772]: E0127 15:23:25.183860 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:445e951df2f21df6d33a466f75917e0f6103052ae751ae11887136e8ab165922\\\"\"" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-l8d48" podUID="c09741c3-6bae-487a-9b4c-7c9f01d8c5bf" Jan 27 15:23:25 crc kubenswrapper[4772]: E0127 15:23:25.184879 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:fa46fc14710961e6b4a76a3522dca3aa3cfa71436c7cf7ade533d3712822f327\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-ww79v" podUID="e76712a7-ebf6-4f04-a52c-c8d2bacb87f7" Jan 27 15:23:25 crc kubenswrapper[4772]: E0127 15:23:25.184927 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/telemetry-operator@sha256:1f1fea3b7df89b81756eab8e6f4c9bed01ab7e949a6ce2d7692c260f41dfbc20\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-799bc87c89-k2l8k" podUID="e4a99865-64a7-49e5-bdce-ff929105fc0d" Jan 27 15:23:25 crc kubenswrapper[4772]: E0127 15:23:25.185000 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:c8dde42dafd41026ed2e4cfc26efc0fff63c4ba9d31326ae7dc644ccceaafa9d\\\"\"" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-ln7xf" podUID="6242683c-24ad-4e22-a7b3-8463e07388c2" Jan 27 15:23:25 crc kubenswrapper[4772]: E0127 15:23:25.185374 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/watcher-operator@sha256:162fb83ed76cbf5d44ba057fbeee02a9182fdf02346afadb3e16b2e3627e1940\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6c9bb4b66c-ws2mh" podUID="783285f4-2e9d-4af5-b017-32676e7d1b01" Jan 27 15:23:25 crc kubenswrapper[4772]: E0127 15:23:25.185603 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:013c0ad82d21a21c7eece5cd4b5d5c4b8eb410b6671ac33a6f3fb78c8510811d\\\"\"" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-vwnwk" podUID="c3d2f06d-4dd3-49a8-a0a0-54a83cc3f4e8" Jan 27 15:23:25 crc kubenswrapper[4772]: I0127 15:23:25.843531 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e7465bd0-3b6e-4199-9ee6-28b512198847-cert\") pod \"infra-operator-controller-manager-7d75bc88d5-t54fr\" (UID: \"e7465bd0-3b6e-4199-9ee6-28b512198847\") " pod="openstack-operators/infra-operator-controller-manager-7d75bc88d5-t54fr" Jan 27 15:23:25 crc kubenswrapper[4772]: E0127 15:23:25.843676 4772 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 27 15:23:25 crc kubenswrapper[4772]: E0127 15:23:25.843723 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e7465bd0-3b6e-4199-9ee6-28b512198847-cert podName:e7465bd0-3b6e-4199-9ee6-28b512198847 nodeName:}" failed. No retries permitted until 2026-01-27 15:23:29.843708996 +0000 UTC m=+995.824318094 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e7465bd0-3b6e-4199-9ee6-28b512198847-cert") pod "infra-operator-controller-manager-7d75bc88d5-t54fr" (UID: "e7465bd0-3b6e-4199-9ee6-28b512198847") : secret "infra-operator-webhook-server-cert" not found Jan 27 15:23:26 crc kubenswrapper[4772]: I0127 15:23:26.249767 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1389813b-42ea-433f-820c-e5b8b41713d7-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854q994c\" (UID: \"1389813b-42ea-433f-820c-e5b8b41713d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854q994c" Jan 27 15:23:26 crc kubenswrapper[4772]: E0127 15:23:26.249948 4772 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 27 15:23:26 crc kubenswrapper[4772]: E0127 15:23:26.250644 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1389813b-42ea-433f-820c-e5b8b41713d7-cert podName:1389813b-42ea-433f-820c-e5b8b41713d7 nodeName:}" failed. No retries permitted until 2026-01-27 15:23:30.250194235 +0000 UTC m=+996.230803333 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1389813b-42ea-433f-820c-e5b8b41713d7-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854q994c" (UID: "1389813b-42ea-433f-820c-e5b8b41713d7") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 27 15:23:26 crc kubenswrapper[4772]: I0127 15:23:26.453139 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-metrics-certs\") pod \"openstack-operator-controller-manager-ff554fc88-clt4p\" (UID: \"8087d8d3-d2f6-4bca-abec-f5b5335f26fa\") " pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:23:26 crc kubenswrapper[4772]: I0127 15:23:26.453276 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-webhook-certs\") pod \"openstack-operator-controller-manager-ff554fc88-clt4p\" (UID: \"8087d8d3-d2f6-4bca-abec-f5b5335f26fa\") " pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:23:26 crc kubenswrapper[4772]: E0127 15:23:26.453305 4772 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 27 15:23:26 crc kubenswrapper[4772]: E0127 15:23:26.453379 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-metrics-certs podName:8087d8d3-d2f6-4bca-abec-f5b5335f26fa nodeName:}" failed. No retries permitted until 2026-01-27 15:23:30.453359077 +0000 UTC m=+996.433968175 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-metrics-certs") pod "openstack-operator-controller-manager-ff554fc88-clt4p" (UID: "8087d8d3-d2f6-4bca-abec-f5b5335f26fa") : secret "metrics-server-cert" not found Jan 27 15:23:26 crc kubenswrapper[4772]: E0127 15:23:26.453413 4772 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 27 15:23:26 crc kubenswrapper[4772]: E0127 15:23:26.453473 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-webhook-certs podName:8087d8d3-d2f6-4bca-abec-f5b5335f26fa nodeName:}" failed. No retries permitted until 2026-01-27 15:23:30.45345915 +0000 UTC m=+996.434068248 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-webhook-certs") pod "openstack-operator-controller-manager-ff554fc88-clt4p" (UID: "8087d8d3-d2f6-4bca-abec-f5b5335f26fa") : secret "webhook-server-cert" not found Jan 27 15:23:29 crc kubenswrapper[4772]: I0127 15:23:29.903501 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e7465bd0-3b6e-4199-9ee6-28b512198847-cert\") pod \"infra-operator-controller-manager-7d75bc88d5-t54fr\" (UID: \"e7465bd0-3b6e-4199-9ee6-28b512198847\") " pod="openstack-operators/infra-operator-controller-manager-7d75bc88d5-t54fr" Jan 27 15:23:29 crc kubenswrapper[4772]: E0127 15:23:29.903934 4772 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 27 15:23:29 crc kubenswrapper[4772]: E0127 15:23:29.904082 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e7465bd0-3b6e-4199-9ee6-28b512198847-cert podName:e7465bd0-3b6e-4199-9ee6-28b512198847 nodeName:}" failed. No retries permitted until 2026-01-27 15:23:37.904062851 +0000 UTC m=+1003.884671959 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e7465bd0-3b6e-4199-9ee6-28b512198847-cert") pod "infra-operator-controller-manager-7d75bc88d5-t54fr" (UID: "e7465bd0-3b6e-4199-9ee6-28b512198847") : secret "infra-operator-webhook-server-cert" not found Jan 27 15:23:30 crc kubenswrapper[4772]: I0127 15:23:30.308051 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1389813b-42ea-433f-820c-e5b8b41713d7-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854q994c\" (UID: \"1389813b-42ea-433f-820c-e5b8b41713d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854q994c" Jan 27 15:23:30 crc kubenswrapper[4772]: E0127 15:23:30.308560 4772 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 27 15:23:30 crc kubenswrapper[4772]: E0127 15:23:30.308781 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1389813b-42ea-433f-820c-e5b8b41713d7-cert podName:1389813b-42ea-433f-820c-e5b8b41713d7 nodeName:}" failed. No retries permitted until 2026-01-27 15:23:38.308753177 +0000 UTC m=+1004.289362295 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1389813b-42ea-433f-820c-e5b8b41713d7-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854q994c" (UID: "1389813b-42ea-433f-820c-e5b8b41713d7") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 27 15:23:30 crc kubenswrapper[4772]: I0127 15:23:30.512711 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-webhook-certs\") pod \"openstack-operator-controller-manager-ff554fc88-clt4p\" (UID: \"8087d8d3-d2f6-4bca-abec-f5b5335f26fa\") " pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:23:30 crc kubenswrapper[4772]: I0127 15:23:30.512856 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-metrics-certs\") pod \"openstack-operator-controller-manager-ff554fc88-clt4p\" (UID: \"8087d8d3-d2f6-4bca-abec-f5b5335f26fa\") " pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:23:30 crc kubenswrapper[4772]: E0127 15:23:30.512901 4772 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 27 15:23:30 crc kubenswrapper[4772]: E0127 15:23:30.512978 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-webhook-certs podName:8087d8d3-d2f6-4bca-abec-f5b5335f26fa nodeName:}" failed. No retries permitted until 2026-01-27 15:23:38.512961229 +0000 UTC m=+1004.493570327 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-webhook-certs") pod "openstack-operator-controller-manager-ff554fc88-clt4p" (UID: "8087d8d3-d2f6-4bca-abec-f5b5335f26fa") : secret "webhook-server-cert" not found Jan 27 15:23:30 crc kubenswrapper[4772]: E0127 15:23:30.513095 4772 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 27 15:23:30 crc kubenswrapper[4772]: E0127 15:23:30.513146 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-metrics-certs podName:8087d8d3-d2f6-4bca-abec-f5b5335f26fa nodeName:}" failed. No retries permitted until 2026-01-27 15:23:38.513131184 +0000 UTC m=+1004.493740282 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-metrics-certs") pod "openstack-operator-controller-manager-ff554fc88-clt4p" (UID: "8087d8d3-d2f6-4bca-abec-f5b5335f26fa") : secret "metrics-server-cert" not found Jan 27 15:23:35 crc kubenswrapper[4772]: I0127 15:23:35.214550 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zvzxk"] Jan 27 15:23:35 crc kubenswrapper[4772]: I0127 15:23:35.216605 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zvzxk" Jan 27 15:23:35 crc kubenswrapper[4772]: I0127 15:23:35.228250 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zvzxk"] Jan 27 15:23:35 crc kubenswrapper[4772]: I0127 15:23:35.391156 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f2e732b-dbc8-423e-8a37-100a97dad4f0-utilities\") pod \"community-operators-zvzxk\" (UID: \"0f2e732b-dbc8-423e-8a37-100a97dad4f0\") " pod="openshift-marketplace/community-operators-zvzxk" Jan 27 15:23:35 crc kubenswrapper[4772]: I0127 15:23:35.391383 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f2e732b-dbc8-423e-8a37-100a97dad4f0-catalog-content\") pod \"community-operators-zvzxk\" (UID: \"0f2e732b-dbc8-423e-8a37-100a97dad4f0\") " pod="openshift-marketplace/community-operators-zvzxk" Jan 27 15:23:35 crc kubenswrapper[4772]: I0127 15:23:35.391590 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29mzp\" (UniqueName: \"kubernetes.io/projected/0f2e732b-dbc8-423e-8a37-100a97dad4f0-kube-api-access-29mzp\") pod \"community-operators-zvzxk\" (UID: \"0f2e732b-dbc8-423e-8a37-100a97dad4f0\") " pod="openshift-marketplace/community-operators-zvzxk" Jan 27 15:23:35 crc kubenswrapper[4772]: I0127 15:23:35.492806 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29mzp\" (UniqueName: \"kubernetes.io/projected/0f2e732b-dbc8-423e-8a37-100a97dad4f0-kube-api-access-29mzp\") pod \"community-operators-zvzxk\" (UID: \"0f2e732b-dbc8-423e-8a37-100a97dad4f0\") " pod="openshift-marketplace/community-operators-zvzxk" Jan 27 15:23:35 crc kubenswrapper[4772]: I0127 15:23:35.492891 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f2e732b-dbc8-423e-8a37-100a97dad4f0-utilities\") pod \"community-operators-zvzxk\" (UID: \"0f2e732b-dbc8-423e-8a37-100a97dad4f0\") " pod="openshift-marketplace/community-operators-zvzxk" Jan 27 15:23:35 crc kubenswrapper[4772]: I0127 15:23:35.492962 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f2e732b-dbc8-423e-8a37-100a97dad4f0-catalog-content\") pod \"community-operators-zvzxk\" (UID: \"0f2e732b-dbc8-423e-8a37-100a97dad4f0\") " pod="openshift-marketplace/community-operators-zvzxk" Jan 27 15:23:35 crc kubenswrapper[4772]: I0127 15:23:35.493525 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f2e732b-dbc8-423e-8a37-100a97dad4f0-catalog-content\") pod \"community-operators-zvzxk\" (UID: \"0f2e732b-dbc8-423e-8a37-100a97dad4f0\") " pod="openshift-marketplace/community-operators-zvzxk" Jan 27 15:23:35 crc kubenswrapper[4772]: I0127 15:23:35.493596 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f2e732b-dbc8-423e-8a37-100a97dad4f0-utilities\") pod \"community-operators-zvzxk\" (UID: \"0f2e732b-dbc8-423e-8a37-100a97dad4f0\") " pod="openshift-marketplace/community-operators-zvzxk" Jan 27 15:23:35 crc kubenswrapper[4772]: I0127 15:23:35.533300 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29mzp\" (UniqueName: \"kubernetes.io/projected/0f2e732b-dbc8-423e-8a37-100a97dad4f0-kube-api-access-29mzp\") pod \"community-operators-zvzxk\" (UID: \"0f2e732b-dbc8-423e-8a37-100a97dad4f0\") " pod="openshift-marketplace/community-operators-zvzxk" Jan 27 15:23:35 crc kubenswrapper[4772]: I0127 15:23:35.575270 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zvzxk" Jan 27 15:23:37 crc kubenswrapper[4772]: E0127 15:23:37.393777 4772 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/lmiccini/glance-operator@sha256:bc45409dff26aca6bd982684cfaf093548adb6a71928f5257fe60ab5535dda39" Jan 27 15:23:37 crc kubenswrapper[4772]: E0127 15:23:37.394481 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/lmiccini/glance-operator@sha256:bc45409dff26aca6bd982684cfaf093548adb6a71928f5257fe60ab5535dda39,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-snjkj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-67dd55ff59-hgscb_openstack-operators(fb300814-fca7-4419-ac6e-c08b33edd4be): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 27 15:23:37 crc kubenswrapper[4772]: E0127 15:23:37.395864 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-67dd55ff59-hgscb" podUID="fb300814-fca7-4419-ac6e-c08b33edd4be" Jan 27 15:23:37 crc kubenswrapper[4772]: E0127 15:23:37.910779 4772 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/lmiccini/ironic-operator@sha256:30e2224475338d3a02d617ae147dc7dc09867cce4ac3543b313a1923c46299fa" Jan 27 15:23:37 crc kubenswrapper[4772]: E0127 15:23:37.910990 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/lmiccini/ironic-operator@sha256:30e2224475338d3a02d617ae147dc7dc09867cce4ac3543b313a1923c46299fa,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zs77k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-768b776ffb-sxbjn_openstack-operators(2df85221-33ed-49be-949c-516810279e4d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 27 15:23:37 crc kubenswrapper[4772]: E0127 15:23:37.912226 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-768b776ffb-sxbjn" podUID="2df85221-33ed-49be-949c-516810279e4d" Jan 27 15:23:37 crc kubenswrapper[4772]: I0127 15:23:37.933180 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e7465bd0-3b6e-4199-9ee6-28b512198847-cert\") pod \"infra-operator-controller-manager-7d75bc88d5-t54fr\" (UID: \"e7465bd0-3b6e-4199-9ee6-28b512198847\") " pod="openstack-operators/infra-operator-controller-manager-7d75bc88d5-t54fr" Jan 27 15:23:37 crc kubenswrapper[4772]: E0127 15:23:37.933353 4772 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 27 15:23:37 crc kubenswrapper[4772]: E0127 15:23:37.933434 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e7465bd0-3b6e-4199-9ee6-28b512198847-cert podName:e7465bd0-3b6e-4199-9ee6-28b512198847 nodeName:}" failed. No retries permitted until 2026-01-27 15:23:53.933414493 +0000 UTC m=+1019.914023591 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e7465bd0-3b6e-4199-9ee6-28b512198847-cert") pod "infra-operator-controller-manager-7d75bc88d5-t54fr" (UID: "e7465bd0-3b6e-4199-9ee6-28b512198847") : secret "infra-operator-webhook-server-cert" not found Jan 27 15:23:38 crc kubenswrapper[4772]: E0127 15:23:38.287597 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/ironic-operator@sha256:30e2224475338d3a02d617ae147dc7dc09867cce4ac3543b313a1923c46299fa\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-768b776ffb-sxbjn" podUID="2df85221-33ed-49be-949c-516810279e4d" Jan 27 15:23:38 crc kubenswrapper[4772]: E0127 15:23:38.288052 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/glance-operator@sha256:bc45409dff26aca6bd982684cfaf093548adb6a71928f5257fe60ab5535dda39\\\"\"" pod="openstack-operators/glance-operator-controller-manager-67dd55ff59-hgscb" podUID="fb300814-fca7-4419-ac6e-c08b33edd4be" Jan 27 15:23:38 crc kubenswrapper[4772]: I0127 15:23:38.338757 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1389813b-42ea-433f-820c-e5b8b41713d7-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854q994c\" (UID: \"1389813b-42ea-433f-820c-e5b8b41713d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854q994c" Jan 27 15:23:38 crc kubenswrapper[4772]: E0127 15:23:38.338970 4772 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 27 15:23:38 crc kubenswrapper[4772]: E0127 15:23:38.339053 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1389813b-42ea-433f-820c-e5b8b41713d7-cert podName:1389813b-42ea-433f-820c-e5b8b41713d7 nodeName:}" failed. No retries permitted until 2026-01-27 15:23:54.339031986 +0000 UTC m=+1020.319641124 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1389813b-42ea-433f-820c-e5b8b41713d7-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854q994c" (UID: "1389813b-42ea-433f-820c-e5b8b41713d7") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 27 15:23:38 crc kubenswrapper[4772]: I0127 15:23:38.542127 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-webhook-certs\") pod \"openstack-operator-controller-manager-ff554fc88-clt4p\" (UID: \"8087d8d3-d2f6-4bca-abec-f5b5335f26fa\") " pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:23:38 crc kubenswrapper[4772]: I0127 15:23:38.542227 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-metrics-certs\") pod \"openstack-operator-controller-manager-ff554fc88-clt4p\" (UID: \"8087d8d3-d2f6-4bca-abec-f5b5335f26fa\") " pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:23:38 crc kubenswrapper[4772]: E0127 15:23:38.542399 4772 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 27 15:23:38 crc kubenswrapper[4772]: E0127 15:23:38.542447 4772 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 27 15:23:38 crc kubenswrapper[4772]: E0127 15:23:38.542464 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-metrics-certs podName:8087d8d3-d2f6-4bca-abec-f5b5335f26fa nodeName:}" failed. No retries permitted until 2026-01-27 15:23:54.542447266 +0000 UTC m=+1020.523056364 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-metrics-certs") pod "openstack-operator-controller-manager-ff554fc88-clt4p" (UID: "8087d8d3-d2f6-4bca-abec-f5b5335f26fa") : secret "metrics-server-cert" not found Jan 27 15:23:38 crc kubenswrapper[4772]: E0127 15:23:38.542537 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-webhook-certs podName:8087d8d3-d2f6-4bca-abec-f5b5335f26fa nodeName:}" failed. No retries permitted until 2026-01-27 15:23:54.542517988 +0000 UTC m=+1020.523127086 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-webhook-certs") pod "openstack-operator-controller-manager-ff554fc88-clt4p" (UID: "8087d8d3-d2f6-4bca-abec-f5b5335f26fa") : secret "webhook-server-cert" not found Jan 27 15:23:39 crc kubenswrapper[4772]: E0127 15:23:39.585718 4772 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/lmiccini/designate-operator@sha256:d26a32730ba8b64e98f68194bd1a766aadc942392b24fa6a2cf1c136969dd99f" Jan 27 15:23:39 crc kubenswrapper[4772]: E0127 15:23:39.585930 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/lmiccini/designate-operator@sha256:d26a32730ba8b64e98f68194bd1a766aadc942392b24fa6a2cf1c136969dd99f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4xv2c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-77554cdc5c-tkr6j_openstack-operators(d395f105-54f0-4497-a119-57802be313a3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 27 15:23:39 crc kubenswrapper[4772]: E0127 15:23:39.599204 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-77554cdc5c-tkr6j" podUID="d395f105-54f0-4497-a119-57802be313a3" Jan 27 15:23:40 crc kubenswrapper[4772]: E0127 15:23:40.082940 4772 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:b673f00227298dcfa89abb46f8296a0825add42da41e8a4bf4dd13367c738d84" Jan 27 15:23:40 crc kubenswrapper[4772]: E0127 15:23:40.083132 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:b673f00227298dcfa89abb46f8296a0825add42da41e8a4bf4dd13367c738d84,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t8n6m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-6b9fb5fdcb-zhd82_openstack-operators(b73c175a-e89e-434f-996a-65c1140bb8dd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 27 15:23:40 crc kubenswrapper[4772]: E0127 15:23:40.084344 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-zhd82" podUID="b73c175a-e89e-434f-996a-65c1140bb8dd" Jan 27 15:23:40 crc kubenswrapper[4772]: E0127 15:23:40.299184 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:b673f00227298dcfa89abb46f8296a0825add42da41e8a4bf4dd13367c738d84\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-zhd82" podUID="b73c175a-e89e-434f-996a-65c1140bb8dd" Jan 27 15:23:40 crc kubenswrapper[4772]: E0127 15:23:40.301094 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/designate-operator@sha256:d26a32730ba8b64e98f68194bd1a766aadc942392b24fa6a2cf1c136969dd99f\\\"\"" pod="openstack-operators/designate-operator-controller-manager-77554cdc5c-tkr6j" podUID="d395f105-54f0-4497-a119-57802be313a3" Jan 27 15:23:41 crc kubenswrapper[4772]: E0127 15:23:41.582661 4772 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/lmiccini/keystone-operator@sha256:008a2e338430e7dd513f81f66320cc5c1332c332a3191b537d75786489d7f487" Jan 27 15:23:41 crc kubenswrapper[4772]: E0127 15:23:41.583127 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/lmiccini/keystone-operator@sha256:008a2e338430e7dd513f81f66320cc5c1332c332a3191b537d75786489d7f487,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dnnt7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-55f684fd56-wzjrz_openstack-operators(783d8159-e67a-4796-83d8-4eff27d79505): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 27 15:23:41 crc kubenswrapper[4772]: E0127 15:23:41.584333 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-55f684fd56-wzjrz" podUID="783d8159-e67a-4796-83d8-4eff27d79505" Jan 27 15:23:42 crc kubenswrapper[4772]: E0127 15:23:42.331463 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/keystone-operator@sha256:008a2e338430e7dd513f81f66320cc5c1332c332a3191b537d75786489d7f487\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-55f684fd56-wzjrz" podUID="783d8159-e67a-4796-83d8-4eff27d79505" Jan 27 15:23:43 crc kubenswrapper[4772]: E0127 15:23:43.574359 4772 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Jan 27 15:23:43 crc kubenswrapper[4772]: E0127 15:23:43.574538 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t57ck,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-h9297_openstack-operators(abaf1142-1b7c-4987-8a9d-c91e6456c4a5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 27 15:23:43 crc kubenswrapper[4772]: E0127 15:23:43.575677 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h9297" podUID="abaf1142-1b7c-4987-8a9d-c91e6456c4a5" Jan 27 15:23:44 crc kubenswrapper[4772]: E0127 15:23:44.018874 4772 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/lmiccini/nova-operator@sha256:a992613466db3478a00c20c28639c4a12f6326aa52c40a418d1ec40038c83b61" Jan 27 15:23:44 crc kubenswrapper[4772]: E0127 15:23:44.019068 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/lmiccini/nova-operator@sha256:a992613466db3478a00c20c28639c4a12f6326aa52c40a418d1ec40038c83b61,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gbb4j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-ddcbfd695-6wltn_openstack-operators(e7fc5297-101a-496e-a7c6-e7296e08a5af): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 27 15:23:44 crc kubenswrapper[4772]: E0127 15:23:44.020463 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-ddcbfd695-6wltn" podUID="e7fc5297-101a-496e-a7c6-e7296e08a5af" Jan 27 15:23:44 crc kubenswrapper[4772]: E0127 15:23:44.342505 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/nova-operator@sha256:a992613466db3478a00c20c28639c4a12f6326aa52c40a418d1ec40038c83b61\\\"\"" pod="openstack-operators/nova-operator-controller-manager-ddcbfd695-6wltn" podUID="e7fc5297-101a-496e-a7c6-e7296e08a5af" Jan 27 15:23:44 crc kubenswrapper[4772]: E0127 15:23:44.344678 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h9297" podUID="abaf1142-1b7c-4987-8a9d-c91e6456c4a5" Jan 27 15:23:44 crc kubenswrapper[4772]: I0127 15:23:44.405847 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zvzxk"] Jan 27 15:23:45 crc kubenswrapper[4772]: W0127 15:23:45.544589 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f2e732b_dbc8_423e_8a37_100a97dad4f0.slice/crio-0a190d2f48d8673469fb8f939468a4f2e791b7ede4b24560514885c4a4400b46 WatchSource:0}: Error finding container 0a190d2f48d8673469fb8f939468a4f2e791b7ede4b24560514885c4a4400b46: Status 404 returned error can't find the container with id 0a190d2f48d8673469fb8f939468a4f2e791b7ede4b24560514885c4a4400b46 Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.369605 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7ffd8d76d4-gcpj4" event={"ID":"0a88aa66-b634-44ee-8e5b-bfeacb765e57","Type":"ContainerStarted","Data":"61c1a5d21ba7cf1ceb1dae1e67a1af2da7cfe582c92cd7a092dab87ddc196a3e"} Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.369890 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-7ffd8d76d4-gcpj4" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.371467 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-l8d48" event={"ID":"c09741c3-6bae-487a-9b4c-7c9f01d8c5bf","Type":"ContainerStarted","Data":"c228aa40cd004f4965b5cdfda9f435ff3667ece11ee27b9b4ff3a639d6972431"} Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.371610 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-l8d48" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.372668 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-655bf9cfbb-cgh7j" event={"ID":"fde95124-892b-411a-ba05-fa70927c8838","Type":"ContainerStarted","Data":"2ddf9e7acdd771dd121c6f9d9dd9eee33de1eb9a1f7294bef8ef1c515b49ac9c"} Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.372735 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-655bf9cfbb-cgh7j" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.373963 4772 generic.go:334] "Generic (PLEG): container finished" podID="0f2e732b-dbc8-423e-8a37-100a97dad4f0" containerID="8b01102f5cfb35617b002369451b43259452694e6501916f5c0f4f257352eaee" exitCode=0 Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.373995 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zvzxk" event={"ID":"0f2e732b-dbc8-423e-8a37-100a97dad4f0","Type":"ContainerDied","Data":"8b01102f5cfb35617b002369451b43259452694e6501916f5c0f4f257352eaee"} Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.374023 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zvzxk" event={"ID":"0f2e732b-dbc8-423e-8a37-100a97dad4f0","Type":"ContainerStarted","Data":"0a190d2f48d8673469fb8f939468a4f2e791b7ede4b24560514885c4a4400b46"} Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.375276 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c9bb4b66c-ws2mh" event={"ID":"783285f4-2e9d-4af5-b017-32676e7d1b01","Type":"ContainerStarted","Data":"822d53489c2003e235600c078a444653771377e0334bc7cc3aedeac351b843f4"} Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.375396 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6c9bb4b66c-ws2mh" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.376977 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-ln7xf" event={"ID":"6242683c-24ad-4e22-a7b3-8463e07388c2","Type":"ContainerStarted","Data":"f00bbc75046fbf41d8ec762fd40a7980cd00ad96650abfc4ac7521f91bdee7b4"} Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.377116 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-ln7xf" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.378301 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-575ffb885b-mtd9d" event={"ID":"4c63a702-50b9-42f3-858e-7e27da0a8d8f","Type":"ContainerStarted","Data":"ad20cd623ff931fc39aedbefd9e6d702c1605209542c84765b9843c8623057c5"} Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.378413 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-575ffb885b-mtd9d" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.379526 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-ww79v" event={"ID":"e76712a7-ebf6-4f04-a52c-c8d2bacb87f7","Type":"ContainerStarted","Data":"4f465583215e8b9affb764b7283ff1474fcd8e3110d6c381e44d1b80d0cbfa4c"} Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.379894 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-ww79v" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.381208 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7875d7675-ktfbt" event={"ID":"80584c24-3c75-4624-802f-e608f640eeaa","Type":"ContainerStarted","Data":"407a631765036d42cd7fa2e38f2ab5d80a6360a56ce7e0782be0431d4c1e0500"} Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.381542 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7875d7675-ktfbt" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.382905 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-799bc87c89-k2l8k" event={"ID":"e4a99865-64a7-49e5-bdce-ff929105fc0d","Type":"ContainerStarted","Data":"e1c3b7fb626f8b787658cc8869351224487fc8c81ad9666b3ac00319adb88188"} Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.383254 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-799bc87c89-k2l8k" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.384264 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-65ff799cfd-t42n9" event={"ID":"674f4da6-f50d-4bab-808d-56ab3b9e2cb4","Type":"ContainerStarted","Data":"c9918d5f6cf717de51ccdc59db2e1329ce487f5634a8a46443e865ca86ced277"} Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.384604 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-65ff799cfd-t42n9" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.386581 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-849fcfbb6b-tvrx9" event={"ID":"27ec5082-c170-465b-b3a3-1f27a545fd71","Type":"ContainerStarted","Data":"917d957605680558b921d84fee15cf764cfec3333e2ce31f348d64b0a4296781"} Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.386954 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-849fcfbb6b-tvrx9" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.387955 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-vwnwk" event={"ID":"c3d2f06d-4dd3-49a8-a0a0-54a83cc3f4e8","Type":"ContainerStarted","Data":"82ce2efa073bd7645b0a3f77e0b34b013a044b94a5f60b09cc16f705a749a7d7"} Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.388360 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-vwnwk" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.389477 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-jcb4p" event={"ID":"e85aef3a-e235-473c-94cc-1f6237798b3e","Type":"ContainerStarted","Data":"281f7228e075798fa1e0b5437241536e28c11266d9ca7a4a73d0ca4d9a81eff6"} Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.389806 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-jcb4p" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.408824 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-7ffd8d76d4-gcpj4" podStartSLOduration=4.9120735159999995 podStartE2EDuration="24.408808832s" podCreationTimestamp="2026-01-27 15:23:22 +0000 UTC" firstStartedPulling="2026-01-27 15:23:23.644001303 +0000 UTC m=+989.624610401" lastFinishedPulling="2026-01-27 15:23:43.140736629 +0000 UTC m=+1009.121345717" observedRunningTime="2026-01-27 15:23:46.388755665 +0000 UTC m=+1012.369364753" watchObservedRunningTime="2026-01-27 15:23:46.408808832 +0000 UTC m=+1012.389417930" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.476251 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-655bf9cfbb-cgh7j" podStartSLOduration=5.607960751 podStartE2EDuration="25.476235361s" podCreationTimestamp="2026-01-27 15:23:21 +0000 UTC" firstStartedPulling="2026-01-27 15:23:23.273490369 +0000 UTC m=+989.254099467" lastFinishedPulling="2026-01-27 15:23:43.141764959 +0000 UTC m=+1009.122374077" observedRunningTime="2026-01-27 15:23:46.474860431 +0000 UTC m=+1012.455469529" watchObservedRunningTime="2026-01-27 15:23:46.476235361 +0000 UTC m=+1012.456844449" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.478620 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-849fcfbb6b-tvrx9" podStartSLOduration=4.5739294919999995 podStartE2EDuration="24.478609409s" podCreationTimestamp="2026-01-27 15:23:22 +0000 UTC" firstStartedPulling="2026-01-27 15:23:23.237508554 +0000 UTC m=+989.218117652" lastFinishedPulling="2026-01-27 15:23:43.142188471 +0000 UTC m=+1009.122797569" observedRunningTime="2026-01-27 15:23:46.453541508 +0000 UTC m=+1012.434150606" watchObservedRunningTime="2026-01-27 15:23:46.478609409 +0000 UTC m=+1012.459218497" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.533795 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-799bc87c89-k2l8k" podStartSLOduration=2.779634278 podStartE2EDuration="24.533777325s" podCreationTimestamp="2026-01-27 15:23:22 +0000 UTC" firstStartedPulling="2026-01-27 15:23:23.870489526 +0000 UTC m=+989.851098624" lastFinishedPulling="2026-01-27 15:23:45.624632573 +0000 UTC m=+1011.605241671" observedRunningTime="2026-01-27 15:23:46.528152454 +0000 UTC m=+1012.508761552" watchObservedRunningTime="2026-01-27 15:23:46.533777325 +0000 UTC m=+1012.514386423" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.564776 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-l8d48" podStartSLOduration=2.743378806 podStartE2EDuration="24.564761826s" podCreationTimestamp="2026-01-27 15:23:22 +0000 UTC" firstStartedPulling="2026-01-27 15:23:23.768355339 +0000 UTC m=+989.748964437" lastFinishedPulling="2026-01-27 15:23:45.589738359 +0000 UTC m=+1011.570347457" observedRunningTime="2026-01-27 15:23:46.559620739 +0000 UTC m=+1012.540229837" watchObservedRunningTime="2026-01-27 15:23:46.564761826 +0000 UTC m=+1012.545370924" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.581793 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6c9bb4b66c-ws2mh" podStartSLOduration=2.756173785 podStartE2EDuration="24.581777636s" podCreationTimestamp="2026-01-27 15:23:22 +0000 UTC" firstStartedPulling="2026-01-27 15:23:23.828021705 +0000 UTC m=+989.808630823" lastFinishedPulling="2026-01-27 15:23:45.653625576 +0000 UTC m=+1011.634234674" observedRunningTime="2026-01-27 15:23:46.576415362 +0000 UTC m=+1012.557024450" watchObservedRunningTime="2026-01-27 15:23:46.581777636 +0000 UTC m=+1012.562386734" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.602800 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7875d7675-ktfbt" podStartSLOduration=5.106107756 podStartE2EDuration="24.60278324s" podCreationTimestamp="2026-01-27 15:23:22 +0000 UTC" firstStartedPulling="2026-01-27 15:23:23.644444366 +0000 UTC m=+989.625053474" lastFinishedPulling="2026-01-27 15:23:43.14111986 +0000 UTC m=+1009.121728958" observedRunningTime="2026-01-27 15:23:46.5972263 +0000 UTC m=+1012.577835398" watchObservedRunningTime="2026-01-27 15:23:46.60278324 +0000 UTC m=+1012.583392338" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.655342 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-ww79v" podStartSLOduration=2.7231234840000003 podStartE2EDuration="24.655328061s" podCreationTimestamp="2026-01-27 15:23:22 +0000 UTC" firstStartedPulling="2026-01-27 15:23:23.761383409 +0000 UTC m=+989.741992507" lastFinishedPulling="2026-01-27 15:23:45.693587986 +0000 UTC m=+1011.674197084" observedRunningTime="2026-01-27 15:23:46.618843072 +0000 UTC m=+1012.599452170" watchObservedRunningTime="2026-01-27 15:23:46.655328061 +0000 UTC m=+1012.635937159" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.656861 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-vwnwk" podStartSLOduration=2.72923875 podStartE2EDuration="24.656856215s" podCreationTimestamp="2026-01-27 15:23:22 +0000 UTC" firstStartedPulling="2026-01-27 15:23:23.753976736 +0000 UTC m=+989.734585834" lastFinishedPulling="2026-01-27 15:23:45.681594201 +0000 UTC m=+1011.662203299" observedRunningTime="2026-01-27 15:23:46.650005348 +0000 UTC m=+1012.630614446" watchObservedRunningTime="2026-01-27 15:23:46.656856215 +0000 UTC m=+1012.637465313" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.671361 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-ln7xf" podStartSLOduration=2.893819282 podStartE2EDuration="24.671343111s" podCreationTimestamp="2026-01-27 15:23:22 +0000 UTC" firstStartedPulling="2026-01-27 15:23:23.867438728 +0000 UTC m=+989.848047836" lastFinishedPulling="2026-01-27 15:23:45.644962567 +0000 UTC m=+1011.625571665" observedRunningTime="2026-01-27 15:23:46.667506121 +0000 UTC m=+1012.648115219" watchObservedRunningTime="2026-01-27 15:23:46.671343111 +0000 UTC m=+1012.651952209" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.686221 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-65ff799cfd-t42n9" podStartSLOduration=5.401818595 podStartE2EDuration="25.686205589s" podCreationTimestamp="2026-01-27 15:23:21 +0000 UTC" firstStartedPulling="2026-01-27 15:23:22.856336515 +0000 UTC m=+988.836945613" lastFinishedPulling="2026-01-27 15:23:43.140723509 +0000 UTC m=+1009.121332607" observedRunningTime="2026-01-27 15:23:46.681860904 +0000 UTC m=+1012.662470002" watchObservedRunningTime="2026-01-27 15:23:46.686205589 +0000 UTC m=+1012.666814687" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.711325 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-575ffb885b-mtd9d" podStartSLOduration=6.010916579 podStartE2EDuration="25.711310421s" podCreationTimestamp="2026-01-27 15:23:21 +0000 UTC" firstStartedPulling="2026-01-27 15:23:23.440838761 +0000 UTC m=+989.421447859" lastFinishedPulling="2026-01-27 15:23:43.141232603 +0000 UTC m=+1009.121841701" observedRunningTime="2026-01-27 15:23:46.709488638 +0000 UTC m=+1012.690097736" watchObservedRunningTime="2026-01-27 15:23:46.711310421 +0000 UTC m=+1012.691919519" Jan 27 15:23:46 crc kubenswrapper[4772]: I0127 15:23:46.728557 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-jcb4p" podStartSLOduration=6.096477309 podStartE2EDuration="25.728535636s" podCreationTimestamp="2026-01-27 15:23:21 +0000 UTC" firstStartedPulling="2026-01-27 15:23:23.50964998 +0000 UTC m=+989.490259078" lastFinishedPulling="2026-01-27 15:23:43.141708277 +0000 UTC m=+1009.122317405" observedRunningTime="2026-01-27 15:23:46.72381218 +0000 UTC m=+1012.704421278" watchObservedRunningTime="2026-01-27 15:23:46.728535636 +0000 UTC m=+1012.709144724" Jan 27 15:23:47 crc kubenswrapper[4772]: I0127 15:23:47.436870 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zvzxk" event={"ID":"0f2e732b-dbc8-423e-8a37-100a97dad4f0","Type":"ContainerStarted","Data":"bb5b9739a5d8e52b4be45c2313787179349baaa953432e2d37f9e46057ccb4e4"} Jan 27 15:23:48 crc kubenswrapper[4772]: I0127 15:23:48.444962 4772 generic.go:334] "Generic (PLEG): container finished" podID="0f2e732b-dbc8-423e-8a37-100a97dad4f0" containerID="bb5b9739a5d8e52b4be45c2313787179349baaa953432e2d37f9e46057ccb4e4" exitCode=0 Jan 27 15:23:48 crc kubenswrapper[4772]: I0127 15:23:48.445034 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zvzxk" event={"ID":"0f2e732b-dbc8-423e-8a37-100a97dad4f0","Type":"ContainerDied","Data":"bb5b9739a5d8e52b4be45c2313787179349baaa953432e2d37f9e46057ccb4e4"} Jan 27 15:23:49 crc kubenswrapper[4772]: I0127 15:23:49.452424 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-768b776ffb-sxbjn" event={"ID":"2df85221-33ed-49be-949c-516810279e4d","Type":"ContainerStarted","Data":"d23ba4f41b49c304e45cc62ff09182ed6204bc15d9389cc2650c71ceffe46b93"} Jan 27 15:23:49 crc kubenswrapper[4772]: I0127 15:23:49.452895 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-768b776ffb-sxbjn" Jan 27 15:23:49 crc kubenswrapper[4772]: I0127 15:23:49.453986 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zvzxk" event={"ID":"0f2e732b-dbc8-423e-8a37-100a97dad4f0","Type":"ContainerStarted","Data":"5f0e331299d6605e44535e575ebf4cf5cfa36757bfa41569f0da713738316ff3"} Jan 27 15:23:49 crc kubenswrapper[4772]: I0127 15:23:49.470700 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-768b776ffb-sxbjn" podStartSLOduration=2.823026481 podStartE2EDuration="28.470679415s" podCreationTimestamp="2026-01-27 15:23:21 +0000 UTC" firstStartedPulling="2026-01-27 15:23:23.429531486 +0000 UTC m=+989.410140574" lastFinishedPulling="2026-01-27 15:23:49.07718442 +0000 UTC m=+1015.057793508" observedRunningTime="2026-01-27 15:23:49.467955917 +0000 UTC m=+1015.448565025" watchObservedRunningTime="2026-01-27 15:23:49.470679415 +0000 UTC m=+1015.451288523" Jan 27 15:23:49 crc kubenswrapper[4772]: I0127 15:23:49.490103 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zvzxk" podStartSLOduration=12.029214242 podStartE2EDuration="14.490085213s" podCreationTimestamp="2026-01-27 15:23:35 +0000 UTC" firstStartedPulling="2026-01-27 15:23:46.374858156 +0000 UTC m=+1012.355467254" lastFinishedPulling="2026-01-27 15:23:48.835729117 +0000 UTC m=+1014.816338225" observedRunningTime="2026-01-27 15:23:49.484814772 +0000 UTC m=+1015.465423870" watchObservedRunningTime="2026-01-27 15:23:49.490085213 +0000 UTC m=+1015.470694311" Jan 27 15:23:52 crc kubenswrapper[4772]: I0127 15:23:52.092732 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-65ff799cfd-t42n9" Jan 27 15:23:52 crc kubenswrapper[4772]: I0127 15:23:52.121286 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-655bf9cfbb-cgh7j" Jan 27 15:23:52 crc kubenswrapper[4772]: I0127 15:23:52.200850 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-575ffb885b-mtd9d" Jan 27 15:23:52 crc kubenswrapper[4772]: I0127 15:23:52.331863 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-849fcfbb6b-tvrx9" Jan 27 15:23:52 crc kubenswrapper[4772]: I0127 15:23:52.525154 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-jcb4p" Jan 27 15:23:52 crc kubenswrapper[4772]: I0127 15:23:52.659985 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-7ffd8d76d4-gcpj4" Jan 27 15:23:52 crc kubenswrapper[4772]: I0127 15:23:52.820823 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7875d7675-ktfbt" Jan 27 15:23:52 crc kubenswrapper[4772]: I0127 15:23:52.962575 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-ww79v" Jan 27 15:23:52 crc kubenswrapper[4772]: I0127 15:23:52.992831 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-vwnwk" Jan 27 15:23:53 crc kubenswrapper[4772]: I0127 15:23:53.023211 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-l8d48" Jan 27 15:23:53 crc kubenswrapper[4772]: I0127 15:23:53.047419 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-ln7xf" Jan 27 15:23:53 crc kubenswrapper[4772]: I0127 15:23:53.109207 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-799bc87c89-k2l8k" Jan 27 15:23:53 crc kubenswrapper[4772]: I0127 15:23:53.166821 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6c9bb4b66c-ws2mh" Jan 27 15:23:53 crc kubenswrapper[4772]: I0127 15:23:53.487491 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55f684fd56-wzjrz" event={"ID":"783d8159-e67a-4796-83d8-4eff27d79505","Type":"ContainerStarted","Data":"eaecf2e53d15afef34bf272d64352d27aa2ae2b41f14db1894bdd3e36d5c565c"} Jan 27 15:23:53 crc kubenswrapper[4772]: I0127 15:23:53.488062 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-55f684fd56-wzjrz" Jan 27 15:23:53 crc kubenswrapper[4772]: I0127 15:23:53.489336 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-77554cdc5c-tkr6j" event={"ID":"d395f105-54f0-4497-a119-57802be313a3","Type":"ContainerStarted","Data":"cf7963476a36c8cfaba13cac3e2de907e23c7953a1ff59b8219b9b83cab2617c"} Jan 27 15:23:53 crc kubenswrapper[4772]: I0127 15:23:53.489675 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-77554cdc5c-tkr6j" Jan 27 15:23:53 crc kubenswrapper[4772]: I0127 15:23:53.526148 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-55f684fd56-wzjrz" podStartSLOduration=1.877660611 podStartE2EDuration="31.526128699s" podCreationTimestamp="2026-01-27 15:23:22 +0000 UTC" firstStartedPulling="2026-01-27 15:23:23.45923593 +0000 UTC m=+989.439845038" lastFinishedPulling="2026-01-27 15:23:53.107704028 +0000 UTC m=+1019.088313126" observedRunningTime="2026-01-27 15:23:53.517369407 +0000 UTC m=+1019.497978505" watchObservedRunningTime="2026-01-27 15:23:53.526128699 +0000 UTC m=+1019.506737797" Jan 27 15:23:53 crc kubenswrapper[4772]: I0127 15:23:53.539778 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-77554cdc5c-tkr6j" podStartSLOduration=2.752698378 podStartE2EDuration="32.539757901s" podCreationTimestamp="2026-01-27 15:23:21 +0000 UTC" firstStartedPulling="2026-01-27 15:23:23.322454617 +0000 UTC m=+989.303063715" lastFinishedPulling="2026-01-27 15:23:53.10951414 +0000 UTC m=+1019.090123238" observedRunningTime="2026-01-27 15:23:53.533570303 +0000 UTC m=+1019.514179401" watchObservedRunningTime="2026-01-27 15:23:53.539757901 +0000 UTC m=+1019.520367009" Jan 27 15:23:54 crc kubenswrapper[4772]: I0127 15:23:54.007947 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e7465bd0-3b6e-4199-9ee6-28b512198847-cert\") pod \"infra-operator-controller-manager-7d75bc88d5-t54fr\" (UID: \"e7465bd0-3b6e-4199-9ee6-28b512198847\") " pod="openstack-operators/infra-operator-controller-manager-7d75bc88d5-t54fr" Jan 27 15:23:54 crc kubenswrapper[4772]: I0127 15:23:54.014317 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e7465bd0-3b6e-4199-9ee6-28b512198847-cert\") pod \"infra-operator-controller-manager-7d75bc88d5-t54fr\" (UID: \"e7465bd0-3b6e-4199-9ee6-28b512198847\") " pod="openstack-operators/infra-operator-controller-manager-7d75bc88d5-t54fr" Jan 27 15:23:54 crc kubenswrapper[4772]: I0127 15:23:54.306352 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7d75bc88d5-t54fr" Jan 27 15:23:54 crc kubenswrapper[4772]: I0127 15:23:54.416319 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1389813b-42ea-433f-820c-e5b8b41713d7-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854q994c\" (UID: \"1389813b-42ea-433f-820c-e5b8b41713d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854q994c" Jan 27 15:23:54 crc kubenswrapper[4772]: I0127 15:23:54.441814 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1389813b-42ea-433f-820c-e5b8b41713d7-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854q994c\" (UID: \"1389813b-42ea-433f-820c-e5b8b41713d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854q994c" Jan 27 15:23:54 crc kubenswrapper[4772]: I0127 15:23:54.548020 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7d75bc88d5-t54fr"] Jan 27 15:23:54 crc kubenswrapper[4772]: W0127 15:23:54.556380 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7465bd0_3b6e_4199_9ee6_28b512198847.slice/crio-f24a8e1d9d29185f462c50004e6568f62134d9ff11f32a055852526a91794abe WatchSource:0}: Error finding container f24a8e1d9d29185f462c50004e6568f62134d9ff11f32a055852526a91794abe: Status 404 returned error can't find the container with id f24a8e1d9d29185f462c50004e6568f62134d9ff11f32a055852526a91794abe Jan 27 15:23:54 crc kubenswrapper[4772]: I0127 15:23:54.619409 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-webhook-certs\") pod \"openstack-operator-controller-manager-ff554fc88-clt4p\" (UID: \"8087d8d3-d2f6-4bca-abec-f5b5335f26fa\") " pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:23:54 crc kubenswrapper[4772]: I0127 15:23:54.620282 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-metrics-certs\") pod \"openstack-operator-controller-manager-ff554fc88-clt4p\" (UID: \"8087d8d3-d2f6-4bca-abec-f5b5335f26fa\") " pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:23:54 crc kubenswrapper[4772]: I0127 15:23:54.623364 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-webhook-certs\") pod \"openstack-operator-controller-manager-ff554fc88-clt4p\" (UID: \"8087d8d3-d2f6-4bca-abec-f5b5335f26fa\") " pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:23:54 crc kubenswrapper[4772]: I0127 15:23:54.623467 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8087d8d3-d2f6-4bca-abec-f5b5335f26fa-metrics-certs\") pod \"openstack-operator-controller-manager-ff554fc88-clt4p\" (UID: \"8087d8d3-d2f6-4bca-abec-f5b5335f26fa\") " pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:23:54 crc kubenswrapper[4772]: I0127 15:23:54.631289 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-tbw52" Jan 27 15:23:54 crc kubenswrapper[4772]: I0127 15:23:54.639447 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854q994c" Jan 27 15:23:54 crc kubenswrapper[4772]: I0127 15:23:54.724650 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-xvq6z" Jan 27 15:23:54 crc kubenswrapper[4772]: I0127 15:23:54.733531 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:23:54 crc kubenswrapper[4772]: I0127 15:23:54.887932 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854q994c"] Jan 27 15:23:55 crc kubenswrapper[4772]: I0127 15:23:55.244443 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p"] Jan 27 15:23:55 crc kubenswrapper[4772]: W0127 15:23:55.249439 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8087d8d3_d2f6_4bca_abec_f5b5335f26fa.slice/crio-ecb882c6aa6eabd50ee7443107232aa8adca0db5010033006e18a3e863cfe81e WatchSource:0}: Error finding container ecb882c6aa6eabd50ee7443107232aa8adca0db5010033006e18a3e863cfe81e: Status 404 returned error can't find the container with id ecb882c6aa6eabd50ee7443107232aa8adca0db5010033006e18a3e863cfe81e Jan 27 15:23:55 crc kubenswrapper[4772]: I0127 15:23:55.508686 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" event={"ID":"8087d8d3-d2f6-4bca-abec-f5b5335f26fa","Type":"ContainerStarted","Data":"ecb882c6aa6eabd50ee7443107232aa8adca0db5010033006e18a3e863cfe81e"} Jan 27 15:23:55 crc kubenswrapper[4772]: I0127 15:23:55.509842 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854q994c" event={"ID":"1389813b-42ea-433f-820c-e5b8b41713d7","Type":"ContainerStarted","Data":"4837d19db0cf6c82d296ae460cfe0da09cf8817ae0253d8c591124098258e70c"} Jan 27 15:23:55 crc kubenswrapper[4772]: I0127 15:23:55.510683 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7d75bc88d5-t54fr" event={"ID":"e7465bd0-3b6e-4199-9ee6-28b512198847","Type":"ContainerStarted","Data":"f24a8e1d9d29185f462c50004e6568f62134d9ff11f32a055852526a91794abe"} Jan 27 15:23:55 crc kubenswrapper[4772]: I0127 15:23:55.575962 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zvzxk" Jan 27 15:23:55 crc kubenswrapper[4772]: I0127 15:23:55.576395 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zvzxk" Jan 27 15:23:55 crc kubenswrapper[4772]: I0127 15:23:55.618398 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zvzxk" Jan 27 15:23:56 crc kubenswrapper[4772]: I0127 15:23:56.518894 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-67dd55ff59-hgscb" event={"ID":"fb300814-fca7-4419-ac6e-c08b33edd4be","Type":"ContainerStarted","Data":"a1d99a9a6663762b7be744c7bdf3a480d1c91959ec97aafb336a67f1b0ee6844"} Jan 27 15:23:56 crc kubenswrapper[4772]: I0127 15:23:56.519389 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-67dd55ff59-hgscb" Jan 27 15:23:56 crc kubenswrapper[4772]: I0127 15:23:56.524634 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" event={"ID":"8087d8d3-d2f6-4bca-abec-f5b5335f26fa","Type":"ContainerStarted","Data":"a213e3d616c938ea6d49b4267e9e03149d9dca6824648a0b735cb41f8091a8e9"} Jan 27 15:23:56 crc kubenswrapper[4772]: I0127 15:23:56.524831 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:23:56 crc kubenswrapper[4772]: I0127 15:23:56.527142 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-zhd82" event={"ID":"b73c175a-e89e-434f-996a-65c1140bb8dd","Type":"ContainerStarted","Data":"02bfcf7788d46afaa4c7f3622e3d45b5db87fa3d03d86c7f62f02d3febf429bb"} Jan 27 15:23:56 crc kubenswrapper[4772]: I0127 15:23:56.527413 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-zhd82" Jan 27 15:23:56 crc kubenswrapper[4772]: I0127 15:23:56.564663 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-67dd55ff59-hgscb" podStartSLOduration=4.957935609 podStartE2EDuration="35.564644311s" podCreationTimestamp="2026-01-27 15:23:21 +0000 UTC" firstStartedPulling="2026-01-27 15:23:23.456972435 +0000 UTC m=+989.437581533" lastFinishedPulling="2026-01-27 15:23:54.063681137 +0000 UTC m=+1020.044290235" observedRunningTime="2026-01-27 15:23:56.538222262 +0000 UTC m=+1022.518831370" watchObservedRunningTime="2026-01-27 15:23:56.564644311 +0000 UTC m=+1022.545253409" Jan 27 15:23:56 crc kubenswrapper[4772]: I0127 15:23:56.568998 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" podStartSLOduration=34.568979976 podStartE2EDuration="34.568979976s" podCreationTimestamp="2026-01-27 15:23:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:23:56.561690477 +0000 UTC m=+1022.542299575" watchObservedRunningTime="2026-01-27 15:23:56.568979976 +0000 UTC m=+1022.549589074" Jan 27 15:23:56 crc kubenswrapper[4772]: I0127 15:23:56.580955 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-zhd82" podStartSLOduration=2.312210718 podStartE2EDuration="34.58093623s" podCreationTimestamp="2026-01-27 15:23:22 +0000 UTC" firstStartedPulling="2026-01-27 15:23:23.646224387 +0000 UTC m=+989.626833485" lastFinishedPulling="2026-01-27 15:23:55.914949909 +0000 UTC m=+1021.895558997" observedRunningTime="2026-01-27 15:23:56.579421636 +0000 UTC m=+1022.560030744" watchObservedRunningTime="2026-01-27 15:23:56.58093623 +0000 UTC m=+1022.561545328" Jan 27 15:23:56 crc kubenswrapper[4772]: I0127 15:23:56.595458 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zvzxk" Jan 27 15:23:56 crc kubenswrapper[4772]: I0127 15:23:56.650719 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zvzxk"] Jan 27 15:23:58 crc kubenswrapper[4772]: I0127 15:23:58.549305 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zvzxk" podUID="0f2e732b-dbc8-423e-8a37-100a97dad4f0" containerName="registry-server" containerID="cri-o://5f0e331299d6605e44535e575ebf4cf5cfa36757bfa41569f0da713738316ff3" gracePeriod=2 Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.423911 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zvzxk" Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.556730 4772 generic.go:334] "Generic (PLEG): container finished" podID="0f2e732b-dbc8-423e-8a37-100a97dad4f0" containerID="5f0e331299d6605e44535e575ebf4cf5cfa36757bfa41569f0da713738316ff3" exitCode=0 Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.556784 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zvzxk" Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.556833 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zvzxk" event={"ID":"0f2e732b-dbc8-423e-8a37-100a97dad4f0","Type":"ContainerDied","Data":"5f0e331299d6605e44535e575ebf4cf5cfa36757bfa41569f0da713738316ff3"} Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.556892 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zvzxk" event={"ID":"0f2e732b-dbc8-423e-8a37-100a97dad4f0","Type":"ContainerDied","Data":"0a190d2f48d8673469fb8f939468a4f2e791b7ede4b24560514885c4a4400b46"} Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.556917 4772 scope.go:117] "RemoveContainer" containerID="5f0e331299d6605e44535e575ebf4cf5cfa36757bfa41569f0da713738316ff3" Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.558796 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7d75bc88d5-t54fr" event={"ID":"e7465bd0-3b6e-4199-9ee6-28b512198847","Type":"ContainerStarted","Data":"823b951e47f8646f1f24d273e1575265c88f88f8998dec9dcd2d7cb961c19b15"} Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.558908 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-7d75bc88d5-t54fr" Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.559950 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854q994c" event={"ID":"1389813b-42ea-433f-820c-e5b8b41713d7","Type":"ContainerStarted","Data":"81da4d1513c5f4df54640f0fdd172eaf3bb9691a0fbc6c32b00d9a205863ffa4"} Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.560347 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854q994c" Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.583746 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h9297" event={"ID":"abaf1142-1b7c-4987-8a9d-c91e6456c4a5","Type":"ContainerStarted","Data":"0f00940bbe45fb05e0bcfd9ef0cccf5f41c4f0e85045a5ade4492f77f252f39c"} Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.586414 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-ddcbfd695-6wltn" event={"ID":"e7fc5297-101a-496e-a7c6-e7296e08a5af","Type":"ContainerStarted","Data":"a306aee00569da20b313031ca959854386d11c011488757e6666ddceb26d0651"} Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.586607 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-ddcbfd695-6wltn" Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.590580 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-7d75bc88d5-t54fr" podStartSLOduration=34.257688451 podStartE2EDuration="38.590566271s" podCreationTimestamp="2026-01-27 15:23:21 +0000 UTC" firstStartedPulling="2026-01-27 15:23:54.557633771 +0000 UTC m=+1020.538242869" lastFinishedPulling="2026-01-27 15:23:58.890511551 +0000 UTC m=+1024.871120689" observedRunningTime="2026-01-27 15:23:59.587450502 +0000 UTC m=+1025.568059600" watchObservedRunningTime="2026-01-27 15:23:59.590566271 +0000 UTC m=+1025.571175359" Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.598317 4772 scope.go:117] "RemoveContainer" containerID="bb5b9739a5d8e52b4be45c2313787179349baaa953432e2d37f9e46057ccb4e4" Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.598848 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29mzp\" (UniqueName: \"kubernetes.io/projected/0f2e732b-dbc8-423e-8a37-100a97dad4f0-kube-api-access-29mzp\") pod \"0f2e732b-dbc8-423e-8a37-100a97dad4f0\" (UID: \"0f2e732b-dbc8-423e-8a37-100a97dad4f0\") " Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.598943 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f2e732b-dbc8-423e-8a37-100a97dad4f0-catalog-content\") pod \"0f2e732b-dbc8-423e-8a37-100a97dad4f0\" (UID: \"0f2e732b-dbc8-423e-8a37-100a97dad4f0\") " Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.599078 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f2e732b-dbc8-423e-8a37-100a97dad4f0-utilities\") pod \"0f2e732b-dbc8-423e-8a37-100a97dad4f0\" (UID: \"0f2e732b-dbc8-423e-8a37-100a97dad4f0\") " Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.600185 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f2e732b-dbc8-423e-8a37-100a97dad4f0-utilities" (OuterVolumeSpecName: "utilities") pod "0f2e732b-dbc8-423e-8a37-100a97dad4f0" (UID: "0f2e732b-dbc8-423e-8a37-100a97dad4f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.607475 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f2e732b-dbc8-423e-8a37-100a97dad4f0-kube-api-access-29mzp" (OuterVolumeSpecName: "kube-api-access-29mzp") pod "0f2e732b-dbc8-423e-8a37-100a97dad4f0" (UID: "0f2e732b-dbc8-423e-8a37-100a97dad4f0"). InnerVolumeSpecName "kube-api-access-29mzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.649938 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854q994c" podStartSLOduration=33.647529301 podStartE2EDuration="37.649919768s" podCreationTimestamp="2026-01-27 15:23:22 +0000 UTC" firstStartedPulling="2026-01-27 15:23:54.892987874 +0000 UTC m=+1020.873596972" lastFinishedPulling="2026-01-27 15:23:58.895378341 +0000 UTC m=+1024.875987439" observedRunningTime="2026-01-27 15:23:59.617493166 +0000 UTC m=+1025.598102264" watchObservedRunningTime="2026-01-27 15:23:59.649919768 +0000 UTC m=+1025.630528866" Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.669517 4772 scope.go:117] "RemoveContainer" containerID="8b01102f5cfb35617b002369451b43259452694e6501916f5c0f4f257352eaee" Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.673822 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-ddcbfd695-6wltn" podStartSLOduration=2.426742101 podStartE2EDuration="37.673806745s" podCreationTimestamp="2026-01-27 15:23:22 +0000 UTC" firstStartedPulling="2026-01-27 15:23:23.643432197 +0000 UTC m=+989.624041295" lastFinishedPulling="2026-01-27 15:23:58.890496831 +0000 UTC m=+1024.871105939" observedRunningTime="2026-01-27 15:23:59.66981234 +0000 UTC m=+1025.650421438" watchObservedRunningTime="2026-01-27 15:23:59.673806745 +0000 UTC m=+1025.654415843" Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.692528 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f2e732b-dbc8-423e-8a37-100a97dad4f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0f2e732b-dbc8-423e-8a37-100a97dad4f0" (UID: "0f2e732b-dbc8-423e-8a37-100a97dad4f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.704147 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29mzp\" (UniqueName: \"kubernetes.io/projected/0f2e732b-dbc8-423e-8a37-100a97dad4f0-kube-api-access-29mzp\") on node \"crc\" DevicePath \"\"" Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.704199 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f2e732b-dbc8-423e-8a37-100a97dad4f0-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.704208 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f2e732b-dbc8-423e-8a37-100a97dad4f0-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.707629 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h9297" podStartSLOduration=2.6547782680000003 podStartE2EDuration="37.707618757s" podCreationTimestamp="2026-01-27 15:23:22 +0000 UTC" firstStartedPulling="2026-01-27 15:23:23.838590769 +0000 UTC m=+989.819199867" lastFinishedPulling="2026-01-27 15:23:58.891431258 +0000 UTC m=+1024.872040356" observedRunningTime="2026-01-27 15:23:59.70737155 +0000 UTC m=+1025.687980648" watchObservedRunningTime="2026-01-27 15:23:59.707618757 +0000 UTC m=+1025.688227845" Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.725205 4772 scope.go:117] "RemoveContainer" containerID="5f0e331299d6605e44535e575ebf4cf5cfa36757bfa41569f0da713738316ff3" Jan 27 15:23:59 crc kubenswrapper[4772]: E0127 15:23:59.725696 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f0e331299d6605e44535e575ebf4cf5cfa36757bfa41569f0da713738316ff3\": container with ID starting with 5f0e331299d6605e44535e575ebf4cf5cfa36757bfa41569f0da713738316ff3 not found: ID does not exist" containerID="5f0e331299d6605e44535e575ebf4cf5cfa36757bfa41569f0da713738316ff3" Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.725724 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f0e331299d6605e44535e575ebf4cf5cfa36757bfa41569f0da713738316ff3"} err="failed to get container status \"5f0e331299d6605e44535e575ebf4cf5cfa36757bfa41569f0da713738316ff3\": rpc error: code = NotFound desc = could not find container \"5f0e331299d6605e44535e575ebf4cf5cfa36757bfa41569f0da713738316ff3\": container with ID starting with 5f0e331299d6605e44535e575ebf4cf5cfa36757bfa41569f0da713738316ff3 not found: ID does not exist" Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.725742 4772 scope.go:117] "RemoveContainer" containerID="bb5b9739a5d8e52b4be45c2313787179349baaa953432e2d37f9e46057ccb4e4" Jan 27 15:23:59 crc kubenswrapper[4772]: E0127 15:23:59.730285 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb5b9739a5d8e52b4be45c2313787179349baaa953432e2d37f9e46057ccb4e4\": container with ID starting with bb5b9739a5d8e52b4be45c2313787179349baaa953432e2d37f9e46057ccb4e4 not found: ID does not exist" containerID="bb5b9739a5d8e52b4be45c2313787179349baaa953432e2d37f9e46057ccb4e4" Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.730331 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb5b9739a5d8e52b4be45c2313787179349baaa953432e2d37f9e46057ccb4e4"} err="failed to get container status \"bb5b9739a5d8e52b4be45c2313787179349baaa953432e2d37f9e46057ccb4e4\": rpc error: code = NotFound desc = could not find container \"bb5b9739a5d8e52b4be45c2313787179349baaa953432e2d37f9e46057ccb4e4\": container with ID starting with bb5b9739a5d8e52b4be45c2313787179349baaa953432e2d37f9e46057ccb4e4 not found: ID does not exist" Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.730356 4772 scope.go:117] "RemoveContainer" containerID="8b01102f5cfb35617b002369451b43259452694e6501916f5c0f4f257352eaee" Jan 27 15:23:59 crc kubenswrapper[4772]: E0127 15:23:59.732117 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b01102f5cfb35617b002369451b43259452694e6501916f5c0f4f257352eaee\": container with ID starting with 8b01102f5cfb35617b002369451b43259452694e6501916f5c0f4f257352eaee not found: ID does not exist" containerID="8b01102f5cfb35617b002369451b43259452694e6501916f5c0f4f257352eaee" Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.732137 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b01102f5cfb35617b002369451b43259452694e6501916f5c0f4f257352eaee"} err="failed to get container status \"8b01102f5cfb35617b002369451b43259452694e6501916f5c0f4f257352eaee\": rpc error: code = NotFound desc = could not find container \"8b01102f5cfb35617b002369451b43259452694e6501916f5c0f4f257352eaee\": container with ID starting with 8b01102f5cfb35617b002369451b43259452694e6501916f5c0f4f257352eaee not found: ID does not exist" Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.880253 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zvzxk"] Jan 27 15:23:59 crc kubenswrapper[4772]: I0127 15:23:59.884707 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zvzxk"] Jan 27 15:24:00 crc kubenswrapper[4772]: I0127 15:24:00.674403 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f2e732b-dbc8-423e-8a37-100a97dad4f0" path="/var/lib/kubelet/pods/0f2e732b-dbc8-423e-8a37-100a97dad4f0/volumes" Jan 27 15:24:02 crc kubenswrapper[4772]: I0127 15:24:02.136229 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-77554cdc5c-tkr6j" Jan 27 15:24:02 crc kubenswrapper[4772]: I0127 15:24:02.154460 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-67dd55ff59-hgscb" Jan 27 15:24:02 crc kubenswrapper[4772]: I0127 15:24:02.317471 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-768b776ffb-sxbjn" Jan 27 15:24:02 crc kubenswrapper[4772]: I0127 15:24:02.583660 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-55f684fd56-wzjrz" Jan 27 15:24:02 crc kubenswrapper[4772]: I0127 15:24:02.770636 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-zhd82" Jan 27 15:24:04 crc kubenswrapper[4772]: I0127 15:24:04.316341 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-7d75bc88d5-t54fr" Jan 27 15:24:04 crc kubenswrapper[4772]: I0127 15:24:04.649791 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854q994c" Jan 27 15:24:04 crc kubenswrapper[4772]: I0127 15:24:04.747006 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-ff554fc88-clt4p" Jan 27 15:24:12 crc kubenswrapper[4772]: I0127 15:24:12.716403 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-ddcbfd695-6wltn" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.714008 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-t8dxn"] Jan 27 15:24:29 crc kubenswrapper[4772]: E0127 15:24:29.716491 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f2e732b-dbc8-423e-8a37-100a97dad4f0" containerName="registry-server" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.716599 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f2e732b-dbc8-423e-8a37-100a97dad4f0" containerName="registry-server" Jan 27 15:24:29 crc kubenswrapper[4772]: E0127 15:24:29.716685 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f2e732b-dbc8-423e-8a37-100a97dad4f0" containerName="extract-content" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.716752 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f2e732b-dbc8-423e-8a37-100a97dad4f0" containerName="extract-content" Jan 27 15:24:29 crc kubenswrapper[4772]: E0127 15:24:29.716822 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f2e732b-dbc8-423e-8a37-100a97dad4f0" containerName="extract-utilities" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.716903 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f2e732b-dbc8-423e-8a37-100a97dad4f0" containerName="extract-utilities" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.717138 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f2e732b-dbc8-423e-8a37-100a97dad4f0" containerName="registry-server" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.718738 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-t8dxn" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.721182 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.721216 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-cnqhx" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.721285 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.723238 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.729777 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-t8dxn"] Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.783487 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6ll7\" (UniqueName: \"kubernetes.io/projected/9e884850-6b45-4657-8c8c-fa8ccdec648d-kube-api-access-h6ll7\") pod \"dnsmasq-dns-675f4bcbfc-t8dxn\" (UID: \"9e884850-6b45-4657-8c8c-fa8ccdec648d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-t8dxn" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.783851 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e884850-6b45-4657-8c8c-fa8ccdec648d-config\") pod \"dnsmasq-dns-675f4bcbfc-t8dxn\" (UID: \"9e884850-6b45-4657-8c8c-fa8ccdec648d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-t8dxn" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.784254 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-bmnxs"] Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.791102 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-bmnxs" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.793158 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-bmnxs"] Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.799552 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.889236 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/580fdc18-8bdc-4a16-89a4-efd7df1b8a17-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-bmnxs\" (UID: \"580fdc18-8bdc-4a16-89a4-efd7df1b8a17\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bmnxs" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.889583 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2msw8\" (UniqueName: \"kubernetes.io/projected/580fdc18-8bdc-4a16-89a4-efd7df1b8a17-kube-api-access-2msw8\") pod \"dnsmasq-dns-78dd6ddcc-bmnxs\" (UID: \"580fdc18-8bdc-4a16-89a4-efd7df1b8a17\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bmnxs" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.889705 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6ll7\" (UniqueName: \"kubernetes.io/projected/9e884850-6b45-4657-8c8c-fa8ccdec648d-kube-api-access-h6ll7\") pod \"dnsmasq-dns-675f4bcbfc-t8dxn\" (UID: \"9e884850-6b45-4657-8c8c-fa8ccdec648d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-t8dxn" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.889860 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e884850-6b45-4657-8c8c-fa8ccdec648d-config\") pod \"dnsmasq-dns-675f4bcbfc-t8dxn\" (UID: \"9e884850-6b45-4657-8c8c-fa8ccdec648d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-t8dxn" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.889979 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/580fdc18-8bdc-4a16-89a4-efd7df1b8a17-config\") pod \"dnsmasq-dns-78dd6ddcc-bmnxs\" (UID: \"580fdc18-8bdc-4a16-89a4-efd7df1b8a17\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bmnxs" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.891328 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e884850-6b45-4657-8c8c-fa8ccdec648d-config\") pod \"dnsmasq-dns-675f4bcbfc-t8dxn\" (UID: \"9e884850-6b45-4657-8c8c-fa8ccdec648d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-t8dxn" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.911248 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6ll7\" (UniqueName: \"kubernetes.io/projected/9e884850-6b45-4657-8c8c-fa8ccdec648d-kube-api-access-h6ll7\") pod \"dnsmasq-dns-675f4bcbfc-t8dxn\" (UID: \"9e884850-6b45-4657-8c8c-fa8ccdec648d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-t8dxn" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.991114 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/580fdc18-8bdc-4a16-89a4-efd7df1b8a17-config\") pod \"dnsmasq-dns-78dd6ddcc-bmnxs\" (UID: \"580fdc18-8bdc-4a16-89a4-efd7df1b8a17\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bmnxs" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.991508 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/580fdc18-8bdc-4a16-89a4-efd7df1b8a17-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-bmnxs\" (UID: \"580fdc18-8bdc-4a16-89a4-efd7df1b8a17\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bmnxs" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.991585 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2msw8\" (UniqueName: \"kubernetes.io/projected/580fdc18-8bdc-4a16-89a4-efd7df1b8a17-kube-api-access-2msw8\") pod \"dnsmasq-dns-78dd6ddcc-bmnxs\" (UID: \"580fdc18-8bdc-4a16-89a4-efd7df1b8a17\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bmnxs" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.991920 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/580fdc18-8bdc-4a16-89a4-efd7df1b8a17-config\") pod \"dnsmasq-dns-78dd6ddcc-bmnxs\" (UID: \"580fdc18-8bdc-4a16-89a4-efd7df1b8a17\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bmnxs" Jan 27 15:24:29 crc kubenswrapper[4772]: I0127 15:24:29.992461 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/580fdc18-8bdc-4a16-89a4-efd7df1b8a17-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-bmnxs\" (UID: \"580fdc18-8bdc-4a16-89a4-efd7df1b8a17\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bmnxs" Jan 27 15:24:30 crc kubenswrapper[4772]: I0127 15:24:30.015036 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2msw8\" (UniqueName: \"kubernetes.io/projected/580fdc18-8bdc-4a16-89a4-efd7df1b8a17-kube-api-access-2msw8\") pod \"dnsmasq-dns-78dd6ddcc-bmnxs\" (UID: \"580fdc18-8bdc-4a16-89a4-efd7df1b8a17\") " pod="openstack/dnsmasq-dns-78dd6ddcc-bmnxs" Jan 27 15:24:30 crc kubenswrapper[4772]: I0127 15:24:30.035883 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-t8dxn" Jan 27 15:24:30 crc kubenswrapper[4772]: I0127 15:24:30.114224 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-bmnxs" Jan 27 15:24:30 crc kubenswrapper[4772]: I0127 15:24:30.343587 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-t8dxn"] Jan 27 15:24:30 crc kubenswrapper[4772]: I0127 15:24:30.356400 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 15:24:30 crc kubenswrapper[4772]: I0127 15:24:30.391681 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-bmnxs"] Jan 27 15:24:30 crc kubenswrapper[4772]: W0127 15:24:30.393493 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod580fdc18_8bdc_4a16_89a4_efd7df1b8a17.slice/crio-621419c58a31ba2a4aca778faec82cf28803b219ef85445b2df7f5a1babfe7d1 WatchSource:0}: Error finding container 621419c58a31ba2a4aca778faec82cf28803b219ef85445b2df7f5a1babfe7d1: Status 404 returned error can't find the container with id 621419c58a31ba2a4aca778faec82cf28803b219ef85445b2df7f5a1babfe7d1 Jan 27 15:24:30 crc kubenswrapper[4772]: I0127 15:24:30.877922 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-bmnxs" event={"ID":"580fdc18-8bdc-4a16-89a4-efd7df1b8a17","Type":"ContainerStarted","Data":"621419c58a31ba2a4aca778faec82cf28803b219ef85445b2df7f5a1babfe7d1"} Jan 27 15:24:30 crc kubenswrapper[4772]: I0127 15:24:30.878927 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-t8dxn" event={"ID":"9e884850-6b45-4657-8c8c-fa8ccdec648d","Type":"ContainerStarted","Data":"2c9714095ed3f7fd054930137f9e21b4fdd5954c595ecb091f834dc4c0fc5111"} Jan 27 15:24:31 crc kubenswrapper[4772]: I0127 15:24:31.421237 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-t8dxn"] Jan 27 15:24:31 crc kubenswrapper[4772]: I0127 15:24:31.438220 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-5wcr9"] Jan 27 15:24:31 crc kubenswrapper[4772]: I0127 15:24:31.439338 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-5wcr9" Jan 27 15:24:31 crc kubenswrapper[4772]: I0127 15:24:31.450646 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-5wcr9"] Jan 27 15:24:31 crc kubenswrapper[4772]: I0127 15:24:31.516542 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ece4b345-2aab-4ee3-a116-366d6b8d7bff-dns-svc\") pod \"dnsmasq-dns-666b6646f7-5wcr9\" (UID: \"ece4b345-2aab-4ee3-a116-366d6b8d7bff\") " pod="openstack/dnsmasq-dns-666b6646f7-5wcr9" Jan 27 15:24:31 crc kubenswrapper[4772]: I0127 15:24:31.516641 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmnqn\" (UniqueName: \"kubernetes.io/projected/ece4b345-2aab-4ee3-a116-366d6b8d7bff-kube-api-access-xmnqn\") pod \"dnsmasq-dns-666b6646f7-5wcr9\" (UID: \"ece4b345-2aab-4ee3-a116-366d6b8d7bff\") " pod="openstack/dnsmasq-dns-666b6646f7-5wcr9" Jan 27 15:24:31 crc kubenswrapper[4772]: I0127 15:24:31.516676 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece4b345-2aab-4ee3-a116-366d6b8d7bff-config\") pod \"dnsmasq-dns-666b6646f7-5wcr9\" (UID: \"ece4b345-2aab-4ee3-a116-366d6b8d7bff\") " pod="openstack/dnsmasq-dns-666b6646f7-5wcr9" Jan 27 15:24:31 crc kubenswrapper[4772]: I0127 15:24:31.617713 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ece4b345-2aab-4ee3-a116-366d6b8d7bff-dns-svc\") pod \"dnsmasq-dns-666b6646f7-5wcr9\" (UID: \"ece4b345-2aab-4ee3-a116-366d6b8d7bff\") " pod="openstack/dnsmasq-dns-666b6646f7-5wcr9" Jan 27 15:24:31 crc kubenswrapper[4772]: I0127 15:24:31.617797 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmnqn\" (UniqueName: \"kubernetes.io/projected/ece4b345-2aab-4ee3-a116-366d6b8d7bff-kube-api-access-xmnqn\") pod \"dnsmasq-dns-666b6646f7-5wcr9\" (UID: \"ece4b345-2aab-4ee3-a116-366d6b8d7bff\") " pod="openstack/dnsmasq-dns-666b6646f7-5wcr9" Jan 27 15:24:31 crc kubenswrapper[4772]: I0127 15:24:31.617826 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece4b345-2aab-4ee3-a116-366d6b8d7bff-config\") pod \"dnsmasq-dns-666b6646f7-5wcr9\" (UID: \"ece4b345-2aab-4ee3-a116-366d6b8d7bff\") " pod="openstack/dnsmasq-dns-666b6646f7-5wcr9" Jan 27 15:24:31 crc kubenswrapper[4772]: I0127 15:24:31.618779 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece4b345-2aab-4ee3-a116-366d6b8d7bff-config\") pod \"dnsmasq-dns-666b6646f7-5wcr9\" (UID: \"ece4b345-2aab-4ee3-a116-366d6b8d7bff\") " pod="openstack/dnsmasq-dns-666b6646f7-5wcr9" Jan 27 15:24:31 crc kubenswrapper[4772]: I0127 15:24:31.619298 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ece4b345-2aab-4ee3-a116-366d6b8d7bff-dns-svc\") pod \"dnsmasq-dns-666b6646f7-5wcr9\" (UID: \"ece4b345-2aab-4ee3-a116-366d6b8d7bff\") " pod="openstack/dnsmasq-dns-666b6646f7-5wcr9" Jan 27 15:24:31 crc kubenswrapper[4772]: I0127 15:24:31.636028 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmnqn\" (UniqueName: \"kubernetes.io/projected/ece4b345-2aab-4ee3-a116-366d6b8d7bff-kube-api-access-xmnqn\") pod \"dnsmasq-dns-666b6646f7-5wcr9\" (UID: \"ece4b345-2aab-4ee3-a116-366d6b8d7bff\") " pod="openstack/dnsmasq-dns-666b6646f7-5wcr9" Jan 27 15:24:31 crc kubenswrapper[4772]: I0127 15:24:31.757619 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-5wcr9" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.233948 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-bmnxs"] Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.265382 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-rsdjb"] Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.266417 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-rsdjb" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.281695 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-rsdjb"] Jan 27 15:24:32 crc kubenswrapper[4772]: W0127 15:24:32.323751 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podece4b345_2aab_4ee3_a116_366d6b8d7bff.slice/crio-69b728c61ccbcb54532eee4001562892bf3d12d85c2e89da36c5a93f0401b107 WatchSource:0}: Error finding container 69b728c61ccbcb54532eee4001562892bf3d12d85c2e89da36c5a93f0401b107: Status 404 returned error can't find the container with id 69b728c61ccbcb54532eee4001562892bf3d12d85c2e89da36c5a93f0401b107 Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.323932 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-5wcr9"] Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.335431 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/477337bf-a24a-44fd-9c46-38d2e1566b18-config\") pod \"dnsmasq-dns-57d769cc4f-rsdjb\" (UID: \"477337bf-a24a-44fd-9c46-38d2e1566b18\") " pod="openstack/dnsmasq-dns-57d769cc4f-rsdjb" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.335491 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnjbj\" (UniqueName: \"kubernetes.io/projected/477337bf-a24a-44fd-9c46-38d2e1566b18-kube-api-access-lnjbj\") pod \"dnsmasq-dns-57d769cc4f-rsdjb\" (UID: \"477337bf-a24a-44fd-9c46-38d2e1566b18\") " pod="openstack/dnsmasq-dns-57d769cc4f-rsdjb" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.335535 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/477337bf-a24a-44fd-9c46-38d2e1566b18-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-rsdjb\" (UID: \"477337bf-a24a-44fd-9c46-38d2e1566b18\") " pod="openstack/dnsmasq-dns-57d769cc4f-rsdjb" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.436896 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/477337bf-a24a-44fd-9c46-38d2e1566b18-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-rsdjb\" (UID: \"477337bf-a24a-44fd-9c46-38d2e1566b18\") " pod="openstack/dnsmasq-dns-57d769cc4f-rsdjb" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.436984 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/477337bf-a24a-44fd-9c46-38d2e1566b18-config\") pod \"dnsmasq-dns-57d769cc4f-rsdjb\" (UID: \"477337bf-a24a-44fd-9c46-38d2e1566b18\") " pod="openstack/dnsmasq-dns-57d769cc4f-rsdjb" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.437021 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnjbj\" (UniqueName: \"kubernetes.io/projected/477337bf-a24a-44fd-9c46-38d2e1566b18-kube-api-access-lnjbj\") pod \"dnsmasq-dns-57d769cc4f-rsdjb\" (UID: \"477337bf-a24a-44fd-9c46-38d2e1566b18\") " pod="openstack/dnsmasq-dns-57d769cc4f-rsdjb" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.438252 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/477337bf-a24a-44fd-9c46-38d2e1566b18-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-rsdjb\" (UID: \"477337bf-a24a-44fd-9c46-38d2e1566b18\") " pod="openstack/dnsmasq-dns-57d769cc4f-rsdjb" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.439002 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/477337bf-a24a-44fd-9c46-38d2e1566b18-config\") pod \"dnsmasq-dns-57d769cc4f-rsdjb\" (UID: \"477337bf-a24a-44fd-9c46-38d2e1566b18\") " pod="openstack/dnsmasq-dns-57d769cc4f-rsdjb" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.466333 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnjbj\" (UniqueName: \"kubernetes.io/projected/477337bf-a24a-44fd-9c46-38d2e1566b18-kube-api-access-lnjbj\") pod \"dnsmasq-dns-57d769cc4f-rsdjb\" (UID: \"477337bf-a24a-44fd-9c46-38d2e1566b18\") " pod="openstack/dnsmasq-dns-57d769cc4f-rsdjb" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.584879 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.586415 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.590490 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.590726 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.593567 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-rsdjb" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.594183 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.594370 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.594404 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.594512 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-6zl48" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.595665 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.597513 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.639888 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.639952 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/508c3d5b-212a-46da-9a55-de3f35d7019b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.640000 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/508c3d5b-212a-46da-9a55-de3f35d7019b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.640030 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/508c3d5b-212a-46da-9a55-de3f35d7019b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.640050 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.640090 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/508c3d5b-212a-46da-9a55-de3f35d7019b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.640118 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8h8d\" (UniqueName: \"kubernetes.io/projected/508c3d5b-212a-46da-9a55-de3f35d7019b-kube-api-access-l8h8d\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.640142 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.640188 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/508c3d5b-212a-46da-9a55-de3f35d7019b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.641068 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/508c3d5b-212a-46da-9a55-de3f35d7019b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.641141 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-config-data\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.743473 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-config-data\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.743584 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.743633 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/508c3d5b-212a-46da-9a55-de3f35d7019b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.743678 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/508c3d5b-212a-46da-9a55-de3f35d7019b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.743717 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/508c3d5b-212a-46da-9a55-de3f35d7019b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.743737 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.743787 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/508c3d5b-212a-46da-9a55-de3f35d7019b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.743813 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8h8d\" (UniqueName: \"kubernetes.io/projected/508c3d5b-212a-46da-9a55-de3f35d7019b-kube-api-access-l8h8d\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.743830 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.744503 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/508c3d5b-212a-46da-9a55-de3f35d7019b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.744543 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/508c3d5b-212a-46da-9a55-de3f35d7019b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.744596 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-config-data\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.744393 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.744469 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/508c3d5b-212a-46da-9a55-de3f35d7019b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.745033 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.745096 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.745473 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/508c3d5b-212a-46da-9a55-de3f35d7019b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.748061 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/508c3d5b-212a-46da-9a55-de3f35d7019b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.748673 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/508c3d5b-212a-46da-9a55-de3f35d7019b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.749489 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/508c3d5b-212a-46da-9a55-de3f35d7019b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.749587 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/508c3d5b-212a-46da-9a55-de3f35d7019b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.762939 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8h8d\" (UniqueName: \"kubernetes.io/projected/508c3d5b-212a-46da-9a55-de3f35d7019b-kube-api-access-l8h8d\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.788877 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " pod="openstack/rabbitmq-server-0" Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.902902 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-5wcr9" event={"ID":"ece4b345-2aab-4ee3-a116-366d6b8d7bff","Type":"ContainerStarted","Data":"69b728c61ccbcb54532eee4001562892bf3d12d85c2e89da36c5a93f0401b107"} Jan 27 15:24:32 crc kubenswrapper[4772]: I0127 15:24:32.914613 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.123967 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-rsdjb"] Jan 27 15:24:33 crc kubenswrapper[4772]: W0127 15:24:33.152490 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod477337bf_a24a_44fd_9c46_38d2e1566b18.slice/crio-e3d43f62abf3e7b0d500727260573ad5c5ab13345a55b4ce49116d69dfd50dd4 WatchSource:0}: Error finding container e3d43f62abf3e7b0d500727260573ad5c5ab13345a55b4ce49116d69dfd50dd4: Status 404 returned error can't find the container with id e3d43f62abf3e7b0d500727260573ad5c5ab13345a55b4ce49116d69dfd50dd4 Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.430721 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.432944 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.438371 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.438653 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.438907 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.439191 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.439354 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.439525 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.439691 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-frtbn" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.455299 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.455365 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.455399 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.455430 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.455450 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gbh9\" (UniqueName: \"kubernetes.io/projected/76fdbdb1-d48a-4cd1-8372-78887671dce8-kube-api-access-9gbh9\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.455483 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/76fdbdb1-d48a-4cd1-8372-78887671dce8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.455518 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.455549 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.455583 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/76fdbdb1-d48a-4cd1-8372-78887671dce8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.455612 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.455637 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.464000 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.469717 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 27 15:24:33 crc kubenswrapper[4772]: W0127 15:24:33.492524 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod508c3d5b_212a_46da_9a55_de3f35d7019b.slice/crio-044e360ab5ed48dba1c044f12dafd0e510d6847bb09f3238ce3b8c8d2130f226 WatchSource:0}: Error finding container 044e360ab5ed48dba1c044f12dafd0e510d6847bb09f3238ce3b8c8d2130f226: Status 404 returned error can't find the container with id 044e360ab5ed48dba1c044f12dafd0e510d6847bb09f3238ce3b8c8d2130f226 Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.556227 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.556280 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/76fdbdb1-d48a-4cd1-8372-78887671dce8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.556316 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.556339 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.556406 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.556448 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.556474 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.556502 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.556528 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gbh9\" (UniqueName: \"kubernetes.io/projected/76fdbdb1-d48a-4cd1-8372-78887671dce8-kube-api-access-9gbh9\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.556561 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/76fdbdb1-d48a-4cd1-8372-78887671dce8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.556593 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.557205 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.557353 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.557495 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.557515 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.557747 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.558291 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.562850 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/76fdbdb1-d48a-4cd1-8372-78887671dce8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.564251 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.565294 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/76fdbdb1-d48a-4cd1-8372-78887671dce8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.573452 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.574262 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gbh9\" (UniqueName: \"kubernetes.io/projected/76fdbdb1-d48a-4cd1-8372-78887671dce8-kube-api-access-9gbh9\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.582799 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.764747 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.917785 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-rsdjb" event={"ID":"477337bf-a24a-44fd-9c46-38d2e1566b18","Type":"ContainerStarted","Data":"e3d43f62abf3e7b0d500727260573ad5c5ab13345a55b4ce49116d69dfd50dd4"} Jan 27 15:24:33 crc kubenswrapper[4772]: I0127 15:24:33.919955 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"508c3d5b-212a-46da-9a55-de3f35d7019b","Type":"ContainerStarted","Data":"044e360ab5ed48dba1c044f12dafd0e510d6847bb09f3238ce3b8c8d2130f226"} Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.282662 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 27 15:24:34 crc kubenswrapper[4772]: W0127 15:24:34.297432 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76fdbdb1_d48a_4cd1_8372_78887671dce8.slice/crio-09e6c8b66552c99b1f924df5f88d4156d8a5bb2bf8b6bbb8e0fc50cdfa96e1ad WatchSource:0}: Error finding container 09e6c8b66552c99b1f924df5f88d4156d8a5bb2bf8b6bbb8e0fc50cdfa96e1ad: Status 404 returned error can't find the container with id 09e6c8b66552c99b1f924df5f88d4156d8a5bb2bf8b6bbb8e0fc50cdfa96e1ad Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.774645 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.775783 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.781609 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.783760 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.784475 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-4dfv4" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.787359 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.787684 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.788678 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.879756 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b1515626-5d79-408d-abc1-cb92abd58f3f-config-data-default\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " pod="openstack/openstack-galera-0" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.879814 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b1515626-5d79-408d-abc1-cb92abd58f3f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " pod="openstack/openstack-galera-0" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.879834 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rxbj\" (UniqueName: \"kubernetes.io/projected/b1515626-5d79-408d-abc1-cb92abd58f3f-kube-api-access-4rxbj\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " pod="openstack/openstack-galera-0" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.879867 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1515626-5d79-408d-abc1-cb92abd58f3f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " pod="openstack/openstack-galera-0" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.879888 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b1515626-5d79-408d-abc1-cb92abd58f3f-kolla-config\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " pod="openstack/openstack-galera-0" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.879927 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1515626-5d79-408d-abc1-cb92abd58f3f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " pod="openstack/openstack-galera-0" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.879965 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1515626-5d79-408d-abc1-cb92abd58f3f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " pod="openstack/openstack-galera-0" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.879989 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " pod="openstack/openstack-galera-0" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.935848 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"76fdbdb1-d48a-4cd1-8372-78887671dce8","Type":"ContainerStarted","Data":"09e6c8b66552c99b1f924df5f88d4156d8a5bb2bf8b6bbb8e0fc50cdfa96e1ad"} Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.982087 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1515626-5d79-408d-abc1-cb92abd58f3f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " pod="openstack/openstack-galera-0" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.982148 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1515626-5d79-408d-abc1-cb92abd58f3f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " pod="openstack/openstack-galera-0" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.982186 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " pod="openstack/openstack-galera-0" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.982244 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b1515626-5d79-408d-abc1-cb92abd58f3f-config-data-default\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " pod="openstack/openstack-galera-0" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.982288 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b1515626-5d79-408d-abc1-cb92abd58f3f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " pod="openstack/openstack-galera-0" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.982308 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rxbj\" (UniqueName: \"kubernetes.io/projected/b1515626-5d79-408d-abc1-cb92abd58f3f-kube-api-access-4rxbj\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " pod="openstack/openstack-galera-0" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.982351 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1515626-5d79-408d-abc1-cb92abd58f3f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " pod="openstack/openstack-galera-0" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.982382 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b1515626-5d79-408d-abc1-cb92abd58f3f-kolla-config\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " pod="openstack/openstack-galera-0" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.983711 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-galera-0" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.983775 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b1515626-5d79-408d-abc1-cb92abd58f3f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " pod="openstack/openstack-galera-0" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.984264 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b1515626-5d79-408d-abc1-cb92abd58f3f-kolla-config\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " pod="openstack/openstack-galera-0" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.984445 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1515626-5d79-408d-abc1-cb92abd58f3f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " pod="openstack/openstack-galera-0" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.986047 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b1515626-5d79-408d-abc1-cb92abd58f3f-config-data-default\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " pod="openstack/openstack-galera-0" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.988800 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1515626-5d79-408d-abc1-cb92abd58f3f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " pod="openstack/openstack-galera-0" Jan 27 15:24:34 crc kubenswrapper[4772]: I0127 15:24:34.989853 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1515626-5d79-408d-abc1-cb92abd58f3f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " pod="openstack/openstack-galera-0" Jan 27 15:24:35 crc kubenswrapper[4772]: I0127 15:24:35.002404 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rxbj\" (UniqueName: \"kubernetes.io/projected/b1515626-5d79-408d-abc1-cb92abd58f3f-kube-api-access-4rxbj\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " pod="openstack/openstack-galera-0" Jan 27 15:24:35 crc kubenswrapper[4772]: I0127 15:24:35.019907 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " pod="openstack/openstack-galera-0" Jan 27 15:24:35 crc kubenswrapper[4772]: I0127 15:24:35.118901 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Jan 27 15:24:35 crc kubenswrapper[4772]: I0127 15:24:35.974650 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.255685 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.257217 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.261211 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.261804 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.262290 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-xm8qg" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.268879 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.270206 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.429294 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvhlp\" (UniqueName: \"kubernetes.io/projected/cf619242-7348-4de4-a37e-8ebdc4ca54d7-kube-api-access-kvhlp\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.429384 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf619242-7348-4de4-a37e-8ebdc4ca54d7-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.429491 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.429529 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf619242-7348-4de4-a37e-8ebdc4ca54d7-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.429553 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf619242-7348-4de4-a37e-8ebdc4ca54d7-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.429620 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cf619242-7348-4de4-a37e-8ebdc4ca54d7-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.429685 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/cf619242-7348-4de4-a37e-8ebdc4ca54d7-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.429732 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/cf619242-7348-4de4-a37e-8ebdc4ca54d7-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.532430 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cf619242-7348-4de4-a37e-8ebdc4ca54d7-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.532484 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/cf619242-7348-4de4-a37e-8ebdc4ca54d7-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.532512 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/cf619242-7348-4de4-a37e-8ebdc4ca54d7-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.532545 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvhlp\" (UniqueName: \"kubernetes.io/projected/cf619242-7348-4de4-a37e-8ebdc4ca54d7-kube-api-access-kvhlp\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.532585 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf619242-7348-4de4-a37e-8ebdc4ca54d7-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.532657 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.532673 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf619242-7348-4de4-a37e-8ebdc4ca54d7-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.532688 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf619242-7348-4de4-a37e-8ebdc4ca54d7-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.533217 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cf619242-7348-4de4-a37e-8ebdc4ca54d7-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.533603 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.534058 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/cf619242-7348-4de4-a37e-8ebdc4ca54d7-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.534420 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf619242-7348-4de4-a37e-8ebdc4ca54d7-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.536102 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/cf619242-7348-4de4-a37e-8ebdc4ca54d7-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.557230 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf619242-7348-4de4-a37e-8ebdc4ca54d7-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.560293 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf619242-7348-4de4-a37e-8ebdc4ca54d7-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.563281 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.573231 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.574875 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.576850 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.577242 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.577355 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-5t8xl" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.589446 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvhlp\" (UniqueName: \"kubernetes.io/projected/cf619242-7348-4de4-a37e-8ebdc4ca54d7-kube-api-access-kvhlp\") pod \"openstack-cell1-galera-0\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.606980 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.643204 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-kolla-config\") pod \"memcached-0\" (UID: \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\") " pod="openstack/memcached-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.643596 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-memcached-tls-certs\") pod \"memcached-0\" (UID: \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\") " pod="openstack/memcached-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.643642 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-combined-ca-bundle\") pod \"memcached-0\" (UID: \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\") " pod="openstack/memcached-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.643673 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-config-data\") pod \"memcached-0\" (UID: \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\") " pod="openstack/memcached-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.643726 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r75qc\" (UniqueName: \"kubernetes.io/projected/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-kube-api-access-r75qc\") pod \"memcached-0\" (UID: \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\") " pod="openstack/memcached-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.748229 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-combined-ca-bundle\") pod \"memcached-0\" (UID: \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\") " pod="openstack/memcached-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.748283 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-config-data\") pod \"memcached-0\" (UID: \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\") " pod="openstack/memcached-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.748329 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r75qc\" (UniqueName: \"kubernetes.io/projected/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-kube-api-access-r75qc\") pod \"memcached-0\" (UID: \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\") " pod="openstack/memcached-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.748378 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-kolla-config\") pod \"memcached-0\" (UID: \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\") " pod="openstack/memcached-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.748410 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-memcached-tls-certs\") pod \"memcached-0\" (UID: \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\") " pod="openstack/memcached-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.749820 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-config-data\") pod \"memcached-0\" (UID: \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\") " pod="openstack/memcached-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.755391 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-kolla-config\") pod \"memcached-0\" (UID: \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\") " pod="openstack/memcached-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.764840 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-memcached-tls-certs\") pod \"memcached-0\" (UID: \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\") " pod="openstack/memcached-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.766997 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r75qc\" (UniqueName: \"kubernetes.io/projected/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-kube-api-access-r75qc\") pod \"memcached-0\" (UID: \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\") " pod="openstack/memcached-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.772348 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-combined-ca-bundle\") pod \"memcached-0\" (UID: \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\") " pod="openstack/memcached-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.888196 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.952508 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b1515626-5d79-408d-abc1-cb92abd58f3f","Type":"ContainerStarted","Data":"e699d423eedfd6502021873114f8ac6157951b5b24e3387e2b6a5c652a5f6465"} Jan 27 15:24:36 crc kubenswrapper[4772]: I0127 15:24:36.957216 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Jan 27 15:24:37 crc kubenswrapper[4772]: I0127 15:24:37.449491 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 27 15:24:37 crc kubenswrapper[4772]: I0127 15:24:37.562768 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Jan 27 15:24:38 crc kubenswrapper[4772]: I0127 15:24:38.209096 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Jan 27 15:24:38 crc kubenswrapper[4772]: I0127 15:24:38.210500 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 27 15:24:38 crc kubenswrapper[4772]: I0127 15:24:38.218623 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-p4qsl" Jan 27 15:24:38 crc kubenswrapper[4772]: I0127 15:24:38.219005 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 27 15:24:38 crc kubenswrapper[4772]: I0127 15:24:38.290206 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znb2g\" (UniqueName: \"kubernetes.io/projected/1ef66151-0ea7-4696-9db0-7b6665731670-kube-api-access-znb2g\") pod \"kube-state-metrics-0\" (UID: \"1ef66151-0ea7-4696-9db0-7b6665731670\") " pod="openstack/kube-state-metrics-0" Jan 27 15:24:38 crc kubenswrapper[4772]: I0127 15:24:38.393969 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znb2g\" (UniqueName: \"kubernetes.io/projected/1ef66151-0ea7-4696-9db0-7b6665731670-kube-api-access-znb2g\") pod \"kube-state-metrics-0\" (UID: \"1ef66151-0ea7-4696-9db0-7b6665731670\") " pod="openstack/kube-state-metrics-0" Jan 27 15:24:38 crc kubenswrapper[4772]: I0127 15:24:38.447864 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znb2g\" (UniqueName: \"kubernetes.io/projected/1ef66151-0ea7-4696-9db0-7b6665731670-kube-api-access-znb2g\") pod \"kube-state-metrics-0\" (UID: \"1ef66151-0ea7-4696-9db0-7b6665731670\") " pod="openstack/kube-state-metrics-0" Jan 27 15:24:38 crc kubenswrapper[4772]: I0127 15:24:38.590607 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.763317 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-gxjzh"] Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.764765 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gxjzh" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.767327 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.768539 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.771097 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-cpjv6" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.776684 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-cqx7r"] Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.780311 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.784731 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gxjzh"] Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.796499 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-cqx7r"] Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.860258 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/38ebd422-35c5-4682-8a4d-ca9073728d7c-var-run\") pod \"ovn-controller-ovs-cqx7r\" (UID: \"38ebd422-35c5-4682-8a4d-ca9073728d7c\") " pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.860410 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/220011f2-8778-4a14-82d4-33a07bd33379-var-run\") pod \"ovn-controller-gxjzh\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " pod="openstack/ovn-controller-gxjzh" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.860457 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/38ebd422-35c5-4682-8a4d-ca9073728d7c-etc-ovs\") pod \"ovn-controller-ovs-cqx7r\" (UID: \"38ebd422-35c5-4682-8a4d-ca9073728d7c\") " pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.860531 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/38ebd422-35c5-4682-8a4d-ca9073728d7c-var-lib\") pod \"ovn-controller-ovs-cqx7r\" (UID: \"38ebd422-35c5-4682-8a4d-ca9073728d7c\") " pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.860566 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/220011f2-8778-4a14-82d4-33a07bd33379-ovn-controller-tls-certs\") pod \"ovn-controller-gxjzh\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " pod="openstack/ovn-controller-gxjzh" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.860585 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/220011f2-8778-4a14-82d4-33a07bd33379-combined-ca-bundle\") pod \"ovn-controller-gxjzh\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " pod="openstack/ovn-controller-gxjzh" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.860720 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/220011f2-8778-4a14-82d4-33a07bd33379-var-log-ovn\") pod \"ovn-controller-gxjzh\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " pod="openstack/ovn-controller-gxjzh" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.860884 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvchv\" (UniqueName: \"kubernetes.io/projected/38ebd422-35c5-4682-8a4d-ca9073728d7c-kube-api-access-zvchv\") pod \"ovn-controller-ovs-cqx7r\" (UID: \"38ebd422-35c5-4682-8a4d-ca9073728d7c\") " pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.861050 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/38ebd422-35c5-4682-8a4d-ca9073728d7c-scripts\") pod \"ovn-controller-ovs-cqx7r\" (UID: \"38ebd422-35c5-4682-8a4d-ca9073728d7c\") " pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.861211 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/38ebd422-35c5-4682-8a4d-ca9073728d7c-var-log\") pod \"ovn-controller-ovs-cqx7r\" (UID: \"38ebd422-35c5-4682-8a4d-ca9073728d7c\") " pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.861367 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/220011f2-8778-4a14-82d4-33a07bd33379-var-run-ovn\") pod \"ovn-controller-gxjzh\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " pod="openstack/ovn-controller-gxjzh" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.861418 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sclc\" (UniqueName: \"kubernetes.io/projected/220011f2-8778-4a14-82d4-33a07bd33379-kube-api-access-5sclc\") pod \"ovn-controller-gxjzh\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " pod="openstack/ovn-controller-gxjzh" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.862160 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/220011f2-8778-4a14-82d4-33a07bd33379-scripts\") pod \"ovn-controller-gxjzh\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " pod="openstack/ovn-controller-gxjzh" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.963347 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvchv\" (UniqueName: \"kubernetes.io/projected/38ebd422-35c5-4682-8a4d-ca9073728d7c-kube-api-access-zvchv\") pod \"ovn-controller-ovs-cqx7r\" (UID: \"38ebd422-35c5-4682-8a4d-ca9073728d7c\") " pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.963399 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/38ebd422-35c5-4682-8a4d-ca9073728d7c-scripts\") pod \"ovn-controller-ovs-cqx7r\" (UID: \"38ebd422-35c5-4682-8a4d-ca9073728d7c\") " pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.963430 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/38ebd422-35c5-4682-8a4d-ca9073728d7c-var-log\") pod \"ovn-controller-ovs-cqx7r\" (UID: \"38ebd422-35c5-4682-8a4d-ca9073728d7c\") " pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.963446 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/220011f2-8778-4a14-82d4-33a07bd33379-var-run-ovn\") pod \"ovn-controller-gxjzh\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " pod="openstack/ovn-controller-gxjzh" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.963461 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sclc\" (UniqueName: \"kubernetes.io/projected/220011f2-8778-4a14-82d4-33a07bd33379-kube-api-access-5sclc\") pod \"ovn-controller-gxjzh\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " pod="openstack/ovn-controller-gxjzh" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.963496 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/220011f2-8778-4a14-82d4-33a07bd33379-scripts\") pod \"ovn-controller-gxjzh\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " pod="openstack/ovn-controller-gxjzh" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.963516 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/38ebd422-35c5-4682-8a4d-ca9073728d7c-var-run\") pod \"ovn-controller-ovs-cqx7r\" (UID: \"38ebd422-35c5-4682-8a4d-ca9073728d7c\") " pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.963538 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/220011f2-8778-4a14-82d4-33a07bd33379-var-run\") pod \"ovn-controller-gxjzh\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " pod="openstack/ovn-controller-gxjzh" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.963562 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/38ebd422-35c5-4682-8a4d-ca9073728d7c-etc-ovs\") pod \"ovn-controller-ovs-cqx7r\" (UID: \"38ebd422-35c5-4682-8a4d-ca9073728d7c\") " pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.963604 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/38ebd422-35c5-4682-8a4d-ca9073728d7c-var-lib\") pod \"ovn-controller-ovs-cqx7r\" (UID: \"38ebd422-35c5-4682-8a4d-ca9073728d7c\") " pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.963625 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/220011f2-8778-4a14-82d4-33a07bd33379-ovn-controller-tls-certs\") pod \"ovn-controller-gxjzh\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " pod="openstack/ovn-controller-gxjzh" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.963641 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/220011f2-8778-4a14-82d4-33a07bd33379-combined-ca-bundle\") pod \"ovn-controller-gxjzh\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " pod="openstack/ovn-controller-gxjzh" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.963665 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/220011f2-8778-4a14-82d4-33a07bd33379-var-log-ovn\") pod \"ovn-controller-gxjzh\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " pod="openstack/ovn-controller-gxjzh" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.964197 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/220011f2-8778-4a14-82d4-33a07bd33379-var-run-ovn\") pod \"ovn-controller-gxjzh\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " pod="openstack/ovn-controller-gxjzh" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.964265 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/220011f2-8778-4a14-82d4-33a07bd33379-var-log-ovn\") pod \"ovn-controller-gxjzh\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " pod="openstack/ovn-controller-gxjzh" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.964310 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/220011f2-8778-4a14-82d4-33a07bd33379-var-run\") pod \"ovn-controller-gxjzh\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " pod="openstack/ovn-controller-gxjzh" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.964335 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/38ebd422-35c5-4682-8a4d-ca9073728d7c-var-run\") pod \"ovn-controller-ovs-cqx7r\" (UID: \"38ebd422-35c5-4682-8a4d-ca9073728d7c\") " pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.964405 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/38ebd422-35c5-4682-8a4d-ca9073728d7c-var-log\") pod \"ovn-controller-ovs-cqx7r\" (UID: \"38ebd422-35c5-4682-8a4d-ca9073728d7c\") " pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.964435 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/38ebd422-35c5-4682-8a4d-ca9073728d7c-etc-ovs\") pod \"ovn-controller-ovs-cqx7r\" (UID: \"38ebd422-35c5-4682-8a4d-ca9073728d7c\") " pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.964944 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/38ebd422-35c5-4682-8a4d-ca9073728d7c-var-lib\") pod \"ovn-controller-ovs-cqx7r\" (UID: \"38ebd422-35c5-4682-8a4d-ca9073728d7c\") " pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.966308 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/220011f2-8778-4a14-82d4-33a07bd33379-scripts\") pod \"ovn-controller-gxjzh\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " pod="openstack/ovn-controller-gxjzh" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.966789 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/38ebd422-35c5-4682-8a4d-ca9073728d7c-scripts\") pod \"ovn-controller-ovs-cqx7r\" (UID: \"38ebd422-35c5-4682-8a4d-ca9073728d7c\") " pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.969601 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/220011f2-8778-4a14-82d4-33a07bd33379-combined-ca-bundle\") pod \"ovn-controller-gxjzh\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " pod="openstack/ovn-controller-gxjzh" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.971450 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/220011f2-8778-4a14-82d4-33a07bd33379-ovn-controller-tls-certs\") pod \"ovn-controller-gxjzh\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " pod="openstack/ovn-controller-gxjzh" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.980060 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvchv\" (UniqueName: \"kubernetes.io/projected/38ebd422-35c5-4682-8a4d-ca9073728d7c-kube-api-access-zvchv\") pod \"ovn-controller-ovs-cqx7r\" (UID: \"38ebd422-35c5-4682-8a4d-ca9073728d7c\") " pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:24:41 crc kubenswrapper[4772]: I0127 15:24:41.985352 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sclc\" (UniqueName: \"kubernetes.io/projected/220011f2-8778-4a14-82d4-33a07bd33379-kube-api-access-5sclc\") pod \"ovn-controller-gxjzh\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " pod="openstack/ovn-controller-gxjzh" Jan 27 15:24:42 crc kubenswrapper[4772]: I0127 15:24:42.090006 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gxjzh" Jan 27 15:24:42 crc kubenswrapper[4772]: I0127 15:24:42.108344 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.222713 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.224418 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.226634 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.226639 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.226709 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.228837 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.230022 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-z6lp4" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.237097 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.396381 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.396430 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc34a3a4-ad0b-4154-82c9-728227b19732-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.396455 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc34a3a4-ad0b-4154-82c9-728227b19732-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.396477 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dc34a3a4-ad0b-4154-82c9-728227b19732-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.396527 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8wt7\" (UniqueName: \"kubernetes.io/projected/dc34a3a4-ad0b-4154-82c9-728227b19732-kube-api-access-g8wt7\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.396557 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc34a3a4-ad0b-4154-82c9-728227b19732-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.396577 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc34a3a4-ad0b-4154-82c9-728227b19732-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.396601 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc34a3a4-ad0b-4154-82c9-728227b19732-config\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.498260 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dc34a3a4-ad0b-4154-82c9-728227b19732-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.498367 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8wt7\" (UniqueName: \"kubernetes.io/projected/dc34a3a4-ad0b-4154-82c9-728227b19732-kube-api-access-g8wt7\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.498414 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc34a3a4-ad0b-4154-82c9-728227b19732-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.498446 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc34a3a4-ad0b-4154-82c9-728227b19732-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.498475 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc34a3a4-ad0b-4154-82c9-728227b19732-config\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.498530 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.498549 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc34a3a4-ad0b-4154-82c9-728227b19732-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.498569 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc34a3a4-ad0b-4154-82c9-728227b19732-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.498977 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.499581 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dc34a3a4-ad0b-4154-82c9-728227b19732-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.499916 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc34a3a4-ad0b-4154-82c9-728227b19732-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.500290 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc34a3a4-ad0b-4154-82c9-728227b19732-config\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.503475 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc34a3a4-ad0b-4154-82c9-728227b19732-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.507983 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc34a3a4-ad0b-4154-82c9-728227b19732-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.508917 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc34a3a4-ad0b-4154-82c9-728227b19732-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.517508 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8wt7\" (UniqueName: \"kubernetes.io/projected/dc34a3a4-ad0b-4154-82c9-728227b19732-kube-api-access-g8wt7\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.518949 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:44 crc kubenswrapper[4772]: I0127 15:24:44.567300 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.608627 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.610324 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.612542 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-mvxxf" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.612775 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.612905 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.612948 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.625411 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.720401 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.720460 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-config\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.720578 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.720626 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.720658 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.720800 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.720857 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.720881 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdkbw\" (UniqueName: \"kubernetes.io/projected/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-kube-api-access-vdkbw\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.823315 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.823362 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-config\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.823405 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.823448 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.823466 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.823506 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.823527 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.823544 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdkbw\" (UniqueName: \"kubernetes.io/projected/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-kube-api-access-vdkbw\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.824916 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.825805 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.825984 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.826255 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-config\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.833863 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.835272 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.836902 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.843102 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdkbw\" (UniqueName: \"kubernetes.io/projected/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-kube-api-access-vdkbw\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.855773 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:45 crc kubenswrapper[4772]: I0127 15:24:45.932411 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Jan 27 15:24:46 crc kubenswrapper[4772]: I0127 15:24:46.058872 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"cf619242-7348-4de4-a37e-8ebdc4ca54d7","Type":"ContainerStarted","Data":"9858c0fc9167c8fdb9fe56212a74207375b7ea71449891249cf75618c47eff4b"} Jan 27 15:24:46 crc kubenswrapper[4772]: I0127 15:24:46.060927 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66","Type":"ContainerStarted","Data":"022c0f29ec3ec9ea31194094e372dfed87fe074f880cb471419a54885eeba246"} Jan 27 15:24:58 crc kubenswrapper[4772]: E0127 15:24:58.466694 4772 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Jan 27 15:24:58 crc kubenswrapper[4772]: E0127 15:24:58.467644 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4rxbj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(b1515626-5d79-408d-abc1-cb92abd58f3f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 27 15:24:58 crc kubenswrapper[4772]: E0127 15:24:58.468895 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="b1515626-5d79-408d-abc1-cb92abd58f3f" Jan 27 15:24:59 crc kubenswrapper[4772]: E0127 15:24:59.152787 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="b1515626-5d79-408d-abc1-cb92abd58f3f" Jan 27 15:24:59 crc kubenswrapper[4772]: E0127 15:24:59.488318 4772 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Jan 27 15:24:59 crc kubenswrapper[4772]: E0127 15:24:59.488536 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9gbh9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(76fdbdb1-d48a-4cd1-8372-78887671dce8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 27 15:24:59 crc kubenswrapper[4772]: E0127 15:24:59.489693 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="76fdbdb1-d48a-4cd1-8372-78887671dce8" Jan 27 15:24:59 crc kubenswrapper[4772]: E0127 15:24:59.496048 4772 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Jan 27 15:24:59 crc kubenswrapper[4772]: E0127 15:24:59.496242 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l8h8d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(508c3d5b-212a-46da-9a55-de3f35d7019b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 27 15:24:59 crc kubenswrapper[4772]: E0127 15:24:59.497474 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="508c3d5b-212a-46da-9a55-de3f35d7019b" Jan 27 15:24:59 crc kubenswrapper[4772]: I0127 15:24:59.805120 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gxjzh"] Jan 27 15:25:00 crc kubenswrapper[4772]: E0127 15:25:00.163784 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="508c3d5b-212a-46da-9a55-de3f35d7019b" Jan 27 15:25:00 crc kubenswrapper[4772]: E0127 15:25:00.164064 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="76fdbdb1-d48a-4cd1-8372-78887671dce8" Jan 27 15:25:00 crc kubenswrapper[4772]: E0127 15:25:00.476588 4772 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Jan 27 15:25:00 crc kubenswrapper[4772]: E0127 15:25:00.476876 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xmnqn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-5wcr9_openstack(ece4b345-2aab-4ee3-a116-366d6b8d7bff): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 27 15:25:00 crc kubenswrapper[4772]: E0127 15:25:00.478740 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-5wcr9" podUID="ece4b345-2aab-4ee3-a116-366d6b8d7bff" Jan 27 15:25:01 crc kubenswrapper[4772]: W0127 15:25:01.165014 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod220011f2_8778_4a14_82d4_33a07bd33379.slice/crio-3cb1a1a1b7113cd35f8e36164d72f2c95422afc48122fb52f34747897808a62b WatchSource:0}: Error finding container 3cb1a1a1b7113cd35f8e36164d72f2c95422afc48122fb52f34747897808a62b: Status 404 returned error can't find the container with id 3cb1a1a1b7113cd35f8e36164d72f2c95422afc48122fb52f34747897808a62b Jan 27 15:25:01 crc kubenswrapper[4772]: E0127 15:25:01.174828 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-5wcr9" podUID="ece4b345-2aab-4ee3-a116-366d6b8d7bff" Jan 27 15:25:01 crc kubenswrapper[4772]: E0127 15:25:01.190528 4772 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Jan 27 15:25:01 crc kubenswrapper[4772]: E0127 15:25:01.190734 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2msw8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-bmnxs_openstack(580fdc18-8bdc-4a16-89a4-efd7df1b8a17): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 27 15:25:01 crc kubenswrapper[4772]: E0127 15:25:01.191929 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-bmnxs" podUID="580fdc18-8bdc-4a16-89a4-efd7df1b8a17" Jan 27 15:25:01 crc kubenswrapper[4772]: E0127 15:25:01.217785 4772 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Jan 27 15:25:01 crc kubenswrapper[4772]: E0127 15:25:01.217936 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h6ll7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-t8dxn_openstack(9e884850-6b45-4657-8c8c-fa8ccdec648d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 27 15:25:01 crc kubenswrapper[4772]: E0127 15:25:01.219310 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-t8dxn" podUID="9e884850-6b45-4657-8c8c-fa8ccdec648d" Jan 27 15:25:01 crc kubenswrapper[4772]: E0127 15:25:01.229088 4772 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Jan 27 15:25:01 crc kubenswrapper[4772]: E0127 15:25:01.229948 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lnjbj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-rsdjb_openstack(477337bf-a24a-44fd-9c46-38d2e1566b18): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 27 15:25:01 crc kubenswrapper[4772]: E0127 15:25:01.231248 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-rsdjb" podUID="477337bf-a24a-44fd-9c46-38d2e1566b18" Jan 27 15:25:01 crc kubenswrapper[4772]: I0127 15:25:01.699825 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 27 15:25:01 crc kubenswrapper[4772]: I0127 15:25:01.776400 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 27 15:25:01 crc kubenswrapper[4772]: W0127 15:25:01.787938 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4270ab9b_f4a9_4d48_9cc2_f25152ee5fb2.slice/crio-2d9f9f123f138892540800ef23f48dae96e200e8a0b42b345d3f87addf089f7e WatchSource:0}: Error finding container 2d9f9f123f138892540800ef23f48dae96e200e8a0b42b345d3f87addf089f7e: Status 404 returned error can't find the container with id 2d9f9f123f138892540800ef23f48dae96e200e8a0b42b345d3f87addf089f7e Jan 27 15:25:01 crc kubenswrapper[4772]: I0127 15:25:01.900852 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-cqx7r"] Jan 27 15:25:01 crc kubenswrapper[4772]: W0127 15:25:01.910701 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38ebd422_35c5_4682_8a4d_ca9073728d7c.slice/crio-b4ae3e61c086f91c9c3a7442484ecc85a4bdf545d39601e45239a3351393b9ff WatchSource:0}: Error finding container b4ae3e61c086f91c9c3a7442484ecc85a4bdf545d39601e45239a3351393b9ff: Status 404 returned error can't find the container with id b4ae3e61c086f91c9c3a7442484ecc85a4bdf545d39601e45239a3351393b9ff Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.179001 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1ef66151-0ea7-4696-9db0-7b6665731670","Type":"ContainerStarted","Data":"fabcd309d9b92ca01d4a1240a11210e76a8e365a872f6471e0b9d641c3e1ff39"} Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.180359 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gxjzh" event={"ID":"220011f2-8778-4a14-82d4-33a07bd33379","Type":"ContainerStarted","Data":"3cb1a1a1b7113cd35f8e36164d72f2c95422afc48122fb52f34747897808a62b"} Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.181794 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"cf619242-7348-4de4-a37e-8ebdc4ca54d7","Type":"ContainerStarted","Data":"0d7ac15f647607d8d8b9ab55f639b5ec78749485b0e54cbc048e0727ed5dbce0"} Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.183225 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66","Type":"ContainerStarted","Data":"faf687181014b14838de86572705cbe5952bdabca1b3fad7e35afc3ce6238c0f"} Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.183427 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.184744 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2","Type":"ContainerStarted","Data":"2d9f9f123f138892540800ef23f48dae96e200e8a0b42b345d3f87addf089f7e"} Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.185752 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-cqx7r" event={"ID":"38ebd422-35c5-4682-8a4d-ca9073728d7c","Type":"ContainerStarted","Data":"b4ae3e61c086f91c9c3a7442484ecc85a4bdf545d39601e45239a3351393b9ff"} Jan 27 15:25:02 crc kubenswrapper[4772]: E0127 15:25:02.187139 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-rsdjb" podUID="477337bf-a24a-44fd-9c46-38d2e1566b18" Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.281433 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=10.862511100999999 podStartE2EDuration="26.281414542s" podCreationTimestamp="2026-01-27 15:24:36 +0000 UTC" firstStartedPulling="2026-01-27 15:24:45.836621882 +0000 UTC m=+1071.817230980" lastFinishedPulling="2026-01-27 15:25:01.255525333 +0000 UTC m=+1087.236134421" observedRunningTime="2026-01-27 15:25:02.275698408 +0000 UTC m=+1088.256307506" watchObservedRunningTime="2026-01-27 15:25:02.281414542 +0000 UTC m=+1088.262023640" Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.620475 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-t8dxn" Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.625044 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-bmnxs" Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.772698 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.776213 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6ll7\" (UniqueName: \"kubernetes.io/projected/9e884850-6b45-4657-8c8c-fa8ccdec648d-kube-api-access-h6ll7\") pod \"9e884850-6b45-4657-8c8c-fa8ccdec648d\" (UID: \"9e884850-6b45-4657-8c8c-fa8ccdec648d\") " Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.776278 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e884850-6b45-4657-8c8c-fa8ccdec648d-config\") pod \"9e884850-6b45-4657-8c8c-fa8ccdec648d\" (UID: \"9e884850-6b45-4657-8c8c-fa8ccdec648d\") " Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.776310 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/580fdc18-8bdc-4a16-89a4-efd7df1b8a17-dns-svc\") pod \"580fdc18-8bdc-4a16-89a4-efd7df1b8a17\" (UID: \"580fdc18-8bdc-4a16-89a4-efd7df1b8a17\") " Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.776409 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2msw8\" (UniqueName: \"kubernetes.io/projected/580fdc18-8bdc-4a16-89a4-efd7df1b8a17-kube-api-access-2msw8\") pod \"580fdc18-8bdc-4a16-89a4-efd7df1b8a17\" (UID: \"580fdc18-8bdc-4a16-89a4-efd7df1b8a17\") " Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.776477 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/580fdc18-8bdc-4a16-89a4-efd7df1b8a17-config\") pod \"580fdc18-8bdc-4a16-89a4-efd7df1b8a17\" (UID: \"580fdc18-8bdc-4a16-89a4-efd7df1b8a17\") " Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.776651 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e884850-6b45-4657-8c8c-fa8ccdec648d-config" (OuterVolumeSpecName: "config") pod "9e884850-6b45-4657-8c8c-fa8ccdec648d" (UID: "9e884850-6b45-4657-8c8c-fa8ccdec648d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.776835 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e884850-6b45-4657-8c8c-fa8ccdec648d-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.777152 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/580fdc18-8bdc-4a16-89a4-efd7df1b8a17-config" (OuterVolumeSpecName: "config") pod "580fdc18-8bdc-4a16-89a4-efd7df1b8a17" (UID: "580fdc18-8bdc-4a16-89a4-efd7df1b8a17"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.777701 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/580fdc18-8bdc-4a16-89a4-efd7df1b8a17-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "580fdc18-8bdc-4a16-89a4-efd7df1b8a17" (UID: "580fdc18-8bdc-4a16-89a4-efd7df1b8a17"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.783771 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/580fdc18-8bdc-4a16-89a4-efd7df1b8a17-kube-api-access-2msw8" (OuterVolumeSpecName: "kube-api-access-2msw8") pod "580fdc18-8bdc-4a16-89a4-efd7df1b8a17" (UID: "580fdc18-8bdc-4a16-89a4-efd7df1b8a17"). InnerVolumeSpecName "kube-api-access-2msw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.784348 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e884850-6b45-4657-8c8c-fa8ccdec648d-kube-api-access-h6ll7" (OuterVolumeSpecName: "kube-api-access-h6ll7") pod "9e884850-6b45-4657-8c8c-fa8ccdec648d" (UID: "9e884850-6b45-4657-8c8c-fa8ccdec648d"). InnerVolumeSpecName "kube-api-access-h6ll7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.877980 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2msw8\" (UniqueName: \"kubernetes.io/projected/580fdc18-8bdc-4a16-89a4-efd7df1b8a17-kube-api-access-2msw8\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.878015 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/580fdc18-8bdc-4a16-89a4-efd7df1b8a17-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.878025 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6ll7\" (UniqueName: \"kubernetes.io/projected/9e884850-6b45-4657-8c8c-fa8ccdec648d-kube-api-access-h6ll7\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:02 crc kubenswrapper[4772]: I0127 15:25:02.878034 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/580fdc18-8bdc-4a16-89a4-efd7df1b8a17-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:03 crc kubenswrapper[4772]: W0127 15:25:03.112449 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc34a3a4_ad0b_4154_82c9_728227b19732.slice/crio-3fdbe52ce9493d11baa4cae24955832bd683139a5196f3cc920d2743d2fdf8c2 WatchSource:0}: Error finding container 3fdbe52ce9493d11baa4cae24955832bd683139a5196f3cc920d2743d2fdf8c2: Status 404 returned error can't find the container with id 3fdbe52ce9493d11baa4cae24955832bd683139a5196f3cc920d2743d2fdf8c2 Jan 27 15:25:03 crc kubenswrapper[4772]: I0127 15:25:03.199590 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-t8dxn" event={"ID":"9e884850-6b45-4657-8c8c-fa8ccdec648d","Type":"ContainerDied","Data":"2c9714095ed3f7fd054930137f9e21b4fdd5954c595ecb091f834dc4c0fc5111"} Jan 27 15:25:03 crc kubenswrapper[4772]: I0127 15:25:03.199640 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-t8dxn" Jan 27 15:25:03 crc kubenswrapper[4772]: I0127 15:25:03.202921 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-bmnxs" event={"ID":"580fdc18-8bdc-4a16-89a4-efd7df1b8a17","Type":"ContainerDied","Data":"621419c58a31ba2a4aca778faec82cf28803b219ef85445b2df7f5a1babfe7d1"} Jan 27 15:25:03 crc kubenswrapper[4772]: I0127 15:25:03.202937 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-bmnxs" Jan 27 15:25:03 crc kubenswrapper[4772]: I0127 15:25:03.204281 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"dc34a3a4-ad0b-4154-82c9-728227b19732","Type":"ContainerStarted","Data":"3fdbe52ce9493d11baa4cae24955832bd683139a5196f3cc920d2743d2fdf8c2"} Jan 27 15:25:03 crc kubenswrapper[4772]: I0127 15:25:03.268121 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-t8dxn"] Jan 27 15:25:03 crc kubenswrapper[4772]: I0127 15:25:03.281125 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-t8dxn"] Jan 27 15:25:03 crc kubenswrapper[4772]: I0127 15:25:03.299160 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-bmnxs"] Jan 27 15:25:03 crc kubenswrapper[4772]: I0127 15:25:03.305065 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-bmnxs"] Jan 27 15:25:04 crc kubenswrapper[4772]: I0127 15:25:04.673721 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="580fdc18-8bdc-4a16-89a4-efd7df1b8a17" path="/var/lib/kubelet/pods/580fdc18-8bdc-4a16-89a4-efd7df1b8a17/volumes" Jan 27 15:25:04 crc kubenswrapper[4772]: I0127 15:25:04.674445 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e884850-6b45-4657-8c8c-fa8ccdec648d" path="/var/lib/kubelet/pods/9e884850-6b45-4657-8c8c-fa8ccdec648d/volumes" Jan 27 15:25:05 crc kubenswrapper[4772]: I0127 15:25:05.219492 4772 generic.go:334] "Generic (PLEG): container finished" podID="cf619242-7348-4de4-a37e-8ebdc4ca54d7" containerID="0d7ac15f647607d8d8b9ab55f639b5ec78749485b0e54cbc048e0727ed5dbce0" exitCode=0 Jan 27 15:25:05 crc kubenswrapper[4772]: I0127 15:25:05.219559 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"cf619242-7348-4de4-a37e-8ebdc4ca54d7","Type":"ContainerDied","Data":"0d7ac15f647607d8d8b9ab55f639b5ec78749485b0e54cbc048e0727ed5dbce0"} Jan 27 15:25:06 crc kubenswrapper[4772]: I0127 15:25:06.228549 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gxjzh" event={"ID":"220011f2-8778-4a14-82d4-33a07bd33379","Type":"ContainerStarted","Data":"afc8ab10fea0840566de64c53bc97d22454ee25e120ead660e5999b0da009daf"} Jan 27 15:25:06 crc kubenswrapper[4772]: I0127 15:25:06.229078 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-gxjzh" Jan 27 15:25:06 crc kubenswrapper[4772]: I0127 15:25:06.230528 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"cf619242-7348-4de4-a37e-8ebdc4ca54d7","Type":"ContainerStarted","Data":"2e743dfaa62b788cb68a4d553d64cf9affaf8ef6e4da1308fddf4dc259167b69"} Jan 27 15:25:06 crc kubenswrapper[4772]: I0127 15:25:06.232050 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"dc34a3a4-ad0b-4154-82c9-728227b19732","Type":"ContainerStarted","Data":"abdb27873ea97363386820f9e29ffa55e1d51fea483628db2a98ad8d2f8fc593"} Jan 27 15:25:06 crc kubenswrapper[4772]: I0127 15:25:06.233789 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2","Type":"ContainerStarted","Data":"abb84f069b7ba6556a04c96fbef42abc5bac570c75f402b32a5f9f20ac96046c"} Jan 27 15:25:06 crc kubenswrapper[4772]: I0127 15:25:06.234985 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-cqx7r" event={"ID":"38ebd422-35c5-4682-8a4d-ca9073728d7c","Type":"ContainerStarted","Data":"c2ec2d9ef51a12150ebe6df637e29030ff2b622c19a7ada45c6cd396c44b8636"} Jan 27 15:25:06 crc kubenswrapper[4772]: I0127 15:25:06.236053 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1ef66151-0ea7-4696-9db0-7b6665731670","Type":"ContainerStarted","Data":"e93f9f446173d4fd985d40db28827a7f313c9dbe0522a2d3003fa93c8ac7de5e"} Jan 27 15:25:06 crc kubenswrapper[4772]: I0127 15:25:06.236213 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Jan 27 15:25:06 crc kubenswrapper[4772]: I0127 15:25:06.266630 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=24.074172213 podStartE2EDuration="28.266607566s" podCreationTimestamp="2026-01-27 15:24:38 +0000 UTC" firstStartedPulling="2026-01-27 15:25:01.726247428 +0000 UTC m=+1087.706856526" lastFinishedPulling="2026-01-27 15:25:05.918682781 +0000 UTC m=+1091.899291879" observedRunningTime="2026-01-27 15:25:06.261946682 +0000 UTC m=+1092.242555790" watchObservedRunningTime="2026-01-27 15:25:06.266607566 +0000 UTC m=+1092.247216654" Jan 27 15:25:06 crc kubenswrapper[4772]: I0127 15:25:06.267068 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-gxjzh" podStartSLOduration=21.34043953 podStartE2EDuration="25.267059189s" podCreationTimestamp="2026-01-27 15:24:41 +0000 UTC" firstStartedPulling="2026-01-27 15:25:01.175025458 +0000 UTC m=+1087.155634556" lastFinishedPulling="2026-01-27 15:25:05.101645117 +0000 UTC m=+1091.082254215" observedRunningTime="2026-01-27 15:25:06.249475433 +0000 UTC m=+1092.230084531" watchObservedRunningTime="2026-01-27 15:25:06.267059189 +0000 UTC m=+1092.247668287" Jan 27 15:25:06 crc kubenswrapper[4772]: I0127 15:25:06.299215 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=15.891828915 podStartE2EDuration="31.299196043s" podCreationTimestamp="2026-01-27 15:24:35 +0000 UTC" firstStartedPulling="2026-01-27 15:24:45.836844049 +0000 UTC m=+1071.817453137" lastFinishedPulling="2026-01-27 15:25:01.244211167 +0000 UTC m=+1087.224820265" observedRunningTime="2026-01-27 15:25:06.296615769 +0000 UTC m=+1092.277224897" watchObservedRunningTime="2026-01-27 15:25:06.299196043 +0000 UTC m=+1092.279805141" Jan 27 15:25:06 crc kubenswrapper[4772]: I0127 15:25:06.889650 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Jan 27 15:25:06 crc kubenswrapper[4772]: I0127 15:25:06.889716 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Jan 27 15:25:06 crc kubenswrapper[4772]: I0127 15:25:06.961543 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Jan 27 15:25:07 crc kubenswrapper[4772]: I0127 15:25:07.247400 4772 generic.go:334] "Generic (PLEG): container finished" podID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerID="c2ec2d9ef51a12150ebe6df637e29030ff2b622c19a7ada45c6cd396c44b8636" exitCode=0 Jan 27 15:25:07 crc kubenswrapper[4772]: I0127 15:25:07.247551 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-cqx7r" event={"ID":"38ebd422-35c5-4682-8a4d-ca9073728d7c","Type":"ContainerDied","Data":"c2ec2d9ef51a12150ebe6df637e29030ff2b622c19a7ada45c6cd396c44b8636"} Jan 27 15:25:08 crc kubenswrapper[4772]: I0127 15:25:08.259817 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-cqx7r" event={"ID":"38ebd422-35c5-4682-8a4d-ca9073728d7c","Type":"ContainerStarted","Data":"d6579efc0c6a14eb40a1349e6b5e9e288881435286dc3fff811374b436b48c91"} Jan 27 15:25:08 crc kubenswrapper[4772]: I0127 15:25:08.260229 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-cqx7r" event={"ID":"38ebd422-35c5-4682-8a4d-ca9073728d7c","Type":"ContainerStarted","Data":"4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b"} Jan 27 15:25:08 crc kubenswrapper[4772]: I0127 15:25:08.260265 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:25:08 crc kubenswrapper[4772]: I0127 15:25:08.260283 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:25:08 crc kubenswrapper[4772]: I0127 15:25:08.282960 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-cqx7r" podStartSLOduration=24.094741048 podStartE2EDuration="27.282942575s" podCreationTimestamp="2026-01-27 15:24:41 +0000 UTC" firstStartedPulling="2026-01-27 15:25:01.913176273 +0000 UTC m=+1087.893785371" lastFinishedPulling="2026-01-27 15:25:05.1013778 +0000 UTC m=+1091.081986898" observedRunningTime="2026-01-27 15:25:08.277226331 +0000 UTC m=+1094.257835429" watchObservedRunningTime="2026-01-27 15:25:08.282942575 +0000 UTC m=+1094.263551673" Jan 27 15:25:08 crc kubenswrapper[4772]: I0127 15:25:08.449929 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-5wcr9"] Jan 27 15:25:08 crc kubenswrapper[4772]: I0127 15:25:08.498753 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-62ktv"] Jan 27 15:25:08 crc kubenswrapper[4772]: I0127 15:25:08.500121 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-62ktv" Jan 27 15:25:08 crc kubenswrapper[4772]: I0127 15:25:08.522680 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-62ktv"] Jan 27 15:25:08 crc kubenswrapper[4772]: I0127 15:25:08.577666 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8aead2c0-bb19-4542-8736-67943c23f0c0-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-62ktv\" (UID: \"8aead2c0-bb19-4542-8736-67943c23f0c0\") " pod="openstack/dnsmasq-dns-7cb5889db5-62ktv" Jan 27 15:25:08 crc kubenswrapper[4772]: I0127 15:25:08.577732 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m449k\" (UniqueName: \"kubernetes.io/projected/8aead2c0-bb19-4542-8736-67943c23f0c0-kube-api-access-m449k\") pod \"dnsmasq-dns-7cb5889db5-62ktv\" (UID: \"8aead2c0-bb19-4542-8736-67943c23f0c0\") " pod="openstack/dnsmasq-dns-7cb5889db5-62ktv" Jan 27 15:25:08 crc kubenswrapper[4772]: I0127 15:25:08.577768 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8aead2c0-bb19-4542-8736-67943c23f0c0-config\") pod \"dnsmasq-dns-7cb5889db5-62ktv\" (UID: \"8aead2c0-bb19-4542-8736-67943c23f0c0\") " pod="openstack/dnsmasq-dns-7cb5889db5-62ktv" Jan 27 15:25:08 crc kubenswrapper[4772]: I0127 15:25:08.679643 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8aead2c0-bb19-4542-8736-67943c23f0c0-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-62ktv\" (UID: \"8aead2c0-bb19-4542-8736-67943c23f0c0\") " pod="openstack/dnsmasq-dns-7cb5889db5-62ktv" Jan 27 15:25:08 crc kubenswrapper[4772]: I0127 15:25:08.679779 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m449k\" (UniqueName: \"kubernetes.io/projected/8aead2c0-bb19-4542-8736-67943c23f0c0-kube-api-access-m449k\") pod \"dnsmasq-dns-7cb5889db5-62ktv\" (UID: \"8aead2c0-bb19-4542-8736-67943c23f0c0\") " pod="openstack/dnsmasq-dns-7cb5889db5-62ktv" Jan 27 15:25:08 crc kubenswrapper[4772]: I0127 15:25:08.680115 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8aead2c0-bb19-4542-8736-67943c23f0c0-config\") pod \"dnsmasq-dns-7cb5889db5-62ktv\" (UID: \"8aead2c0-bb19-4542-8736-67943c23f0c0\") " pod="openstack/dnsmasq-dns-7cb5889db5-62ktv" Jan 27 15:25:08 crc kubenswrapper[4772]: I0127 15:25:08.681158 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8aead2c0-bb19-4542-8736-67943c23f0c0-config\") pod \"dnsmasq-dns-7cb5889db5-62ktv\" (UID: \"8aead2c0-bb19-4542-8736-67943c23f0c0\") " pod="openstack/dnsmasq-dns-7cb5889db5-62ktv" Jan 27 15:25:08 crc kubenswrapper[4772]: I0127 15:25:08.681848 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8aead2c0-bb19-4542-8736-67943c23f0c0-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-62ktv\" (UID: \"8aead2c0-bb19-4542-8736-67943c23f0c0\") " pod="openstack/dnsmasq-dns-7cb5889db5-62ktv" Jan 27 15:25:08 crc kubenswrapper[4772]: I0127 15:25:08.701778 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m449k\" (UniqueName: \"kubernetes.io/projected/8aead2c0-bb19-4542-8736-67943c23f0c0-kube-api-access-m449k\") pod \"dnsmasq-dns-7cb5889db5-62ktv\" (UID: \"8aead2c0-bb19-4542-8736-67943c23f0c0\") " pod="openstack/dnsmasq-dns-7cb5889db5-62ktv" Jan 27 15:25:08 crc kubenswrapper[4772]: I0127 15:25:08.826621 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-62ktv" Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.620860 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.627526 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.630071 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.630379 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.630601 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-g7gch" Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.630857 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.655518 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.701377 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-etc-swift\") pod \"swift-storage-0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " pod="openstack/swift-storage-0" Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.701456 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/3ef68955-b80c-4732-9e87-0bec53d0b3a0-lock\") pod \"swift-storage-0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " pod="openstack/swift-storage-0" Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.701548 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxlv4\" (UniqueName: \"kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-kube-api-access-mxlv4\") pod \"swift-storage-0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " pod="openstack/swift-storage-0" Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.701708 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " pod="openstack/swift-storage-0" Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.701767 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/3ef68955-b80c-4732-9e87-0bec53d0b3a0-cache\") pod \"swift-storage-0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " pod="openstack/swift-storage-0" Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.701798 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ef68955-b80c-4732-9e87-0bec53d0b3a0-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " pod="openstack/swift-storage-0" Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.802805 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " pod="openstack/swift-storage-0" Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.802865 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/3ef68955-b80c-4732-9e87-0bec53d0b3a0-cache\") pod \"swift-storage-0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " pod="openstack/swift-storage-0" Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.802891 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ef68955-b80c-4732-9e87-0bec53d0b3a0-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " pod="openstack/swift-storage-0" Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.802990 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-etc-swift\") pod \"swift-storage-0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " pod="openstack/swift-storage-0" Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.803018 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/3ef68955-b80c-4732-9e87-0bec53d0b3a0-lock\") pod \"swift-storage-0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " pod="openstack/swift-storage-0" Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.803080 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxlv4\" (UniqueName: \"kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-kube-api-access-mxlv4\") pod \"swift-storage-0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " pod="openstack/swift-storage-0" Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.803286 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/swift-storage-0" Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.804148 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/3ef68955-b80c-4732-9e87-0bec53d0b3a0-lock\") pod \"swift-storage-0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " pod="openstack/swift-storage-0" Jan 27 15:25:09 crc kubenswrapper[4772]: E0127 15:25:09.804279 4772 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 27 15:25:09 crc kubenswrapper[4772]: E0127 15:25:09.804300 4772 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 27 15:25:09 crc kubenswrapper[4772]: E0127 15:25:09.804352 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-etc-swift podName:3ef68955-b80c-4732-9e87-0bec53d0b3a0 nodeName:}" failed. No retries permitted until 2026-01-27 15:25:10.304330393 +0000 UTC m=+1096.284939491 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-etc-swift") pod "swift-storage-0" (UID: "3ef68955-b80c-4732-9e87-0bec53d0b3a0") : configmap "swift-ring-files" not found Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.804851 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/3ef68955-b80c-4732-9e87-0bec53d0b3a0-cache\") pod \"swift-storage-0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " pod="openstack/swift-storage-0" Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.822178 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ef68955-b80c-4732-9e87-0bec53d0b3a0-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " pod="openstack/swift-storage-0" Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.822247 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxlv4\" (UniqueName: \"kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-kube-api-access-mxlv4\") pod \"swift-storage-0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " pod="openstack/swift-storage-0" Jan 27 15:25:09 crc kubenswrapper[4772]: I0127 15:25:09.834661 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " pod="openstack/swift-storage-0" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.025761 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-5wcr9" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.107601 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ece4b345-2aab-4ee3-a116-366d6b8d7bff-dns-svc\") pod \"ece4b345-2aab-4ee3-a116-366d6b8d7bff\" (UID: \"ece4b345-2aab-4ee3-a116-366d6b8d7bff\") " Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.107657 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmnqn\" (UniqueName: \"kubernetes.io/projected/ece4b345-2aab-4ee3-a116-366d6b8d7bff-kube-api-access-xmnqn\") pod \"ece4b345-2aab-4ee3-a116-366d6b8d7bff\" (UID: \"ece4b345-2aab-4ee3-a116-366d6b8d7bff\") " Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.107764 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece4b345-2aab-4ee3-a116-366d6b8d7bff-config\") pod \"ece4b345-2aab-4ee3-a116-366d6b8d7bff\" (UID: \"ece4b345-2aab-4ee3-a116-366d6b8d7bff\") " Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.111954 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ece4b345-2aab-4ee3-a116-366d6b8d7bff-config" (OuterVolumeSpecName: "config") pod "ece4b345-2aab-4ee3-a116-366d6b8d7bff" (UID: "ece4b345-2aab-4ee3-a116-366d6b8d7bff"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.112423 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ece4b345-2aab-4ee3-a116-366d6b8d7bff-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ece4b345-2aab-4ee3-a116-366d6b8d7bff" (UID: "ece4b345-2aab-4ee3-a116-366d6b8d7bff"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.113350 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ece4b345-2aab-4ee3-a116-366d6b8d7bff-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.113371 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ece4b345-2aab-4ee3-a116-366d6b8d7bff-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.117556 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ece4b345-2aab-4ee3-a116-366d6b8d7bff-kube-api-access-xmnqn" (OuterVolumeSpecName: "kube-api-access-xmnqn") pod "ece4b345-2aab-4ee3-a116-366d6b8d7bff" (UID: "ece4b345-2aab-4ee3-a116-366d6b8d7bff"). InnerVolumeSpecName "kube-api-access-xmnqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.201588 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-d4llz"] Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.203098 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.208425 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.211980 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.212183 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.215572 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-d4llz"] Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.215580 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmnqn\" (UniqueName: \"kubernetes.io/projected/ece4b345-2aab-4ee3-a116-366d6b8d7bff-kube-api-access-xmnqn\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.272929 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-5wcr9" event={"ID":"ece4b345-2aab-4ee3-a116-366d6b8d7bff","Type":"ContainerDied","Data":"69b728c61ccbcb54532eee4001562892bf3d12d85c2e89da36c5a93f0401b107"} Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.272942 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-5wcr9" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.278199 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"dc34a3a4-ad0b-4154-82c9-728227b19732","Type":"ContainerStarted","Data":"fc9316436d0d6826797760ceeb255662e4eca1649200864d00e921be6f2e6eed"} Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.283244 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2","Type":"ContainerStarted","Data":"9ab2ac6bce7a8071ec2b4cecbc76933f6c63344bca73557900280dd89a9b1ef4"} Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.295157 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=20.431232725 podStartE2EDuration="27.295134036s" podCreationTimestamp="2026-01-27 15:24:43 +0000 UTC" firstStartedPulling="2026-01-27 15:25:03.114376344 +0000 UTC m=+1089.094985442" lastFinishedPulling="2026-01-27 15:25:09.978277655 +0000 UTC m=+1095.958886753" observedRunningTime="2026-01-27 15:25:10.29492261 +0000 UTC m=+1096.275531708" watchObservedRunningTime="2026-01-27 15:25:10.295134036 +0000 UTC m=+1096.275743134" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.315419 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=18.127424414 podStartE2EDuration="26.315400069s" podCreationTimestamp="2026-01-27 15:24:44 +0000 UTC" firstStartedPulling="2026-01-27 15:25:01.790571558 +0000 UTC m=+1087.771180656" lastFinishedPulling="2026-01-27 15:25:09.978547213 +0000 UTC m=+1095.959156311" observedRunningTime="2026-01-27 15:25:10.312557167 +0000 UTC m=+1096.293166265" watchObservedRunningTime="2026-01-27 15:25:10.315400069 +0000 UTC m=+1096.296009167" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.316585 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-dispersionconf\") pod \"swift-ring-rebalance-d4llz\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.316634 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-ring-data-devices\") pod \"swift-ring-rebalance-d4llz\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.316753 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-combined-ca-bundle\") pod \"swift-ring-rebalance-d4llz\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.316903 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-swiftconf\") pod \"swift-ring-rebalance-d4llz\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.317026 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-etc-swift\") pod \"swift-ring-rebalance-d4llz\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.317067 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvbq9\" (UniqueName: \"kubernetes.io/projected/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-kube-api-access-mvbq9\") pod \"swift-ring-rebalance-d4llz\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.317102 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-scripts\") pod \"swift-ring-rebalance-d4llz\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.317133 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-etc-swift\") pod \"swift-storage-0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " pod="openstack/swift-storage-0" Jan 27 15:25:10 crc kubenswrapper[4772]: E0127 15:25:10.317372 4772 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 27 15:25:10 crc kubenswrapper[4772]: E0127 15:25:10.317393 4772 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 27 15:25:10 crc kubenswrapper[4772]: E0127 15:25:10.317432 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-etc-swift podName:3ef68955-b80c-4732-9e87-0bec53d0b3a0 nodeName:}" failed. No retries permitted until 2026-01-27 15:25:11.317418117 +0000 UTC m=+1097.298027215 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-etc-swift") pod "swift-storage-0" (UID: "3ef68955-b80c-4732-9e87-0bec53d0b3a0") : configmap "swift-ring-files" not found Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.351280 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-5wcr9"] Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.358475 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-5wcr9"] Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.398725 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-62ktv"] Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.419060 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-combined-ca-bundle\") pod \"swift-ring-rebalance-d4llz\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.419124 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-swiftconf\") pod \"swift-ring-rebalance-d4llz\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.419209 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-etc-swift\") pod \"swift-ring-rebalance-d4llz\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.419231 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvbq9\" (UniqueName: \"kubernetes.io/projected/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-kube-api-access-mvbq9\") pod \"swift-ring-rebalance-d4llz\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.420101 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-scripts\") pod \"swift-ring-rebalance-d4llz\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.419723 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-etc-swift\") pod \"swift-ring-rebalance-d4llz\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.420277 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-dispersionconf\") pod \"swift-ring-rebalance-d4llz\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.420341 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-ring-data-devices\") pod \"swift-ring-rebalance-d4llz\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.420704 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-scripts\") pod \"swift-ring-rebalance-d4llz\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.420986 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-ring-data-devices\") pod \"swift-ring-rebalance-d4llz\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.426642 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-swiftconf\") pod \"swift-ring-rebalance-d4llz\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.428918 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-combined-ca-bundle\") pod \"swift-ring-rebalance-d4llz\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.431302 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-dispersionconf\") pod \"swift-ring-rebalance-d4llz\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.438825 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvbq9\" (UniqueName: \"kubernetes.io/projected/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-kube-api-access-mvbq9\") pod \"swift-ring-rebalance-d4llz\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.529193 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.675526 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ece4b345-2aab-4ee3-a116-366d6b8d7bff" path="/var/lib/kubelet/pods/ece4b345-2aab-4ee3-a116-366d6b8d7bff/volumes" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.933041 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Jan 27 15:25:10 crc kubenswrapper[4772]: I0127 15:25:10.975017 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-d4llz"] Jan 27 15:25:10 crc kubenswrapper[4772]: W0127 15:25:10.997109 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2313c291_4eb5_4b79_ad9b_b04cd06a1ee9.slice/crio-1fe7af5e79ef4da92cef0b65e4f6ca7d519d839fcd9321afbd7bf43485be0f39 WatchSource:0}: Error finding container 1fe7af5e79ef4da92cef0b65e4f6ca7d519d839fcd9321afbd7bf43485be0f39: Status 404 returned error can't find the container with id 1fe7af5e79ef4da92cef0b65e4f6ca7d519d839fcd9321afbd7bf43485be0f39 Jan 27 15:25:11 crc kubenswrapper[4772]: I0127 15:25:11.010335 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Jan 27 15:25:11 crc kubenswrapper[4772]: I0127 15:25:11.097423 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Jan 27 15:25:11 crc kubenswrapper[4772]: I0127 15:25:11.290399 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b1515626-5d79-408d-abc1-cb92abd58f3f","Type":"ContainerStarted","Data":"3f38ceb1ab131833479b9e418df05230599249523d60e8e11929add232b03e39"} Jan 27 15:25:11 crc kubenswrapper[4772]: I0127 15:25:11.291906 4772 generic.go:334] "Generic (PLEG): container finished" podID="8aead2c0-bb19-4542-8736-67943c23f0c0" containerID="932ddd9664bf5058afe13fff46657f7b87008d96d4495fb6b027b04295f1af73" exitCode=0 Jan 27 15:25:11 crc kubenswrapper[4772]: I0127 15:25:11.291940 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-62ktv" event={"ID":"8aead2c0-bb19-4542-8736-67943c23f0c0","Type":"ContainerDied","Data":"932ddd9664bf5058afe13fff46657f7b87008d96d4495fb6b027b04295f1af73"} Jan 27 15:25:11 crc kubenswrapper[4772]: I0127 15:25:11.291970 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-62ktv" event={"ID":"8aead2c0-bb19-4542-8736-67943c23f0c0","Type":"ContainerStarted","Data":"80889f4358a04bf7fac97ea1ad80ad55c76754b98341035087589aff0399485d"} Jan 27 15:25:11 crc kubenswrapper[4772]: I0127 15:25:11.293128 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-d4llz" event={"ID":"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9","Type":"ContainerStarted","Data":"1fe7af5e79ef4da92cef0b65e4f6ca7d519d839fcd9321afbd7bf43485be0f39"} Jan 27 15:25:11 crc kubenswrapper[4772]: I0127 15:25:11.337297 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-etc-swift\") pod \"swift-storage-0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " pod="openstack/swift-storage-0" Jan 27 15:25:11 crc kubenswrapper[4772]: E0127 15:25:11.337793 4772 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 27 15:25:11 crc kubenswrapper[4772]: E0127 15:25:11.337822 4772 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 27 15:25:11 crc kubenswrapper[4772]: E0127 15:25:11.337859 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-etc-swift podName:3ef68955-b80c-4732-9e87-0bec53d0b3a0 nodeName:}" failed. No retries permitted until 2026-01-27 15:25:13.337844118 +0000 UTC m=+1099.318453216 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-etc-swift") pod "swift-storage-0" (UID: "3ef68955-b80c-4732-9e87-0bec53d0b3a0") : configmap "swift-ring-files" not found Jan 27 15:25:11 crc kubenswrapper[4772]: I0127 15:25:11.567947 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Jan 27 15:25:11 crc kubenswrapper[4772]: I0127 15:25:11.607702 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.058934 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.059049 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.300152 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-62ktv" event={"ID":"8aead2c0-bb19-4542-8736-67943c23f0c0","Type":"ContainerStarted","Data":"80133493ee73cefd49c95662c6d25ef4d357fe61d09a830267d9876af454f579"} Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.300687 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.320895 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb5889db5-62ktv" podStartSLOduration=3.719444193 podStartE2EDuration="4.320880106s" podCreationTimestamp="2026-01-27 15:25:08 +0000 UTC" firstStartedPulling="2026-01-27 15:25:10.398788127 +0000 UTC m=+1096.379397225" lastFinishedPulling="2026-01-27 15:25:11.00022401 +0000 UTC m=+1096.980833138" observedRunningTime="2026-01-27 15:25:12.318157117 +0000 UTC m=+1098.298766205" watchObservedRunningTime="2026-01-27 15:25:12.320880106 +0000 UTC m=+1098.301489204" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.343419 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.612127 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-rsdjb"] Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.632524 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-vrz8v"] Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.633834 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.636075 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.647152 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-vrz8v"] Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.686187 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-vqpfg"] Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.689128 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-vqpfg"] Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.689312 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-vqpfg" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.692030 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.763534 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t22gz\" (UniqueName: \"kubernetes.io/projected/003a41cd-8661-4d0a-a5b7-4e06e02d3785-kube-api-access-t22gz\") pod \"dnsmasq-dns-6c89d5d749-vrz8v\" (UID: \"003a41cd-8661-4d0a-a5b7-4e06e02d3785\") " pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.763662 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/003a41cd-8661-4d0a-a5b7-4e06e02d3785-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-vrz8v\" (UID: \"003a41cd-8661-4d0a-a5b7-4e06e02d3785\") " pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.763713 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/003a41cd-8661-4d0a-a5b7-4e06e02d3785-config\") pod \"dnsmasq-dns-6c89d5d749-vrz8v\" (UID: \"003a41cd-8661-4d0a-a5b7-4e06e02d3785\") " pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.763812 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/003a41cd-8661-4d0a-a5b7-4e06e02d3785-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-vrz8v\" (UID: \"003a41cd-8661-4d0a-a5b7-4e06e02d3785\") " pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.822142 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-62ktv"] Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.853222 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-tltm6"] Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.863268 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-tltm6" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.864762 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a490a71b-c33d-4c94-9592-f97d1d315e81-combined-ca-bundle\") pod \"ovn-controller-metrics-vqpfg\" (UID: \"a490a71b-c33d-4c94-9592-f97d1d315e81\") " pod="openstack/ovn-controller-metrics-vqpfg" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.864804 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t22gz\" (UniqueName: \"kubernetes.io/projected/003a41cd-8661-4d0a-a5b7-4e06e02d3785-kube-api-access-t22gz\") pod \"dnsmasq-dns-6c89d5d749-vrz8v\" (UID: \"003a41cd-8661-4d0a-a5b7-4e06e02d3785\") " pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.864825 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a490a71b-c33d-4c94-9592-f97d1d315e81-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vqpfg\" (UID: \"a490a71b-c33d-4c94-9592-f97d1d315e81\") " pod="openstack/ovn-controller-metrics-vqpfg" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.864844 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a490a71b-c33d-4c94-9592-f97d1d315e81-ovs-rundir\") pod \"ovn-controller-metrics-vqpfg\" (UID: \"a490a71b-c33d-4c94-9592-f97d1d315e81\") " pod="openstack/ovn-controller-metrics-vqpfg" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.864865 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmhcm\" (UniqueName: \"kubernetes.io/projected/a490a71b-c33d-4c94-9592-f97d1d315e81-kube-api-access-dmhcm\") pod \"ovn-controller-metrics-vqpfg\" (UID: \"a490a71b-c33d-4c94-9592-f97d1d315e81\") " pod="openstack/ovn-controller-metrics-vqpfg" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.864901 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/003a41cd-8661-4d0a-a5b7-4e06e02d3785-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-vrz8v\" (UID: \"003a41cd-8661-4d0a-a5b7-4e06e02d3785\") " pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.864935 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/003a41cd-8661-4d0a-a5b7-4e06e02d3785-config\") pod \"dnsmasq-dns-6c89d5d749-vrz8v\" (UID: \"003a41cd-8661-4d0a-a5b7-4e06e02d3785\") " pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.864957 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a490a71b-c33d-4c94-9592-f97d1d315e81-ovn-rundir\") pod \"ovn-controller-metrics-vqpfg\" (UID: \"a490a71b-c33d-4c94-9592-f97d1d315e81\") " pod="openstack/ovn-controller-metrics-vqpfg" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.864987 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a490a71b-c33d-4c94-9592-f97d1d315e81-config\") pod \"ovn-controller-metrics-vqpfg\" (UID: \"a490a71b-c33d-4c94-9592-f97d1d315e81\") " pod="openstack/ovn-controller-metrics-vqpfg" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.865013 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/003a41cd-8661-4d0a-a5b7-4e06e02d3785-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-vrz8v\" (UID: \"003a41cd-8661-4d0a-a5b7-4e06e02d3785\") " pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.865833 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/003a41cd-8661-4d0a-a5b7-4e06e02d3785-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-vrz8v\" (UID: \"003a41cd-8661-4d0a-a5b7-4e06e02d3785\") " pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.865836 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.866004 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tltm6"] Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.866073 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/003a41cd-8661-4d0a-a5b7-4e06e02d3785-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-vrz8v\" (UID: \"003a41cd-8661-4d0a-a5b7-4e06e02d3785\") " pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.866855 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/003a41cd-8661-4d0a-a5b7-4e06e02d3785-config\") pod \"dnsmasq-dns-6c89d5d749-vrz8v\" (UID: \"003a41cd-8661-4d0a-a5b7-4e06e02d3785\") " pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.905825 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t22gz\" (UniqueName: \"kubernetes.io/projected/003a41cd-8661-4d0a-a5b7-4e06e02d3785-kube-api-access-t22gz\") pod \"dnsmasq-dns-6c89d5d749-vrz8v\" (UID: \"003a41cd-8661-4d0a-a5b7-4e06e02d3785\") " pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.933383 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.959076 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.966182 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a490a71b-c33d-4c94-9592-f97d1d315e81-ovn-rundir\") pod \"ovn-controller-metrics-vqpfg\" (UID: \"a490a71b-c33d-4c94-9592-f97d1d315e81\") " pod="openstack/ovn-controller-metrics-vqpfg" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.966290 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-tltm6\" (UID: \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\") " pod="openstack/dnsmasq-dns-698758b865-tltm6" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.966324 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a490a71b-c33d-4c94-9592-f97d1d315e81-config\") pod \"ovn-controller-metrics-vqpfg\" (UID: \"a490a71b-c33d-4c94-9592-f97d1d315e81\") " pod="openstack/ovn-controller-metrics-vqpfg" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.966356 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-dns-svc\") pod \"dnsmasq-dns-698758b865-tltm6\" (UID: \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\") " pod="openstack/dnsmasq-dns-698758b865-tltm6" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.966396 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-config\") pod \"dnsmasq-dns-698758b865-tltm6\" (UID: \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\") " pod="openstack/dnsmasq-dns-698758b865-tltm6" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.966428 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a490a71b-c33d-4c94-9592-f97d1d315e81-combined-ca-bundle\") pod \"ovn-controller-metrics-vqpfg\" (UID: \"a490a71b-c33d-4c94-9592-f97d1d315e81\") " pod="openstack/ovn-controller-metrics-vqpfg" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.966449 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a490a71b-c33d-4c94-9592-f97d1d315e81-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vqpfg\" (UID: \"a490a71b-c33d-4c94-9592-f97d1d315e81\") " pod="openstack/ovn-controller-metrics-vqpfg" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.966468 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a490a71b-c33d-4c94-9592-f97d1d315e81-ovs-rundir\") pod \"ovn-controller-metrics-vqpfg\" (UID: \"a490a71b-c33d-4c94-9592-f97d1d315e81\") " pod="openstack/ovn-controller-metrics-vqpfg" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.966486 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmhcm\" (UniqueName: \"kubernetes.io/projected/a490a71b-c33d-4c94-9592-f97d1d315e81-kube-api-access-dmhcm\") pod \"ovn-controller-metrics-vqpfg\" (UID: \"a490a71b-c33d-4c94-9592-f97d1d315e81\") " pod="openstack/ovn-controller-metrics-vqpfg" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.966513 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-tltm6\" (UID: \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\") " pod="openstack/dnsmasq-dns-698758b865-tltm6" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.966531 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2pzt\" (UniqueName: \"kubernetes.io/projected/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-kube-api-access-p2pzt\") pod \"dnsmasq-dns-698758b865-tltm6\" (UID: \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\") " pod="openstack/dnsmasq-dns-698758b865-tltm6" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.966553 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a490a71b-c33d-4c94-9592-f97d1d315e81-ovn-rundir\") pod \"ovn-controller-metrics-vqpfg\" (UID: \"a490a71b-c33d-4c94-9592-f97d1d315e81\") " pod="openstack/ovn-controller-metrics-vqpfg" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.967291 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a490a71b-c33d-4c94-9592-f97d1d315e81-ovs-rundir\") pod \"ovn-controller-metrics-vqpfg\" (UID: \"a490a71b-c33d-4c94-9592-f97d1d315e81\") " pod="openstack/ovn-controller-metrics-vqpfg" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.967681 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a490a71b-c33d-4c94-9592-f97d1d315e81-config\") pod \"ovn-controller-metrics-vqpfg\" (UID: \"a490a71b-c33d-4c94-9592-f97d1d315e81\") " pod="openstack/ovn-controller-metrics-vqpfg" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.970302 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a490a71b-c33d-4c94-9592-f97d1d315e81-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vqpfg\" (UID: \"a490a71b-c33d-4c94-9592-f97d1d315e81\") " pod="openstack/ovn-controller-metrics-vqpfg" Jan 27 15:25:12 crc kubenswrapper[4772]: I0127 15:25:12.970424 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a490a71b-c33d-4c94-9592-f97d1d315e81-combined-ca-bundle\") pod \"ovn-controller-metrics-vqpfg\" (UID: \"a490a71b-c33d-4c94-9592-f97d1d315e81\") " pod="openstack/ovn-controller-metrics-vqpfg" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.029004 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.052789 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmhcm\" (UniqueName: \"kubernetes.io/projected/a490a71b-c33d-4c94-9592-f97d1d315e81-kube-api-access-dmhcm\") pod \"ovn-controller-metrics-vqpfg\" (UID: \"a490a71b-c33d-4c94-9592-f97d1d315e81\") " pod="openstack/ovn-controller-metrics-vqpfg" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.077057 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-dns-svc\") pod \"dnsmasq-dns-698758b865-tltm6\" (UID: \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\") " pod="openstack/dnsmasq-dns-698758b865-tltm6" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.077121 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-config\") pod \"dnsmasq-dns-698758b865-tltm6\" (UID: \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\") " pod="openstack/dnsmasq-dns-698758b865-tltm6" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.077191 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-tltm6\" (UID: \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\") " pod="openstack/dnsmasq-dns-698758b865-tltm6" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.077212 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2pzt\" (UniqueName: \"kubernetes.io/projected/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-kube-api-access-p2pzt\") pod \"dnsmasq-dns-698758b865-tltm6\" (UID: \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\") " pod="openstack/dnsmasq-dns-698758b865-tltm6" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.077261 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-tltm6\" (UID: \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\") " pod="openstack/dnsmasq-dns-698758b865-tltm6" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.077976 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-tltm6\" (UID: \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\") " pod="openstack/dnsmasq-dns-698758b865-tltm6" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.078475 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-dns-svc\") pod \"dnsmasq-dns-698758b865-tltm6\" (UID: \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\") " pod="openstack/dnsmasq-dns-698758b865-tltm6" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.078988 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-config\") pod \"dnsmasq-dns-698758b865-tltm6\" (UID: \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\") " pod="openstack/dnsmasq-dns-698758b865-tltm6" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.079495 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-tltm6\" (UID: \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\") " pod="openstack/dnsmasq-dns-698758b865-tltm6" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.116602 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2pzt\" (UniqueName: \"kubernetes.io/projected/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-kube-api-access-p2pzt\") pod \"dnsmasq-dns-698758b865-tltm6\" (UID: \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\") " pod="openstack/dnsmasq-dns-698758b865-tltm6" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.201225 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-tltm6" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.308455 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb5889db5-62ktv" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.326655 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-vqpfg" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.347230 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.381319 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-etc-swift\") pod \"swift-storage-0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " pod="openstack/swift-storage-0" Jan 27 15:25:13 crc kubenswrapper[4772]: E0127 15:25:13.382375 4772 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 27 15:25:13 crc kubenswrapper[4772]: E0127 15:25:13.382404 4772 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 27 15:25:13 crc kubenswrapper[4772]: E0127 15:25:13.382461 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-etc-swift podName:3ef68955-b80c-4732-9e87-0bec53d0b3a0 nodeName:}" failed. No retries permitted until 2026-01-27 15:25:17.382441331 +0000 UTC m=+1103.363050449 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-etc-swift") pod "swift-storage-0" (UID: "3ef68955-b80c-4732-9e87-0bec53d0b3a0") : configmap "swift-ring-files" not found Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.515849 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.518308 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.521342 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.521561 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-frj8c" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.521722 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.521876 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.528980 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.583900 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " pod="openstack/ovn-northd-0" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.583945 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " pod="openstack/ovn-northd-0" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.583969 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-429qb\" (UniqueName: \"kubernetes.io/projected/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-kube-api-access-429qb\") pod \"ovn-northd-0\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " pod="openstack/ovn-northd-0" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.584040 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-scripts\") pod \"ovn-northd-0\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " pod="openstack/ovn-northd-0" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.584069 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " pod="openstack/ovn-northd-0" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.584111 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-config\") pod \"ovn-northd-0\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " pod="openstack/ovn-northd-0" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.584129 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " pod="openstack/ovn-northd-0" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.685630 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " pod="openstack/ovn-northd-0" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.685684 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " pod="openstack/ovn-northd-0" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.685714 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-429qb\" (UniqueName: \"kubernetes.io/projected/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-kube-api-access-429qb\") pod \"ovn-northd-0\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " pod="openstack/ovn-northd-0" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.685811 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-scripts\") pod \"ovn-northd-0\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " pod="openstack/ovn-northd-0" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.685851 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " pod="openstack/ovn-northd-0" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.685914 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-config\") pod \"ovn-northd-0\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " pod="openstack/ovn-northd-0" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.685938 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " pod="openstack/ovn-northd-0" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.686029 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " pod="openstack/ovn-northd-0" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.686663 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-scripts\") pod \"ovn-northd-0\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " pod="openstack/ovn-northd-0" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.690815 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " pod="openstack/ovn-northd-0" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.691556 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " pod="openstack/ovn-northd-0" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.691719 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-config\") pod \"ovn-northd-0\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " pod="openstack/ovn-northd-0" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.691783 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " pod="openstack/ovn-northd-0" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.703109 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-429qb\" (UniqueName: \"kubernetes.io/projected/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-kube-api-access-429qb\") pod \"ovn-northd-0\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " pod="openstack/ovn-northd-0" Jan 27 15:25:13 crc kubenswrapper[4772]: I0127 15:25:13.839708 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Jan 27 15:25:14 crc kubenswrapper[4772]: I0127 15:25:14.316105 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb5889db5-62ktv" podUID="8aead2c0-bb19-4542-8736-67943c23f0c0" containerName="dnsmasq-dns" containerID="cri-o://80133493ee73cefd49c95662c6d25ef4d357fe61d09a830267d9876af454f579" gracePeriod=10 Jan 27 15:25:14 crc kubenswrapper[4772]: I0127 15:25:14.716673 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-rsdjb" Jan 27 15:25:14 crc kubenswrapper[4772]: I0127 15:25:14.816008 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/477337bf-a24a-44fd-9c46-38d2e1566b18-config\") pod \"477337bf-a24a-44fd-9c46-38d2e1566b18\" (UID: \"477337bf-a24a-44fd-9c46-38d2e1566b18\") " Jan 27 15:25:14 crc kubenswrapper[4772]: I0127 15:25:14.816375 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnjbj\" (UniqueName: \"kubernetes.io/projected/477337bf-a24a-44fd-9c46-38d2e1566b18-kube-api-access-lnjbj\") pod \"477337bf-a24a-44fd-9c46-38d2e1566b18\" (UID: \"477337bf-a24a-44fd-9c46-38d2e1566b18\") " Jan 27 15:25:14 crc kubenswrapper[4772]: I0127 15:25:14.816412 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/477337bf-a24a-44fd-9c46-38d2e1566b18-dns-svc\") pod \"477337bf-a24a-44fd-9c46-38d2e1566b18\" (UID: \"477337bf-a24a-44fd-9c46-38d2e1566b18\") " Jan 27 15:25:14 crc kubenswrapper[4772]: I0127 15:25:14.820442 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/477337bf-a24a-44fd-9c46-38d2e1566b18-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "477337bf-a24a-44fd-9c46-38d2e1566b18" (UID: "477337bf-a24a-44fd-9c46-38d2e1566b18"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:14 crc kubenswrapper[4772]: I0127 15:25:14.820730 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/477337bf-a24a-44fd-9c46-38d2e1566b18-config" (OuterVolumeSpecName: "config") pod "477337bf-a24a-44fd-9c46-38d2e1566b18" (UID: "477337bf-a24a-44fd-9c46-38d2e1566b18"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:14 crc kubenswrapper[4772]: I0127 15:25:14.827277 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/477337bf-a24a-44fd-9c46-38d2e1566b18-kube-api-access-lnjbj" (OuterVolumeSpecName: "kube-api-access-lnjbj") pod "477337bf-a24a-44fd-9c46-38d2e1566b18" (UID: "477337bf-a24a-44fd-9c46-38d2e1566b18"). InnerVolumeSpecName "kube-api-access-lnjbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:25:14 crc kubenswrapper[4772]: I0127 15:25:14.858591 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-62ktv" Jan 27 15:25:14 crc kubenswrapper[4772]: I0127 15:25:14.918794 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m449k\" (UniqueName: \"kubernetes.io/projected/8aead2c0-bb19-4542-8736-67943c23f0c0-kube-api-access-m449k\") pod \"8aead2c0-bb19-4542-8736-67943c23f0c0\" (UID: \"8aead2c0-bb19-4542-8736-67943c23f0c0\") " Jan 27 15:25:14 crc kubenswrapper[4772]: I0127 15:25:14.918884 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8aead2c0-bb19-4542-8736-67943c23f0c0-config\") pod \"8aead2c0-bb19-4542-8736-67943c23f0c0\" (UID: \"8aead2c0-bb19-4542-8736-67943c23f0c0\") " Jan 27 15:25:14 crc kubenswrapper[4772]: I0127 15:25:14.918978 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8aead2c0-bb19-4542-8736-67943c23f0c0-dns-svc\") pod \"8aead2c0-bb19-4542-8736-67943c23f0c0\" (UID: \"8aead2c0-bb19-4542-8736-67943c23f0c0\") " Jan 27 15:25:14 crc kubenswrapper[4772]: I0127 15:25:14.919415 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/477337bf-a24a-44fd-9c46-38d2e1566b18-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:14 crc kubenswrapper[4772]: I0127 15:25:14.919433 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnjbj\" (UniqueName: \"kubernetes.io/projected/477337bf-a24a-44fd-9c46-38d2e1566b18-kube-api-access-lnjbj\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:14 crc kubenswrapper[4772]: I0127 15:25:14.919445 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/477337bf-a24a-44fd-9c46-38d2e1566b18-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:14 crc kubenswrapper[4772]: I0127 15:25:14.927125 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8aead2c0-bb19-4542-8736-67943c23f0c0-kube-api-access-m449k" (OuterVolumeSpecName: "kube-api-access-m449k") pod "8aead2c0-bb19-4542-8736-67943c23f0c0" (UID: "8aead2c0-bb19-4542-8736-67943c23f0c0"). InnerVolumeSpecName "kube-api-access-m449k". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:25:14 crc kubenswrapper[4772]: I0127 15:25:14.954683 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8aead2c0-bb19-4542-8736-67943c23f0c0-config" (OuterVolumeSpecName: "config") pod "8aead2c0-bb19-4542-8736-67943c23f0c0" (UID: "8aead2c0-bb19-4542-8736-67943c23f0c0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:14 crc kubenswrapper[4772]: I0127 15:25:14.973535 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8aead2c0-bb19-4542-8736-67943c23f0c0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8aead2c0-bb19-4542-8736-67943c23f0c0" (UID: "8aead2c0-bb19-4542-8736-67943c23f0c0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.020849 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8aead2c0-bb19-4542-8736-67943c23f0c0-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.020882 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m449k\" (UniqueName: \"kubernetes.io/projected/8aead2c0-bb19-4542-8736-67943c23f0c0-kube-api-access-m449k\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.020915 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8aead2c0-bb19-4542-8736-67943c23f0c0-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.139052 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.200980 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-vqpfg"] Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.209054 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tltm6"] Jan 27 15:25:15 crc kubenswrapper[4772]: W0127 15:25:15.225494 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda490a71b_c33d_4c94_9592_f97d1d315e81.slice/crio-b17e4736d7a1350c0c68f20fb3327f8519a43cb3b16a163a3b8e79d710328aca WatchSource:0}: Error finding container b17e4736d7a1350c0c68f20fb3327f8519a43cb3b16a163a3b8e79d710328aca: Status 404 returned error can't find the container with id b17e4736d7a1350c0c68f20fb3327f8519a43cb3b16a163a3b8e79d710328aca Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.325756 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-vrz8v"] Jan 27 15:25:15 crc kubenswrapper[4772]: W0127 15:25:15.336632 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod003a41cd_8661_4d0a_a5b7_4e06e02d3785.slice/crio-268aefd3f8b14bdf425ab8056006b626765f75c0f77a51699e073edbd88f6c5f WatchSource:0}: Error finding container 268aefd3f8b14bdf425ab8056006b626765f75c0f77a51699e073edbd88f6c5f: Status 404 returned error can't find the container with id 268aefd3f8b14bdf425ab8056006b626765f75c0f77a51699e073edbd88f6c5f Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.336789 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938","Type":"ContainerStarted","Data":"383ce19f4879446a46975b1e3757ca75d5dbab13e103b56af11750ee3019f6bc"} Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.341926 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-4jb8m"] Jan 27 15:25:15 crc kubenswrapper[4772]: E0127 15:25:15.346114 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aead2c0-bb19-4542-8736-67943c23f0c0" containerName="init" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.346156 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aead2c0-bb19-4542-8736-67943c23f0c0" containerName="init" Jan 27 15:25:15 crc kubenswrapper[4772]: E0127 15:25:15.346257 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aead2c0-bb19-4542-8736-67943c23f0c0" containerName="dnsmasq-dns" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.346267 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aead2c0-bb19-4542-8736-67943c23f0c0" containerName="dnsmasq-dns" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.347631 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="8aead2c0-bb19-4542-8736-67943c23f0c0" containerName="dnsmasq-dns" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.363319 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-4jb8m"] Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.363356 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-d4llz" event={"ID":"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9","Type":"ContainerStarted","Data":"a5599751ce46331dd2a224ba692cd6619979f4eb0205e3a54352eb587e777c31"} Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.363485 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-4jb8m" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.366267 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.367736 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tltm6" event={"ID":"01d2ace8-4fbb-4f53-aa31-7557dbaabcce","Type":"ContainerStarted","Data":"67ecd74afacd326820a12dc1cfdc76a179790d7bc04cff09eef9f1e0a03e5d5e"} Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.369770 4772 generic.go:334] "Generic (PLEG): container finished" podID="b1515626-5d79-408d-abc1-cb92abd58f3f" containerID="3f38ceb1ab131833479b9e418df05230599249523d60e8e11929add232b03e39" exitCode=0 Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.369890 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b1515626-5d79-408d-abc1-cb92abd58f3f","Type":"ContainerDied","Data":"3f38ceb1ab131833479b9e418df05230599249523d60e8e11929add232b03e39"} Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.372845 4772 generic.go:334] "Generic (PLEG): container finished" podID="8aead2c0-bb19-4542-8736-67943c23f0c0" containerID="80133493ee73cefd49c95662c6d25ef4d357fe61d09a830267d9876af454f579" exitCode=0 Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.372945 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-62ktv" event={"ID":"8aead2c0-bb19-4542-8736-67943c23f0c0","Type":"ContainerDied","Data":"80133493ee73cefd49c95662c6d25ef4d357fe61d09a830267d9876af454f579"} Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.372975 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-62ktv" event={"ID":"8aead2c0-bb19-4542-8736-67943c23f0c0","Type":"ContainerDied","Data":"80889f4358a04bf7fac97ea1ad80ad55c76754b98341035087589aff0399485d"} Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.373021 4772 scope.go:117] "RemoveContainer" containerID="80133493ee73cefd49c95662c6d25ef4d357fe61d09a830267d9876af454f579" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.373243 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-62ktv" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.378245 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-rsdjb" event={"ID":"477337bf-a24a-44fd-9c46-38d2e1566b18","Type":"ContainerDied","Data":"e3d43f62abf3e7b0d500727260573ad5c5ab13345a55b4ce49116d69dfd50dd4"} Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.378317 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-rsdjb" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.389453 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-vqpfg" event={"ID":"a490a71b-c33d-4c94-9592-f97d1d315e81","Type":"ContainerStarted","Data":"b17e4736d7a1350c0c68f20fb3327f8519a43cb3b16a163a3b8e79d710328aca"} Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.398791 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-d4llz" podStartSLOduration=1.746414787 podStartE2EDuration="5.39876811s" podCreationTimestamp="2026-01-27 15:25:10 +0000 UTC" firstStartedPulling="2026-01-27 15:25:10.99954297 +0000 UTC m=+1096.980152068" lastFinishedPulling="2026-01-27 15:25:14.651896293 +0000 UTC m=+1100.632505391" observedRunningTime="2026-01-27 15:25:15.389581466 +0000 UTC m=+1101.370190584" watchObservedRunningTime="2026-01-27 15:25:15.39876811 +0000 UTC m=+1101.379377208" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.403774 4772 scope.go:117] "RemoveContainer" containerID="932ddd9664bf5058afe13fff46657f7b87008d96d4495fb6b027b04295f1af73" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.426818 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0a9e15e-9947-44be-872f-20072b41a7fc-operator-scripts\") pod \"root-account-create-update-4jb8m\" (UID: \"e0a9e15e-9947-44be-872f-20072b41a7fc\") " pod="openstack/root-account-create-update-4jb8m" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.426905 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flptk\" (UniqueName: \"kubernetes.io/projected/e0a9e15e-9947-44be-872f-20072b41a7fc-kube-api-access-flptk\") pod \"root-account-create-update-4jb8m\" (UID: \"e0a9e15e-9947-44be-872f-20072b41a7fc\") " pod="openstack/root-account-create-update-4jb8m" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.480500 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-rsdjb"] Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.485313 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-rsdjb"] Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.494415 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-62ktv"] Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.500244 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-62ktv"] Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.511377 4772 scope.go:117] "RemoveContainer" containerID="80133493ee73cefd49c95662c6d25ef4d357fe61d09a830267d9876af454f579" Jan 27 15:25:15 crc kubenswrapper[4772]: E0127 15:25:15.512503 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80133493ee73cefd49c95662c6d25ef4d357fe61d09a830267d9876af454f579\": container with ID starting with 80133493ee73cefd49c95662c6d25ef4d357fe61d09a830267d9876af454f579 not found: ID does not exist" containerID="80133493ee73cefd49c95662c6d25ef4d357fe61d09a830267d9876af454f579" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.512569 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80133493ee73cefd49c95662c6d25ef4d357fe61d09a830267d9876af454f579"} err="failed to get container status \"80133493ee73cefd49c95662c6d25ef4d357fe61d09a830267d9876af454f579\": rpc error: code = NotFound desc = could not find container \"80133493ee73cefd49c95662c6d25ef4d357fe61d09a830267d9876af454f579\": container with ID starting with 80133493ee73cefd49c95662c6d25ef4d357fe61d09a830267d9876af454f579 not found: ID does not exist" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.512602 4772 scope.go:117] "RemoveContainer" containerID="932ddd9664bf5058afe13fff46657f7b87008d96d4495fb6b027b04295f1af73" Jan 27 15:25:15 crc kubenswrapper[4772]: E0127 15:25:15.513275 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"932ddd9664bf5058afe13fff46657f7b87008d96d4495fb6b027b04295f1af73\": container with ID starting with 932ddd9664bf5058afe13fff46657f7b87008d96d4495fb6b027b04295f1af73 not found: ID does not exist" containerID="932ddd9664bf5058afe13fff46657f7b87008d96d4495fb6b027b04295f1af73" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.513328 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"932ddd9664bf5058afe13fff46657f7b87008d96d4495fb6b027b04295f1af73"} err="failed to get container status \"932ddd9664bf5058afe13fff46657f7b87008d96d4495fb6b027b04295f1af73\": rpc error: code = NotFound desc = could not find container \"932ddd9664bf5058afe13fff46657f7b87008d96d4495fb6b027b04295f1af73\": container with ID starting with 932ddd9664bf5058afe13fff46657f7b87008d96d4495fb6b027b04295f1af73 not found: ID does not exist" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.528358 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0a9e15e-9947-44be-872f-20072b41a7fc-operator-scripts\") pod \"root-account-create-update-4jb8m\" (UID: \"e0a9e15e-9947-44be-872f-20072b41a7fc\") " pod="openstack/root-account-create-update-4jb8m" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.528441 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flptk\" (UniqueName: \"kubernetes.io/projected/e0a9e15e-9947-44be-872f-20072b41a7fc-kube-api-access-flptk\") pod \"root-account-create-update-4jb8m\" (UID: \"e0a9e15e-9947-44be-872f-20072b41a7fc\") " pod="openstack/root-account-create-update-4jb8m" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.529680 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0a9e15e-9947-44be-872f-20072b41a7fc-operator-scripts\") pod \"root-account-create-update-4jb8m\" (UID: \"e0a9e15e-9947-44be-872f-20072b41a7fc\") " pod="openstack/root-account-create-update-4jb8m" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.555932 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flptk\" (UniqueName: \"kubernetes.io/projected/e0a9e15e-9947-44be-872f-20072b41a7fc-kube-api-access-flptk\") pod \"root-account-create-update-4jb8m\" (UID: \"e0a9e15e-9947-44be-872f-20072b41a7fc\") " pod="openstack/root-account-create-update-4jb8m" Jan 27 15:25:15 crc kubenswrapper[4772]: I0127 15:25:15.734437 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-4jb8m" Jan 27 15:25:16 crc kubenswrapper[4772]: I0127 15:25:16.179532 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-4jb8m"] Jan 27 15:25:16 crc kubenswrapper[4772]: W0127 15:25:16.302639 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0a9e15e_9947_44be_872f_20072b41a7fc.slice/crio-821dcdb092c7d4fb686ec29c55b5cbeddc94342c181cf08bb0e05a5c49e2d37d WatchSource:0}: Error finding container 821dcdb092c7d4fb686ec29c55b5cbeddc94342c181cf08bb0e05a5c49e2d37d: Status 404 returned error can't find the container with id 821dcdb092c7d4fb686ec29c55b5cbeddc94342c181cf08bb0e05a5c49e2d37d Jan 27 15:25:16 crc kubenswrapper[4772]: I0127 15:25:16.399645 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-4jb8m" event={"ID":"e0a9e15e-9947-44be-872f-20072b41a7fc","Type":"ContainerStarted","Data":"821dcdb092c7d4fb686ec29c55b5cbeddc94342c181cf08bb0e05a5c49e2d37d"} Jan 27 15:25:16 crc kubenswrapper[4772]: I0127 15:25:16.401048 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"76fdbdb1-d48a-4cd1-8372-78887671dce8","Type":"ContainerStarted","Data":"d53d0dfba4b0af64ac6186cc8eb8efb21a1ec89a66c075c5a53ab1db9987e594"} Jan 27 15:25:16 crc kubenswrapper[4772]: I0127 15:25:16.426551 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-vqpfg" event={"ID":"a490a71b-c33d-4c94-9592-f97d1d315e81","Type":"ContainerStarted","Data":"b93ad84c922746d427d3e2a2deb04a875a239fcafbecb5146ae05b1b11e36a09"} Jan 27 15:25:16 crc kubenswrapper[4772]: I0127 15:25:16.430742 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"508c3d5b-212a-46da-9a55-de3f35d7019b","Type":"ContainerStarted","Data":"900401625caff4c2d87fe06884c7dcba7f46fdc58e9213b1a6cc2cf36d383e52"} Jan 27 15:25:16 crc kubenswrapper[4772]: I0127 15:25:16.433330 4772 generic.go:334] "Generic (PLEG): container finished" podID="01d2ace8-4fbb-4f53-aa31-7557dbaabcce" containerID="a76c09aaadcee4723d7ef767396afbe7396ff3e3af040a33171b3953859d1cba" exitCode=0 Jan 27 15:25:16 crc kubenswrapper[4772]: I0127 15:25:16.433390 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tltm6" event={"ID":"01d2ace8-4fbb-4f53-aa31-7557dbaabcce","Type":"ContainerDied","Data":"a76c09aaadcee4723d7ef767396afbe7396ff3e3af040a33171b3953859d1cba"} Jan 27 15:25:16 crc kubenswrapper[4772]: I0127 15:25:16.439508 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b1515626-5d79-408d-abc1-cb92abd58f3f","Type":"ContainerStarted","Data":"21613f2614f5809a9a792371fe2e685753a1a2fc6ea2f8fa7dcc2390d4bafda0"} Jan 27 15:25:16 crc kubenswrapper[4772]: I0127 15:25:16.456364 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-vqpfg" podStartSLOduration=4.456346971 podStartE2EDuration="4.456346971s" podCreationTimestamp="2026-01-27 15:25:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:25:16.444552281 +0000 UTC m=+1102.425161369" watchObservedRunningTime="2026-01-27 15:25:16.456346971 +0000 UTC m=+1102.436956069" Jan 27 15:25:16 crc kubenswrapper[4772]: I0127 15:25:16.456910 4772 generic.go:334] "Generic (PLEG): container finished" podID="003a41cd-8661-4d0a-a5b7-4e06e02d3785" containerID="5a47e1d72c8f5eacedc4878d5823874f52e03c455d9958e9166510df3dfd80ce" exitCode=0 Jan 27 15:25:16 crc kubenswrapper[4772]: I0127 15:25:16.456972 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" event={"ID":"003a41cd-8661-4d0a-a5b7-4e06e02d3785","Type":"ContainerDied","Data":"5a47e1d72c8f5eacedc4878d5823874f52e03c455d9958e9166510df3dfd80ce"} Jan 27 15:25:16 crc kubenswrapper[4772]: I0127 15:25:16.456998 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" event={"ID":"003a41cd-8661-4d0a-a5b7-4e06e02d3785","Type":"ContainerStarted","Data":"268aefd3f8b14bdf425ab8056006b626765f75c0f77a51699e073edbd88f6c5f"} Jan 27 15:25:16 crc kubenswrapper[4772]: I0127 15:25:16.477445 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371993.377354 podStartE2EDuration="43.477422867s" podCreationTimestamp="2026-01-27 15:24:33 +0000 UTC" firstStartedPulling="2026-01-27 15:24:35.998493379 +0000 UTC m=+1061.979102477" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:25:16.474594455 +0000 UTC m=+1102.455203553" watchObservedRunningTime="2026-01-27 15:25:16.477422867 +0000 UTC m=+1102.458031965" Jan 27 15:25:16 crc kubenswrapper[4772]: I0127 15:25:16.676192 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="477337bf-a24a-44fd-9c46-38d2e1566b18" path="/var/lib/kubelet/pods/477337bf-a24a-44fd-9c46-38d2e1566b18/volumes" Jan 27 15:25:16 crc kubenswrapper[4772]: I0127 15:25:16.676769 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8aead2c0-bb19-4542-8736-67943c23f0c0" path="/var/lib/kubelet/pods/8aead2c0-bb19-4542-8736-67943c23f0c0/volumes" Jan 27 15:25:17 crc kubenswrapper[4772]: I0127 15:25:17.469281 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" event={"ID":"003a41cd-8661-4d0a-a5b7-4e06e02d3785","Type":"ContainerStarted","Data":"36cd1ff546856c238a871962177bd25e90c208878c18e039ee4f0f710207ad78"} Jan 27 15:25:17 crc kubenswrapper[4772]: I0127 15:25:17.469739 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" Jan 27 15:25:17 crc kubenswrapper[4772]: I0127 15:25:17.470684 4772 generic.go:334] "Generic (PLEG): container finished" podID="e0a9e15e-9947-44be-872f-20072b41a7fc" containerID="6d6c94667c0ae61eab0c4931fc95c11f862c674ae06fd177d824e395ced6b9a6" exitCode=0 Jan 27 15:25:17 crc kubenswrapper[4772]: I0127 15:25:17.470741 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-4jb8m" event={"ID":"e0a9e15e-9947-44be-872f-20072b41a7fc","Type":"ContainerDied","Data":"6d6c94667c0ae61eab0c4931fc95c11f862c674ae06fd177d824e395ced6b9a6"} Jan 27 15:25:17 crc kubenswrapper[4772]: I0127 15:25:17.472622 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938","Type":"ContainerStarted","Data":"b1542ba131aec1cffd5520f2969b843d3aa12fe7b4cd60022addce3e73977b99"} Jan 27 15:25:17 crc kubenswrapper[4772]: I0127 15:25:17.472653 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938","Type":"ContainerStarted","Data":"f351431c9793a13f48f307e65178046dd4ccdc52ebd7ba269a580599ff0da01a"} Jan 27 15:25:17 crc kubenswrapper[4772]: I0127 15:25:17.472727 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Jan 27 15:25:17 crc kubenswrapper[4772]: I0127 15:25:17.474520 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tltm6" event={"ID":"01d2ace8-4fbb-4f53-aa31-7557dbaabcce","Type":"ContainerStarted","Data":"d1b5117c10f9331477f591f10a624b08ae6968087cc1bb15580ee055f80a719c"} Jan 27 15:25:17 crc kubenswrapper[4772]: I0127 15:25:17.476204 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-etc-swift\") pod \"swift-storage-0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " pod="openstack/swift-storage-0" Jan 27 15:25:17 crc kubenswrapper[4772]: E0127 15:25:17.476388 4772 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 27 15:25:17 crc kubenswrapper[4772]: E0127 15:25:17.476434 4772 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 27 15:25:17 crc kubenswrapper[4772]: E0127 15:25:17.476485 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-etc-swift podName:3ef68955-b80c-4732-9e87-0bec53d0b3a0 nodeName:}" failed. No retries permitted until 2026-01-27 15:25:25.476470244 +0000 UTC m=+1111.457079342 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-etc-swift") pod "swift-storage-0" (UID: "3ef68955-b80c-4732-9e87-0bec53d0b3a0") : configmap "swift-ring-files" not found Jan 27 15:25:17 crc kubenswrapper[4772]: I0127 15:25:17.490183 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" podStartSLOduration=5.490142838 podStartE2EDuration="5.490142838s" podCreationTimestamp="2026-01-27 15:25:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:25:17.487635196 +0000 UTC m=+1103.468244314" watchObservedRunningTime="2026-01-27 15:25:17.490142838 +0000 UTC m=+1103.470751936" Jan 27 15:25:17 crc kubenswrapper[4772]: I0127 15:25:17.519481 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.3742372400000002 podStartE2EDuration="4.519459681s" podCreationTimestamp="2026-01-27 15:25:13 +0000 UTC" firstStartedPulling="2026-01-27 15:25:15.216403316 +0000 UTC m=+1101.197012414" lastFinishedPulling="2026-01-27 15:25:16.361625757 +0000 UTC m=+1102.342234855" observedRunningTime="2026-01-27 15:25:17.514025254 +0000 UTC m=+1103.494634352" watchObservedRunningTime="2026-01-27 15:25:17.519459681 +0000 UTC m=+1103.500068779" Jan 27 15:25:17 crc kubenswrapper[4772]: I0127 15:25:17.532067 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-tltm6" podStartSLOduration=5.532037612 podStartE2EDuration="5.532037612s" podCreationTimestamp="2026-01-27 15:25:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:25:17.530470797 +0000 UTC m=+1103.511079905" watchObservedRunningTime="2026-01-27 15:25:17.532037612 +0000 UTC m=+1103.512646710" Jan 27 15:25:18 crc kubenswrapper[4772]: I0127 15:25:18.201312 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-tltm6" Jan 27 15:25:18 crc kubenswrapper[4772]: I0127 15:25:18.595363 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Jan 27 15:25:18 crc kubenswrapper[4772]: I0127 15:25:18.826322 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-4jb8m" Jan 27 15:25:19 crc kubenswrapper[4772]: I0127 15:25:19.002493 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0a9e15e-9947-44be-872f-20072b41a7fc-operator-scripts\") pod \"e0a9e15e-9947-44be-872f-20072b41a7fc\" (UID: \"e0a9e15e-9947-44be-872f-20072b41a7fc\") " Jan 27 15:25:19 crc kubenswrapper[4772]: I0127 15:25:19.002563 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flptk\" (UniqueName: \"kubernetes.io/projected/e0a9e15e-9947-44be-872f-20072b41a7fc-kube-api-access-flptk\") pod \"e0a9e15e-9947-44be-872f-20072b41a7fc\" (UID: \"e0a9e15e-9947-44be-872f-20072b41a7fc\") " Jan 27 15:25:19 crc kubenswrapper[4772]: I0127 15:25:19.003339 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0a9e15e-9947-44be-872f-20072b41a7fc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e0a9e15e-9947-44be-872f-20072b41a7fc" (UID: "e0a9e15e-9947-44be-872f-20072b41a7fc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:19 crc kubenswrapper[4772]: I0127 15:25:19.022792 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0a9e15e-9947-44be-872f-20072b41a7fc-kube-api-access-flptk" (OuterVolumeSpecName: "kube-api-access-flptk") pod "e0a9e15e-9947-44be-872f-20072b41a7fc" (UID: "e0a9e15e-9947-44be-872f-20072b41a7fc"). InnerVolumeSpecName "kube-api-access-flptk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:25:19 crc kubenswrapper[4772]: I0127 15:25:19.104754 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flptk\" (UniqueName: \"kubernetes.io/projected/e0a9e15e-9947-44be-872f-20072b41a7fc-kube-api-access-flptk\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:19 crc kubenswrapper[4772]: I0127 15:25:19.104795 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0a9e15e-9947-44be-872f-20072b41a7fc-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:19 crc kubenswrapper[4772]: I0127 15:25:19.488441 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-4jb8m" event={"ID":"e0a9e15e-9947-44be-872f-20072b41a7fc","Type":"ContainerDied","Data":"821dcdb092c7d4fb686ec29c55b5cbeddc94342c181cf08bb0e05a5c49e2d37d"} Jan 27 15:25:19 crc kubenswrapper[4772]: I0127 15:25:19.488499 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="821dcdb092c7d4fb686ec29c55b5cbeddc94342c181cf08bb0e05a5c49e2d37d" Jan 27 15:25:19 crc kubenswrapper[4772]: I0127 15:25:19.488458 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-4jb8m" Jan 27 15:25:22 crc kubenswrapper[4772]: I0127 15:25:22.961355 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" Jan 27 15:25:23 crc kubenswrapper[4772]: I0127 15:25:23.203057 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-tltm6" Jan 27 15:25:23 crc kubenswrapper[4772]: I0127 15:25:23.251906 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-vrz8v"] Jan 27 15:25:23 crc kubenswrapper[4772]: I0127 15:25:23.513795 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" podUID="003a41cd-8661-4d0a-a5b7-4e06e02d3785" containerName="dnsmasq-dns" containerID="cri-o://36cd1ff546856c238a871962177bd25e90c208878c18e039ee4f0f710207ad78" gracePeriod=10 Jan 27 15:25:24 crc kubenswrapper[4772]: I0127 15:25:24.529839 4772 generic.go:334] "Generic (PLEG): container finished" podID="003a41cd-8661-4d0a-a5b7-4e06e02d3785" containerID="36cd1ff546856c238a871962177bd25e90c208878c18e039ee4f0f710207ad78" exitCode=0 Jan 27 15:25:24 crc kubenswrapper[4772]: I0127 15:25:24.529933 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" event={"ID":"003a41cd-8661-4d0a-a5b7-4e06e02d3785","Type":"ContainerDied","Data":"36cd1ff546856c238a871962177bd25e90c208878c18e039ee4f0f710207ad78"} Jan 27 15:25:24 crc kubenswrapper[4772]: I0127 15:25:24.833978 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" Jan 27 15:25:24 crc kubenswrapper[4772]: I0127 15:25:24.913692 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t22gz\" (UniqueName: \"kubernetes.io/projected/003a41cd-8661-4d0a-a5b7-4e06e02d3785-kube-api-access-t22gz\") pod \"003a41cd-8661-4d0a-a5b7-4e06e02d3785\" (UID: \"003a41cd-8661-4d0a-a5b7-4e06e02d3785\") " Jan 27 15:25:24 crc kubenswrapper[4772]: I0127 15:25:24.913753 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/003a41cd-8661-4d0a-a5b7-4e06e02d3785-dns-svc\") pod \"003a41cd-8661-4d0a-a5b7-4e06e02d3785\" (UID: \"003a41cd-8661-4d0a-a5b7-4e06e02d3785\") " Jan 27 15:25:24 crc kubenswrapper[4772]: I0127 15:25:24.913836 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/003a41cd-8661-4d0a-a5b7-4e06e02d3785-ovsdbserver-sb\") pod \"003a41cd-8661-4d0a-a5b7-4e06e02d3785\" (UID: \"003a41cd-8661-4d0a-a5b7-4e06e02d3785\") " Jan 27 15:25:24 crc kubenswrapper[4772]: I0127 15:25:24.913952 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/003a41cd-8661-4d0a-a5b7-4e06e02d3785-config\") pod \"003a41cd-8661-4d0a-a5b7-4e06e02d3785\" (UID: \"003a41cd-8661-4d0a-a5b7-4e06e02d3785\") " Jan 27 15:25:24 crc kubenswrapper[4772]: I0127 15:25:24.929737 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/003a41cd-8661-4d0a-a5b7-4e06e02d3785-kube-api-access-t22gz" (OuterVolumeSpecName: "kube-api-access-t22gz") pod "003a41cd-8661-4d0a-a5b7-4e06e02d3785" (UID: "003a41cd-8661-4d0a-a5b7-4e06e02d3785"). InnerVolumeSpecName "kube-api-access-t22gz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:25:24 crc kubenswrapper[4772]: I0127 15:25:24.962010 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/003a41cd-8661-4d0a-a5b7-4e06e02d3785-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "003a41cd-8661-4d0a-a5b7-4e06e02d3785" (UID: "003a41cd-8661-4d0a-a5b7-4e06e02d3785"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:24 crc kubenswrapper[4772]: I0127 15:25:24.962561 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/003a41cd-8661-4d0a-a5b7-4e06e02d3785-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "003a41cd-8661-4d0a-a5b7-4e06e02d3785" (UID: "003a41cd-8661-4d0a-a5b7-4e06e02d3785"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:24 crc kubenswrapper[4772]: I0127 15:25:24.980875 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/003a41cd-8661-4d0a-a5b7-4e06e02d3785-config" (OuterVolumeSpecName: "config") pod "003a41cd-8661-4d0a-a5b7-4e06e02d3785" (UID: "003a41cd-8661-4d0a-a5b7-4e06e02d3785"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:25 crc kubenswrapper[4772]: I0127 15:25:25.016070 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/003a41cd-8661-4d0a-a5b7-4e06e02d3785-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:25 crc kubenswrapper[4772]: I0127 15:25:25.016106 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t22gz\" (UniqueName: \"kubernetes.io/projected/003a41cd-8661-4d0a-a5b7-4e06e02d3785-kube-api-access-t22gz\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:25 crc kubenswrapper[4772]: I0127 15:25:25.016123 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/003a41cd-8661-4d0a-a5b7-4e06e02d3785-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:25 crc kubenswrapper[4772]: I0127 15:25:25.016134 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/003a41cd-8661-4d0a-a5b7-4e06e02d3785-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:25 crc kubenswrapper[4772]: I0127 15:25:25.119678 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Jan 27 15:25:25 crc kubenswrapper[4772]: I0127 15:25:25.119748 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Jan 27 15:25:25 crc kubenswrapper[4772]: I0127 15:25:25.524579 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-etc-swift\") pod \"swift-storage-0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " pod="openstack/swift-storage-0" Jan 27 15:25:25 crc kubenswrapper[4772]: I0127 15:25:25.530345 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-etc-swift\") pod \"swift-storage-0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " pod="openstack/swift-storage-0" Jan 27 15:25:25 crc kubenswrapper[4772]: I0127 15:25:25.540184 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" event={"ID":"003a41cd-8661-4d0a-a5b7-4e06e02d3785","Type":"ContainerDied","Data":"268aefd3f8b14bdf425ab8056006b626765f75c0f77a51699e073edbd88f6c5f"} Jan 27 15:25:25 crc kubenswrapper[4772]: I0127 15:25:25.540471 4772 scope.go:117] "RemoveContainer" containerID="36cd1ff546856c238a871962177bd25e90c208878c18e039ee4f0f710207ad78" Jan 27 15:25:25 crc kubenswrapper[4772]: I0127 15:25:25.540294 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-vrz8v" Jan 27 15:25:25 crc kubenswrapper[4772]: I0127 15:25:25.559867 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Jan 27 15:25:25 crc kubenswrapper[4772]: I0127 15:25:25.574612 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-vrz8v"] Jan 27 15:25:25 crc kubenswrapper[4772]: I0127 15:25:25.577202 4772 scope.go:117] "RemoveContainer" containerID="5a47e1d72c8f5eacedc4878d5823874f52e03c455d9958e9166510df3dfd80ce" Jan 27 15:25:25 crc kubenswrapper[4772]: I0127 15:25:25.581335 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-vrz8v"] Jan 27 15:25:26 crc kubenswrapper[4772]: I0127 15:25:26.097049 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Jan 27 15:25:26 crc kubenswrapper[4772]: I0127 15:25:26.553114 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerStarted","Data":"7e2686f92b31392fd2420828f9959abe37458794a1d13beae3bf48377776f704"} Jan 27 15:25:26 crc kubenswrapper[4772]: I0127 15:25:26.554383 4772 generic.go:334] "Generic (PLEG): container finished" podID="2313c291-4eb5-4b79-ad9b-b04cd06a1ee9" containerID="a5599751ce46331dd2a224ba692cd6619979f4eb0205e3a54352eb587e777c31" exitCode=0 Jan 27 15:25:26 crc kubenswrapper[4772]: I0127 15:25:26.554426 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-d4llz" event={"ID":"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9","Type":"ContainerDied","Data":"a5599751ce46331dd2a224ba692cd6619979f4eb0205e3a54352eb587e777c31"} Jan 27 15:25:26 crc kubenswrapper[4772]: I0127 15:25:26.672560 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="003a41cd-8661-4d0a-a5b7-4e06e02d3785" path="/var/lib/kubelet/pods/003a41cd-8661-4d0a-a5b7-4e06e02d3785/volumes" Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.005629 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.066926 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvbq9\" (UniqueName: \"kubernetes.io/projected/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-kube-api-access-mvbq9\") pod \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.066981 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-scripts\") pod \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.070317 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-combined-ca-bundle\") pod \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.070374 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-ring-data-devices\") pod \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.070406 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-dispersionconf\") pod \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.070459 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-swiftconf\") pod \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.070666 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-etc-swift\") pod \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\" (UID: \"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9\") " Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.071064 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-kube-api-access-mvbq9" (OuterVolumeSpecName: "kube-api-access-mvbq9") pod "2313c291-4eb5-4b79-ad9b-b04cd06a1ee9" (UID: "2313c291-4eb5-4b79-ad9b-b04cd06a1ee9"). InnerVolumeSpecName "kube-api-access-mvbq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.071484 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvbq9\" (UniqueName: \"kubernetes.io/projected/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-kube-api-access-mvbq9\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.072039 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "2313c291-4eb5-4b79-ad9b-b04cd06a1ee9" (UID: "2313c291-4eb5-4b79-ad9b-b04cd06a1ee9"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.072300 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "2313c291-4eb5-4b79-ad9b-b04cd06a1ee9" (UID: "2313c291-4eb5-4b79-ad9b-b04cd06a1ee9"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.077569 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "2313c291-4eb5-4b79-ad9b-b04cd06a1ee9" (UID: "2313c291-4eb5-4b79-ad9b-b04cd06a1ee9"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.091233 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "2313c291-4eb5-4b79-ad9b-b04cd06a1ee9" (UID: "2313c291-4eb5-4b79-ad9b-b04cd06a1ee9"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.093315 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-scripts" (OuterVolumeSpecName: "scripts") pod "2313c291-4eb5-4b79-ad9b-b04cd06a1ee9" (UID: "2313c291-4eb5-4b79-ad9b-b04cd06a1ee9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.096160 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2313c291-4eb5-4b79-ad9b-b04cd06a1ee9" (UID: "2313c291-4eb5-4b79-ad9b-b04cd06a1ee9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.172939 4772 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-swiftconf\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.172972 4772 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-etc-swift\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.172982 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.172990 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.173000 4772 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-ring-data-devices\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.173008 4772 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9-dispersionconf\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.570725 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerStarted","Data":"c3f602f5b8fe5f978c40989adc1d0130c6aaae0dce0fc13d5e34bbe819e8eccb"} Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.571055 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerStarted","Data":"5f271cd2dcb6b658cde722402c5b2945c28f4d7486cab8c56e064081779416a1"} Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.571065 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerStarted","Data":"d35aa807e61d39133b8319305719556fcfa6889495c80253864eaf2dc48a450b"} Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.572555 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-d4llz" event={"ID":"2313c291-4eb5-4b79-ad9b-b04cd06a1ee9","Type":"ContainerDied","Data":"1fe7af5e79ef4da92cef0b65e4f6ca7d519d839fcd9321afbd7bf43485be0f39"} Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.572609 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fe7af5e79ef4da92cef0b65e4f6ca7d519d839fcd9321afbd7bf43485be0f39" Jan 27 15:25:28 crc kubenswrapper[4772]: I0127 15:25:28.572638 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-d4llz" Jan 27 15:25:29 crc kubenswrapper[4772]: I0127 15:25:29.043010 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Jan 27 15:25:29 crc kubenswrapper[4772]: I0127 15:25:29.135667 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Jan 27 15:25:29 crc kubenswrapper[4772]: I0127 15:25:29.582320 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerStarted","Data":"ac32767b3784713a66fbfe32a337398a7461aa8ffad58bbfea7ccf6e3c4ee19d"} Jan 27 15:25:30 crc kubenswrapper[4772]: I0127 15:25:30.594359 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerStarted","Data":"99c9f47c0720632dfecbfc5e9152885ab96d751677b561767c79f0a032ca5cf5"} Jan 27 15:25:30 crc kubenswrapper[4772]: I0127 15:25:30.595769 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerStarted","Data":"0c6f6ecf89a4947c23560538762ca73dfe5e13c4acb04e206d91772a3cfc9c49"} Jan 27 15:25:30 crc kubenswrapper[4772]: I0127 15:25:30.595788 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerStarted","Data":"94e4c588a745acb16ce919a52f7150cf54119c1c41e94c9e658206e6b58958ed"} Jan 27 15:25:30 crc kubenswrapper[4772]: I0127 15:25:30.595815 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerStarted","Data":"494d3ebaeddb756bf375d2bc394a4b4086ee3e25d9a76747552d41c1f40a9737"} Jan 27 15:25:32 crc kubenswrapper[4772]: I0127 15:25:32.617208 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerStarted","Data":"b0a7c137687a720a7d8c3f84cc586f4b9d3bde7c9bc9e2e0c83a325c2ae23322"} Jan 27 15:25:32 crc kubenswrapper[4772]: I0127 15:25:32.617779 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerStarted","Data":"8bbb31c1be222187b0e9b27f07c1ac0fe66d8ad583df4ff6b26fec62ab98cf87"} Jan 27 15:25:32 crc kubenswrapper[4772]: I0127 15:25:32.617793 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerStarted","Data":"71b4242b9081be055bfb8bd2db6959d32259cd0c3ee2b95ddde1c1d2154be74b"} Jan 27 15:25:32 crc kubenswrapper[4772]: I0127 15:25:32.617802 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerStarted","Data":"bc57f117c387fb10832190ea21f63cdb319308d9390292395fb515e28966d217"} Jan 27 15:25:33 crc kubenswrapper[4772]: I0127 15:25:33.634729 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerStarted","Data":"0b50101071feccad5793667a8f4849d22482c6d522fac228c249d69d6d557cdf"} Jan 27 15:25:33 crc kubenswrapper[4772]: I0127 15:25:33.634779 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerStarted","Data":"8d889567d10b3e8868d76680ff442da2a14216919aae766c356918ec9960b9a4"} Jan 27 15:25:33 crc kubenswrapper[4772]: I0127 15:25:33.634792 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerStarted","Data":"c1cf3012e8501ba3a809e028a1ab49c960d95fb090a04b4dbca6cd01d2de9524"} Jan 27 15:25:33 crc kubenswrapper[4772]: I0127 15:25:33.677558 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=20.308887431 podStartE2EDuration="25.67753161s" podCreationTimestamp="2026-01-27 15:25:08 +0000 UTC" firstStartedPulling="2026-01-27 15:25:26.106914043 +0000 UTC m=+1112.087523161" lastFinishedPulling="2026-01-27 15:25:31.475558242 +0000 UTC m=+1117.456167340" observedRunningTime="2026-01-27 15:25:33.669697814 +0000 UTC m=+1119.650306922" watchObservedRunningTime="2026-01-27 15:25:33.67753161 +0000 UTC m=+1119.658140698" Jan 27 15:25:33 crc kubenswrapper[4772]: I0127 15:25:33.745705 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-4jb8m"] Jan 27 15:25:33 crc kubenswrapper[4772]: I0127 15:25:33.753886 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-4jb8m"] Jan 27 15:25:33 crc kubenswrapper[4772]: I0127 15:25:33.825311 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-547fc"] Jan 27 15:25:33 crc kubenswrapper[4772]: E0127 15:25:33.825658 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="003a41cd-8661-4d0a-a5b7-4e06e02d3785" containerName="init" Jan 27 15:25:33 crc kubenswrapper[4772]: I0127 15:25:33.825674 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="003a41cd-8661-4d0a-a5b7-4e06e02d3785" containerName="init" Jan 27 15:25:33 crc kubenswrapper[4772]: E0127 15:25:33.825694 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="003a41cd-8661-4d0a-a5b7-4e06e02d3785" containerName="dnsmasq-dns" Jan 27 15:25:33 crc kubenswrapper[4772]: I0127 15:25:33.825701 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="003a41cd-8661-4d0a-a5b7-4e06e02d3785" containerName="dnsmasq-dns" Jan 27 15:25:33 crc kubenswrapper[4772]: E0127 15:25:33.825710 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0a9e15e-9947-44be-872f-20072b41a7fc" containerName="mariadb-account-create-update" Jan 27 15:25:33 crc kubenswrapper[4772]: I0127 15:25:33.825716 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0a9e15e-9947-44be-872f-20072b41a7fc" containerName="mariadb-account-create-update" Jan 27 15:25:33 crc kubenswrapper[4772]: E0127 15:25:33.825726 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2313c291-4eb5-4b79-ad9b-b04cd06a1ee9" containerName="swift-ring-rebalance" Jan 27 15:25:33 crc kubenswrapper[4772]: I0127 15:25:33.825732 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="2313c291-4eb5-4b79-ad9b-b04cd06a1ee9" containerName="swift-ring-rebalance" Jan 27 15:25:33 crc kubenswrapper[4772]: I0127 15:25:33.825885 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="2313c291-4eb5-4b79-ad9b-b04cd06a1ee9" containerName="swift-ring-rebalance" Jan 27 15:25:33 crc kubenswrapper[4772]: I0127 15:25:33.825897 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0a9e15e-9947-44be-872f-20072b41a7fc" containerName="mariadb-account-create-update" Jan 27 15:25:33 crc kubenswrapper[4772]: I0127 15:25:33.825906 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="003a41cd-8661-4d0a-a5b7-4e06e02d3785" containerName="dnsmasq-dns" Jan 27 15:25:33 crc kubenswrapper[4772]: I0127 15:25:33.826427 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-547fc" Jan 27 15:25:33 crc kubenswrapper[4772]: I0127 15:25:33.831295 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Jan 27 15:25:33 crc kubenswrapper[4772]: I0127 15:25:33.840766 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-547fc"] Jan 27 15:25:33 crc kubenswrapper[4772]: I0127 15:25:33.915442 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Jan 27 15:25:33 crc kubenswrapper[4772]: I0127 15:25:33.972775 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4q5jf\" (UniqueName: \"kubernetes.io/projected/dd1eaad6-cd29-4189-8ecd-62b7658e69ef-kube-api-access-4q5jf\") pod \"root-account-create-update-547fc\" (UID: \"dd1eaad6-cd29-4189-8ecd-62b7658e69ef\") " pod="openstack/root-account-create-update-547fc" Jan 27 15:25:33 crc kubenswrapper[4772]: I0127 15:25:33.972868 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd1eaad6-cd29-4189-8ecd-62b7658e69ef-operator-scripts\") pod \"root-account-create-update-547fc\" (UID: \"dd1eaad6-cd29-4189-8ecd-62b7658e69ef\") " pod="openstack/root-account-create-update-547fc" Jan 27 15:25:33 crc kubenswrapper[4772]: I0127 15:25:33.993774 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-lll86"] Jan 27 15:25:33 crc kubenswrapper[4772]: I0127 15:25:33.995433 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-lll86" Jan 27 15:25:33 crc kubenswrapper[4772]: I0127 15:25:33.998027 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.012762 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-lll86"] Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.074027 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd1eaad6-cd29-4189-8ecd-62b7658e69ef-operator-scripts\") pod \"root-account-create-update-547fc\" (UID: \"dd1eaad6-cd29-4189-8ecd-62b7658e69ef\") " pod="openstack/root-account-create-update-547fc" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.074238 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4q5jf\" (UniqueName: \"kubernetes.io/projected/dd1eaad6-cd29-4189-8ecd-62b7658e69ef-kube-api-access-4q5jf\") pod \"root-account-create-update-547fc\" (UID: \"dd1eaad6-cd29-4189-8ecd-62b7658e69ef\") " pod="openstack/root-account-create-update-547fc" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.074954 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd1eaad6-cd29-4189-8ecd-62b7658e69ef-operator-scripts\") pod \"root-account-create-update-547fc\" (UID: \"dd1eaad6-cd29-4189-8ecd-62b7658e69ef\") " pod="openstack/root-account-create-update-547fc" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.096990 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4q5jf\" (UniqueName: \"kubernetes.io/projected/dd1eaad6-cd29-4189-8ecd-62b7658e69ef-kube-api-access-4q5jf\") pod \"root-account-create-update-547fc\" (UID: \"dd1eaad6-cd29-4189-8ecd-62b7658e69ef\") " pod="openstack/root-account-create-update-547fc" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.140983 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-547fc" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.175861 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-lll86\" (UID: \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\") " pod="openstack/dnsmasq-dns-77585f5f8c-lll86" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.175984 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-config\") pod \"dnsmasq-dns-77585f5f8c-lll86\" (UID: \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\") " pod="openstack/dnsmasq-dns-77585f5f8c-lll86" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.176041 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-lll86\" (UID: \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\") " pod="openstack/dnsmasq-dns-77585f5f8c-lll86" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.176060 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-lll86\" (UID: \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\") " pod="openstack/dnsmasq-dns-77585f5f8c-lll86" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.176134 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-lll86\" (UID: \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\") " pod="openstack/dnsmasq-dns-77585f5f8c-lll86" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.176187 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtsjh\" (UniqueName: \"kubernetes.io/projected/b6aa637d-4418-4fa4-8a26-249446d2fb3f-kube-api-access-gtsjh\") pod \"dnsmasq-dns-77585f5f8c-lll86\" (UID: \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\") " pod="openstack/dnsmasq-dns-77585f5f8c-lll86" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.277537 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-config\") pod \"dnsmasq-dns-77585f5f8c-lll86\" (UID: \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\") " pod="openstack/dnsmasq-dns-77585f5f8c-lll86" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.277997 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-lll86\" (UID: \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\") " pod="openstack/dnsmasq-dns-77585f5f8c-lll86" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.278040 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-lll86\" (UID: \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\") " pod="openstack/dnsmasq-dns-77585f5f8c-lll86" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.278124 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-lll86\" (UID: \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\") " pod="openstack/dnsmasq-dns-77585f5f8c-lll86" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.278220 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtsjh\" (UniqueName: \"kubernetes.io/projected/b6aa637d-4418-4fa4-8a26-249446d2fb3f-kube-api-access-gtsjh\") pod \"dnsmasq-dns-77585f5f8c-lll86\" (UID: \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\") " pod="openstack/dnsmasq-dns-77585f5f8c-lll86" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.278307 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-lll86\" (UID: \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\") " pod="openstack/dnsmasq-dns-77585f5f8c-lll86" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.278950 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-config\") pod \"dnsmasq-dns-77585f5f8c-lll86\" (UID: \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\") " pod="openstack/dnsmasq-dns-77585f5f8c-lll86" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.283845 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-lll86\" (UID: \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\") " pod="openstack/dnsmasq-dns-77585f5f8c-lll86" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.284042 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-lll86\" (UID: \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\") " pod="openstack/dnsmasq-dns-77585f5f8c-lll86" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.285638 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-lll86\" (UID: \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\") " pod="openstack/dnsmasq-dns-77585f5f8c-lll86" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.285663 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-lll86\" (UID: \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\") " pod="openstack/dnsmasq-dns-77585f5f8c-lll86" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.298253 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtsjh\" (UniqueName: \"kubernetes.io/projected/b6aa637d-4418-4fa4-8a26-249446d2fb3f-kube-api-access-gtsjh\") pod \"dnsmasq-dns-77585f5f8c-lll86\" (UID: \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\") " pod="openstack/dnsmasq-dns-77585f5f8c-lll86" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.324124 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-lll86" Jan 27 15:25:34 crc kubenswrapper[4772]: W0127 15:25:34.558871 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd1eaad6_cd29_4189_8ecd_62b7658e69ef.slice/crio-ba7e59e7bcd568f0fef4ca50eaae4a57d525926c5e0abdd0b3a8025ace15dd12 WatchSource:0}: Error finding container ba7e59e7bcd568f0fef4ca50eaae4a57d525926c5e0abdd0b3a8025ace15dd12: Status 404 returned error can't find the container with id ba7e59e7bcd568f0fef4ca50eaae4a57d525926c5e0abdd0b3a8025ace15dd12 Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.562458 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-547fc"] Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.647672 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-547fc" event={"ID":"dd1eaad6-cd29-4189-8ecd-62b7658e69ef","Type":"ContainerStarted","Data":"ba7e59e7bcd568f0fef4ca50eaae4a57d525926c5e0abdd0b3a8025ace15dd12"} Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.677979 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0a9e15e-9947-44be-872f-20072b41a7fc" path="/var/lib/kubelet/pods/e0a9e15e-9947-44be-872f-20072b41a7fc/volumes" Jan 27 15:25:34 crc kubenswrapper[4772]: I0127 15:25:34.782993 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-lll86"] Jan 27 15:25:34 crc kubenswrapper[4772]: W0127 15:25:34.783338 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6aa637d_4418_4fa4_8a26_249446d2fb3f.slice/crio-353300bf1914ec8c1fafaa4dfe7633842f95697653e6f9ec7954d70422c9cfbd WatchSource:0}: Error finding container 353300bf1914ec8c1fafaa4dfe7633842f95697653e6f9ec7954d70422c9cfbd: Status 404 returned error can't find the container with id 353300bf1914ec8c1fafaa4dfe7633842f95697653e6f9ec7954d70422c9cfbd Jan 27 15:25:35 crc kubenswrapper[4772]: I0127 15:25:35.656140 4772 generic.go:334] "Generic (PLEG): container finished" podID="dd1eaad6-cd29-4189-8ecd-62b7658e69ef" containerID="7ce8beebc480cca9e2ff0700b901cda6f6e2d53f77d8edbfd7e337a2359ae80a" exitCode=0 Jan 27 15:25:35 crc kubenswrapper[4772]: I0127 15:25:35.656335 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-547fc" event={"ID":"dd1eaad6-cd29-4189-8ecd-62b7658e69ef","Type":"ContainerDied","Data":"7ce8beebc480cca9e2ff0700b901cda6f6e2d53f77d8edbfd7e337a2359ae80a"} Jan 27 15:25:35 crc kubenswrapper[4772]: I0127 15:25:35.658114 4772 generic.go:334] "Generic (PLEG): container finished" podID="b6aa637d-4418-4fa4-8a26-249446d2fb3f" containerID="26a0610819b472b19e1babe3f9b5893ac7bd92b0c9047d536f0dadb42db99a12" exitCode=0 Jan 27 15:25:35 crc kubenswrapper[4772]: I0127 15:25:35.658147 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-lll86" event={"ID":"b6aa637d-4418-4fa4-8a26-249446d2fb3f","Type":"ContainerDied","Data":"26a0610819b472b19e1babe3f9b5893ac7bd92b0c9047d536f0dadb42db99a12"} Jan 27 15:25:35 crc kubenswrapper[4772]: I0127 15:25:35.658214 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-lll86" event={"ID":"b6aa637d-4418-4fa4-8a26-249446d2fb3f","Type":"ContainerStarted","Data":"353300bf1914ec8c1fafaa4dfe7633842f95697653e6f9ec7954d70422c9cfbd"} Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.473715 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-nmvpf"] Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.475281 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-nmvpf" Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.486601 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-nmvpf"] Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.589475 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-4c0e-account-create-update-w9dkg"] Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.591746 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4c0e-account-create-update-w9dkg" Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.594834 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.610065 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-4c0e-account-create-update-w9dkg"] Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.622711 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zzs9\" (UniqueName: \"kubernetes.io/projected/bbad3a30-e11d-4ae8-9c42-e06b6382c6de-kube-api-access-5zzs9\") pod \"keystone-db-create-nmvpf\" (UID: \"bbad3a30-e11d-4ae8-9c42-e06b6382c6de\") " pod="openstack/keystone-db-create-nmvpf" Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.622933 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbad3a30-e11d-4ae8-9c42-e06b6382c6de-operator-scripts\") pod \"keystone-db-create-nmvpf\" (UID: \"bbad3a30-e11d-4ae8-9c42-e06b6382c6de\") " pod="openstack/keystone-db-create-nmvpf" Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.674302 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-lll86" event={"ID":"b6aa637d-4418-4fa4-8a26-249446d2fb3f","Type":"ContainerStarted","Data":"ded8f7e741d736bdfe8cef79d54407ecbfa8926bb6d56e27836f39ea6ec4c8ef"} Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.690300 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-lll86" podStartSLOduration=3.690277525 podStartE2EDuration="3.690277525s" podCreationTimestamp="2026-01-27 15:25:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:25:36.685607421 +0000 UTC m=+1122.666216539" watchObservedRunningTime="2026-01-27 15:25:36.690277525 +0000 UTC m=+1122.670886633" Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.724023 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzq4h\" (UniqueName: \"kubernetes.io/projected/ef900211-2a44-498c-adb6-fec1abcba5ec-kube-api-access-kzq4h\") pod \"keystone-4c0e-account-create-update-w9dkg\" (UID: \"ef900211-2a44-498c-adb6-fec1abcba5ec\") " pod="openstack/keystone-4c0e-account-create-update-w9dkg" Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.724096 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zzs9\" (UniqueName: \"kubernetes.io/projected/bbad3a30-e11d-4ae8-9c42-e06b6382c6de-kube-api-access-5zzs9\") pod \"keystone-db-create-nmvpf\" (UID: \"bbad3a30-e11d-4ae8-9c42-e06b6382c6de\") " pod="openstack/keystone-db-create-nmvpf" Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.724357 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbad3a30-e11d-4ae8-9c42-e06b6382c6de-operator-scripts\") pod \"keystone-db-create-nmvpf\" (UID: \"bbad3a30-e11d-4ae8-9c42-e06b6382c6de\") " pod="openstack/keystone-db-create-nmvpf" Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.724466 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef900211-2a44-498c-adb6-fec1abcba5ec-operator-scripts\") pod \"keystone-4c0e-account-create-update-w9dkg\" (UID: \"ef900211-2a44-498c-adb6-fec1abcba5ec\") " pod="openstack/keystone-4c0e-account-create-update-w9dkg" Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.725150 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbad3a30-e11d-4ae8-9c42-e06b6382c6de-operator-scripts\") pod \"keystone-db-create-nmvpf\" (UID: \"bbad3a30-e11d-4ae8-9c42-e06b6382c6de\") " pod="openstack/keystone-db-create-nmvpf" Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.741660 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zzs9\" (UniqueName: \"kubernetes.io/projected/bbad3a30-e11d-4ae8-9c42-e06b6382c6de-kube-api-access-5zzs9\") pod \"keystone-db-create-nmvpf\" (UID: \"bbad3a30-e11d-4ae8-9c42-e06b6382c6de\") " pod="openstack/keystone-db-create-nmvpf" Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.795312 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-cg94r"] Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.796862 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cg94r" Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.804797 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-cg94r"] Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.826291 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef900211-2a44-498c-adb6-fec1abcba5ec-operator-scripts\") pod \"keystone-4c0e-account-create-update-w9dkg\" (UID: \"ef900211-2a44-498c-adb6-fec1abcba5ec\") " pod="openstack/keystone-4c0e-account-create-update-w9dkg" Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.826626 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzq4h\" (UniqueName: \"kubernetes.io/projected/ef900211-2a44-498c-adb6-fec1abcba5ec-kube-api-access-kzq4h\") pod \"keystone-4c0e-account-create-update-w9dkg\" (UID: \"ef900211-2a44-498c-adb6-fec1abcba5ec\") " pod="openstack/keystone-4c0e-account-create-update-w9dkg" Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.827259 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef900211-2a44-498c-adb6-fec1abcba5ec-operator-scripts\") pod \"keystone-4c0e-account-create-update-w9dkg\" (UID: \"ef900211-2a44-498c-adb6-fec1abcba5ec\") " pod="openstack/keystone-4c0e-account-create-update-w9dkg" Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.836773 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-nmvpf" Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.846139 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzq4h\" (UniqueName: \"kubernetes.io/projected/ef900211-2a44-498c-adb6-fec1abcba5ec-kube-api-access-kzq4h\") pod \"keystone-4c0e-account-create-update-w9dkg\" (UID: \"ef900211-2a44-498c-adb6-fec1abcba5ec\") " pod="openstack/keystone-4c0e-account-create-update-w9dkg" Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.908974 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-e8b1-account-create-update-8rlww"] Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.910551 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4c0e-account-create-update-w9dkg" Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.911237 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e8b1-account-create-update-8rlww" Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.913808 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.916279 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-e8b1-account-create-update-8rlww"] Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.927864 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b68551f-119d-4d84-9c91-20e013018b7a-operator-scripts\") pod \"placement-db-create-cg94r\" (UID: \"2b68551f-119d-4d84-9c91-20e013018b7a\") " pod="openstack/placement-db-create-cg94r" Jan 27 15:25:36 crc kubenswrapper[4772]: I0127 15:25:36.928016 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2fg6\" (UniqueName: \"kubernetes.io/projected/2b68551f-119d-4d84-9c91-20e013018b7a-kube-api-access-s2fg6\") pod \"placement-db-create-cg94r\" (UID: \"2b68551f-119d-4d84-9c91-20e013018b7a\") " pod="openstack/placement-db-create-cg94r" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.029447 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pssdb\" (UniqueName: \"kubernetes.io/projected/9cbda9cc-3ec5-4193-a7fb-ff06bdd20846-kube-api-access-pssdb\") pod \"placement-e8b1-account-create-update-8rlww\" (UID: \"9cbda9cc-3ec5-4193-a7fb-ff06bdd20846\") " pod="openstack/placement-e8b1-account-create-update-8rlww" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.029838 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2fg6\" (UniqueName: \"kubernetes.io/projected/2b68551f-119d-4d84-9c91-20e013018b7a-kube-api-access-s2fg6\") pod \"placement-db-create-cg94r\" (UID: \"2b68551f-119d-4d84-9c91-20e013018b7a\") " pod="openstack/placement-db-create-cg94r" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.029935 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b68551f-119d-4d84-9c91-20e013018b7a-operator-scripts\") pod \"placement-db-create-cg94r\" (UID: \"2b68551f-119d-4d84-9c91-20e013018b7a\") " pod="openstack/placement-db-create-cg94r" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.030078 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cbda9cc-3ec5-4193-a7fb-ff06bdd20846-operator-scripts\") pod \"placement-e8b1-account-create-update-8rlww\" (UID: \"9cbda9cc-3ec5-4193-a7fb-ff06bdd20846\") " pod="openstack/placement-e8b1-account-create-update-8rlww" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.030956 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b68551f-119d-4d84-9c91-20e013018b7a-operator-scripts\") pod \"placement-db-create-cg94r\" (UID: \"2b68551f-119d-4d84-9c91-20e013018b7a\") " pod="openstack/placement-db-create-cg94r" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.047660 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-547fc" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.051865 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2fg6\" (UniqueName: \"kubernetes.io/projected/2b68551f-119d-4d84-9c91-20e013018b7a-kube-api-access-s2fg6\") pod \"placement-db-create-cg94r\" (UID: \"2b68551f-119d-4d84-9c91-20e013018b7a\") " pod="openstack/placement-db-create-cg94r" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.102719 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-dpr42"] Jan 27 15:25:37 crc kubenswrapper[4772]: E0127 15:25:37.103126 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd1eaad6-cd29-4189-8ecd-62b7658e69ef" containerName="mariadb-account-create-update" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.103151 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd1eaad6-cd29-4189-8ecd-62b7658e69ef" containerName="mariadb-account-create-update" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.103386 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd1eaad6-cd29-4189-8ecd-62b7658e69ef" containerName="mariadb-account-create-update" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.104036 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dpr42" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.117892 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-dpr42"] Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.124783 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cg94r" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.133184 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cbda9cc-3ec5-4193-a7fb-ff06bdd20846-operator-scripts\") pod \"placement-e8b1-account-create-update-8rlww\" (UID: \"9cbda9cc-3ec5-4193-a7fb-ff06bdd20846\") " pod="openstack/placement-e8b1-account-create-update-8rlww" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.133269 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pssdb\" (UniqueName: \"kubernetes.io/projected/9cbda9cc-3ec5-4193-a7fb-ff06bdd20846-kube-api-access-pssdb\") pod \"placement-e8b1-account-create-update-8rlww\" (UID: \"9cbda9cc-3ec5-4193-a7fb-ff06bdd20846\") " pod="openstack/placement-e8b1-account-create-update-8rlww" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.134022 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cbda9cc-3ec5-4193-a7fb-ff06bdd20846-operator-scripts\") pod \"placement-e8b1-account-create-update-8rlww\" (UID: \"9cbda9cc-3ec5-4193-a7fb-ff06bdd20846\") " pod="openstack/placement-e8b1-account-create-update-8rlww" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.152608 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pssdb\" (UniqueName: \"kubernetes.io/projected/9cbda9cc-3ec5-4193-a7fb-ff06bdd20846-kube-api-access-pssdb\") pod \"placement-e8b1-account-create-update-8rlww\" (UID: \"9cbda9cc-3ec5-4193-a7fb-ff06bdd20846\") " pod="openstack/placement-e8b1-account-create-update-8rlww" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.159954 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.178153 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-gxjzh" podUID="220011f2-8778-4a14-82d4-33a07bd33379" containerName="ovn-controller" probeResult="failure" output=< Jan 27 15:25:37 crc kubenswrapper[4772]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Jan 27 15:25:37 crc kubenswrapper[4772]: > Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.206742 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-a648-account-create-update-qhx8z"] Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.208967 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a648-account-create-update-qhx8z" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.211603 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.213960 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.229858 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-a648-account-create-update-qhx8z"] Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.236864 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4q5jf\" (UniqueName: \"kubernetes.io/projected/dd1eaad6-cd29-4189-8ecd-62b7658e69ef-kube-api-access-4q5jf\") pod \"dd1eaad6-cd29-4189-8ecd-62b7658e69ef\" (UID: \"dd1eaad6-cd29-4189-8ecd-62b7658e69ef\") " Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.237041 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd1eaad6-cd29-4189-8ecd-62b7658e69ef-operator-scripts\") pod \"dd1eaad6-cd29-4189-8ecd-62b7658e69ef\" (UID: \"dd1eaad6-cd29-4189-8ecd-62b7658e69ef\") " Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.237550 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ww95\" (UniqueName: \"kubernetes.io/projected/af586fb2-38ff-4e17-86bc-a7793cb3ac45-kube-api-access-6ww95\") pod \"glance-db-create-dpr42\" (UID: \"af586fb2-38ff-4e17-86bc-a7793cb3ac45\") " pod="openstack/glance-db-create-dpr42" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.237591 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af586fb2-38ff-4e17-86bc-a7793cb3ac45-operator-scripts\") pod \"glance-db-create-dpr42\" (UID: \"af586fb2-38ff-4e17-86bc-a7793cb3ac45\") " pod="openstack/glance-db-create-dpr42" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.238071 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd1eaad6-cd29-4189-8ecd-62b7658e69ef-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dd1eaad6-cd29-4189-8ecd-62b7658e69ef" (UID: "dd1eaad6-cd29-4189-8ecd-62b7658e69ef"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.246101 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd1eaad6-cd29-4189-8ecd-62b7658e69ef-kube-api-access-4q5jf" (OuterVolumeSpecName: "kube-api-access-4q5jf") pod "dd1eaad6-cd29-4189-8ecd-62b7658e69ef" (UID: "dd1eaad6-cd29-4189-8ecd-62b7658e69ef"). InnerVolumeSpecName "kube-api-access-4q5jf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.345301 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmjw4\" (UniqueName: \"kubernetes.io/projected/752279e5-88ff-469d-a4db-2942659c7e24-kube-api-access-vmjw4\") pod \"glance-a648-account-create-update-qhx8z\" (UID: \"752279e5-88ff-469d-a4db-2942659c7e24\") " pod="openstack/glance-a648-account-create-update-qhx8z" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.345726 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/752279e5-88ff-469d-a4db-2942659c7e24-operator-scripts\") pod \"glance-a648-account-create-update-qhx8z\" (UID: \"752279e5-88ff-469d-a4db-2942659c7e24\") " pod="openstack/glance-a648-account-create-update-qhx8z" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.345811 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ww95\" (UniqueName: \"kubernetes.io/projected/af586fb2-38ff-4e17-86bc-a7793cb3ac45-kube-api-access-6ww95\") pod \"glance-db-create-dpr42\" (UID: \"af586fb2-38ff-4e17-86bc-a7793cb3ac45\") " pod="openstack/glance-db-create-dpr42" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.345843 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af586fb2-38ff-4e17-86bc-a7793cb3ac45-operator-scripts\") pod \"glance-db-create-dpr42\" (UID: \"af586fb2-38ff-4e17-86bc-a7793cb3ac45\") " pod="openstack/glance-db-create-dpr42" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.345973 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4q5jf\" (UniqueName: \"kubernetes.io/projected/dd1eaad6-cd29-4189-8ecd-62b7658e69ef-kube-api-access-4q5jf\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.345992 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd1eaad6-cd29-4189-8ecd-62b7658e69ef-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.346826 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af586fb2-38ff-4e17-86bc-a7793cb3ac45-operator-scripts\") pod \"glance-db-create-dpr42\" (UID: \"af586fb2-38ff-4e17-86bc-a7793cb3ac45\") " pod="openstack/glance-db-create-dpr42" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.361931 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-nmvpf"] Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.365976 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ww95\" (UniqueName: \"kubernetes.io/projected/af586fb2-38ff-4e17-86bc-a7793cb3ac45-kube-api-access-6ww95\") pod \"glance-db-create-dpr42\" (UID: \"af586fb2-38ff-4e17-86bc-a7793cb3ac45\") " pod="openstack/glance-db-create-dpr42" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.366647 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e8b1-account-create-update-8rlww" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.430458 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dpr42" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.448139 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/752279e5-88ff-469d-a4db-2942659c7e24-operator-scripts\") pod \"glance-a648-account-create-update-qhx8z\" (UID: \"752279e5-88ff-469d-a4db-2942659c7e24\") " pod="openstack/glance-a648-account-create-update-qhx8z" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.448384 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmjw4\" (UniqueName: \"kubernetes.io/projected/752279e5-88ff-469d-a4db-2942659c7e24-kube-api-access-vmjw4\") pod \"glance-a648-account-create-update-qhx8z\" (UID: \"752279e5-88ff-469d-a4db-2942659c7e24\") " pod="openstack/glance-a648-account-create-update-qhx8z" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.449587 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/752279e5-88ff-469d-a4db-2942659c7e24-operator-scripts\") pod \"glance-a648-account-create-update-qhx8z\" (UID: \"752279e5-88ff-469d-a4db-2942659c7e24\") " pod="openstack/glance-a648-account-create-update-qhx8z" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.453064 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-gxjzh-config-c4mg2"] Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.454117 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gxjzh-config-c4mg2" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.461625 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.467916 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gxjzh-config-c4mg2"] Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.495032 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmjw4\" (UniqueName: \"kubernetes.io/projected/752279e5-88ff-469d-a4db-2942659c7e24-kube-api-access-vmjw4\") pod \"glance-a648-account-create-update-qhx8z\" (UID: \"752279e5-88ff-469d-a4db-2942659c7e24\") " pod="openstack/glance-a648-account-create-update-qhx8z" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.537445 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-4c0e-account-create-update-w9dkg"] Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.550901 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a648-account-create-update-qhx8z" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.652872 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-additional-scripts\") pod \"ovn-controller-gxjzh-config-c4mg2\" (UID: \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\") " pod="openstack/ovn-controller-gxjzh-config-c4mg2" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.652920 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pd4n\" (UniqueName: \"kubernetes.io/projected/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-kube-api-access-2pd4n\") pod \"ovn-controller-gxjzh-config-c4mg2\" (UID: \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\") " pod="openstack/ovn-controller-gxjzh-config-c4mg2" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.652990 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-var-run-ovn\") pod \"ovn-controller-gxjzh-config-c4mg2\" (UID: \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\") " pod="openstack/ovn-controller-gxjzh-config-c4mg2" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.653064 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-var-log-ovn\") pod \"ovn-controller-gxjzh-config-c4mg2\" (UID: \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\") " pod="openstack/ovn-controller-gxjzh-config-c4mg2" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.653094 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-var-run\") pod \"ovn-controller-gxjzh-config-c4mg2\" (UID: \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\") " pod="openstack/ovn-controller-gxjzh-config-c4mg2" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.653110 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-scripts\") pod \"ovn-controller-gxjzh-config-c4mg2\" (UID: \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\") " pod="openstack/ovn-controller-gxjzh-config-c4mg2" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.655646 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-cg94r"] Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.692382 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-547fc" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.692377 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-547fc" event={"ID":"dd1eaad6-cd29-4189-8ecd-62b7658e69ef","Type":"ContainerDied","Data":"ba7e59e7bcd568f0fef4ca50eaae4a57d525926c5e0abdd0b3a8025ace15dd12"} Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.694354 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba7e59e7bcd568f0fef4ca50eaae4a57d525926c5e0abdd0b3a8025ace15dd12" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.699866 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cg94r" event={"ID":"2b68551f-119d-4d84-9c91-20e013018b7a","Type":"ContainerStarted","Data":"69a5b31d0e44865250d8bac44f85dd2b9adb78c9d00444c5fa3b2797795ff8b2"} Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.702271 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4c0e-account-create-update-w9dkg" event={"ID":"ef900211-2a44-498c-adb6-fec1abcba5ec","Type":"ContainerStarted","Data":"04dbba6fb62087c40aa12d14c3493253436c9c1687fdba32ecfb0a887e347e49"} Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.705720 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-nmvpf" event={"ID":"bbad3a30-e11d-4ae8-9c42-e06b6382c6de","Type":"ContainerStarted","Data":"31d9e486da5aa706768e022e398d969ef41f15c9db5b579c83d50ae160db05a7"} Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.705787 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-nmvpf" event={"ID":"bbad3a30-e11d-4ae8-9c42-e06b6382c6de","Type":"ContainerStarted","Data":"8df1e3e2195f10d57dae3b8ddab8a9f5430c67dd9cc313305d8caf2dd324bd9e"} Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.706939 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-lll86" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.726455 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-nmvpf" podStartSLOduration=1.726386395 podStartE2EDuration="1.726386395s" podCreationTimestamp="2026-01-27 15:25:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:25:37.72273344 +0000 UTC m=+1123.703342538" watchObservedRunningTime="2026-01-27 15:25:37.726386395 +0000 UTC m=+1123.706995493" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.755067 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-var-run-ovn\") pod \"ovn-controller-gxjzh-config-c4mg2\" (UID: \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\") " pod="openstack/ovn-controller-gxjzh-config-c4mg2" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.755124 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-var-log-ovn\") pod \"ovn-controller-gxjzh-config-c4mg2\" (UID: \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\") " pod="openstack/ovn-controller-gxjzh-config-c4mg2" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.755146 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-var-run\") pod \"ovn-controller-gxjzh-config-c4mg2\" (UID: \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\") " pod="openstack/ovn-controller-gxjzh-config-c4mg2" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.755179 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-scripts\") pod \"ovn-controller-gxjzh-config-c4mg2\" (UID: \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\") " pod="openstack/ovn-controller-gxjzh-config-c4mg2" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.755249 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-additional-scripts\") pod \"ovn-controller-gxjzh-config-c4mg2\" (UID: \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\") " pod="openstack/ovn-controller-gxjzh-config-c4mg2" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.755282 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pd4n\" (UniqueName: \"kubernetes.io/projected/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-kube-api-access-2pd4n\") pod \"ovn-controller-gxjzh-config-c4mg2\" (UID: \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\") " pod="openstack/ovn-controller-gxjzh-config-c4mg2" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.756114 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-var-run-ovn\") pod \"ovn-controller-gxjzh-config-c4mg2\" (UID: \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\") " pod="openstack/ovn-controller-gxjzh-config-c4mg2" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.756180 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-var-log-ovn\") pod \"ovn-controller-gxjzh-config-c4mg2\" (UID: \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\") " pod="openstack/ovn-controller-gxjzh-config-c4mg2" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.756220 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-var-run\") pod \"ovn-controller-gxjzh-config-c4mg2\" (UID: \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\") " pod="openstack/ovn-controller-gxjzh-config-c4mg2" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.758217 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-scripts\") pod \"ovn-controller-gxjzh-config-c4mg2\" (UID: \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\") " pod="openstack/ovn-controller-gxjzh-config-c4mg2" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.758711 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-additional-scripts\") pod \"ovn-controller-gxjzh-config-c4mg2\" (UID: \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\") " pod="openstack/ovn-controller-gxjzh-config-c4mg2" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.777635 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pd4n\" (UniqueName: \"kubernetes.io/projected/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-kube-api-access-2pd4n\") pod \"ovn-controller-gxjzh-config-c4mg2\" (UID: \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\") " pod="openstack/ovn-controller-gxjzh-config-c4mg2" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.781720 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gxjzh-config-c4mg2" Jan 27 15:25:37 crc kubenswrapper[4772]: I0127 15:25:37.954051 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-e8b1-account-create-update-8rlww"] Jan 27 15:25:38 crc kubenswrapper[4772]: I0127 15:25:38.061067 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-dpr42"] Jan 27 15:25:38 crc kubenswrapper[4772]: I0127 15:25:38.124039 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-a648-account-create-update-qhx8z"] Jan 27 15:25:38 crc kubenswrapper[4772]: W0127 15:25:38.138692 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod752279e5_88ff_469d_a4db_2942659c7e24.slice/crio-71e0dffca8f1bcda6eb557e7af38812a722313603f4af46f0a96549446e9a419 WatchSource:0}: Error finding container 71e0dffca8f1bcda6eb557e7af38812a722313603f4af46f0a96549446e9a419: Status 404 returned error can't find the container with id 71e0dffca8f1bcda6eb557e7af38812a722313603f4af46f0a96549446e9a419 Jan 27 15:25:38 crc kubenswrapper[4772]: I0127 15:25:38.276610 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gxjzh-config-c4mg2"] Jan 27 15:25:38 crc kubenswrapper[4772]: I0127 15:25:38.718026 4772 generic.go:334] "Generic (PLEG): container finished" podID="bbad3a30-e11d-4ae8-9c42-e06b6382c6de" containerID="31d9e486da5aa706768e022e398d969ef41f15c9db5b579c83d50ae160db05a7" exitCode=0 Jan 27 15:25:38 crc kubenswrapper[4772]: I0127 15:25:38.718104 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-nmvpf" event={"ID":"bbad3a30-e11d-4ae8-9c42-e06b6382c6de","Type":"ContainerDied","Data":"31d9e486da5aa706768e022e398d969ef41f15c9db5b579c83d50ae160db05a7"} Jan 27 15:25:38 crc kubenswrapper[4772]: I0127 15:25:38.719850 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cg94r" event={"ID":"2b68551f-119d-4d84-9c91-20e013018b7a","Type":"ContainerStarted","Data":"6aa60721dd7c09b05e3a663482308f5a6da188370cc19651da9a73a40e00696f"} Jan 27 15:25:38 crc kubenswrapper[4772]: I0127 15:25:38.722076 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gxjzh-config-c4mg2" event={"ID":"e7333f8a-0a54-4dec-8e7a-c7a648d2a841","Type":"ContainerStarted","Data":"d22be9ecfb9cc0389dd0f2e64dbdb2f980e40813484563a0a652ad657fd8f5b7"} Jan 27 15:25:38 crc kubenswrapper[4772]: I0127 15:25:38.722136 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gxjzh-config-c4mg2" event={"ID":"e7333f8a-0a54-4dec-8e7a-c7a648d2a841","Type":"ContainerStarted","Data":"ea92769dda2095af7844e2ac2e1dec699541d4477733f6b0a3693467b5238a7b"} Jan 27 15:25:38 crc kubenswrapper[4772]: I0127 15:25:38.723897 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e8b1-account-create-update-8rlww" event={"ID":"9cbda9cc-3ec5-4193-a7fb-ff06bdd20846","Type":"ContainerStarted","Data":"e068687fbbe8ba2bc884327a323113a2f9b397134b3783cc71145217f0aced72"} Jan 27 15:25:38 crc kubenswrapper[4772]: I0127 15:25:38.723949 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e8b1-account-create-update-8rlww" event={"ID":"9cbda9cc-3ec5-4193-a7fb-ff06bdd20846","Type":"ContainerStarted","Data":"b7b0b5ce95f1f8f79fe79dcceda26c46edd68768fdca8bfa61a8c002836d8e7e"} Jan 27 15:25:38 crc kubenswrapper[4772]: I0127 15:25:38.725721 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a648-account-create-update-qhx8z" event={"ID":"752279e5-88ff-469d-a4db-2942659c7e24","Type":"ContainerStarted","Data":"fcb62876ceaa2921dde5172c985a61fd201c04281f9d06dbb383e8128d91c935"} Jan 27 15:25:38 crc kubenswrapper[4772]: I0127 15:25:38.725760 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a648-account-create-update-qhx8z" event={"ID":"752279e5-88ff-469d-a4db-2942659c7e24","Type":"ContainerStarted","Data":"71e0dffca8f1bcda6eb557e7af38812a722313603f4af46f0a96549446e9a419"} Jan 27 15:25:38 crc kubenswrapper[4772]: I0127 15:25:38.728666 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dpr42" event={"ID":"af586fb2-38ff-4e17-86bc-a7793cb3ac45","Type":"ContainerStarted","Data":"d569280ad66a5087c9e0aa7b8abe04a7d97361bee2ca7b7c30646e77734ba51d"} Jan 27 15:25:38 crc kubenswrapper[4772]: I0127 15:25:38.728738 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dpr42" event={"ID":"af586fb2-38ff-4e17-86bc-a7793cb3ac45","Type":"ContainerStarted","Data":"44aa414f58e36e4ce3bdd0cdcc25ef0840eb5c07319ac311b07bd27876572c79"} Jan 27 15:25:38 crc kubenswrapper[4772]: I0127 15:25:38.730493 4772 generic.go:334] "Generic (PLEG): container finished" podID="ef900211-2a44-498c-adb6-fec1abcba5ec" containerID="37cb21cfa353006443b3a1e31571db32c636cbf5e0c7a880cb766a2b91769826" exitCode=0 Jan 27 15:25:38 crc kubenswrapper[4772]: I0127 15:25:38.730551 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4c0e-account-create-update-w9dkg" event={"ID":"ef900211-2a44-498c-adb6-fec1abcba5ec","Type":"ContainerDied","Data":"37cb21cfa353006443b3a1e31571db32c636cbf5e0c7a880cb766a2b91769826"} Jan 27 15:25:38 crc kubenswrapper[4772]: I0127 15:25:38.765853 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-cg94r" podStartSLOduration=2.765830132 podStartE2EDuration="2.765830132s" podCreationTimestamp="2026-01-27 15:25:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:25:38.75848066 +0000 UTC m=+1124.739089768" watchObservedRunningTime="2026-01-27 15:25:38.765830132 +0000 UTC m=+1124.746439230" Jan 27 15:25:38 crc kubenswrapper[4772]: I0127 15:25:38.778687 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-dpr42" podStartSLOduration=1.778661981 podStartE2EDuration="1.778661981s" podCreationTimestamp="2026-01-27 15:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:25:38.774770559 +0000 UTC m=+1124.755379657" watchObservedRunningTime="2026-01-27 15:25:38.778661981 +0000 UTC m=+1124.759271079" Jan 27 15:25:38 crc kubenswrapper[4772]: I0127 15:25:38.813639 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-e8b1-account-create-update-8rlww" podStartSLOduration=2.813615948 podStartE2EDuration="2.813615948s" podCreationTimestamp="2026-01-27 15:25:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:25:38.810433447 +0000 UTC m=+1124.791042555" watchObservedRunningTime="2026-01-27 15:25:38.813615948 +0000 UTC m=+1124.794225046" Jan 27 15:25:38 crc kubenswrapper[4772]: I0127 15:25:38.817399 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-a648-account-create-update-qhx8z" podStartSLOduration=1.8173877470000002 podStartE2EDuration="1.817387747s" podCreationTimestamp="2026-01-27 15:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:25:38.796878626 +0000 UTC m=+1124.777487724" watchObservedRunningTime="2026-01-27 15:25:38.817387747 +0000 UTC m=+1124.797996845" Jan 27 15:25:39 crc kubenswrapper[4772]: I0127 15:25:39.740458 4772 generic.go:334] "Generic (PLEG): container finished" podID="af586fb2-38ff-4e17-86bc-a7793cb3ac45" containerID="d569280ad66a5087c9e0aa7b8abe04a7d97361bee2ca7b7c30646e77734ba51d" exitCode=0 Jan 27 15:25:39 crc kubenswrapper[4772]: I0127 15:25:39.740532 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dpr42" event={"ID":"af586fb2-38ff-4e17-86bc-a7793cb3ac45","Type":"ContainerDied","Data":"d569280ad66a5087c9e0aa7b8abe04a7d97361bee2ca7b7c30646e77734ba51d"} Jan 27 15:25:39 crc kubenswrapper[4772]: I0127 15:25:39.743459 4772 generic.go:334] "Generic (PLEG): container finished" podID="2b68551f-119d-4d84-9c91-20e013018b7a" containerID="6aa60721dd7c09b05e3a663482308f5a6da188370cc19651da9a73a40e00696f" exitCode=0 Jan 27 15:25:39 crc kubenswrapper[4772]: I0127 15:25:39.743508 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cg94r" event={"ID":"2b68551f-119d-4d84-9c91-20e013018b7a","Type":"ContainerDied","Data":"6aa60721dd7c09b05e3a663482308f5a6da188370cc19651da9a73a40e00696f"} Jan 27 15:25:39 crc kubenswrapper[4772]: I0127 15:25:39.745072 4772 generic.go:334] "Generic (PLEG): container finished" podID="e7333f8a-0a54-4dec-8e7a-c7a648d2a841" containerID="d22be9ecfb9cc0389dd0f2e64dbdb2f980e40813484563a0a652ad657fd8f5b7" exitCode=0 Jan 27 15:25:39 crc kubenswrapper[4772]: I0127 15:25:39.745248 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gxjzh-config-c4mg2" event={"ID":"e7333f8a-0a54-4dec-8e7a-c7a648d2a841","Type":"ContainerDied","Data":"d22be9ecfb9cc0389dd0f2e64dbdb2f980e40813484563a0a652ad657fd8f5b7"} Jan 27 15:25:39 crc kubenswrapper[4772]: E0127 15:25:39.782984 4772 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf586fb2_38ff_4e17_86bc_a7793cb3ac45.slice/crio-d569280ad66a5087c9e0aa7b8abe04a7d97361bee2ca7b7c30646e77734ba51d.scope\": RecentStats: unable to find data in memory cache]" Jan 27 15:25:40 crc kubenswrapper[4772]: I0127 15:25:40.332255 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4c0e-account-create-update-w9dkg" Jan 27 15:25:40 crc kubenswrapper[4772]: I0127 15:25:40.339274 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-nmvpf" Jan 27 15:25:40 crc kubenswrapper[4772]: I0127 15:25:40.515290 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zzs9\" (UniqueName: \"kubernetes.io/projected/bbad3a30-e11d-4ae8-9c42-e06b6382c6de-kube-api-access-5zzs9\") pod \"bbad3a30-e11d-4ae8-9c42-e06b6382c6de\" (UID: \"bbad3a30-e11d-4ae8-9c42-e06b6382c6de\") " Jan 27 15:25:40 crc kubenswrapper[4772]: I0127 15:25:40.515505 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzq4h\" (UniqueName: \"kubernetes.io/projected/ef900211-2a44-498c-adb6-fec1abcba5ec-kube-api-access-kzq4h\") pod \"ef900211-2a44-498c-adb6-fec1abcba5ec\" (UID: \"ef900211-2a44-498c-adb6-fec1abcba5ec\") " Jan 27 15:25:40 crc kubenswrapper[4772]: I0127 15:25:40.515607 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef900211-2a44-498c-adb6-fec1abcba5ec-operator-scripts\") pod \"ef900211-2a44-498c-adb6-fec1abcba5ec\" (UID: \"ef900211-2a44-498c-adb6-fec1abcba5ec\") " Jan 27 15:25:40 crc kubenswrapper[4772]: I0127 15:25:40.515645 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbad3a30-e11d-4ae8-9c42-e06b6382c6de-operator-scripts\") pod \"bbad3a30-e11d-4ae8-9c42-e06b6382c6de\" (UID: \"bbad3a30-e11d-4ae8-9c42-e06b6382c6de\") " Jan 27 15:25:40 crc kubenswrapper[4772]: I0127 15:25:40.516473 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbad3a30-e11d-4ae8-9c42-e06b6382c6de-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bbad3a30-e11d-4ae8-9c42-e06b6382c6de" (UID: "bbad3a30-e11d-4ae8-9c42-e06b6382c6de"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:40 crc kubenswrapper[4772]: I0127 15:25:40.516560 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef900211-2a44-498c-adb6-fec1abcba5ec-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ef900211-2a44-498c-adb6-fec1abcba5ec" (UID: "ef900211-2a44-498c-adb6-fec1abcba5ec"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:40 crc kubenswrapper[4772]: I0127 15:25:40.521695 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbad3a30-e11d-4ae8-9c42-e06b6382c6de-kube-api-access-5zzs9" (OuterVolumeSpecName: "kube-api-access-5zzs9") pod "bbad3a30-e11d-4ae8-9c42-e06b6382c6de" (UID: "bbad3a30-e11d-4ae8-9c42-e06b6382c6de"). InnerVolumeSpecName "kube-api-access-5zzs9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:25:40 crc kubenswrapper[4772]: I0127 15:25:40.522338 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef900211-2a44-498c-adb6-fec1abcba5ec-kube-api-access-kzq4h" (OuterVolumeSpecName: "kube-api-access-kzq4h") pod "ef900211-2a44-498c-adb6-fec1abcba5ec" (UID: "ef900211-2a44-498c-adb6-fec1abcba5ec"). InnerVolumeSpecName "kube-api-access-kzq4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:25:40 crc kubenswrapper[4772]: I0127 15:25:40.618339 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzq4h\" (UniqueName: \"kubernetes.io/projected/ef900211-2a44-498c-adb6-fec1abcba5ec-kube-api-access-kzq4h\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:40 crc kubenswrapper[4772]: I0127 15:25:40.618380 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef900211-2a44-498c-adb6-fec1abcba5ec-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:40 crc kubenswrapper[4772]: I0127 15:25:40.618398 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbad3a30-e11d-4ae8-9c42-e06b6382c6de-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:40 crc kubenswrapper[4772]: I0127 15:25:40.618412 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zzs9\" (UniqueName: \"kubernetes.io/projected/bbad3a30-e11d-4ae8-9c42-e06b6382c6de-kube-api-access-5zzs9\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:40 crc kubenswrapper[4772]: I0127 15:25:40.759123 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-4c0e-account-create-update-w9dkg" event={"ID":"ef900211-2a44-498c-adb6-fec1abcba5ec","Type":"ContainerDied","Data":"04dbba6fb62087c40aa12d14c3493253436c9c1687fdba32ecfb0a887e347e49"} Jan 27 15:25:40 crc kubenswrapper[4772]: I0127 15:25:40.759194 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04dbba6fb62087c40aa12d14c3493253436c9c1687fdba32ecfb0a887e347e49" Jan 27 15:25:40 crc kubenswrapper[4772]: I0127 15:25:40.760254 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4c0e-account-create-update-w9dkg" Jan 27 15:25:40 crc kubenswrapper[4772]: I0127 15:25:40.761420 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-nmvpf" event={"ID":"bbad3a30-e11d-4ae8-9c42-e06b6382c6de","Type":"ContainerDied","Data":"8df1e3e2195f10d57dae3b8ddab8a9f5430c67dd9cc313305d8caf2dd324bd9e"} Jan 27 15:25:40 crc kubenswrapper[4772]: I0127 15:25:40.761429 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-nmvpf" Jan 27 15:25:40 crc kubenswrapper[4772]: I0127 15:25:40.761441 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8df1e3e2195f10d57dae3b8ddab8a9f5430c67dd9cc313305d8caf2dd324bd9e" Jan 27 15:25:40 crc kubenswrapper[4772]: I0127 15:25:40.762856 4772 generic.go:334] "Generic (PLEG): container finished" podID="9cbda9cc-3ec5-4193-a7fb-ff06bdd20846" containerID="e068687fbbe8ba2bc884327a323113a2f9b397134b3783cc71145217f0aced72" exitCode=0 Jan 27 15:25:40 crc kubenswrapper[4772]: I0127 15:25:40.762945 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e8b1-account-create-update-8rlww" event={"ID":"9cbda9cc-3ec5-4193-a7fb-ff06bdd20846","Type":"ContainerDied","Data":"e068687fbbe8ba2bc884327a323113a2f9b397134b3783cc71145217f0aced72"} Jan 27 15:25:40 crc kubenswrapper[4772]: I0127 15:25:40.766127 4772 generic.go:334] "Generic (PLEG): container finished" podID="752279e5-88ff-469d-a4db-2942659c7e24" containerID="fcb62876ceaa2921dde5172c985a61fd201c04281f9d06dbb383e8128d91c935" exitCode=0 Jan 27 15:25:40 crc kubenswrapper[4772]: I0127 15:25:40.766257 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a648-account-create-update-qhx8z" event={"ID":"752279e5-88ff-469d-a4db-2942659c7e24","Type":"ContainerDied","Data":"fcb62876ceaa2921dde5172c985a61fd201c04281f9d06dbb383e8128d91c935"} Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.040124 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dpr42" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.232373 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ww95\" (UniqueName: \"kubernetes.io/projected/af586fb2-38ff-4e17-86bc-a7793cb3ac45-kube-api-access-6ww95\") pod \"af586fb2-38ff-4e17-86bc-a7793cb3ac45\" (UID: \"af586fb2-38ff-4e17-86bc-a7793cb3ac45\") " Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.232723 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af586fb2-38ff-4e17-86bc-a7793cb3ac45-operator-scripts\") pod \"af586fb2-38ff-4e17-86bc-a7793cb3ac45\" (UID: \"af586fb2-38ff-4e17-86bc-a7793cb3ac45\") " Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.233290 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af586fb2-38ff-4e17-86bc-a7793cb3ac45-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "af586fb2-38ff-4e17-86bc-a7793cb3ac45" (UID: "af586fb2-38ff-4e17-86bc-a7793cb3ac45"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.236661 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af586fb2-38ff-4e17-86bc-a7793cb3ac45-kube-api-access-6ww95" (OuterVolumeSpecName: "kube-api-access-6ww95") pod "af586fb2-38ff-4e17-86bc-a7793cb3ac45" (UID: "af586fb2-38ff-4e17-86bc-a7793cb3ac45"). InnerVolumeSpecName "kube-api-access-6ww95". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.283663 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cg94r" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.288713 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gxjzh-config-c4mg2" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.334297 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af586fb2-38ff-4e17-86bc-a7793cb3ac45-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.334340 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ww95\" (UniqueName: \"kubernetes.io/projected/af586fb2-38ff-4e17-86bc-a7793cb3ac45-kube-api-access-6ww95\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.435631 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-scripts\") pod \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\" (UID: \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\") " Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.435712 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pd4n\" (UniqueName: \"kubernetes.io/projected/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-kube-api-access-2pd4n\") pod \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\" (UID: \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\") " Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.435729 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-var-run-ovn\") pod \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\" (UID: \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\") " Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.435755 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2fg6\" (UniqueName: \"kubernetes.io/projected/2b68551f-119d-4d84-9c91-20e013018b7a-kube-api-access-s2fg6\") pod \"2b68551f-119d-4d84-9c91-20e013018b7a\" (UID: \"2b68551f-119d-4d84-9c91-20e013018b7a\") " Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.435813 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-var-run\") pod \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\" (UID: \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\") " Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.435892 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-additional-scripts\") pod \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\" (UID: \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\") " Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.435915 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-var-log-ovn\") pod \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\" (UID: \"e7333f8a-0a54-4dec-8e7a-c7a648d2a841\") " Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.435947 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b68551f-119d-4d84-9c91-20e013018b7a-operator-scripts\") pod \"2b68551f-119d-4d84-9c91-20e013018b7a\" (UID: \"2b68551f-119d-4d84-9c91-20e013018b7a\") " Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.435944 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "e7333f8a-0a54-4dec-8e7a-c7a648d2a841" (UID: "e7333f8a-0a54-4dec-8e7a-c7a648d2a841"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.435975 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-var-run" (OuterVolumeSpecName: "var-run") pod "e7333f8a-0a54-4dec-8e7a-c7a648d2a841" (UID: "e7333f8a-0a54-4dec-8e7a-c7a648d2a841"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.436077 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "e7333f8a-0a54-4dec-8e7a-c7a648d2a841" (UID: "e7333f8a-0a54-4dec-8e7a-c7a648d2a841"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.436417 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b68551f-119d-4d84-9c91-20e013018b7a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2b68551f-119d-4d84-9c91-20e013018b7a" (UID: "2b68551f-119d-4d84-9c91-20e013018b7a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.436609 4772 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-var-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.436627 4772 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-var-run\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.436635 4772 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-var-log-ovn\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.436644 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b68551f-119d-4d84-9c91-20e013018b7a-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.436796 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "e7333f8a-0a54-4dec-8e7a-c7a648d2a841" (UID: "e7333f8a-0a54-4dec-8e7a-c7a648d2a841"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.437085 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-scripts" (OuterVolumeSpecName: "scripts") pod "e7333f8a-0a54-4dec-8e7a-c7a648d2a841" (UID: "e7333f8a-0a54-4dec-8e7a-c7a648d2a841"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.444414 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-kube-api-access-2pd4n" (OuterVolumeSpecName: "kube-api-access-2pd4n") pod "e7333f8a-0a54-4dec-8e7a-c7a648d2a841" (UID: "e7333f8a-0a54-4dec-8e7a-c7a648d2a841"). InnerVolumeSpecName "kube-api-access-2pd4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.444563 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b68551f-119d-4d84-9c91-20e013018b7a-kube-api-access-s2fg6" (OuterVolumeSpecName: "kube-api-access-s2fg6") pod "2b68551f-119d-4d84-9c91-20e013018b7a" (UID: "2b68551f-119d-4d84-9c91-20e013018b7a"). InnerVolumeSpecName "kube-api-access-s2fg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.538554 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2fg6\" (UniqueName: \"kubernetes.io/projected/2b68551f-119d-4d84-9c91-20e013018b7a-kube-api-access-s2fg6\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.538598 4772 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-additional-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.538616 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.538628 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pd4n\" (UniqueName: \"kubernetes.io/projected/e7333f8a-0a54-4dec-8e7a-c7a648d2a841-kube-api-access-2pd4n\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.774733 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dpr42" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.774726 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dpr42" event={"ID":"af586fb2-38ff-4e17-86bc-a7793cb3ac45","Type":"ContainerDied","Data":"44aa414f58e36e4ce3bdd0cdcc25ef0840eb5c07319ac311b07bd27876572c79"} Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.774842 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44aa414f58e36e4ce3bdd0cdcc25ef0840eb5c07319ac311b07bd27876572c79" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.776214 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cg94r" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.776319 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cg94r" event={"ID":"2b68551f-119d-4d84-9c91-20e013018b7a","Type":"ContainerDied","Data":"69a5b31d0e44865250d8bac44f85dd2b9adb78c9d00444c5fa3b2797795ff8b2"} Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.776391 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69a5b31d0e44865250d8bac44f85dd2b9adb78c9d00444c5fa3b2797795ff8b2" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.777693 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gxjzh-config-c4mg2" event={"ID":"e7333f8a-0a54-4dec-8e7a-c7a648d2a841","Type":"ContainerDied","Data":"ea92769dda2095af7844e2ac2e1dec699541d4477733f6b0a3693467b5238a7b"} Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.777737 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea92769dda2095af7844e2ac2e1dec699541d4477733f6b0a3693467b5238a7b" Jan 27 15:25:41 crc kubenswrapper[4772]: I0127 15:25:41.777844 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gxjzh-config-c4mg2" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.054658 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a648-account-create-update-qhx8z" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.058139 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.058197 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.127811 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-gxjzh" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.147340 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e8b1-account-create-update-8rlww" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.257298 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmjw4\" (UniqueName: \"kubernetes.io/projected/752279e5-88ff-469d-a4db-2942659c7e24-kube-api-access-vmjw4\") pod \"752279e5-88ff-469d-a4db-2942659c7e24\" (UID: \"752279e5-88ff-469d-a4db-2942659c7e24\") " Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.257368 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/752279e5-88ff-469d-a4db-2942659c7e24-operator-scripts\") pod \"752279e5-88ff-469d-a4db-2942659c7e24\" (UID: \"752279e5-88ff-469d-a4db-2942659c7e24\") " Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.257561 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cbda9cc-3ec5-4193-a7fb-ff06bdd20846-operator-scripts\") pod \"9cbda9cc-3ec5-4193-a7fb-ff06bdd20846\" (UID: \"9cbda9cc-3ec5-4193-a7fb-ff06bdd20846\") " Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.257681 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pssdb\" (UniqueName: \"kubernetes.io/projected/9cbda9cc-3ec5-4193-a7fb-ff06bdd20846-kube-api-access-pssdb\") pod \"9cbda9cc-3ec5-4193-a7fb-ff06bdd20846\" (UID: \"9cbda9cc-3ec5-4193-a7fb-ff06bdd20846\") " Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.258071 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/752279e5-88ff-469d-a4db-2942659c7e24-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "752279e5-88ff-469d-a4db-2942659c7e24" (UID: "752279e5-88ff-469d-a4db-2942659c7e24"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.258101 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cbda9cc-3ec5-4193-a7fb-ff06bdd20846-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9cbda9cc-3ec5-4193-a7fb-ff06bdd20846" (UID: "9cbda9cc-3ec5-4193-a7fb-ff06bdd20846"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.258437 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/752279e5-88ff-469d-a4db-2942659c7e24-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.258459 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cbda9cc-3ec5-4193-a7fb-ff06bdd20846-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.262108 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/752279e5-88ff-469d-a4db-2942659c7e24-kube-api-access-vmjw4" (OuterVolumeSpecName: "kube-api-access-vmjw4") pod "752279e5-88ff-469d-a4db-2942659c7e24" (UID: "752279e5-88ff-469d-a4db-2942659c7e24"). InnerVolumeSpecName "kube-api-access-vmjw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.262236 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cbda9cc-3ec5-4193-a7fb-ff06bdd20846-kube-api-access-pssdb" (OuterVolumeSpecName: "kube-api-access-pssdb") pod "9cbda9cc-3ec5-4193-a7fb-ff06bdd20846" (UID: "9cbda9cc-3ec5-4193-a7fb-ff06bdd20846"). InnerVolumeSpecName "kube-api-access-pssdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.365237 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pssdb\" (UniqueName: \"kubernetes.io/projected/9cbda9cc-3ec5-4193-a7fb-ff06bdd20846-kube-api-access-pssdb\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.365275 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmjw4\" (UniqueName: \"kubernetes.io/projected/752279e5-88ff-469d-a4db-2942659c7e24-kube-api-access-vmjw4\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.397409 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-gxjzh-config-c4mg2"] Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.407136 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-gxjzh-config-c4mg2"] Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.543661 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-gxjzh-config-gzr9r"] Jan 27 15:25:42 crc kubenswrapper[4772]: E0127 15:25:42.544016 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cbda9cc-3ec5-4193-a7fb-ff06bdd20846" containerName="mariadb-account-create-update" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.544034 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cbda9cc-3ec5-4193-a7fb-ff06bdd20846" containerName="mariadb-account-create-update" Jan 27 15:25:42 crc kubenswrapper[4772]: E0127 15:25:42.544053 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="752279e5-88ff-469d-a4db-2942659c7e24" containerName="mariadb-account-create-update" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.544058 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="752279e5-88ff-469d-a4db-2942659c7e24" containerName="mariadb-account-create-update" Jan 27 15:25:42 crc kubenswrapper[4772]: E0127 15:25:42.544070 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef900211-2a44-498c-adb6-fec1abcba5ec" containerName="mariadb-account-create-update" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.544077 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef900211-2a44-498c-adb6-fec1abcba5ec" containerName="mariadb-account-create-update" Jan 27 15:25:42 crc kubenswrapper[4772]: E0127 15:25:42.544087 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbad3a30-e11d-4ae8-9c42-e06b6382c6de" containerName="mariadb-database-create" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.544093 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbad3a30-e11d-4ae8-9c42-e06b6382c6de" containerName="mariadb-database-create" Jan 27 15:25:42 crc kubenswrapper[4772]: E0127 15:25:42.544103 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b68551f-119d-4d84-9c91-20e013018b7a" containerName="mariadb-database-create" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.544108 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b68551f-119d-4d84-9c91-20e013018b7a" containerName="mariadb-database-create" Jan 27 15:25:42 crc kubenswrapper[4772]: E0127 15:25:42.544121 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af586fb2-38ff-4e17-86bc-a7793cb3ac45" containerName="mariadb-database-create" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.544128 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="af586fb2-38ff-4e17-86bc-a7793cb3ac45" containerName="mariadb-database-create" Jan 27 15:25:42 crc kubenswrapper[4772]: E0127 15:25:42.544146 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7333f8a-0a54-4dec-8e7a-c7a648d2a841" containerName="ovn-config" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.544152 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7333f8a-0a54-4dec-8e7a-c7a648d2a841" containerName="ovn-config" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.544311 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b68551f-119d-4d84-9c91-20e013018b7a" containerName="mariadb-database-create" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.544324 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="752279e5-88ff-469d-a4db-2942659c7e24" containerName="mariadb-account-create-update" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.544332 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7333f8a-0a54-4dec-8e7a-c7a648d2a841" containerName="ovn-config" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.544346 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef900211-2a44-498c-adb6-fec1abcba5ec" containerName="mariadb-account-create-update" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.544357 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cbda9cc-3ec5-4193-a7fb-ff06bdd20846" containerName="mariadb-account-create-update" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.544367 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbad3a30-e11d-4ae8-9c42-e06b6382c6de" containerName="mariadb-database-create" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.544376 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="af586fb2-38ff-4e17-86bc-a7793cb3ac45" containerName="mariadb-database-create" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.544882 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gxjzh-config-gzr9r" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.547138 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.554095 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gxjzh-config-gzr9r"] Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.572871 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbxmh\" (UniqueName: \"kubernetes.io/projected/b287920c-19e8-47a5-9644-a1651f4c39e9-kube-api-access-xbxmh\") pod \"ovn-controller-gxjzh-config-gzr9r\" (UID: \"b287920c-19e8-47a5-9644-a1651f4c39e9\") " pod="openstack/ovn-controller-gxjzh-config-gzr9r" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.572943 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b287920c-19e8-47a5-9644-a1651f4c39e9-scripts\") pod \"ovn-controller-gxjzh-config-gzr9r\" (UID: \"b287920c-19e8-47a5-9644-a1651f4c39e9\") " pod="openstack/ovn-controller-gxjzh-config-gzr9r" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.572993 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b287920c-19e8-47a5-9644-a1651f4c39e9-additional-scripts\") pod \"ovn-controller-gxjzh-config-gzr9r\" (UID: \"b287920c-19e8-47a5-9644-a1651f4c39e9\") " pod="openstack/ovn-controller-gxjzh-config-gzr9r" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.573061 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b287920c-19e8-47a5-9644-a1651f4c39e9-var-run\") pod \"ovn-controller-gxjzh-config-gzr9r\" (UID: \"b287920c-19e8-47a5-9644-a1651f4c39e9\") " pod="openstack/ovn-controller-gxjzh-config-gzr9r" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.573376 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b287920c-19e8-47a5-9644-a1651f4c39e9-var-run-ovn\") pod \"ovn-controller-gxjzh-config-gzr9r\" (UID: \"b287920c-19e8-47a5-9644-a1651f4c39e9\") " pod="openstack/ovn-controller-gxjzh-config-gzr9r" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.573415 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b287920c-19e8-47a5-9644-a1651f4c39e9-var-log-ovn\") pod \"ovn-controller-gxjzh-config-gzr9r\" (UID: \"b287920c-19e8-47a5-9644-a1651f4c39e9\") " pod="openstack/ovn-controller-gxjzh-config-gzr9r" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.674324 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7333f8a-0a54-4dec-8e7a-c7a648d2a841" path="/var/lib/kubelet/pods/e7333f8a-0a54-4dec-8e7a-c7a648d2a841/volumes" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.674816 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b287920c-19e8-47a5-9644-a1651f4c39e9-var-run-ovn\") pod \"ovn-controller-gxjzh-config-gzr9r\" (UID: \"b287920c-19e8-47a5-9644-a1651f4c39e9\") " pod="openstack/ovn-controller-gxjzh-config-gzr9r" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.674858 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b287920c-19e8-47a5-9644-a1651f4c39e9-var-log-ovn\") pod \"ovn-controller-gxjzh-config-gzr9r\" (UID: \"b287920c-19e8-47a5-9644-a1651f4c39e9\") " pod="openstack/ovn-controller-gxjzh-config-gzr9r" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.674944 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbxmh\" (UniqueName: \"kubernetes.io/projected/b287920c-19e8-47a5-9644-a1651f4c39e9-kube-api-access-xbxmh\") pod \"ovn-controller-gxjzh-config-gzr9r\" (UID: \"b287920c-19e8-47a5-9644-a1651f4c39e9\") " pod="openstack/ovn-controller-gxjzh-config-gzr9r" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.674978 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b287920c-19e8-47a5-9644-a1651f4c39e9-scripts\") pod \"ovn-controller-gxjzh-config-gzr9r\" (UID: \"b287920c-19e8-47a5-9644-a1651f4c39e9\") " pod="openstack/ovn-controller-gxjzh-config-gzr9r" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.675009 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b287920c-19e8-47a5-9644-a1651f4c39e9-additional-scripts\") pod \"ovn-controller-gxjzh-config-gzr9r\" (UID: \"b287920c-19e8-47a5-9644-a1651f4c39e9\") " pod="openstack/ovn-controller-gxjzh-config-gzr9r" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.675046 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b287920c-19e8-47a5-9644-a1651f4c39e9-var-run\") pod \"ovn-controller-gxjzh-config-gzr9r\" (UID: \"b287920c-19e8-47a5-9644-a1651f4c39e9\") " pod="openstack/ovn-controller-gxjzh-config-gzr9r" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.675368 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b287920c-19e8-47a5-9644-a1651f4c39e9-var-run-ovn\") pod \"ovn-controller-gxjzh-config-gzr9r\" (UID: \"b287920c-19e8-47a5-9644-a1651f4c39e9\") " pod="openstack/ovn-controller-gxjzh-config-gzr9r" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.675408 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b287920c-19e8-47a5-9644-a1651f4c39e9-var-run\") pod \"ovn-controller-gxjzh-config-gzr9r\" (UID: \"b287920c-19e8-47a5-9644-a1651f4c39e9\") " pod="openstack/ovn-controller-gxjzh-config-gzr9r" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.675490 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b287920c-19e8-47a5-9644-a1651f4c39e9-var-log-ovn\") pod \"ovn-controller-gxjzh-config-gzr9r\" (UID: \"b287920c-19e8-47a5-9644-a1651f4c39e9\") " pod="openstack/ovn-controller-gxjzh-config-gzr9r" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.676133 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b287920c-19e8-47a5-9644-a1651f4c39e9-additional-scripts\") pod \"ovn-controller-gxjzh-config-gzr9r\" (UID: \"b287920c-19e8-47a5-9644-a1651f4c39e9\") " pod="openstack/ovn-controller-gxjzh-config-gzr9r" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.678295 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b287920c-19e8-47a5-9644-a1651f4c39e9-scripts\") pod \"ovn-controller-gxjzh-config-gzr9r\" (UID: \"b287920c-19e8-47a5-9644-a1651f4c39e9\") " pod="openstack/ovn-controller-gxjzh-config-gzr9r" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.704375 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbxmh\" (UniqueName: \"kubernetes.io/projected/b287920c-19e8-47a5-9644-a1651f4c39e9-kube-api-access-xbxmh\") pod \"ovn-controller-gxjzh-config-gzr9r\" (UID: \"b287920c-19e8-47a5-9644-a1651f4c39e9\") " pod="openstack/ovn-controller-gxjzh-config-gzr9r" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.786032 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e8b1-account-create-update-8rlww" event={"ID":"9cbda9cc-3ec5-4193-a7fb-ff06bdd20846","Type":"ContainerDied","Data":"b7b0b5ce95f1f8f79fe79dcceda26c46edd68768fdca8bfa61a8c002836d8e7e"} Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.786070 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7b0b5ce95f1f8f79fe79dcceda26c46edd68768fdca8bfa61a8c002836d8e7e" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.786245 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e8b1-account-create-update-8rlww" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.787463 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a648-account-create-update-qhx8z" event={"ID":"752279e5-88ff-469d-a4db-2942659c7e24","Type":"ContainerDied","Data":"71e0dffca8f1bcda6eb557e7af38812a722313603f4af46f0a96549446e9a419"} Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.787485 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71e0dffca8f1bcda6eb557e7af38812a722313603f4af46f0a96549446e9a419" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.787564 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a648-account-create-update-qhx8z" Jan 27 15:25:42 crc kubenswrapper[4772]: I0127 15:25:42.873050 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gxjzh-config-gzr9r" Jan 27 15:25:43 crc kubenswrapper[4772]: I0127 15:25:43.311230 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gxjzh-config-gzr9r"] Jan 27 15:25:43 crc kubenswrapper[4772]: W0127 15:25:43.311652 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb287920c_19e8_47a5_9644_a1651f4c39e9.slice/crio-ce6f9ec54fe9d90af32886adae36e9717f225263d8d3fd44e860708fc18d9938 WatchSource:0}: Error finding container ce6f9ec54fe9d90af32886adae36e9717f225263d8d3fd44e860708fc18d9938: Status 404 returned error can't find the container with id ce6f9ec54fe9d90af32886adae36e9717f225263d8d3fd44e860708fc18d9938 Jan 27 15:25:43 crc kubenswrapper[4772]: I0127 15:25:43.802091 4772 generic.go:334] "Generic (PLEG): container finished" podID="b287920c-19e8-47a5-9644-a1651f4c39e9" containerID="666a2855e8df449d0b2a9f22d64efe41fc16e80a56e57924cea7c6f56eb00af0" exitCode=0 Jan 27 15:25:43 crc kubenswrapper[4772]: I0127 15:25:43.802192 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gxjzh-config-gzr9r" event={"ID":"b287920c-19e8-47a5-9644-a1651f4c39e9","Type":"ContainerDied","Data":"666a2855e8df449d0b2a9f22d64efe41fc16e80a56e57924cea7c6f56eb00af0"} Jan 27 15:25:43 crc kubenswrapper[4772]: I0127 15:25:43.802226 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gxjzh-config-gzr9r" event={"ID":"b287920c-19e8-47a5-9644-a1651f4c39e9","Type":"ContainerStarted","Data":"ce6f9ec54fe9d90af32886adae36e9717f225263d8d3fd44e860708fc18d9938"} Jan 27 15:25:44 crc kubenswrapper[4772]: I0127 15:25:44.326293 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77585f5f8c-lll86" Jan 27 15:25:44 crc kubenswrapper[4772]: I0127 15:25:44.379553 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tltm6"] Jan 27 15:25:44 crc kubenswrapper[4772]: I0127 15:25:44.379853 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-tltm6" podUID="01d2ace8-4fbb-4f53-aa31-7557dbaabcce" containerName="dnsmasq-dns" containerID="cri-o://d1b5117c10f9331477f591f10a624b08ae6968087cc1bb15580ee055f80a719c" gracePeriod=10 Jan 27 15:25:44 crc kubenswrapper[4772]: I0127 15:25:44.811866 4772 generic.go:334] "Generic (PLEG): container finished" podID="01d2ace8-4fbb-4f53-aa31-7557dbaabcce" containerID="d1b5117c10f9331477f591f10a624b08ae6968087cc1bb15580ee055f80a719c" exitCode=0 Jan 27 15:25:44 crc kubenswrapper[4772]: I0127 15:25:44.812052 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tltm6" event={"ID":"01d2ace8-4fbb-4f53-aa31-7557dbaabcce","Type":"ContainerDied","Data":"d1b5117c10f9331477f591f10a624b08ae6968087cc1bb15580ee055f80a719c"} Jan 27 15:25:44 crc kubenswrapper[4772]: I0127 15:25:44.812245 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-tltm6" event={"ID":"01d2ace8-4fbb-4f53-aa31-7557dbaabcce","Type":"ContainerDied","Data":"67ecd74afacd326820a12dc1cfdc76a179790d7bc04cff09eef9f1e0a03e5d5e"} Jan 27 15:25:44 crc kubenswrapper[4772]: I0127 15:25:44.812261 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67ecd74afacd326820a12dc1cfdc76a179790d7bc04cff09eef9f1e0a03e5d5e" Jan 27 15:25:44 crc kubenswrapper[4772]: I0127 15:25:44.826825 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-tltm6" Jan 27 15:25:44 crc kubenswrapper[4772]: I0127 15:25:44.916561 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-ovsdbserver-sb\") pod \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\" (UID: \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\") " Jan 27 15:25:44 crc kubenswrapper[4772]: I0127 15:25:44.916616 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-config\") pod \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\" (UID: \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\") " Jan 27 15:25:44 crc kubenswrapper[4772]: I0127 15:25:44.916658 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-ovsdbserver-nb\") pod \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\" (UID: \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\") " Jan 27 15:25:44 crc kubenswrapper[4772]: I0127 15:25:44.916676 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2pzt\" (UniqueName: \"kubernetes.io/projected/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-kube-api-access-p2pzt\") pod \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\" (UID: \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\") " Jan 27 15:25:44 crc kubenswrapper[4772]: I0127 15:25:44.916697 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-dns-svc\") pod \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\" (UID: \"01d2ace8-4fbb-4f53-aa31-7557dbaabcce\") " Jan 27 15:25:44 crc kubenswrapper[4772]: I0127 15:25:44.936127 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-kube-api-access-p2pzt" (OuterVolumeSpecName: "kube-api-access-p2pzt") pod "01d2ace8-4fbb-4f53-aa31-7557dbaabcce" (UID: "01d2ace8-4fbb-4f53-aa31-7557dbaabcce"). InnerVolumeSpecName "kube-api-access-p2pzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:25:44 crc kubenswrapper[4772]: I0127 15:25:44.964119 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "01d2ace8-4fbb-4f53-aa31-7557dbaabcce" (UID: "01d2ace8-4fbb-4f53-aa31-7557dbaabcce"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:44 crc kubenswrapper[4772]: I0127 15:25:44.977707 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-config" (OuterVolumeSpecName: "config") pod "01d2ace8-4fbb-4f53-aa31-7557dbaabcce" (UID: "01d2ace8-4fbb-4f53-aa31-7557dbaabcce"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:44 crc kubenswrapper[4772]: I0127 15:25:44.984984 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "01d2ace8-4fbb-4f53-aa31-7557dbaabcce" (UID: "01d2ace8-4fbb-4f53-aa31-7557dbaabcce"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:44 crc kubenswrapper[4772]: I0127 15:25:44.986387 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "01d2ace8-4fbb-4f53-aa31-7557dbaabcce" (UID: "01d2ace8-4fbb-4f53-aa31-7557dbaabcce"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.018303 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.018335 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.018346 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2pzt\" (UniqueName: \"kubernetes.io/projected/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-kube-api-access-p2pzt\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.018356 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.018365 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/01d2ace8-4fbb-4f53-aa31-7557dbaabcce-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.053537 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gxjzh-config-gzr9r" Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.119987 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b287920c-19e8-47a5-9644-a1651f4c39e9-additional-scripts\") pod \"b287920c-19e8-47a5-9644-a1651f4c39e9\" (UID: \"b287920c-19e8-47a5-9644-a1651f4c39e9\") " Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.120076 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b287920c-19e8-47a5-9644-a1651f4c39e9-var-run\") pod \"b287920c-19e8-47a5-9644-a1651f4c39e9\" (UID: \"b287920c-19e8-47a5-9644-a1651f4c39e9\") " Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.120184 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b287920c-19e8-47a5-9644-a1651f4c39e9-scripts\") pod \"b287920c-19e8-47a5-9644-a1651f4c39e9\" (UID: \"b287920c-19e8-47a5-9644-a1651f4c39e9\") " Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.120245 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b287920c-19e8-47a5-9644-a1651f4c39e9-var-log-ovn\") pod \"b287920c-19e8-47a5-9644-a1651f4c39e9\" (UID: \"b287920c-19e8-47a5-9644-a1651f4c39e9\") " Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.120301 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbxmh\" (UniqueName: \"kubernetes.io/projected/b287920c-19e8-47a5-9644-a1651f4c39e9-kube-api-access-xbxmh\") pod \"b287920c-19e8-47a5-9644-a1651f4c39e9\" (UID: \"b287920c-19e8-47a5-9644-a1651f4c39e9\") " Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.120342 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b287920c-19e8-47a5-9644-a1651f4c39e9-var-run-ovn\") pod \"b287920c-19e8-47a5-9644-a1651f4c39e9\" (UID: \"b287920c-19e8-47a5-9644-a1651f4c39e9\") " Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.120678 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b287920c-19e8-47a5-9644-a1651f4c39e9-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "b287920c-19e8-47a5-9644-a1651f4c39e9" (UID: "b287920c-19e8-47a5-9644-a1651f4c39e9"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.120720 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b287920c-19e8-47a5-9644-a1651f4c39e9-var-run" (OuterVolumeSpecName: "var-run") pod "b287920c-19e8-47a5-9644-a1651f4c39e9" (UID: "b287920c-19e8-47a5-9644-a1651f4c39e9"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.120790 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b287920c-19e8-47a5-9644-a1651f4c39e9-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "b287920c-19e8-47a5-9644-a1651f4c39e9" (UID: "b287920c-19e8-47a5-9644-a1651f4c39e9"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.120853 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b287920c-19e8-47a5-9644-a1651f4c39e9-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "b287920c-19e8-47a5-9644-a1651f4c39e9" (UID: "b287920c-19e8-47a5-9644-a1651f4c39e9"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.121462 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b287920c-19e8-47a5-9644-a1651f4c39e9-scripts" (OuterVolumeSpecName: "scripts") pod "b287920c-19e8-47a5-9644-a1651f4c39e9" (UID: "b287920c-19e8-47a5-9644-a1651f4c39e9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.124770 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b287920c-19e8-47a5-9644-a1651f4c39e9-kube-api-access-xbxmh" (OuterVolumeSpecName: "kube-api-access-xbxmh") pod "b287920c-19e8-47a5-9644-a1651f4c39e9" (UID: "b287920c-19e8-47a5-9644-a1651f4c39e9"). InnerVolumeSpecName "kube-api-access-xbxmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.222030 4772 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b287920c-19e8-47a5-9644-a1651f4c39e9-var-log-ovn\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.222075 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbxmh\" (UniqueName: \"kubernetes.io/projected/b287920c-19e8-47a5-9644-a1651f4c39e9-kube-api-access-xbxmh\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.222090 4772 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b287920c-19e8-47a5-9644-a1651f4c39e9-var-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.222103 4772 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b287920c-19e8-47a5-9644-a1651f4c39e9-additional-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.222116 4772 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b287920c-19e8-47a5-9644-a1651f4c39e9-var-run\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.222127 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b287920c-19e8-47a5-9644-a1651f4c39e9-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.820910 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-tltm6" Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.821262 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gxjzh-config-gzr9r" event={"ID":"b287920c-19e8-47a5-9644-a1651f4c39e9","Type":"ContainerDied","Data":"ce6f9ec54fe9d90af32886adae36e9717f225263d8d3fd44e860708fc18d9938"} Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.821335 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce6f9ec54fe9d90af32886adae36e9717f225263d8d3fd44e860708fc18d9938" Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.821345 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gxjzh-config-gzr9r" Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.856891 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tltm6"] Jan 27 15:25:45 crc kubenswrapper[4772]: I0127 15:25:45.863317 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-tltm6"] Jan 27 15:25:46 crc kubenswrapper[4772]: I0127 15:25:46.120723 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-gxjzh-config-gzr9r"] Jan 27 15:25:46 crc kubenswrapper[4772]: I0127 15:25:46.127707 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-gxjzh-config-gzr9r"] Jan 27 15:25:46 crc kubenswrapper[4772]: I0127 15:25:46.673793 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01d2ace8-4fbb-4f53-aa31-7557dbaabcce" path="/var/lib/kubelet/pods/01d2ace8-4fbb-4f53-aa31-7557dbaabcce/volumes" Jan 27 15:25:46 crc kubenswrapper[4772]: I0127 15:25:46.674774 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b287920c-19e8-47a5-9644-a1651f4c39e9" path="/var/lib/kubelet/pods/b287920c-19e8-47a5-9644-a1651f4c39e9/volumes" Jan 27 15:25:47 crc kubenswrapper[4772]: I0127 15:25:47.335679 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-vdmv7"] Jan 27 15:25:47 crc kubenswrapper[4772]: E0127 15:25:47.335996 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b287920c-19e8-47a5-9644-a1651f4c39e9" containerName="ovn-config" Jan 27 15:25:47 crc kubenswrapper[4772]: I0127 15:25:47.336009 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="b287920c-19e8-47a5-9644-a1651f4c39e9" containerName="ovn-config" Jan 27 15:25:47 crc kubenswrapper[4772]: E0127 15:25:47.336018 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01d2ace8-4fbb-4f53-aa31-7557dbaabcce" containerName="init" Jan 27 15:25:47 crc kubenswrapper[4772]: I0127 15:25:47.336023 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="01d2ace8-4fbb-4f53-aa31-7557dbaabcce" containerName="init" Jan 27 15:25:47 crc kubenswrapper[4772]: E0127 15:25:47.336046 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01d2ace8-4fbb-4f53-aa31-7557dbaabcce" containerName="dnsmasq-dns" Jan 27 15:25:47 crc kubenswrapper[4772]: I0127 15:25:47.336051 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="01d2ace8-4fbb-4f53-aa31-7557dbaabcce" containerName="dnsmasq-dns" Jan 27 15:25:47 crc kubenswrapper[4772]: I0127 15:25:47.336210 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="b287920c-19e8-47a5-9644-a1651f4c39e9" containerName="ovn-config" Jan 27 15:25:47 crc kubenswrapper[4772]: I0127 15:25:47.336223 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="01d2ace8-4fbb-4f53-aa31-7557dbaabcce" containerName="dnsmasq-dns" Jan 27 15:25:47 crc kubenswrapper[4772]: I0127 15:25:47.336786 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-vdmv7" Jan 27 15:25:47 crc kubenswrapper[4772]: I0127 15:25:47.338907 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Jan 27 15:25:47 crc kubenswrapper[4772]: I0127 15:25:47.338943 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-vd4fn" Jan 27 15:25:47 crc kubenswrapper[4772]: I0127 15:25:47.350278 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-vdmv7"] Jan 27 15:25:47 crc kubenswrapper[4772]: I0127 15:25:47.473967 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/86d0241f-ae16-400f-837c-3b43c904c91e-db-sync-config-data\") pod \"glance-db-sync-vdmv7\" (UID: \"86d0241f-ae16-400f-837c-3b43c904c91e\") " pod="openstack/glance-db-sync-vdmv7" Jan 27 15:25:47 crc kubenswrapper[4772]: I0127 15:25:47.474017 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d0241f-ae16-400f-837c-3b43c904c91e-combined-ca-bundle\") pod \"glance-db-sync-vdmv7\" (UID: \"86d0241f-ae16-400f-837c-3b43c904c91e\") " pod="openstack/glance-db-sync-vdmv7" Jan 27 15:25:47 crc kubenswrapper[4772]: I0127 15:25:47.474058 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d0241f-ae16-400f-837c-3b43c904c91e-config-data\") pod \"glance-db-sync-vdmv7\" (UID: \"86d0241f-ae16-400f-837c-3b43c904c91e\") " pod="openstack/glance-db-sync-vdmv7" Jan 27 15:25:47 crc kubenswrapper[4772]: I0127 15:25:47.474273 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkbkr\" (UniqueName: \"kubernetes.io/projected/86d0241f-ae16-400f-837c-3b43c904c91e-kube-api-access-kkbkr\") pod \"glance-db-sync-vdmv7\" (UID: \"86d0241f-ae16-400f-837c-3b43c904c91e\") " pod="openstack/glance-db-sync-vdmv7" Jan 27 15:25:47 crc kubenswrapper[4772]: I0127 15:25:47.575403 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkbkr\" (UniqueName: \"kubernetes.io/projected/86d0241f-ae16-400f-837c-3b43c904c91e-kube-api-access-kkbkr\") pod \"glance-db-sync-vdmv7\" (UID: \"86d0241f-ae16-400f-837c-3b43c904c91e\") " pod="openstack/glance-db-sync-vdmv7" Jan 27 15:25:47 crc kubenswrapper[4772]: I0127 15:25:47.575488 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/86d0241f-ae16-400f-837c-3b43c904c91e-db-sync-config-data\") pod \"glance-db-sync-vdmv7\" (UID: \"86d0241f-ae16-400f-837c-3b43c904c91e\") " pod="openstack/glance-db-sync-vdmv7" Jan 27 15:25:47 crc kubenswrapper[4772]: I0127 15:25:47.575516 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d0241f-ae16-400f-837c-3b43c904c91e-combined-ca-bundle\") pod \"glance-db-sync-vdmv7\" (UID: \"86d0241f-ae16-400f-837c-3b43c904c91e\") " pod="openstack/glance-db-sync-vdmv7" Jan 27 15:25:47 crc kubenswrapper[4772]: I0127 15:25:47.575563 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d0241f-ae16-400f-837c-3b43c904c91e-config-data\") pod \"glance-db-sync-vdmv7\" (UID: \"86d0241f-ae16-400f-837c-3b43c904c91e\") " pod="openstack/glance-db-sync-vdmv7" Jan 27 15:25:47 crc kubenswrapper[4772]: I0127 15:25:47.580934 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d0241f-ae16-400f-837c-3b43c904c91e-combined-ca-bundle\") pod \"glance-db-sync-vdmv7\" (UID: \"86d0241f-ae16-400f-837c-3b43c904c91e\") " pod="openstack/glance-db-sync-vdmv7" Jan 27 15:25:47 crc kubenswrapper[4772]: I0127 15:25:47.581092 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d0241f-ae16-400f-837c-3b43c904c91e-config-data\") pod \"glance-db-sync-vdmv7\" (UID: \"86d0241f-ae16-400f-837c-3b43c904c91e\") " pod="openstack/glance-db-sync-vdmv7" Jan 27 15:25:47 crc kubenswrapper[4772]: I0127 15:25:47.584347 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/86d0241f-ae16-400f-837c-3b43c904c91e-db-sync-config-data\") pod \"glance-db-sync-vdmv7\" (UID: \"86d0241f-ae16-400f-837c-3b43c904c91e\") " pod="openstack/glance-db-sync-vdmv7" Jan 27 15:25:47 crc kubenswrapper[4772]: I0127 15:25:47.590310 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkbkr\" (UniqueName: \"kubernetes.io/projected/86d0241f-ae16-400f-837c-3b43c904c91e-kube-api-access-kkbkr\") pod \"glance-db-sync-vdmv7\" (UID: \"86d0241f-ae16-400f-837c-3b43c904c91e\") " pod="openstack/glance-db-sync-vdmv7" Jan 27 15:25:47 crc kubenswrapper[4772]: I0127 15:25:47.659549 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-vdmv7" Jan 27 15:25:49 crc kubenswrapper[4772]: I0127 15:25:48.211994 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-vdmv7"] Jan 27 15:25:49 crc kubenswrapper[4772]: W0127 15:25:48.218734 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86d0241f_ae16_400f_837c_3b43c904c91e.slice/crio-a28e441b5d88b4432b8107753f5714c4987db6ac6635bc08e9f396a0b42288ec WatchSource:0}: Error finding container a28e441b5d88b4432b8107753f5714c4987db6ac6635bc08e9f396a0b42288ec: Status 404 returned error can't find the container with id a28e441b5d88b4432b8107753f5714c4987db6ac6635bc08e9f396a0b42288ec Jan 27 15:25:49 crc kubenswrapper[4772]: I0127 15:25:48.852692 4772 generic.go:334] "Generic (PLEG): container finished" podID="508c3d5b-212a-46da-9a55-de3f35d7019b" containerID="900401625caff4c2d87fe06884c7dcba7f46fdc58e9213b1a6cc2cf36d383e52" exitCode=0 Jan 27 15:25:49 crc kubenswrapper[4772]: I0127 15:25:48.852792 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"508c3d5b-212a-46da-9a55-de3f35d7019b","Type":"ContainerDied","Data":"900401625caff4c2d87fe06884c7dcba7f46fdc58e9213b1a6cc2cf36d383e52"} Jan 27 15:25:49 crc kubenswrapper[4772]: I0127 15:25:48.854916 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-vdmv7" event={"ID":"86d0241f-ae16-400f-837c-3b43c904c91e","Type":"ContainerStarted","Data":"a28e441b5d88b4432b8107753f5714c4987db6ac6635bc08e9f396a0b42288ec"} Jan 27 15:25:49 crc kubenswrapper[4772]: I0127 15:25:48.857080 4772 generic.go:334] "Generic (PLEG): container finished" podID="76fdbdb1-d48a-4cd1-8372-78887671dce8" containerID="d53d0dfba4b0af64ac6186cc8eb8efb21a1ec89a66c075c5a53ab1db9987e594" exitCode=0 Jan 27 15:25:49 crc kubenswrapper[4772]: I0127 15:25:48.857108 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"76fdbdb1-d48a-4cd1-8372-78887671dce8","Type":"ContainerDied","Data":"d53d0dfba4b0af64ac6186cc8eb8efb21a1ec89a66c075c5a53ab1db9987e594"} Jan 27 15:25:49 crc kubenswrapper[4772]: I0127 15:25:49.885919 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"76fdbdb1-d48a-4cd1-8372-78887671dce8","Type":"ContainerStarted","Data":"d8699d4e2fb6bcbb97c43048a20aeda8d17be226258e1acddb0364ab41c23e4a"} Jan 27 15:25:49 crc kubenswrapper[4772]: I0127 15:25:49.886375 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:25:49 crc kubenswrapper[4772]: I0127 15:25:49.888897 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"508c3d5b-212a-46da-9a55-de3f35d7019b","Type":"ContainerStarted","Data":"f002759dea4443f7600e0f76f24481c1604449a5ee31bd8aa53171a2121ec4b2"} Jan 27 15:25:49 crc kubenswrapper[4772]: I0127 15:25:49.889152 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Jan 27 15:25:49 crc kubenswrapper[4772]: I0127 15:25:49.914961 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.426593695 podStartE2EDuration="1m17.914942503s" podCreationTimestamp="2026-01-27 15:24:32 +0000 UTC" firstStartedPulling="2026-01-27 15:24:34.314057243 +0000 UTC m=+1060.294666331" lastFinishedPulling="2026-01-27 15:25:14.802406041 +0000 UTC m=+1100.783015139" observedRunningTime="2026-01-27 15:25:49.912626906 +0000 UTC m=+1135.893236034" watchObservedRunningTime="2026-01-27 15:25:49.914942503 +0000 UTC m=+1135.895551601" Jan 27 15:25:49 crc kubenswrapper[4772]: I0127 15:25:49.951986 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=-9223371957.902811 podStartE2EDuration="1m18.951965269s" podCreationTimestamp="2026-01-27 15:24:31 +0000 UTC" firstStartedPulling="2026-01-27 15:24:33.501124377 +0000 UTC m=+1059.481733475" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:25:49.940310154 +0000 UTC m=+1135.920919262" watchObservedRunningTime="2026-01-27 15:25:49.951965269 +0000 UTC m=+1135.932574367" Jan 27 15:26:00 crc kubenswrapper[4772]: I0127 15:26:00.974480 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-vdmv7" event={"ID":"86d0241f-ae16-400f-837c-3b43c904c91e","Type":"ContainerStarted","Data":"317ff691da5e191e31778e1d02f29484703e057687e372739fcbc9dd6f8088d2"} Jan 27 15:26:00 crc kubenswrapper[4772]: I0127 15:26:00.999894 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-vdmv7" podStartSLOduration=2.356557616 podStartE2EDuration="13.999877194s" podCreationTimestamp="2026-01-27 15:25:47 +0000 UTC" firstStartedPulling="2026-01-27 15:25:48.220746314 +0000 UTC m=+1134.201355422" lastFinishedPulling="2026-01-27 15:25:59.864065902 +0000 UTC m=+1145.844675000" observedRunningTime="2026-01-27 15:26:00.99523813 +0000 UTC m=+1146.975847228" watchObservedRunningTime="2026-01-27 15:26:00.999877194 +0000 UTC m=+1146.980486292" Jan 27 15:26:02 crc kubenswrapper[4772]: I0127 15:26:02.917152 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="508c3d5b-212a-46da-9a55-de3f35d7019b" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Jan 27 15:26:03 crc kubenswrapper[4772]: I0127 15:26:03.769453 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:26:12 crc kubenswrapper[4772]: I0127 15:26:12.059390 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:26:12 crc kubenswrapper[4772]: I0127 15:26:12.060293 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:26:12 crc kubenswrapper[4772]: I0127 15:26:12.060392 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:26:12 crc kubenswrapper[4772]: I0127 15:26:12.061883 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ed9bc8d4920540552bc96f7af996996e69c893224418d74c897e7298ed107163"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 15:26:12 crc kubenswrapper[4772]: I0127 15:26:12.062031 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://ed9bc8d4920540552bc96f7af996996e69c893224418d74c897e7298ed107163" gracePeriod=600 Jan 27 15:26:12 crc kubenswrapper[4772]: I0127 15:26:12.917450 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.092435 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="ed9bc8d4920540552bc96f7af996996e69c893224418d74c897e7298ed107163" exitCode=0 Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.092489 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"ed9bc8d4920540552bc96f7af996996e69c893224418d74c897e7298ed107163"} Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.092526 4772 scope.go:117] "RemoveContainer" containerID="c8213e4fa74445d3800c2dbcb45efc3fb34a6f40c3d5ed5845b811a51d3d8497" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.368680 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-v7ncm"] Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.369983 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-v7ncm" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.384269 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-v7ncm"] Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.396653 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-xpbb6"] Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.397837 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-xpbb6" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.426656 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-xpbb6"] Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.488963 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-97c3-account-create-update-xlghl"] Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.489862 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-97c3-account-create-update-xlghl" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.491549 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.507048 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-97c3-account-create-update-xlghl"] Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.526334 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22x22\" (UniqueName: \"kubernetes.io/projected/ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112-kube-api-access-22x22\") pod \"barbican-db-create-xpbb6\" (UID: \"ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112\") " pod="openstack/barbican-db-create-xpbb6" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.526574 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1746148-2e3f-476f-9a1f-f3656d44fb0b-operator-scripts\") pod \"cinder-db-create-v7ncm\" (UID: \"f1746148-2e3f-476f-9a1f-f3656d44fb0b\") " pod="openstack/cinder-db-create-v7ncm" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.526666 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrbpt\" (UniqueName: \"kubernetes.io/projected/f1746148-2e3f-476f-9a1f-f3656d44fb0b-kube-api-access-mrbpt\") pod \"cinder-db-create-v7ncm\" (UID: \"f1746148-2e3f-476f-9a1f-f3656d44fb0b\") " pod="openstack/cinder-db-create-v7ncm" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.526741 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112-operator-scripts\") pod \"barbican-db-create-xpbb6\" (UID: \"ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112\") " pod="openstack/barbican-db-create-xpbb6" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.587942 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-z224f"] Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.589131 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-z224f" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.595643 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-d2a3-account-create-update-hfkkb"] Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.596525 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d2a3-account-create-update-hfkkb" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.611357 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.626141 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-z224f"] Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.628388 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22x22\" (UniqueName: \"kubernetes.io/projected/ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112-kube-api-access-22x22\") pod \"barbican-db-create-xpbb6\" (UID: \"ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112\") " pod="openstack/barbican-db-create-xpbb6" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.629317 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5rsl\" (UniqueName: \"kubernetes.io/projected/af61fb8e-e749-4872-8dc6-c590e4b9787a-kube-api-access-f5rsl\") pod \"barbican-97c3-account-create-update-xlghl\" (UID: \"af61fb8e-e749-4872-8dc6-c590e4b9787a\") " pod="openstack/barbican-97c3-account-create-update-xlghl" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.629582 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af61fb8e-e749-4872-8dc6-c590e4b9787a-operator-scripts\") pod \"barbican-97c3-account-create-update-xlghl\" (UID: \"af61fb8e-e749-4872-8dc6-c590e4b9787a\") " pod="openstack/barbican-97c3-account-create-update-xlghl" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.629739 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1746148-2e3f-476f-9a1f-f3656d44fb0b-operator-scripts\") pod \"cinder-db-create-v7ncm\" (UID: \"f1746148-2e3f-476f-9a1f-f3656d44fb0b\") " pod="openstack/cinder-db-create-v7ncm" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.630569 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1746148-2e3f-476f-9a1f-f3656d44fb0b-operator-scripts\") pod \"cinder-db-create-v7ncm\" (UID: \"f1746148-2e3f-476f-9a1f-f3656d44fb0b\") " pod="openstack/cinder-db-create-v7ncm" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.636751 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrbpt\" (UniqueName: \"kubernetes.io/projected/f1746148-2e3f-476f-9a1f-f3656d44fb0b-kube-api-access-mrbpt\") pod \"cinder-db-create-v7ncm\" (UID: \"f1746148-2e3f-476f-9a1f-f3656d44fb0b\") " pod="openstack/cinder-db-create-v7ncm" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.638052 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112-operator-scripts\") pod \"barbican-db-create-xpbb6\" (UID: \"ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112\") " pod="openstack/barbican-db-create-xpbb6" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.639764 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112-operator-scripts\") pod \"barbican-db-create-xpbb6\" (UID: \"ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112\") " pod="openstack/barbican-db-create-xpbb6" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.654894 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-d2a3-account-create-update-hfkkb"] Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.670665 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrbpt\" (UniqueName: \"kubernetes.io/projected/f1746148-2e3f-476f-9a1f-f3656d44fb0b-kube-api-access-mrbpt\") pod \"cinder-db-create-v7ncm\" (UID: \"f1746148-2e3f-476f-9a1f-f3656d44fb0b\") " pod="openstack/cinder-db-create-v7ncm" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.671317 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22x22\" (UniqueName: \"kubernetes.io/projected/ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112-kube-api-access-22x22\") pod \"barbican-db-create-xpbb6\" (UID: \"ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112\") " pod="openstack/barbican-db-create-xpbb6" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.684330 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-kvb25"] Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.685300 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-kvb25" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.686588 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-v7ncm" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.690442 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.690496 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-bdjsw" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.690504 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.693392 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.700914 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-kvb25"] Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.713570 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-xpbb6" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.742306 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5rsl\" (UniqueName: \"kubernetes.io/projected/af61fb8e-e749-4872-8dc6-c590e4b9787a-kube-api-access-f5rsl\") pod \"barbican-97c3-account-create-update-xlghl\" (UID: \"af61fb8e-e749-4872-8dc6-c590e4b9787a\") " pod="openstack/barbican-97c3-account-create-update-xlghl" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.742346 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af61fb8e-e749-4872-8dc6-c590e4b9787a-operator-scripts\") pod \"barbican-97c3-account-create-update-xlghl\" (UID: \"af61fb8e-e749-4872-8dc6-c590e4b9787a\") " pod="openstack/barbican-97c3-account-create-update-xlghl" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.742393 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgk7m\" (UniqueName: \"kubernetes.io/projected/57a1d71f-3b00-42c0-92c4-a29fb3d4518c-kube-api-access-mgk7m\") pod \"keystone-db-sync-kvb25\" (UID: \"57a1d71f-3b00-42c0-92c4-a29fb3d4518c\") " pod="openstack/keystone-db-sync-kvb25" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.742445 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ffgc\" (UniqueName: \"kubernetes.io/projected/b412abae-93af-4ae0-8cd8-7c0a827da4b3-kube-api-access-9ffgc\") pod \"neutron-db-create-z224f\" (UID: \"b412abae-93af-4ae0-8cd8-7c0a827da4b3\") " pod="openstack/neutron-db-create-z224f" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.742481 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57a1d71f-3b00-42c0-92c4-a29fb3d4518c-combined-ca-bundle\") pod \"keystone-db-sync-kvb25\" (UID: \"57a1d71f-3b00-42c0-92c4-a29fb3d4518c\") " pod="openstack/keystone-db-sync-kvb25" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.742535 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b412abae-93af-4ae0-8cd8-7c0a827da4b3-operator-scripts\") pod \"neutron-db-create-z224f\" (UID: \"b412abae-93af-4ae0-8cd8-7c0a827da4b3\") " pod="openstack/neutron-db-create-z224f" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.742586 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c54b2036-d943-4f0d-b1c4-8a47dfab5099-operator-scripts\") pod \"cinder-d2a3-account-create-update-hfkkb\" (UID: \"c54b2036-d943-4f0d-b1c4-8a47dfab5099\") " pod="openstack/cinder-d2a3-account-create-update-hfkkb" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.742624 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfhpb\" (UniqueName: \"kubernetes.io/projected/c54b2036-d943-4f0d-b1c4-8a47dfab5099-kube-api-access-rfhpb\") pod \"cinder-d2a3-account-create-update-hfkkb\" (UID: \"c54b2036-d943-4f0d-b1c4-8a47dfab5099\") " pod="openstack/cinder-d2a3-account-create-update-hfkkb" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.742644 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57a1d71f-3b00-42c0-92c4-a29fb3d4518c-config-data\") pod \"keystone-db-sync-kvb25\" (UID: \"57a1d71f-3b00-42c0-92c4-a29fb3d4518c\") " pod="openstack/keystone-db-sync-kvb25" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.744232 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af61fb8e-e749-4872-8dc6-c590e4b9787a-operator-scripts\") pod \"barbican-97c3-account-create-update-xlghl\" (UID: \"af61fb8e-e749-4872-8dc6-c590e4b9787a\") " pod="openstack/barbican-97c3-account-create-update-xlghl" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.766790 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5rsl\" (UniqueName: \"kubernetes.io/projected/af61fb8e-e749-4872-8dc6-c590e4b9787a-kube-api-access-f5rsl\") pod \"barbican-97c3-account-create-update-xlghl\" (UID: \"af61fb8e-e749-4872-8dc6-c590e4b9787a\") " pod="openstack/barbican-97c3-account-create-update-xlghl" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.798023 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-ed9a-account-create-update-b7pnl"] Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.799296 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ed9a-account-create-update-b7pnl" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.804445 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-97c3-account-create-update-xlghl" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.804987 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.823206 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ed9a-account-create-update-b7pnl"] Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.845595 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ffgc\" (UniqueName: \"kubernetes.io/projected/b412abae-93af-4ae0-8cd8-7c0a827da4b3-kube-api-access-9ffgc\") pod \"neutron-db-create-z224f\" (UID: \"b412abae-93af-4ae0-8cd8-7c0a827da4b3\") " pod="openstack/neutron-db-create-z224f" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.845663 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57a1d71f-3b00-42c0-92c4-a29fb3d4518c-combined-ca-bundle\") pod \"keystone-db-sync-kvb25\" (UID: \"57a1d71f-3b00-42c0-92c4-a29fb3d4518c\") " pod="openstack/keystone-db-sync-kvb25" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.845742 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b412abae-93af-4ae0-8cd8-7c0a827da4b3-operator-scripts\") pod \"neutron-db-create-z224f\" (UID: \"b412abae-93af-4ae0-8cd8-7c0a827da4b3\") " pod="openstack/neutron-db-create-z224f" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.845808 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c54b2036-d943-4f0d-b1c4-8a47dfab5099-operator-scripts\") pod \"cinder-d2a3-account-create-update-hfkkb\" (UID: \"c54b2036-d943-4f0d-b1c4-8a47dfab5099\") " pod="openstack/cinder-d2a3-account-create-update-hfkkb" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.845855 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b163780a-6dd7-4232-b0da-a22f18d36fcc-operator-scripts\") pod \"neutron-ed9a-account-create-update-b7pnl\" (UID: \"b163780a-6dd7-4232-b0da-a22f18d36fcc\") " pod="openstack/neutron-ed9a-account-create-update-b7pnl" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.845883 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfhpb\" (UniqueName: \"kubernetes.io/projected/c54b2036-d943-4f0d-b1c4-8a47dfab5099-kube-api-access-rfhpb\") pod \"cinder-d2a3-account-create-update-hfkkb\" (UID: \"c54b2036-d943-4f0d-b1c4-8a47dfab5099\") " pod="openstack/cinder-d2a3-account-create-update-hfkkb" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.845907 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t54t7\" (UniqueName: \"kubernetes.io/projected/b163780a-6dd7-4232-b0da-a22f18d36fcc-kube-api-access-t54t7\") pod \"neutron-ed9a-account-create-update-b7pnl\" (UID: \"b163780a-6dd7-4232-b0da-a22f18d36fcc\") " pod="openstack/neutron-ed9a-account-create-update-b7pnl" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.845943 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57a1d71f-3b00-42c0-92c4-a29fb3d4518c-config-data\") pod \"keystone-db-sync-kvb25\" (UID: \"57a1d71f-3b00-42c0-92c4-a29fb3d4518c\") " pod="openstack/keystone-db-sync-kvb25" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.846040 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgk7m\" (UniqueName: \"kubernetes.io/projected/57a1d71f-3b00-42c0-92c4-a29fb3d4518c-kube-api-access-mgk7m\") pod \"keystone-db-sync-kvb25\" (UID: \"57a1d71f-3b00-42c0-92c4-a29fb3d4518c\") " pod="openstack/keystone-db-sync-kvb25" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.850954 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c54b2036-d943-4f0d-b1c4-8a47dfab5099-operator-scripts\") pod \"cinder-d2a3-account-create-update-hfkkb\" (UID: \"c54b2036-d943-4f0d-b1c4-8a47dfab5099\") " pod="openstack/cinder-d2a3-account-create-update-hfkkb" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.852021 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57a1d71f-3b00-42c0-92c4-a29fb3d4518c-config-data\") pod \"keystone-db-sync-kvb25\" (UID: \"57a1d71f-3b00-42c0-92c4-a29fb3d4518c\") " pod="openstack/keystone-db-sync-kvb25" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.856373 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57a1d71f-3b00-42c0-92c4-a29fb3d4518c-combined-ca-bundle\") pod \"keystone-db-sync-kvb25\" (UID: \"57a1d71f-3b00-42c0-92c4-a29fb3d4518c\") " pod="openstack/keystone-db-sync-kvb25" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.856762 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b412abae-93af-4ae0-8cd8-7c0a827da4b3-operator-scripts\") pod \"neutron-db-create-z224f\" (UID: \"b412abae-93af-4ae0-8cd8-7c0a827da4b3\") " pod="openstack/neutron-db-create-z224f" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.870716 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfhpb\" (UniqueName: \"kubernetes.io/projected/c54b2036-d943-4f0d-b1c4-8a47dfab5099-kube-api-access-rfhpb\") pod \"cinder-d2a3-account-create-update-hfkkb\" (UID: \"c54b2036-d943-4f0d-b1c4-8a47dfab5099\") " pod="openstack/cinder-d2a3-account-create-update-hfkkb" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.871339 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ffgc\" (UniqueName: \"kubernetes.io/projected/b412abae-93af-4ae0-8cd8-7c0a827da4b3-kube-api-access-9ffgc\") pod \"neutron-db-create-z224f\" (UID: \"b412abae-93af-4ae0-8cd8-7c0a827da4b3\") " pod="openstack/neutron-db-create-z224f" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.895910 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgk7m\" (UniqueName: \"kubernetes.io/projected/57a1d71f-3b00-42c0-92c4-a29fb3d4518c-kube-api-access-mgk7m\") pod \"keystone-db-sync-kvb25\" (UID: \"57a1d71f-3b00-42c0-92c4-a29fb3d4518c\") " pod="openstack/keystone-db-sync-kvb25" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.921535 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d2a3-account-create-update-hfkkb" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.921576 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-z224f" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.947253 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b163780a-6dd7-4232-b0da-a22f18d36fcc-operator-scripts\") pod \"neutron-ed9a-account-create-update-b7pnl\" (UID: \"b163780a-6dd7-4232-b0da-a22f18d36fcc\") " pod="openstack/neutron-ed9a-account-create-update-b7pnl" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.947301 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t54t7\" (UniqueName: \"kubernetes.io/projected/b163780a-6dd7-4232-b0da-a22f18d36fcc-kube-api-access-t54t7\") pod \"neutron-ed9a-account-create-update-b7pnl\" (UID: \"b163780a-6dd7-4232-b0da-a22f18d36fcc\") " pod="openstack/neutron-ed9a-account-create-update-b7pnl" Jan 27 15:26:13 crc kubenswrapper[4772]: I0127 15:26:13.948409 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b163780a-6dd7-4232-b0da-a22f18d36fcc-operator-scripts\") pod \"neutron-ed9a-account-create-update-b7pnl\" (UID: \"b163780a-6dd7-4232-b0da-a22f18d36fcc\") " pod="openstack/neutron-ed9a-account-create-update-b7pnl" Jan 27 15:26:14 crc kubenswrapper[4772]: I0127 15:26:14.043787 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t54t7\" (UniqueName: \"kubernetes.io/projected/b163780a-6dd7-4232-b0da-a22f18d36fcc-kube-api-access-t54t7\") pod \"neutron-ed9a-account-create-update-b7pnl\" (UID: \"b163780a-6dd7-4232-b0da-a22f18d36fcc\") " pod="openstack/neutron-ed9a-account-create-update-b7pnl" Jan 27 15:26:14 crc kubenswrapper[4772]: I0127 15:26:14.114760 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"1d1c45659af37dbb5fcad6152d119ca4f804c58006a54555795ff000f3b7aea9"} Jan 27 15:26:14 crc kubenswrapper[4772]: I0127 15:26:14.130972 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-kvb25" Jan 27 15:26:14 crc kubenswrapper[4772]: I0127 15:26:14.157550 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ed9a-account-create-update-b7pnl" Jan 27 15:26:14 crc kubenswrapper[4772]: I0127 15:26:14.202690 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-v7ncm"] Jan 27 15:26:14 crc kubenswrapper[4772]: W0127 15:26:14.227122 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1746148_2e3f_476f_9a1f_f3656d44fb0b.slice/crio-d7520f953d62324f0f59051719778147c98517e633ea0fe6fa2fd5195a4cadf3 WatchSource:0}: Error finding container d7520f953d62324f0f59051719778147c98517e633ea0fe6fa2fd5195a4cadf3: Status 404 returned error can't find the container with id d7520f953d62324f0f59051719778147c98517e633ea0fe6fa2fd5195a4cadf3 Jan 27 15:26:14 crc kubenswrapper[4772]: I0127 15:26:14.386419 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-xpbb6"] Jan 27 15:26:14 crc kubenswrapper[4772]: W0127 15:26:14.392664 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf61fb8e_e749_4872_8dc6_c590e4b9787a.slice/crio-4fc5b98afeb27a7f560b6faf0b64da965c312fa1d6668ea8e195f9bf3cf774cd WatchSource:0}: Error finding container 4fc5b98afeb27a7f560b6faf0b64da965c312fa1d6668ea8e195f9bf3cf774cd: Status 404 returned error can't find the container with id 4fc5b98afeb27a7f560b6faf0b64da965c312fa1d6668ea8e195f9bf3cf774cd Jan 27 15:26:14 crc kubenswrapper[4772]: I0127 15:26:14.394063 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-97c3-account-create-update-xlghl"] Jan 27 15:26:14 crc kubenswrapper[4772]: W0127 15:26:14.398873 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee9c9aa3_63e7_49ae_b3f3_f9bc0802f112.slice/crio-7078456546f57835b75bdc01276c8dc7b83b93913928c27cd8b38ff3cf49ddaf WatchSource:0}: Error finding container 7078456546f57835b75bdc01276c8dc7b83b93913928c27cd8b38ff3cf49ddaf: Status 404 returned error can't find the container with id 7078456546f57835b75bdc01276c8dc7b83b93913928c27cd8b38ff3cf49ddaf Jan 27 15:26:14 crc kubenswrapper[4772]: I0127 15:26:14.629682 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-d2a3-account-create-update-hfkkb"] Jan 27 15:26:14 crc kubenswrapper[4772]: I0127 15:26:14.638386 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-z224f"] Jan 27 15:26:14 crc kubenswrapper[4772]: I0127 15:26:14.715112 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-kvb25"] Jan 27 15:26:14 crc kubenswrapper[4772]: I0127 15:26:14.742933 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ed9a-account-create-update-b7pnl"] Jan 27 15:26:14 crc kubenswrapper[4772]: W0127 15:26:14.771487 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb163780a_6dd7_4232_b0da_a22f18d36fcc.slice/crio-e83383c104bd8b761200170a9e0bc2e126e98220c9813a9e33d886b26ab147ee WatchSource:0}: Error finding container e83383c104bd8b761200170a9e0bc2e126e98220c9813a9e33d886b26ab147ee: Status 404 returned error can't find the container with id e83383c104bd8b761200170a9e0bc2e126e98220c9813a9e33d886b26ab147ee Jan 27 15:26:15 crc kubenswrapper[4772]: I0127 15:26:15.122577 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-kvb25" event={"ID":"57a1d71f-3b00-42c0-92c4-a29fb3d4518c","Type":"ContainerStarted","Data":"cc68b23df999b3c351ddd5132639a8352068b638f27d53a94da0ce06a1009dad"} Jan 27 15:26:15 crc kubenswrapper[4772]: I0127 15:26:15.123658 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ed9a-account-create-update-b7pnl" event={"ID":"b163780a-6dd7-4232-b0da-a22f18d36fcc","Type":"ContainerStarted","Data":"e83383c104bd8b761200170a9e0bc2e126e98220c9813a9e33d886b26ab147ee"} Jan 27 15:26:15 crc kubenswrapper[4772]: I0127 15:26:15.124736 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-97c3-account-create-update-xlghl" event={"ID":"af61fb8e-e749-4872-8dc6-c590e4b9787a","Type":"ContainerStarted","Data":"4fc5b98afeb27a7f560b6faf0b64da965c312fa1d6668ea8e195f9bf3cf774cd"} Jan 27 15:26:15 crc kubenswrapper[4772]: I0127 15:26:15.125703 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-v7ncm" event={"ID":"f1746148-2e3f-476f-9a1f-f3656d44fb0b","Type":"ContainerStarted","Data":"d7520f953d62324f0f59051719778147c98517e633ea0fe6fa2fd5195a4cadf3"} Jan 27 15:26:15 crc kubenswrapper[4772]: I0127 15:26:15.131208 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d2a3-account-create-update-hfkkb" event={"ID":"c54b2036-d943-4f0d-b1c4-8a47dfab5099","Type":"ContainerStarted","Data":"08d6532f77d29e47d3d345a3cbcc4de484a6e75463c90eca472f8ccd07be6a84"} Jan 27 15:26:15 crc kubenswrapper[4772]: I0127 15:26:15.132502 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-z224f" event={"ID":"b412abae-93af-4ae0-8cd8-7c0a827da4b3","Type":"ContainerStarted","Data":"96d15bd1f51df9f2151a057a908e64905ec96ea27be5ce6e7029e233334cd82f"} Jan 27 15:26:15 crc kubenswrapper[4772]: I0127 15:26:15.133515 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-xpbb6" event={"ID":"ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112","Type":"ContainerStarted","Data":"7078456546f57835b75bdc01276c8dc7b83b93913928c27cd8b38ff3cf49ddaf"} Jan 27 15:26:16 crc kubenswrapper[4772]: I0127 15:26:16.145102 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d2a3-account-create-update-hfkkb" event={"ID":"c54b2036-d943-4f0d-b1c4-8a47dfab5099","Type":"ContainerStarted","Data":"c454404cb2dabeb6539bab075b0096e5a7ba9d3726f1b7a2ce5d55b30cc778e8"} Jan 27 15:26:16 crc kubenswrapper[4772]: I0127 15:26:16.147750 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-z224f" event={"ID":"b412abae-93af-4ae0-8cd8-7c0a827da4b3","Type":"ContainerStarted","Data":"478f9eb73f50cba542d4259825587e98caddfe9513876ed4823af8b00681f571"} Jan 27 15:26:16 crc kubenswrapper[4772]: I0127 15:26:16.150194 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-xpbb6" event={"ID":"ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112","Type":"ContainerStarted","Data":"7cb0416a54334bdd5699afd4b64397c193035c399e5586172a360ff52cd674f9"} Jan 27 15:26:16 crc kubenswrapper[4772]: I0127 15:26:16.152827 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ed9a-account-create-update-b7pnl" event={"ID":"b163780a-6dd7-4232-b0da-a22f18d36fcc","Type":"ContainerStarted","Data":"a5ffbaeea04257a22f38554ccc4304785fadfe22ac90bb6e3544b162aab10857"} Jan 27 15:26:16 crc kubenswrapper[4772]: I0127 15:26:16.155564 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-97c3-account-create-update-xlghl" event={"ID":"af61fb8e-e749-4872-8dc6-c590e4b9787a","Type":"ContainerStarted","Data":"d265eb93689c326c68ce844d36ec8e13845ff3f6cfb1ed7e88273d0cf4e91cbd"} Jan 27 15:26:16 crc kubenswrapper[4772]: I0127 15:26:16.156706 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-v7ncm" event={"ID":"f1746148-2e3f-476f-9a1f-f3656d44fb0b","Type":"ContainerStarted","Data":"597636ff183f237bb3b639ea5c67c6b5f6f29f40e362b71df3d4ec02eaa6036b"} Jan 27 15:26:16 crc kubenswrapper[4772]: I0127 15:26:16.167952 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-d2a3-account-create-update-hfkkb" podStartSLOduration=3.167932458 podStartE2EDuration="3.167932458s" podCreationTimestamp="2026-01-27 15:26:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:26:16.157425536 +0000 UTC m=+1162.138034634" watchObservedRunningTime="2026-01-27 15:26:16.167932458 +0000 UTC m=+1162.148541556" Jan 27 15:26:16 crc kubenswrapper[4772]: I0127 15:26:16.173847 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-xpbb6" podStartSLOduration=3.173825248 podStartE2EDuration="3.173825248s" podCreationTimestamp="2026-01-27 15:26:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:26:16.173601932 +0000 UTC m=+1162.154211040" watchObservedRunningTime="2026-01-27 15:26:16.173825248 +0000 UTC m=+1162.154434346" Jan 27 15:26:16 crc kubenswrapper[4772]: I0127 15:26:16.195831 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-z224f" podStartSLOduration=3.195813932 podStartE2EDuration="3.195813932s" podCreationTimestamp="2026-01-27 15:26:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:26:16.188831201 +0000 UTC m=+1162.169440299" watchObservedRunningTime="2026-01-27 15:26:16.195813932 +0000 UTC m=+1162.176423030" Jan 27 15:26:16 crc kubenswrapper[4772]: I0127 15:26:16.208747 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-ed9a-account-create-update-b7pnl" podStartSLOduration=3.208728104 podStartE2EDuration="3.208728104s" podCreationTimestamp="2026-01-27 15:26:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:26:16.207526299 +0000 UTC m=+1162.188135397" watchObservedRunningTime="2026-01-27 15:26:16.208728104 +0000 UTC m=+1162.189337202" Jan 27 15:26:16 crc kubenswrapper[4772]: I0127 15:26:16.225158 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-v7ncm" podStartSLOduration=3.225139677 podStartE2EDuration="3.225139677s" podCreationTimestamp="2026-01-27 15:26:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:26:16.218470434 +0000 UTC m=+1162.199079542" watchObservedRunningTime="2026-01-27 15:26:16.225139677 +0000 UTC m=+1162.205748765" Jan 27 15:26:16 crc kubenswrapper[4772]: I0127 15:26:16.253881 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-97c3-account-create-update-xlghl" podStartSLOduration=3.253866454 podStartE2EDuration="3.253866454s" podCreationTimestamp="2026-01-27 15:26:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:26:16.25164821 +0000 UTC m=+1162.232257308" watchObservedRunningTime="2026-01-27 15:26:16.253866454 +0000 UTC m=+1162.234475552" Jan 27 15:26:18 crc kubenswrapper[4772]: I0127 15:26:18.171032 4772 generic.go:334] "Generic (PLEG): container finished" podID="f1746148-2e3f-476f-9a1f-f3656d44fb0b" containerID="597636ff183f237bb3b639ea5c67c6b5f6f29f40e362b71df3d4ec02eaa6036b" exitCode=0 Jan 27 15:26:18 crc kubenswrapper[4772]: I0127 15:26:18.171498 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-v7ncm" event={"ID":"f1746148-2e3f-476f-9a1f-f3656d44fb0b","Type":"ContainerDied","Data":"597636ff183f237bb3b639ea5c67c6b5f6f29f40e362b71df3d4ec02eaa6036b"} Jan 27 15:26:19 crc kubenswrapper[4772]: I0127 15:26:19.184547 4772 generic.go:334] "Generic (PLEG): container finished" podID="ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112" containerID="7cb0416a54334bdd5699afd4b64397c193035c399e5586172a360ff52cd674f9" exitCode=0 Jan 27 15:26:19 crc kubenswrapper[4772]: I0127 15:26:19.184627 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-xpbb6" event={"ID":"ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112","Type":"ContainerDied","Data":"7cb0416a54334bdd5699afd4b64397c193035c399e5586172a360ff52cd674f9"} Jan 27 15:26:21 crc kubenswrapper[4772]: I0127 15:26:21.662138 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-xpbb6" Jan 27 15:26:21 crc kubenswrapper[4772]: I0127 15:26:21.669223 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-v7ncm" Jan 27 15:26:21 crc kubenswrapper[4772]: I0127 15:26:21.709967 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1746148-2e3f-476f-9a1f-f3656d44fb0b-operator-scripts\") pod \"f1746148-2e3f-476f-9a1f-f3656d44fb0b\" (UID: \"f1746148-2e3f-476f-9a1f-f3656d44fb0b\") " Jan 27 15:26:21 crc kubenswrapper[4772]: I0127 15:26:21.710017 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112-operator-scripts\") pod \"ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112\" (UID: \"ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112\") " Jan 27 15:26:21 crc kubenswrapper[4772]: I0127 15:26:21.710050 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrbpt\" (UniqueName: \"kubernetes.io/projected/f1746148-2e3f-476f-9a1f-f3656d44fb0b-kube-api-access-mrbpt\") pod \"f1746148-2e3f-476f-9a1f-f3656d44fb0b\" (UID: \"f1746148-2e3f-476f-9a1f-f3656d44fb0b\") " Jan 27 15:26:21 crc kubenswrapper[4772]: I0127 15:26:21.710302 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22x22\" (UniqueName: \"kubernetes.io/projected/ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112-kube-api-access-22x22\") pod \"ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112\" (UID: \"ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112\") " Jan 27 15:26:21 crc kubenswrapper[4772]: I0127 15:26:21.710933 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112" (UID: "ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:26:21 crc kubenswrapper[4772]: I0127 15:26:21.711670 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1746148-2e3f-476f-9a1f-f3656d44fb0b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f1746148-2e3f-476f-9a1f-f3656d44fb0b" (UID: "f1746148-2e3f-476f-9a1f-f3656d44fb0b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:26:21 crc kubenswrapper[4772]: I0127 15:26:21.717828 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1746148-2e3f-476f-9a1f-f3656d44fb0b-kube-api-access-mrbpt" (OuterVolumeSpecName: "kube-api-access-mrbpt") pod "f1746148-2e3f-476f-9a1f-f3656d44fb0b" (UID: "f1746148-2e3f-476f-9a1f-f3656d44fb0b"). InnerVolumeSpecName "kube-api-access-mrbpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:26:21 crc kubenswrapper[4772]: I0127 15:26:21.718416 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112-kube-api-access-22x22" (OuterVolumeSpecName: "kube-api-access-22x22") pod "ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112" (UID: "ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112"). InnerVolumeSpecName "kube-api-access-22x22". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:26:21 crc kubenswrapper[4772]: I0127 15:26:21.812213 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22x22\" (UniqueName: \"kubernetes.io/projected/ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112-kube-api-access-22x22\") on node \"crc\" DevicePath \"\"" Jan 27 15:26:21 crc kubenswrapper[4772]: I0127 15:26:21.812261 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1746148-2e3f-476f-9a1f-f3656d44fb0b-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:26:21 crc kubenswrapper[4772]: I0127 15:26:21.812274 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:26:21 crc kubenswrapper[4772]: I0127 15:26:21.812288 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrbpt\" (UniqueName: \"kubernetes.io/projected/f1746148-2e3f-476f-9a1f-f3656d44fb0b-kube-api-access-mrbpt\") on node \"crc\" DevicePath \"\"" Jan 27 15:26:22 crc kubenswrapper[4772]: I0127 15:26:22.208784 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-xpbb6" event={"ID":"ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112","Type":"ContainerDied","Data":"7078456546f57835b75bdc01276c8dc7b83b93913928c27cd8b38ff3cf49ddaf"} Jan 27 15:26:22 crc kubenswrapper[4772]: I0127 15:26:22.208809 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-xpbb6" Jan 27 15:26:22 crc kubenswrapper[4772]: I0127 15:26:22.208820 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7078456546f57835b75bdc01276c8dc7b83b93913928c27cd8b38ff3cf49ddaf" Jan 27 15:26:22 crc kubenswrapper[4772]: I0127 15:26:22.210508 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-v7ncm" event={"ID":"f1746148-2e3f-476f-9a1f-f3656d44fb0b","Type":"ContainerDied","Data":"d7520f953d62324f0f59051719778147c98517e633ea0fe6fa2fd5195a4cadf3"} Jan 27 15:26:22 crc kubenswrapper[4772]: I0127 15:26:22.210554 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-v7ncm" Jan 27 15:26:22 crc kubenswrapper[4772]: I0127 15:26:22.210554 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7520f953d62324f0f59051719778147c98517e633ea0fe6fa2fd5195a4cadf3" Jan 27 15:26:23 crc kubenswrapper[4772]: I0127 15:26:23.221313 4772 generic.go:334] "Generic (PLEG): container finished" podID="b412abae-93af-4ae0-8cd8-7c0a827da4b3" containerID="478f9eb73f50cba542d4259825587e98caddfe9513876ed4823af8b00681f571" exitCode=0 Jan 27 15:26:23 crc kubenswrapper[4772]: I0127 15:26:23.221425 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-z224f" event={"ID":"b412abae-93af-4ae0-8cd8-7c0a827da4b3","Type":"ContainerDied","Data":"478f9eb73f50cba542d4259825587e98caddfe9513876ed4823af8b00681f571"} Jan 27 15:26:25 crc kubenswrapper[4772]: I0127 15:26:25.091725 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-z224f" Jan 27 15:26:25 crc kubenswrapper[4772]: I0127 15:26:25.162947 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b412abae-93af-4ae0-8cd8-7c0a827da4b3-operator-scripts\") pod \"b412abae-93af-4ae0-8cd8-7c0a827da4b3\" (UID: \"b412abae-93af-4ae0-8cd8-7c0a827da4b3\") " Jan 27 15:26:25 crc kubenswrapper[4772]: I0127 15:26:25.163033 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ffgc\" (UniqueName: \"kubernetes.io/projected/b412abae-93af-4ae0-8cd8-7c0a827da4b3-kube-api-access-9ffgc\") pod \"b412abae-93af-4ae0-8cd8-7c0a827da4b3\" (UID: \"b412abae-93af-4ae0-8cd8-7c0a827da4b3\") " Jan 27 15:26:25 crc kubenswrapper[4772]: I0127 15:26:25.163572 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b412abae-93af-4ae0-8cd8-7c0a827da4b3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b412abae-93af-4ae0-8cd8-7c0a827da4b3" (UID: "b412abae-93af-4ae0-8cd8-7c0a827da4b3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:26:25 crc kubenswrapper[4772]: I0127 15:26:25.168416 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b412abae-93af-4ae0-8cd8-7c0a827da4b3-kube-api-access-9ffgc" (OuterVolumeSpecName: "kube-api-access-9ffgc") pod "b412abae-93af-4ae0-8cd8-7c0a827da4b3" (UID: "b412abae-93af-4ae0-8cd8-7c0a827da4b3"). InnerVolumeSpecName "kube-api-access-9ffgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:26:25 crc kubenswrapper[4772]: I0127 15:26:25.238251 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-z224f" event={"ID":"b412abae-93af-4ae0-8cd8-7c0a827da4b3","Type":"ContainerDied","Data":"96d15bd1f51df9f2151a057a908e64905ec96ea27be5ce6e7029e233334cd82f"} Jan 27 15:26:25 crc kubenswrapper[4772]: I0127 15:26:25.238294 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96d15bd1f51df9f2151a057a908e64905ec96ea27be5ce6e7029e233334cd82f" Jan 27 15:26:25 crc kubenswrapper[4772]: I0127 15:26:25.238312 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-z224f" Jan 27 15:26:25 crc kubenswrapper[4772]: I0127 15:26:25.265113 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b412abae-93af-4ae0-8cd8-7c0a827da4b3-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:26:25 crc kubenswrapper[4772]: I0127 15:26:25.265148 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ffgc\" (UniqueName: \"kubernetes.io/projected/b412abae-93af-4ae0-8cd8-7c0a827da4b3-kube-api-access-9ffgc\") on node \"crc\" DevicePath \"\"" Jan 27 15:26:26 crc kubenswrapper[4772]: I0127 15:26:26.248925 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-kvb25" event={"ID":"57a1d71f-3b00-42c0-92c4-a29fb3d4518c","Type":"ContainerStarted","Data":"e1a2cafeb608c7919a88b50bf39a141cb90ef87745db78d4f8f6a94522bb8d2e"} Jan 27 15:26:26 crc kubenswrapper[4772]: I0127 15:26:26.267790 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-kvb25" podStartSLOduration=2.667467401 podStartE2EDuration="13.267768321s" podCreationTimestamp="2026-01-27 15:26:13 +0000 UTC" firstStartedPulling="2026-01-27 15:26:14.733393191 +0000 UTC m=+1160.714002289" lastFinishedPulling="2026-01-27 15:26:25.333694111 +0000 UTC m=+1171.314303209" observedRunningTime="2026-01-27 15:26:26.265025792 +0000 UTC m=+1172.245634900" watchObservedRunningTime="2026-01-27 15:26:26.267768321 +0000 UTC m=+1172.248377419" Jan 27 15:26:33 crc kubenswrapper[4772]: I0127 15:26:33.317460 4772 generic.go:334] "Generic (PLEG): container finished" podID="b163780a-6dd7-4232-b0da-a22f18d36fcc" containerID="a5ffbaeea04257a22f38554ccc4304785fadfe22ac90bb6e3544b162aab10857" exitCode=0 Jan 27 15:26:33 crc kubenswrapper[4772]: I0127 15:26:33.317563 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ed9a-account-create-update-b7pnl" event={"ID":"b163780a-6dd7-4232-b0da-a22f18d36fcc","Type":"ContainerDied","Data":"a5ffbaeea04257a22f38554ccc4304785fadfe22ac90bb6e3544b162aab10857"} Jan 27 15:26:33 crc kubenswrapper[4772]: I0127 15:26:33.320750 4772 generic.go:334] "Generic (PLEG): container finished" podID="af61fb8e-e749-4872-8dc6-c590e4b9787a" containerID="d265eb93689c326c68ce844d36ec8e13845ff3f6cfb1ed7e88273d0cf4e91cbd" exitCode=0 Jan 27 15:26:33 crc kubenswrapper[4772]: I0127 15:26:33.320817 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-97c3-account-create-update-xlghl" event={"ID":"af61fb8e-e749-4872-8dc6-c590e4b9787a","Type":"ContainerDied","Data":"d265eb93689c326c68ce844d36ec8e13845ff3f6cfb1ed7e88273d0cf4e91cbd"} Jan 27 15:26:34 crc kubenswrapper[4772]: I0127 15:26:34.778557 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-97c3-account-create-update-xlghl" Jan 27 15:26:34 crc kubenswrapper[4772]: I0127 15:26:34.790432 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ed9a-account-create-update-b7pnl" Jan 27 15:26:34 crc kubenswrapper[4772]: I0127 15:26:34.932477 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b163780a-6dd7-4232-b0da-a22f18d36fcc-operator-scripts\") pod \"b163780a-6dd7-4232-b0da-a22f18d36fcc\" (UID: \"b163780a-6dd7-4232-b0da-a22f18d36fcc\") " Jan 27 15:26:34 crc kubenswrapper[4772]: I0127 15:26:34.932642 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5rsl\" (UniqueName: \"kubernetes.io/projected/af61fb8e-e749-4872-8dc6-c590e4b9787a-kube-api-access-f5rsl\") pod \"af61fb8e-e749-4872-8dc6-c590e4b9787a\" (UID: \"af61fb8e-e749-4872-8dc6-c590e4b9787a\") " Jan 27 15:26:34 crc kubenswrapper[4772]: I0127 15:26:34.932688 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t54t7\" (UniqueName: \"kubernetes.io/projected/b163780a-6dd7-4232-b0da-a22f18d36fcc-kube-api-access-t54t7\") pod \"b163780a-6dd7-4232-b0da-a22f18d36fcc\" (UID: \"b163780a-6dd7-4232-b0da-a22f18d36fcc\") " Jan 27 15:26:34 crc kubenswrapper[4772]: I0127 15:26:34.932748 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af61fb8e-e749-4872-8dc6-c590e4b9787a-operator-scripts\") pod \"af61fb8e-e749-4872-8dc6-c590e4b9787a\" (UID: \"af61fb8e-e749-4872-8dc6-c590e4b9787a\") " Jan 27 15:26:34 crc kubenswrapper[4772]: I0127 15:26:34.933394 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b163780a-6dd7-4232-b0da-a22f18d36fcc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b163780a-6dd7-4232-b0da-a22f18d36fcc" (UID: "b163780a-6dd7-4232-b0da-a22f18d36fcc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:26:34 crc kubenswrapper[4772]: I0127 15:26:34.933508 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af61fb8e-e749-4872-8dc6-c590e4b9787a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "af61fb8e-e749-4872-8dc6-c590e4b9787a" (UID: "af61fb8e-e749-4872-8dc6-c590e4b9787a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:26:34 crc kubenswrapper[4772]: I0127 15:26:34.933927 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b163780a-6dd7-4232-b0da-a22f18d36fcc-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:26:34 crc kubenswrapper[4772]: I0127 15:26:34.933960 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af61fb8e-e749-4872-8dc6-c590e4b9787a-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:26:34 crc kubenswrapper[4772]: I0127 15:26:34.938843 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b163780a-6dd7-4232-b0da-a22f18d36fcc-kube-api-access-t54t7" (OuterVolumeSpecName: "kube-api-access-t54t7") pod "b163780a-6dd7-4232-b0da-a22f18d36fcc" (UID: "b163780a-6dd7-4232-b0da-a22f18d36fcc"). InnerVolumeSpecName "kube-api-access-t54t7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:26:34 crc kubenswrapper[4772]: I0127 15:26:34.938970 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af61fb8e-e749-4872-8dc6-c590e4b9787a-kube-api-access-f5rsl" (OuterVolumeSpecName: "kube-api-access-f5rsl") pod "af61fb8e-e749-4872-8dc6-c590e4b9787a" (UID: "af61fb8e-e749-4872-8dc6-c590e4b9787a"). InnerVolumeSpecName "kube-api-access-f5rsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:26:35 crc kubenswrapper[4772]: I0127 15:26:35.035654 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5rsl\" (UniqueName: \"kubernetes.io/projected/af61fb8e-e749-4872-8dc6-c590e4b9787a-kube-api-access-f5rsl\") on node \"crc\" DevicePath \"\"" Jan 27 15:26:35 crc kubenswrapper[4772]: I0127 15:26:35.035693 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t54t7\" (UniqueName: \"kubernetes.io/projected/b163780a-6dd7-4232-b0da-a22f18d36fcc-kube-api-access-t54t7\") on node \"crc\" DevicePath \"\"" Jan 27 15:26:35 crc kubenswrapper[4772]: I0127 15:26:35.345784 4772 generic.go:334] "Generic (PLEG): container finished" podID="c54b2036-d943-4f0d-b1c4-8a47dfab5099" containerID="c454404cb2dabeb6539bab075b0096e5a7ba9d3726f1b7a2ce5d55b30cc778e8" exitCode=0 Jan 27 15:26:35 crc kubenswrapper[4772]: I0127 15:26:35.345799 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d2a3-account-create-update-hfkkb" event={"ID":"c54b2036-d943-4f0d-b1c4-8a47dfab5099","Type":"ContainerDied","Data":"c454404cb2dabeb6539bab075b0096e5a7ba9d3726f1b7a2ce5d55b30cc778e8"} Jan 27 15:26:35 crc kubenswrapper[4772]: I0127 15:26:35.348624 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ed9a-account-create-update-b7pnl" Jan 27 15:26:35 crc kubenswrapper[4772]: I0127 15:26:35.348627 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ed9a-account-create-update-b7pnl" event={"ID":"b163780a-6dd7-4232-b0da-a22f18d36fcc","Type":"ContainerDied","Data":"e83383c104bd8b761200170a9e0bc2e126e98220c9813a9e33d886b26ab147ee"} Jan 27 15:26:35 crc kubenswrapper[4772]: I0127 15:26:35.348684 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e83383c104bd8b761200170a9e0bc2e126e98220c9813a9e33d886b26ab147ee" Jan 27 15:26:35 crc kubenswrapper[4772]: I0127 15:26:35.350414 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-97c3-account-create-update-xlghl" event={"ID":"af61fb8e-e749-4872-8dc6-c590e4b9787a","Type":"ContainerDied","Data":"4fc5b98afeb27a7f560b6faf0b64da965c312fa1d6668ea8e195f9bf3cf774cd"} Jan 27 15:26:35 crc kubenswrapper[4772]: I0127 15:26:35.350454 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fc5b98afeb27a7f560b6faf0b64da965c312fa1d6668ea8e195f9bf3cf774cd" Jan 27 15:26:35 crc kubenswrapper[4772]: I0127 15:26:35.350487 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-97c3-account-create-update-xlghl" Jan 27 15:26:36 crc kubenswrapper[4772]: I0127 15:26:36.643099 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d2a3-account-create-update-hfkkb" Jan 27 15:26:36 crc kubenswrapper[4772]: I0127 15:26:36.659976 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c54b2036-d943-4f0d-b1c4-8a47dfab5099-operator-scripts\") pod \"c54b2036-d943-4f0d-b1c4-8a47dfab5099\" (UID: \"c54b2036-d943-4f0d-b1c4-8a47dfab5099\") " Jan 27 15:26:36 crc kubenswrapper[4772]: I0127 15:26:36.660046 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfhpb\" (UniqueName: \"kubernetes.io/projected/c54b2036-d943-4f0d-b1c4-8a47dfab5099-kube-api-access-rfhpb\") pod \"c54b2036-d943-4f0d-b1c4-8a47dfab5099\" (UID: \"c54b2036-d943-4f0d-b1c4-8a47dfab5099\") " Jan 27 15:26:36 crc kubenswrapper[4772]: I0127 15:26:36.661396 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c54b2036-d943-4f0d-b1c4-8a47dfab5099-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c54b2036-d943-4f0d-b1c4-8a47dfab5099" (UID: "c54b2036-d943-4f0d-b1c4-8a47dfab5099"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:26:36 crc kubenswrapper[4772]: I0127 15:26:36.667326 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c54b2036-d943-4f0d-b1c4-8a47dfab5099-kube-api-access-rfhpb" (OuterVolumeSpecName: "kube-api-access-rfhpb") pod "c54b2036-d943-4f0d-b1c4-8a47dfab5099" (UID: "c54b2036-d943-4f0d-b1c4-8a47dfab5099"). InnerVolumeSpecName "kube-api-access-rfhpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:26:36 crc kubenswrapper[4772]: I0127 15:26:36.762280 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c54b2036-d943-4f0d-b1c4-8a47dfab5099-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:26:36 crc kubenswrapper[4772]: I0127 15:26:36.762310 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfhpb\" (UniqueName: \"kubernetes.io/projected/c54b2036-d943-4f0d-b1c4-8a47dfab5099-kube-api-access-rfhpb\") on node \"crc\" DevicePath \"\"" Jan 27 15:26:37 crc kubenswrapper[4772]: I0127 15:26:37.371012 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d2a3-account-create-update-hfkkb" event={"ID":"c54b2036-d943-4f0d-b1c4-8a47dfab5099","Type":"ContainerDied","Data":"08d6532f77d29e47d3d345a3cbcc4de484a6e75463c90eca472f8ccd07be6a84"} Jan 27 15:26:37 crc kubenswrapper[4772]: I0127 15:26:37.371066 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08d6532f77d29e47d3d345a3cbcc4de484a6e75463c90eca472f8ccd07be6a84" Jan 27 15:26:37 crc kubenswrapper[4772]: I0127 15:26:37.371084 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d2a3-account-create-update-hfkkb" Jan 27 15:26:45 crc kubenswrapper[4772]: I0127 15:26:45.441134 4772 generic.go:334] "Generic (PLEG): container finished" podID="57a1d71f-3b00-42c0-92c4-a29fb3d4518c" containerID="e1a2cafeb608c7919a88b50bf39a141cb90ef87745db78d4f8f6a94522bb8d2e" exitCode=0 Jan 27 15:26:45 crc kubenswrapper[4772]: I0127 15:26:45.441265 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-kvb25" event={"ID":"57a1d71f-3b00-42c0-92c4-a29fb3d4518c","Type":"ContainerDied","Data":"e1a2cafeb608c7919a88b50bf39a141cb90ef87745db78d4f8f6a94522bb8d2e"} Jan 27 15:26:46 crc kubenswrapper[4772]: I0127 15:26:46.780186 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-kvb25" Jan 27 15:26:46 crc kubenswrapper[4772]: I0127 15:26:46.928376 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgk7m\" (UniqueName: \"kubernetes.io/projected/57a1d71f-3b00-42c0-92c4-a29fb3d4518c-kube-api-access-mgk7m\") pod \"57a1d71f-3b00-42c0-92c4-a29fb3d4518c\" (UID: \"57a1d71f-3b00-42c0-92c4-a29fb3d4518c\") " Jan 27 15:26:46 crc kubenswrapper[4772]: I0127 15:26:46.928418 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57a1d71f-3b00-42c0-92c4-a29fb3d4518c-config-data\") pod \"57a1d71f-3b00-42c0-92c4-a29fb3d4518c\" (UID: \"57a1d71f-3b00-42c0-92c4-a29fb3d4518c\") " Jan 27 15:26:46 crc kubenswrapper[4772]: I0127 15:26:46.928492 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57a1d71f-3b00-42c0-92c4-a29fb3d4518c-combined-ca-bundle\") pod \"57a1d71f-3b00-42c0-92c4-a29fb3d4518c\" (UID: \"57a1d71f-3b00-42c0-92c4-a29fb3d4518c\") " Jan 27 15:26:46 crc kubenswrapper[4772]: I0127 15:26:46.934270 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a1d71f-3b00-42c0-92c4-a29fb3d4518c-kube-api-access-mgk7m" (OuterVolumeSpecName: "kube-api-access-mgk7m") pod "57a1d71f-3b00-42c0-92c4-a29fb3d4518c" (UID: "57a1d71f-3b00-42c0-92c4-a29fb3d4518c"). InnerVolumeSpecName "kube-api-access-mgk7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:26:46 crc kubenswrapper[4772]: I0127 15:26:46.963895 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57a1d71f-3b00-42c0-92c4-a29fb3d4518c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57a1d71f-3b00-42c0-92c4-a29fb3d4518c" (UID: "57a1d71f-3b00-42c0-92c4-a29fb3d4518c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:26:46 crc kubenswrapper[4772]: I0127 15:26:46.979382 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57a1d71f-3b00-42c0-92c4-a29fb3d4518c-config-data" (OuterVolumeSpecName: "config-data") pod "57a1d71f-3b00-42c0-92c4-a29fb3d4518c" (UID: "57a1d71f-3b00-42c0-92c4-a29fb3d4518c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.031025 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgk7m\" (UniqueName: \"kubernetes.io/projected/57a1d71f-3b00-42c0-92c4-a29fb3d4518c-kube-api-access-mgk7m\") on node \"crc\" DevicePath \"\"" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.031080 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57a1d71f-3b00-42c0-92c4-a29fb3d4518c-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.031094 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57a1d71f-3b00-42c0-92c4-a29fb3d4518c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.458749 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-kvb25" event={"ID":"57a1d71f-3b00-42c0-92c4-a29fb3d4518c","Type":"ContainerDied","Data":"cc68b23df999b3c351ddd5132639a8352068b638f27d53a94da0ce06a1009dad"} Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.458791 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc68b23df999b3c351ddd5132639a8352068b638f27d53a94da0ce06a1009dad" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.458795 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-kvb25" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.722443 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-x8hbr"] Jan 27 15:26:47 crc kubenswrapper[4772]: E0127 15:26:47.722769 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af61fb8e-e749-4872-8dc6-c590e4b9787a" containerName="mariadb-account-create-update" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.722785 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="af61fb8e-e749-4872-8dc6-c590e4b9787a" containerName="mariadb-account-create-update" Jan 27 15:26:47 crc kubenswrapper[4772]: E0127 15:26:47.722811 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112" containerName="mariadb-database-create" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.722817 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112" containerName="mariadb-database-create" Jan 27 15:26:47 crc kubenswrapper[4772]: E0127 15:26:47.722828 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1746148-2e3f-476f-9a1f-f3656d44fb0b" containerName="mariadb-database-create" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.722834 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1746148-2e3f-476f-9a1f-f3656d44fb0b" containerName="mariadb-database-create" Jan 27 15:26:47 crc kubenswrapper[4772]: E0127 15:26:47.722847 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c54b2036-d943-4f0d-b1c4-8a47dfab5099" containerName="mariadb-account-create-update" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.722852 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="c54b2036-d943-4f0d-b1c4-8a47dfab5099" containerName="mariadb-account-create-update" Jan 27 15:26:47 crc kubenswrapper[4772]: E0127 15:26:47.722861 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b412abae-93af-4ae0-8cd8-7c0a827da4b3" containerName="mariadb-database-create" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.722868 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="b412abae-93af-4ae0-8cd8-7c0a827da4b3" containerName="mariadb-database-create" Jan 27 15:26:47 crc kubenswrapper[4772]: E0127 15:26:47.722877 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57a1d71f-3b00-42c0-92c4-a29fb3d4518c" containerName="keystone-db-sync" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.722883 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="57a1d71f-3b00-42c0-92c4-a29fb3d4518c" containerName="keystone-db-sync" Jan 27 15:26:47 crc kubenswrapper[4772]: E0127 15:26:47.722900 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b163780a-6dd7-4232-b0da-a22f18d36fcc" containerName="mariadb-account-create-update" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.722906 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="b163780a-6dd7-4232-b0da-a22f18d36fcc" containerName="mariadb-account-create-update" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.723069 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112" containerName="mariadb-database-create" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.723081 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="57a1d71f-3b00-42c0-92c4-a29fb3d4518c" containerName="keystone-db-sync" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.723089 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="b163780a-6dd7-4232-b0da-a22f18d36fcc" containerName="mariadb-account-create-update" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.723100 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="c54b2036-d943-4f0d-b1c4-8a47dfab5099" containerName="mariadb-account-create-update" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.723112 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1746148-2e3f-476f-9a1f-f3656d44fb0b" containerName="mariadb-database-create" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.723122 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="af61fb8e-e749-4872-8dc6-c590e4b9787a" containerName="mariadb-account-create-update" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.723132 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="b412abae-93af-4ae0-8cd8-7c0a827da4b3" containerName="mariadb-database-create" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.724758 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.733933 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-x8hbr"] Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.756752 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-rl8kf"] Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.759080 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rl8kf" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.764159 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.764312 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.764510 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.764620 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.764896 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-bdjsw" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.772857 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rl8kf"] Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.851445 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-dns-svc\") pod \"dnsmasq-dns-55fff446b9-x8hbr\" (UID: \"34ee743a-1628-42e4-a465-0e3957cae089\") " pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.851511 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-credential-keys\") pod \"keystone-bootstrap-rl8kf\" (UID: \"0851ad59-841c-4133-a043-13d2cfdb0803\") " pod="openstack/keystone-bootstrap-rl8kf" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.851553 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx2xr\" (UniqueName: \"kubernetes.io/projected/34ee743a-1628-42e4-a465-0e3957cae089-kube-api-access-tx2xr\") pod \"dnsmasq-dns-55fff446b9-x8hbr\" (UID: \"34ee743a-1628-42e4-a465-0e3957cae089\") " pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.851643 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwktg\" (UniqueName: \"kubernetes.io/projected/0851ad59-841c-4133-a043-13d2cfdb0803-kube-api-access-fwktg\") pod \"keystone-bootstrap-rl8kf\" (UID: \"0851ad59-841c-4133-a043-13d2cfdb0803\") " pod="openstack/keystone-bootstrap-rl8kf" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.851692 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-scripts\") pod \"keystone-bootstrap-rl8kf\" (UID: \"0851ad59-841c-4133-a043-13d2cfdb0803\") " pod="openstack/keystone-bootstrap-rl8kf" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.851731 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-x8hbr\" (UID: \"34ee743a-1628-42e4-a465-0e3957cae089\") " pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.851753 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-x8hbr\" (UID: \"34ee743a-1628-42e4-a465-0e3957cae089\") " pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.851806 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-config\") pod \"dnsmasq-dns-55fff446b9-x8hbr\" (UID: \"34ee743a-1628-42e4-a465-0e3957cae089\") " pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.851866 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-fernet-keys\") pod \"keystone-bootstrap-rl8kf\" (UID: \"0851ad59-841c-4133-a043-13d2cfdb0803\") " pod="openstack/keystone-bootstrap-rl8kf" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.852115 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-combined-ca-bundle\") pod \"keystone-bootstrap-rl8kf\" (UID: \"0851ad59-841c-4133-a043-13d2cfdb0803\") " pod="openstack/keystone-bootstrap-rl8kf" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.852217 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-x8hbr\" (UID: \"34ee743a-1628-42e4-a465-0e3957cae089\") " pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.852388 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-config-data\") pod \"keystone-bootstrap-rl8kf\" (UID: \"0851ad59-841c-4133-a043-13d2cfdb0803\") " pod="openstack/keystone-bootstrap-rl8kf" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.924246 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-v689b"] Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.925840 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-v689b" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.930485 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.930916 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-rjm9r" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.936314 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-v689b"] Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.937749 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.954796 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwktg\" (UniqueName: \"kubernetes.io/projected/0851ad59-841c-4133-a043-13d2cfdb0803-kube-api-access-fwktg\") pod \"keystone-bootstrap-rl8kf\" (UID: \"0851ad59-841c-4133-a043-13d2cfdb0803\") " pod="openstack/keystone-bootstrap-rl8kf" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.954846 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-scripts\") pod \"keystone-bootstrap-rl8kf\" (UID: \"0851ad59-841c-4133-a043-13d2cfdb0803\") " pod="openstack/keystone-bootstrap-rl8kf" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.954873 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-x8hbr\" (UID: \"34ee743a-1628-42e4-a465-0e3957cae089\") " pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.954888 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-x8hbr\" (UID: \"34ee743a-1628-42e4-a465-0e3957cae089\") " pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.954916 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-config\") pod \"dnsmasq-dns-55fff446b9-x8hbr\" (UID: \"34ee743a-1628-42e4-a465-0e3957cae089\") " pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.954942 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-fernet-keys\") pod \"keystone-bootstrap-rl8kf\" (UID: \"0851ad59-841c-4133-a043-13d2cfdb0803\") " pod="openstack/keystone-bootstrap-rl8kf" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.954974 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-combined-ca-bundle\") pod \"keystone-bootstrap-rl8kf\" (UID: \"0851ad59-841c-4133-a043-13d2cfdb0803\") " pod="openstack/keystone-bootstrap-rl8kf" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.954990 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-x8hbr\" (UID: \"34ee743a-1628-42e4-a465-0e3957cae089\") " pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.955018 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-config-data\") pod \"keystone-bootstrap-rl8kf\" (UID: \"0851ad59-841c-4133-a043-13d2cfdb0803\") " pod="openstack/keystone-bootstrap-rl8kf" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.955049 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-dns-svc\") pod \"dnsmasq-dns-55fff446b9-x8hbr\" (UID: \"34ee743a-1628-42e4-a465-0e3957cae089\") " pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.955068 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-credential-keys\") pod \"keystone-bootstrap-rl8kf\" (UID: \"0851ad59-841c-4133-a043-13d2cfdb0803\") " pod="openstack/keystone-bootstrap-rl8kf" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.955087 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx2xr\" (UniqueName: \"kubernetes.io/projected/34ee743a-1628-42e4-a465-0e3957cae089-kube-api-access-tx2xr\") pod \"dnsmasq-dns-55fff446b9-x8hbr\" (UID: \"34ee743a-1628-42e4-a465-0e3957cae089\") " pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.957426 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-x8hbr\" (UID: \"34ee743a-1628-42e4-a465-0e3957cae089\") " pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.959445 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-x8hbr\" (UID: \"34ee743a-1628-42e4-a465-0e3957cae089\") " pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.960209 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-config\") pod \"dnsmasq-dns-55fff446b9-x8hbr\" (UID: \"34ee743a-1628-42e4-a465-0e3957cae089\") " pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.961868 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-combined-ca-bundle\") pod \"keystone-bootstrap-rl8kf\" (UID: \"0851ad59-841c-4133-a043-13d2cfdb0803\") " pod="openstack/keystone-bootstrap-rl8kf" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.962095 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-dns-svc\") pod \"dnsmasq-dns-55fff446b9-x8hbr\" (UID: \"34ee743a-1628-42e4-a465-0e3957cae089\") " pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.962718 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-x8hbr\" (UID: \"34ee743a-1628-42e4-a465-0e3957cae089\") " pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.964032 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-fernet-keys\") pod \"keystone-bootstrap-rl8kf\" (UID: \"0851ad59-841c-4133-a043-13d2cfdb0803\") " pod="openstack/keystone-bootstrap-rl8kf" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.968899 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-config-data\") pod \"keystone-bootstrap-rl8kf\" (UID: \"0851ad59-841c-4133-a043-13d2cfdb0803\") " pod="openstack/keystone-bootstrap-rl8kf" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.972696 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-credential-keys\") pod \"keystone-bootstrap-rl8kf\" (UID: \"0851ad59-841c-4133-a043-13d2cfdb0803\") " pod="openstack/keystone-bootstrap-rl8kf" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.975352 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-scripts\") pod \"keystone-bootstrap-rl8kf\" (UID: \"0851ad59-841c-4133-a043-13d2cfdb0803\") " pod="openstack/keystone-bootstrap-rl8kf" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.984836 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.988149 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.992933 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx2xr\" (UniqueName: \"kubernetes.io/projected/34ee743a-1628-42e4-a465-0e3957cae089-kube-api-access-tx2xr\") pod \"dnsmasq-dns-55fff446b9-x8hbr\" (UID: \"34ee743a-1628-42e4-a465-0e3957cae089\") " pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.993052 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 27 15:26:47 crc kubenswrapper[4772]: I0127 15:26:47.996076 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwktg\" (UniqueName: \"kubernetes.io/projected/0851ad59-841c-4133-a043-13d2cfdb0803-kube-api-access-fwktg\") pod \"keystone-bootstrap-rl8kf\" (UID: \"0851ad59-841c-4133-a043-13d2cfdb0803\") " pod="openstack/keystone-bootstrap-rl8kf" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.006966 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.026186 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.052949 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.056008 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b0625578-3b48-44c7-9082-174fce3a7e74-config\") pod \"neutron-db-sync-v689b\" (UID: \"b0625578-3b48-44c7-9082-174fce3a7e74\") " pod="openstack/neutron-db-sync-v689b" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.056085 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jn9b\" (UniqueName: \"kubernetes.io/projected/b0625578-3b48-44c7-9082-174fce3a7e74-kube-api-access-9jn9b\") pod \"neutron-db-sync-v689b\" (UID: \"b0625578-3b48-44c7-9082-174fce3a7e74\") " pod="openstack/neutron-db-sync-v689b" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.056112 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0625578-3b48-44c7-9082-174fce3a7e74-combined-ca-bundle\") pod \"neutron-db-sync-v689b\" (UID: \"b0625578-3b48-44c7-9082-174fce3a7e74\") " pod="openstack/neutron-db-sync-v689b" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.060540 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-pmk27"] Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.061661 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pmk27" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.063510 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-flljj" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.063846 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.079120 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-pmk27"] Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.079841 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rl8kf" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.115316 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-8l85z"] Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.116635 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8l85z" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.122759 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.122992 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-8nhs4" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.123203 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.131059 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-8l85z"] Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.160147 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de415c6e-4424-49c4-bc9d-076a5b13ab4e-scripts\") pod \"ceilometer-0\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " pod="openstack/ceilometer-0" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.160282 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jn9b\" (UniqueName: \"kubernetes.io/projected/b0625578-3b48-44c7-9082-174fce3a7e74-kube-api-access-9jn9b\") pod \"neutron-db-sync-v689b\" (UID: \"b0625578-3b48-44c7-9082-174fce3a7e74\") " pod="openstack/neutron-db-sync-v689b" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.160320 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0625578-3b48-44c7-9082-174fce3a7e74-combined-ca-bundle\") pod \"neutron-db-sync-v689b\" (UID: \"b0625578-3b48-44c7-9082-174fce3a7e74\") " pod="openstack/neutron-db-sync-v689b" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.160345 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cvd2\" (UniqueName: \"kubernetes.io/projected/5a423229-06be-4934-9715-58105e1af686-kube-api-access-8cvd2\") pod \"barbican-db-sync-pmk27\" (UID: \"5a423229-06be-4934-9715-58105e1af686\") " pod="openstack/barbican-db-sync-pmk27" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.160400 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvxkw\" (UniqueName: \"kubernetes.io/projected/de415c6e-4424-49c4-bc9d-076a5b13ab4e-kube-api-access-fvxkw\") pod \"ceilometer-0\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " pod="openstack/ceilometer-0" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.160467 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de415c6e-4424-49c4-bc9d-076a5b13ab4e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " pod="openstack/ceilometer-0" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.160492 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de415c6e-4424-49c4-bc9d-076a5b13ab4e-run-httpd\") pod \"ceilometer-0\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " pod="openstack/ceilometer-0" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.160518 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a423229-06be-4934-9715-58105e1af686-combined-ca-bundle\") pod \"barbican-db-sync-pmk27\" (UID: \"5a423229-06be-4934-9715-58105e1af686\") " pod="openstack/barbican-db-sync-pmk27" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.160540 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de415c6e-4424-49c4-bc9d-076a5b13ab4e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " pod="openstack/ceilometer-0" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.160566 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de415c6e-4424-49c4-bc9d-076a5b13ab4e-config-data\") pod \"ceilometer-0\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " pod="openstack/ceilometer-0" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.160591 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b0625578-3b48-44c7-9082-174fce3a7e74-config\") pod \"neutron-db-sync-v689b\" (UID: \"b0625578-3b48-44c7-9082-174fce3a7e74\") " pod="openstack/neutron-db-sync-v689b" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.160621 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de415c6e-4424-49c4-bc9d-076a5b13ab4e-log-httpd\") pod \"ceilometer-0\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " pod="openstack/ceilometer-0" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.160652 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5a423229-06be-4934-9715-58105e1af686-db-sync-config-data\") pod \"barbican-db-sync-pmk27\" (UID: \"5a423229-06be-4934-9715-58105e1af686\") " pod="openstack/barbican-db-sync-pmk27" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.166020 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0625578-3b48-44c7-9082-174fce3a7e74-combined-ca-bundle\") pod \"neutron-db-sync-v689b\" (UID: \"b0625578-3b48-44c7-9082-174fce3a7e74\") " pod="openstack/neutron-db-sync-v689b" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.173927 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b0625578-3b48-44c7-9082-174fce3a7e74-config\") pod \"neutron-db-sync-v689b\" (UID: \"b0625578-3b48-44c7-9082-174fce3a7e74\") " pod="openstack/neutron-db-sync-v689b" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.200353 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jn9b\" (UniqueName: \"kubernetes.io/projected/b0625578-3b48-44c7-9082-174fce3a7e74-kube-api-access-9jn9b\") pod \"neutron-db-sync-v689b\" (UID: \"b0625578-3b48-44c7-9082-174fce3a7e74\") " pod="openstack/neutron-db-sync-v689b" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.204838 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-x8hbr"] Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.225560 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-zf2tx"] Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.226584 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zf2tx" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.232496 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.232706 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.232817 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-4tg2g" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.234701 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-zf2tx"] Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.243526 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-8d9vp"] Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.245005 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.246598 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-v689b" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.250951 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-8d9vp"] Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.262656 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de415c6e-4424-49c4-bc9d-076a5b13ab4e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " pod="openstack/ceilometer-0" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.262702 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de415c6e-4424-49c4-bc9d-076a5b13ab4e-run-httpd\") pod \"ceilometer-0\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " pod="openstack/ceilometer-0" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.262723 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a423229-06be-4934-9715-58105e1af686-combined-ca-bundle\") pod \"barbican-db-sync-pmk27\" (UID: \"5a423229-06be-4934-9715-58105e1af686\") " pod="openstack/barbican-db-sync-pmk27" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.262744 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de415c6e-4424-49c4-bc9d-076a5b13ab4e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " pod="openstack/ceilometer-0" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.262766 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de415c6e-4424-49c4-bc9d-076a5b13ab4e-config-data\") pod \"ceilometer-0\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " pod="openstack/ceilometer-0" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.262791 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de415c6e-4424-49c4-bc9d-076a5b13ab4e-log-httpd\") pod \"ceilometer-0\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " pod="openstack/ceilometer-0" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.262809 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5a423229-06be-4934-9715-58105e1af686-db-sync-config-data\") pod \"barbican-db-sync-pmk27\" (UID: \"5a423229-06be-4934-9715-58105e1af686\") " pod="openstack/barbican-db-sync-pmk27" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.262858 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de415c6e-4424-49c4-bc9d-076a5b13ab4e-scripts\") pod \"ceilometer-0\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " pod="openstack/ceilometer-0" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.262883 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ae05919-68bf-43d1-abd9-9908ec287bd0-combined-ca-bundle\") pod \"cinder-db-sync-8l85z\" (UID: \"9ae05919-68bf-43d1-abd9-9908ec287bd0\") " pod="openstack/cinder-db-sync-8l85z" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.262908 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cvd2\" (UniqueName: \"kubernetes.io/projected/5a423229-06be-4934-9715-58105e1af686-kube-api-access-8cvd2\") pod \"barbican-db-sync-pmk27\" (UID: \"5a423229-06be-4934-9715-58105e1af686\") " pod="openstack/barbican-db-sync-pmk27" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.262925 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4tmg\" (UniqueName: \"kubernetes.io/projected/9ae05919-68bf-43d1-abd9-9908ec287bd0-kube-api-access-d4tmg\") pod \"cinder-db-sync-8l85z\" (UID: \"9ae05919-68bf-43d1-abd9-9908ec287bd0\") " pod="openstack/cinder-db-sync-8l85z" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.262950 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ae05919-68bf-43d1-abd9-9908ec287bd0-scripts\") pod \"cinder-db-sync-8l85z\" (UID: \"9ae05919-68bf-43d1-abd9-9908ec287bd0\") " pod="openstack/cinder-db-sync-8l85z" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.262965 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ae05919-68bf-43d1-abd9-9908ec287bd0-config-data\") pod \"cinder-db-sync-8l85z\" (UID: \"9ae05919-68bf-43d1-abd9-9908ec287bd0\") " pod="openstack/cinder-db-sync-8l85z" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.262983 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9ae05919-68bf-43d1-abd9-9908ec287bd0-etc-machine-id\") pod \"cinder-db-sync-8l85z\" (UID: \"9ae05919-68bf-43d1-abd9-9908ec287bd0\") " pod="openstack/cinder-db-sync-8l85z" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.263015 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvxkw\" (UniqueName: \"kubernetes.io/projected/de415c6e-4424-49c4-bc9d-076a5b13ab4e-kube-api-access-fvxkw\") pod \"ceilometer-0\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " pod="openstack/ceilometer-0" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.263032 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9ae05919-68bf-43d1-abd9-9908ec287bd0-db-sync-config-data\") pod \"cinder-db-sync-8l85z\" (UID: \"9ae05919-68bf-43d1-abd9-9908ec287bd0\") " pod="openstack/cinder-db-sync-8l85z" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.265325 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de415c6e-4424-49c4-bc9d-076a5b13ab4e-run-httpd\") pod \"ceilometer-0\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " pod="openstack/ceilometer-0" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.269382 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de415c6e-4424-49c4-bc9d-076a5b13ab4e-log-httpd\") pod \"ceilometer-0\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " pod="openstack/ceilometer-0" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.271159 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de415c6e-4424-49c4-bc9d-076a5b13ab4e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " pod="openstack/ceilometer-0" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.272459 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de415c6e-4424-49c4-bc9d-076a5b13ab4e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " pod="openstack/ceilometer-0" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.277928 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5a423229-06be-4934-9715-58105e1af686-db-sync-config-data\") pod \"barbican-db-sync-pmk27\" (UID: \"5a423229-06be-4934-9715-58105e1af686\") " pod="openstack/barbican-db-sync-pmk27" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.278742 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a423229-06be-4934-9715-58105e1af686-combined-ca-bundle\") pod \"barbican-db-sync-pmk27\" (UID: \"5a423229-06be-4934-9715-58105e1af686\") " pod="openstack/barbican-db-sync-pmk27" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.280410 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de415c6e-4424-49c4-bc9d-076a5b13ab4e-scripts\") pod \"ceilometer-0\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " pod="openstack/ceilometer-0" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.285096 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cvd2\" (UniqueName: \"kubernetes.io/projected/5a423229-06be-4934-9715-58105e1af686-kube-api-access-8cvd2\") pod \"barbican-db-sync-pmk27\" (UID: \"5a423229-06be-4934-9715-58105e1af686\") " pod="openstack/barbican-db-sync-pmk27" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.285994 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvxkw\" (UniqueName: \"kubernetes.io/projected/de415c6e-4424-49c4-bc9d-076a5b13ab4e-kube-api-access-fvxkw\") pod \"ceilometer-0\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " pod="openstack/ceilometer-0" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.293366 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de415c6e-4424-49c4-bc9d-076a5b13ab4e-config-data\") pod \"ceilometer-0\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " pod="openstack/ceilometer-0" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.364721 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9ae05919-68bf-43d1-abd9-9908ec287bd0-etc-machine-id\") pod \"cinder-db-sync-8l85z\" (UID: \"9ae05919-68bf-43d1-abd9-9908ec287bd0\") " pod="openstack/cinder-db-sync-8l85z" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.364806 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmlv8\" (UniqueName: \"kubernetes.io/projected/e329efba-60e3-49c7-81ff-b073be77e34b-kube-api-access-zmlv8\") pod \"dnsmasq-dns-76fcf4b695-8d9vp\" (UID: \"e329efba-60e3-49c7-81ff-b073be77e34b\") " pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.364839 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9ae05919-68bf-43d1-abd9-9908ec287bd0-db-sync-config-data\") pod \"cinder-db-sync-8l85z\" (UID: \"9ae05919-68bf-43d1-abd9-9908ec287bd0\") " pod="openstack/cinder-db-sync-8l85z" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.364868 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-config-data\") pod \"placement-db-sync-zf2tx\" (UID: \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\") " pod="openstack/placement-db-sync-zf2tx" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.364890 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-logs\") pod \"placement-db-sync-zf2tx\" (UID: \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\") " pod="openstack/placement-db-sync-zf2tx" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.364958 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-8d9vp\" (UID: \"e329efba-60e3-49c7-81ff-b073be77e34b\") " pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.364987 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-config\") pod \"dnsmasq-dns-76fcf4b695-8d9vp\" (UID: \"e329efba-60e3-49c7-81ff-b073be77e34b\") " pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.365038 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-8d9vp\" (UID: \"e329efba-60e3-49c7-81ff-b073be77e34b\") " pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.365101 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8x5b\" (UniqueName: \"kubernetes.io/projected/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-kube-api-access-d8x5b\") pod \"placement-db-sync-zf2tx\" (UID: \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\") " pod="openstack/placement-db-sync-zf2tx" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.365131 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-combined-ca-bundle\") pod \"placement-db-sync-zf2tx\" (UID: \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\") " pod="openstack/placement-db-sync-zf2tx" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.365154 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-8d9vp\" (UID: \"e329efba-60e3-49c7-81ff-b073be77e34b\") " pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.365224 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-8d9vp\" (UID: \"e329efba-60e3-49c7-81ff-b073be77e34b\") " pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.365347 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ae05919-68bf-43d1-abd9-9908ec287bd0-combined-ca-bundle\") pod \"cinder-db-sync-8l85z\" (UID: \"9ae05919-68bf-43d1-abd9-9908ec287bd0\") " pod="openstack/cinder-db-sync-8l85z" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.365385 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4tmg\" (UniqueName: \"kubernetes.io/projected/9ae05919-68bf-43d1-abd9-9908ec287bd0-kube-api-access-d4tmg\") pod \"cinder-db-sync-8l85z\" (UID: \"9ae05919-68bf-43d1-abd9-9908ec287bd0\") " pod="openstack/cinder-db-sync-8l85z" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.365425 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ae05919-68bf-43d1-abd9-9908ec287bd0-scripts\") pod \"cinder-db-sync-8l85z\" (UID: \"9ae05919-68bf-43d1-abd9-9908ec287bd0\") " pod="openstack/cinder-db-sync-8l85z" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.365461 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ae05919-68bf-43d1-abd9-9908ec287bd0-config-data\") pod \"cinder-db-sync-8l85z\" (UID: \"9ae05919-68bf-43d1-abd9-9908ec287bd0\") " pod="openstack/cinder-db-sync-8l85z" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.365490 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-scripts\") pod \"placement-db-sync-zf2tx\" (UID: \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\") " pod="openstack/placement-db-sync-zf2tx" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.365599 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9ae05919-68bf-43d1-abd9-9908ec287bd0-etc-machine-id\") pod \"cinder-db-sync-8l85z\" (UID: \"9ae05919-68bf-43d1-abd9-9908ec287bd0\") " pod="openstack/cinder-db-sync-8l85z" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.368512 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.372428 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ae05919-68bf-43d1-abd9-9908ec287bd0-combined-ca-bundle\") pod \"cinder-db-sync-8l85z\" (UID: \"9ae05919-68bf-43d1-abd9-9908ec287bd0\") " pod="openstack/cinder-db-sync-8l85z" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.372447 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9ae05919-68bf-43d1-abd9-9908ec287bd0-db-sync-config-data\") pod \"cinder-db-sync-8l85z\" (UID: \"9ae05919-68bf-43d1-abd9-9908ec287bd0\") " pod="openstack/cinder-db-sync-8l85z" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.386023 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ae05919-68bf-43d1-abd9-9908ec287bd0-config-data\") pod \"cinder-db-sync-8l85z\" (UID: \"9ae05919-68bf-43d1-abd9-9908ec287bd0\") " pod="openstack/cinder-db-sync-8l85z" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.386538 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ae05919-68bf-43d1-abd9-9908ec287bd0-scripts\") pod \"cinder-db-sync-8l85z\" (UID: \"9ae05919-68bf-43d1-abd9-9908ec287bd0\") " pod="openstack/cinder-db-sync-8l85z" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.392373 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4tmg\" (UniqueName: \"kubernetes.io/projected/9ae05919-68bf-43d1-abd9-9908ec287bd0-kube-api-access-d4tmg\") pod \"cinder-db-sync-8l85z\" (UID: \"9ae05919-68bf-43d1-abd9-9908ec287bd0\") " pod="openstack/cinder-db-sync-8l85z" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.441695 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pmk27" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.466884 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-scripts\") pod \"placement-db-sync-zf2tx\" (UID: \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\") " pod="openstack/placement-db-sync-zf2tx" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.466948 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmlv8\" (UniqueName: \"kubernetes.io/projected/e329efba-60e3-49c7-81ff-b073be77e34b-kube-api-access-zmlv8\") pod \"dnsmasq-dns-76fcf4b695-8d9vp\" (UID: \"e329efba-60e3-49c7-81ff-b073be77e34b\") " pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.466971 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-config-data\") pod \"placement-db-sync-zf2tx\" (UID: \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\") " pod="openstack/placement-db-sync-zf2tx" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.466988 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-logs\") pod \"placement-db-sync-zf2tx\" (UID: \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\") " pod="openstack/placement-db-sync-zf2tx" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.467036 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-8d9vp\" (UID: \"e329efba-60e3-49c7-81ff-b073be77e34b\") " pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.467056 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-config\") pod \"dnsmasq-dns-76fcf4b695-8d9vp\" (UID: \"e329efba-60e3-49c7-81ff-b073be77e34b\") " pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.467095 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-8d9vp\" (UID: \"e329efba-60e3-49c7-81ff-b073be77e34b\") " pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.467133 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8x5b\" (UniqueName: \"kubernetes.io/projected/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-kube-api-access-d8x5b\") pod \"placement-db-sync-zf2tx\" (UID: \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\") " pod="openstack/placement-db-sync-zf2tx" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.467151 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-combined-ca-bundle\") pod \"placement-db-sync-zf2tx\" (UID: \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\") " pod="openstack/placement-db-sync-zf2tx" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.467180 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-8d9vp\" (UID: \"e329efba-60e3-49c7-81ff-b073be77e34b\") " pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.467201 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-8d9vp\" (UID: \"e329efba-60e3-49c7-81ff-b073be77e34b\") " pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.467955 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-8d9vp\" (UID: \"e329efba-60e3-49c7-81ff-b073be77e34b\") " pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.469440 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-config\") pod \"dnsmasq-dns-76fcf4b695-8d9vp\" (UID: \"e329efba-60e3-49c7-81ff-b073be77e34b\") " pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.470607 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-logs\") pod \"placement-db-sync-zf2tx\" (UID: \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\") " pod="openstack/placement-db-sync-zf2tx" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.471204 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-8d9vp\" (UID: \"e329efba-60e3-49c7-81ff-b073be77e34b\") " pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.474622 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-8d9vp\" (UID: \"e329efba-60e3-49c7-81ff-b073be77e34b\") " pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.475693 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-scripts\") pod \"placement-db-sync-zf2tx\" (UID: \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\") " pod="openstack/placement-db-sync-zf2tx" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.476592 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-config-data\") pod \"placement-db-sync-zf2tx\" (UID: \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\") " pod="openstack/placement-db-sync-zf2tx" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.476750 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-combined-ca-bundle\") pod \"placement-db-sync-zf2tx\" (UID: \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\") " pod="openstack/placement-db-sync-zf2tx" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.477712 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-8d9vp\" (UID: \"e329efba-60e3-49c7-81ff-b073be77e34b\") " pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.488054 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmlv8\" (UniqueName: \"kubernetes.io/projected/e329efba-60e3-49c7-81ff-b073be77e34b-kube-api-access-zmlv8\") pod \"dnsmasq-dns-76fcf4b695-8d9vp\" (UID: \"e329efba-60e3-49c7-81ff-b073be77e34b\") " pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.491687 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8x5b\" (UniqueName: \"kubernetes.io/projected/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-kube-api-access-d8x5b\") pod \"placement-db-sync-zf2tx\" (UID: \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\") " pod="openstack/placement-db-sync-zf2tx" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.533331 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8l85z" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.557573 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zf2tx" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.573543 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.709982 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-x8hbr"] Jan 27 15:26:48 crc kubenswrapper[4772]: W0127 15:26:48.766573 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34ee743a_1628_42e4_a465_0e3957cae089.slice/crio-e4869b167a5dcad1e31d6b3e55e70bfbc70108cabaa63e1ee42f306d56b633d7 WatchSource:0}: Error finding container e4869b167a5dcad1e31d6b3e55e70bfbc70108cabaa63e1ee42f306d56b633d7: Status 404 returned error can't find the container with id e4869b167a5dcad1e31d6b3e55e70bfbc70108cabaa63e1ee42f306d56b633d7 Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.882645 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rl8kf"] Jan 27 15:26:48 crc kubenswrapper[4772]: I0127 15:26:48.914049 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-v689b"] Jan 27 15:26:49 crc kubenswrapper[4772]: I0127 15:26:49.045810 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:26:49 crc kubenswrapper[4772]: I0127 15:26:49.208055 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-pmk27"] Jan 27 15:26:49 crc kubenswrapper[4772]: W0127 15:26:49.213762 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a423229_06be_4934_9715_58105e1af686.slice/crio-604f080c5b545eb272e78d6599f0497ec22c32b54d41f8331dbefcd9a29b19de WatchSource:0}: Error finding container 604f080c5b545eb272e78d6599f0497ec22c32b54d41f8331dbefcd9a29b19de: Status 404 returned error can't find the container with id 604f080c5b545eb272e78d6599f0497ec22c32b54d41f8331dbefcd9a29b19de Jan 27 15:26:49 crc kubenswrapper[4772]: I0127 15:26:49.318960 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-8l85z"] Jan 27 15:26:49 crc kubenswrapper[4772]: I0127 15:26:49.380724 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-zf2tx"] Jan 27 15:26:49 crc kubenswrapper[4772]: W0127 15:26:49.385128 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode329efba_60e3_49c7_81ff_b073be77e34b.slice/crio-dc924b979634b0cd0c7264ffb70a5a244bf22da4a19a82562f283a10b69f4841 WatchSource:0}: Error finding container dc924b979634b0cd0c7264ffb70a5a244bf22da4a19a82562f283a10b69f4841: Status 404 returned error can't find the container with id dc924b979634b0cd0c7264ffb70a5a244bf22da4a19a82562f283a10b69f4841 Jan 27 15:26:49 crc kubenswrapper[4772]: I0127 15:26:49.389548 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-8d9vp"] Jan 27 15:26:49 crc kubenswrapper[4772]: I0127 15:26:49.490194 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de415c6e-4424-49c4-bc9d-076a5b13ab4e","Type":"ContainerStarted","Data":"a0254acb416eb806ca40cead3274ef3b55185c0cdbabec25da60a2a08040318a"} Jan 27 15:26:49 crc kubenswrapper[4772]: I0127 15:26:49.492706 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zf2tx" event={"ID":"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6","Type":"ContainerStarted","Data":"7dfb28db1e03cbbc36a413590b93de83567a6c9fa02be76267be1180098e9795"} Jan 27 15:26:49 crc kubenswrapper[4772]: I0127 15:26:49.494551 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-v689b" event={"ID":"b0625578-3b48-44c7-9082-174fce3a7e74","Type":"ContainerStarted","Data":"b3123ce803c91e7738d4af911f91769cd0703aad347549f4989b2ccc532f36ea"} Jan 27 15:26:49 crc kubenswrapper[4772]: I0127 15:26:49.495513 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8l85z" event={"ID":"9ae05919-68bf-43d1-abd9-9908ec287bd0","Type":"ContainerStarted","Data":"6c7bfeb67dfdf4e440bd40114d111aab9077e461f93cb6bdda5f337cad29c97d"} Jan 27 15:26:49 crc kubenswrapper[4772]: I0127 15:26:49.498318 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pmk27" event={"ID":"5a423229-06be-4934-9715-58105e1af686","Type":"ContainerStarted","Data":"604f080c5b545eb272e78d6599f0497ec22c32b54d41f8331dbefcd9a29b19de"} Jan 27 15:26:49 crc kubenswrapper[4772]: I0127 15:26:49.500149 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" event={"ID":"e329efba-60e3-49c7-81ff-b073be77e34b","Type":"ContainerStarted","Data":"dc924b979634b0cd0c7264ffb70a5a244bf22da4a19a82562f283a10b69f4841"} Jan 27 15:26:49 crc kubenswrapper[4772]: I0127 15:26:49.509778 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rl8kf" event={"ID":"0851ad59-841c-4133-a043-13d2cfdb0803","Type":"ContainerStarted","Data":"c20becc5003b571ca45e8d820a72a46ddfed0eee505f84347fd06aa34646e7c4"} Jan 27 15:26:49 crc kubenswrapper[4772]: I0127 15:26:49.511224 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" event={"ID":"34ee743a-1628-42e4-a465-0e3957cae089","Type":"ContainerStarted","Data":"e4869b167a5dcad1e31d6b3e55e70bfbc70108cabaa63e1ee42f306d56b633d7"} Jan 27 15:26:50 crc kubenswrapper[4772]: I0127 15:26:50.517909 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:26:50 crc kubenswrapper[4772]: I0127 15:26:50.523439 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rl8kf" event={"ID":"0851ad59-841c-4133-a043-13d2cfdb0803","Type":"ContainerStarted","Data":"c63a10e019701dbe41c4487398c76cb4acdd6a0eda99f6edb9df7d6273b71a27"} Jan 27 15:26:50 crc kubenswrapper[4772]: I0127 15:26:50.530555 4772 generic.go:334] "Generic (PLEG): container finished" podID="34ee743a-1628-42e4-a465-0e3957cae089" containerID="d73a0cc0e4bda3bf05abf4b518f94f7eadee3b5d18eccad0b9033c86816467f5" exitCode=0 Jan 27 15:26:50 crc kubenswrapper[4772]: I0127 15:26:50.530668 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" event={"ID":"34ee743a-1628-42e4-a465-0e3957cae089","Type":"ContainerDied","Data":"d73a0cc0e4bda3bf05abf4b518f94f7eadee3b5d18eccad0b9033c86816467f5"} Jan 27 15:26:50 crc kubenswrapper[4772]: I0127 15:26:50.534894 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-v689b" event={"ID":"b0625578-3b48-44c7-9082-174fce3a7e74","Type":"ContainerStarted","Data":"d2de8b3a1c27ebd01b5c3393c6dcb85d202fe549eef0c41d0f9f318c3b15d219"} Jan 27 15:26:50 crc kubenswrapper[4772]: I0127 15:26:50.536763 4772 generic.go:334] "Generic (PLEG): container finished" podID="e329efba-60e3-49c7-81ff-b073be77e34b" containerID="10196ecd014a671d2bb0c35a007cf89f6cc32f81e1a8290e8f8bb5f8f7575614" exitCode=0 Jan 27 15:26:50 crc kubenswrapper[4772]: I0127 15:26:50.536809 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" event={"ID":"e329efba-60e3-49c7-81ff-b073be77e34b","Type":"ContainerDied","Data":"10196ecd014a671d2bb0c35a007cf89f6cc32f81e1a8290e8f8bb5f8f7575614"} Jan 27 15:26:50 crc kubenswrapper[4772]: I0127 15:26:50.583440 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-rl8kf" podStartSLOduration=3.583396222 podStartE2EDuration="3.583396222s" podCreationTimestamp="2026-01-27 15:26:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:26:50.540015872 +0000 UTC m=+1196.520624970" watchObservedRunningTime="2026-01-27 15:26:50.583396222 +0000 UTC m=+1196.564005330" Jan 27 15:26:50 crc kubenswrapper[4772]: I0127 15:26:50.598509 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-v689b" podStartSLOduration=3.598487167 podStartE2EDuration="3.598487167s" podCreationTimestamp="2026-01-27 15:26:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:26:50.554085038 +0000 UTC m=+1196.534694136" watchObservedRunningTime="2026-01-27 15:26:50.598487167 +0000 UTC m=+1196.579096265" Jan 27 15:26:50 crc kubenswrapper[4772]: I0127 15:26:50.955266 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.133614 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-ovsdbserver-sb\") pod \"34ee743a-1628-42e4-a465-0e3957cae089\" (UID: \"34ee743a-1628-42e4-a465-0e3957cae089\") " Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.133696 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-dns-svc\") pod \"34ee743a-1628-42e4-a465-0e3957cae089\" (UID: \"34ee743a-1628-42e4-a465-0e3957cae089\") " Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.133754 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tx2xr\" (UniqueName: \"kubernetes.io/projected/34ee743a-1628-42e4-a465-0e3957cae089-kube-api-access-tx2xr\") pod \"34ee743a-1628-42e4-a465-0e3957cae089\" (UID: \"34ee743a-1628-42e4-a465-0e3957cae089\") " Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.133986 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-ovsdbserver-nb\") pod \"34ee743a-1628-42e4-a465-0e3957cae089\" (UID: \"34ee743a-1628-42e4-a465-0e3957cae089\") " Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.134017 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-config\") pod \"34ee743a-1628-42e4-a465-0e3957cae089\" (UID: \"34ee743a-1628-42e4-a465-0e3957cae089\") " Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.134042 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-dns-swift-storage-0\") pod \"34ee743a-1628-42e4-a465-0e3957cae089\" (UID: \"34ee743a-1628-42e4-a465-0e3957cae089\") " Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.153357 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34ee743a-1628-42e4-a465-0e3957cae089-kube-api-access-tx2xr" (OuterVolumeSpecName: "kube-api-access-tx2xr") pod "34ee743a-1628-42e4-a465-0e3957cae089" (UID: "34ee743a-1628-42e4-a465-0e3957cae089"). InnerVolumeSpecName "kube-api-access-tx2xr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.175290 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "34ee743a-1628-42e4-a465-0e3957cae089" (UID: "34ee743a-1628-42e4-a465-0e3957cae089"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.182600 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-config" (OuterVolumeSpecName: "config") pod "34ee743a-1628-42e4-a465-0e3957cae089" (UID: "34ee743a-1628-42e4-a465-0e3957cae089"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.186546 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "34ee743a-1628-42e4-a465-0e3957cae089" (UID: "34ee743a-1628-42e4-a465-0e3957cae089"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.191810 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "34ee743a-1628-42e4-a465-0e3957cae089" (UID: "34ee743a-1628-42e4-a465-0e3957cae089"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.236436 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.236467 4772 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.236477 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.236485 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.236493 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tx2xr\" (UniqueName: \"kubernetes.io/projected/34ee743a-1628-42e4-a465-0e3957cae089-kube-api-access-tx2xr\") on node \"crc\" DevicePath \"\"" Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.270155 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "34ee743a-1628-42e4-a465-0e3957cae089" (UID: "34ee743a-1628-42e4-a465-0e3957cae089"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.338022 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/34ee743a-1628-42e4-a465-0e3957cae089-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.568529 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" event={"ID":"34ee743a-1628-42e4-a465-0e3957cae089","Type":"ContainerDied","Data":"e4869b167a5dcad1e31d6b3e55e70bfbc70108cabaa63e1ee42f306d56b633d7"} Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.568541 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-x8hbr" Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.568578 4772 scope.go:117] "RemoveContainer" containerID="d73a0cc0e4bda3bf05abf4b518f94f7eadee3b5d18eccad0b9033c86816467f5" Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.572018 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" event={"ID":"e329efba-60e3-49c7-81ff-b073be77e34b","Type":"ContainerStarted","Data":"0bc2eb78a83f1e9ddf6e0c975669640d497ff3e50951b0eaadaee82dc03caffd"} Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.572224 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.599069 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" podStartSLOduration=3.599051722 podStartE2EDuration="3.599051722s" podCreationTimestamp="2026-01-27 15:26:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:26:51.592971947 +0000 UTC m=+1197.573581045" watchObservedRunningTime="2026-01-27 15:26:51.599051722 +0000 UTC m=+1197.579660820" Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.646414 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-x8hbr"] Jan 27 15:26:51 crc kubenswrapper[4772]: I0127 15:26:51.659768 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-x8hbr"] Jan 27 15:26:52 crc kubenswrapper[4772]: I0127 15:26:52.680528 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34ee743a-1628-42e4-a465-0e3957cae089" path="/var/lib/kubelet/pods/34ee743a-1628-42e4-a465-0e3957cae089/volumes" Jan 27 15:26:53 crc kubenswrapper[4772]: I0127 15:26:53.593806 4772 generic.go:334] "Generic (PLEG): container finished" podID="86d0241f-ae16-400f-837c-3b43c904c91e" containerID="317ff691da5e191e31778e1d02f29484703e057687e372739fcbc9dd6f8088d2" exitCode=0 Jan 27 15:26:53 crc kubenswrapper[4772]: I0127 15:26:53.593866 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-vdmv7" event={"ID":"86d0241f-ae16-400f-837c-3b43c904c91e","Type":"ContainerDied","Data":"317ff691da5e191e31778e1d02f29484703e057687e372739fcbc9dd6f8088d2"} Jan 27 15:26:54 crc kubenswrapper[4772]: I0127 15:26:54.604809 4772 generic.go:334] "Generic (PLEG): container finished" podID="0851ad59-841c-4133-a043-13d2cfdb0803" containerID="c63a10e019701dbe41c4487398c76cb4acdd6a0eda99f6edb9df7d6273b71a27" exitCode=0 Jan 27 15:26:54 crc kubenswrapper[4772]: I0127 15:26:54.605280 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rl8kf" event={"ID":"0851ad59-841c-4133-a043-13d2cfdb0803","Type":"ContainerDied","Data":"c63a10e019701dbe41c4487398c76cb4acdd6a0eda99f6edb9df7d6273b71a27"} Jan 27 15:26:58 crc kubenswrapper[4772]: I0127 15:26:58.575351 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" Jan 27 15:26:58 crc kubenswrapper[4772]: I0127 15:26:58.677717 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-lll86"] Jan 27 15:26:58 crc kubenswrapper[4772]: I0127 15:26:58.678160 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-lll86" podUID="b6aa637d-4418-4fa4-8a26-249446d2fb3f" containerName="dnsmasq-dns" containerID="cri-o://ded8f7e741d736bdfe8cef79d54407ecbfa8926bb6d56e27836f39ea6ec4c8ef" gracePeriod=10 Jan 27 15:26:59 crc kubenswrapper[4772]: I0127 15:26:59.325550 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-lll86" podUID="b6aa637d-4418-4fa4-8a26-249446d2fb3f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.120:5353: connect: connection refused" Jan 27 15:26:59 crc kubenswrapper[4772]: I0127 15:26:59.678062 4772 generic.go:334] "Generic (PLEG): container finished" podID="b6aa637d-4418-4fa4-8a26-249446d2fb3f" containerID="ded8f7e741d736bdfe8cef79d54407ecbfa8926bb6d56e27836f39ea6ec4c8ef" exitCode=0 Jan 27 15:26:59 crc kubenswrapper[4772]: I0127 15:26:59.678119 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-lll86" event={"ID":"b6aa637d-4418-4fa4-8a26-249446d2fb3f","Type":"ContainerDied","Data":"ded8f7e741d736bdfe8cef79d54407ecbfa8926bb6d56e27836f39ea6ec4c8ef"} Jan 27 15:27:01 crc kubenswrapper[4772]: I0127 15:27:01.631856 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-vdmv7" Jan 27 15:27:01 crc kubenswrapper[4772]: I0127 15:27:01.698999 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-vdmv7" event={"ID":"86d0241f-ae16-400f-837c-3b43c904c91e","Type":"ContainerDied","Data":"a28e441b5d88b4432b8107753f5714c4987db6ac6635bc08e9f396a0b42288ec"} Jan 27 15:27:01 crc kubenswrapper[4772]: I0127 15:27:01.699038 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a28e441b5d88b4432b8107753f5714c4987db6ac6635bc08e9f396a0b42288ec" Jan 27 15:27:01 crc kubenswrapper[4772]: I0127 15:27:01.699118 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-vdmv7" Jan 27 15:27:01 crc kubenswrapper[4772]: I0127 15:27:01.762192 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d0241f-ae16-400f-837c-3b43c904c91e-combined-ca-bundle\") pod \"86d0241f-ae16-400f-837c-3b43c904c91e\" (UID: \"86d0241f-ae16-400f-837c-3b43c904c91e\") " Jan 27 15:27:01 crc kubenswrapper[4772]: I0127 15:27:01.762392 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d0241f-ae16-400f-837c-3b43c904c91e-config-data\") pod \"86d0241f-ae16-400f-837c-3b43c904c91e\" (UID: \"86d0241f-ae16-400f-837c-3b43c904c91e\") " Jan 27 15:27:01 crc kubenswrapper[4772]: I0127 15:27:01.762454 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkbkr\" (UniqueName: \"kubernetes.io/projected/86d0241f-ae16-400f-837c-3b43c904c91e-kube-api-access-kkbkr\") pod \"86d0241f-ae16-400f-837c-3b43c904c91e\" (UID: \"86d0241f-ae16-400f-837c-3b43c904c91e\") " Jan 27 15:27:01 crc kubenswrapper[4772]: I0127 15:27:01.762687 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/86d0241f-ae16-400f-837c-3b43c904c91e-db-sync-config-data\") pod \"86d0241f-ae16-400f-837c-3b43c904c91e\" (UID: \"86d0241f-ae16-400f-837c-3b43c904c91e\") " Jan 27 15:27:01 crc kubenswrapper[4772]: I0127 15:27:01.769292 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86d0241f-ae16-400f-837c-3b43c904c91e-kube-api-access-kkbkr" (OuterVolumeSpecName: "kube-api-access-kkbkr") pod "86d0241f-ae16-400f-837c-3b43c904c91e" (UID: "86d0241f-ae16-400f-837c-3b43c904c91e"). InnerVolumeSpecName "kube-api-access-kkbkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:27:01 crc kubenswrapper[4772]: I0127 15:27:01.770051 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86d0241f-ae16-400f-837c-3b43c904c91e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "86d0241f-ae16-400f-837c-3b43c904c91e" (UID: "86d0241f-ae16-400f-837c-3b43c904c91e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:01 crc kubenswrapper[4772]: I0127 15:27:01.794875 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86d0241f-ae16-400f-837c-3b43c904c91e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "86d0241f-ae16-400f-837c-3b43c904c91e" (UID: "86d0241f-ae16-400f-837c-3b43c904c91e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:01 crc kubenswrapper[4772]: I0127 15:27:01.816271 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86d0241f-ae16-400f-837c-3b43c904c91e-config-data" (OuterVolumeSpecName: "config-data") pod "86d0241f-ae16-400f-837c-3b43c904c91e" (UID: "86d0241f-ae16-400f-837c-3b43c904c91e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:01 crc kubenswrapper[4772]: I0127 15:27:01.868407 4772 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/86d0241f-ae16-400f-837c-3b43c904c91e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:01 crc kubenswrapper[4772]: I0127 15:27:01.870806 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d0241f-ae16-400f-837c-3b43c904c91e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:01 crc kubenswrapper[4772]: I0127 15:27:01.870936 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d0241f-ae16-400f-837c-3b43c904c91e-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:01 crc kubenswrapper[4772]: I0127 15:27:01.870998 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkbkr\" (UniqueName: \"kubernetes.io/projected/86d0241f-ae16-400f-837c-3b43c904c91e-kube-api-access-kkbkr\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.149976 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-dqgvx"] Jan 27 15:27:03 crc kubenswrapper[4772]: E0127 15:27:03.150398 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34ee743a-1628-42e4-a465-0e3957cae089" containerName="init" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.150413 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="34ee743a-1628-42e4-a465-0e3957cae089" containerName="init" Jan 27 15:27:03 crc kubenswrapper[4772]: E0127 15:27:03.150440 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86d0241f-ae16-400f-837c-3b43c904c91e" containerName="glance-db-sync" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.150448 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="86d0241f-ae16-400f-837c-3b43c904c91e" containerName="glance-db-sync" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.150630 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="86d0241f-ae16-400f-837c-3b43c904c91e" containerName="glance-db-sync" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.150649 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="34ee743a-1628-42e4-a465-0e3957cae089" containerName="init" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.153683 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.190280 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-dqgvx"] Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.211135 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-dqgvx\" (UID: \"17a547a9-a098-43b7-a153-ad9a137369de\") " pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.211312 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzhkt\" (UniqueName: \"kubernetes.io/projected/17a547a9-a098-43b7-a153-ad9a137369de-kube-api-access-xzhkt\") pod \"dnsmasq-dns-8b5c85b87-dqgvx\" (UID: \"17a547a9-a098-43b7-a153-ad9a137369de\") " pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.211368 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-dqgvx\" (UID: \"17a547a9-a098-43b7-a153-ad9a137369de\") " pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.211393 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-config\") pod \"dnsmasq-dns-8b5c85b87-dqgvx\" (UID: \"17a547a9-a098-43b7-a153-ad9a137369de\") " pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.211431 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-dqgvx\" (UID: \"17a547a9-a098-43b7-a153-ad9a137369de\") " pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.211462 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-dqgvx\" (UID: \"17a547a9-a098-43b7-a153-ad9a137369de\") " pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.313669 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-dqgvx\" (UID: \"17a547a9-a098-43b7-a153-ad9a137369de\") " pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.313783 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzhkt\" (UniqueName: \"kubernetes.io/projected/17a547a9-a098-43b7-a153-ad9a137369de-kube-api-access-xzhkt\") pod \"dnsmasq-dns-8b5c85b87-dqgvx\" (UID: \"17a547a9-a098-43b7-a153-ad9a137369de\") " pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.313813 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-dqgvx\" (UID: \"17a547a9-a098-43b7-a153-ad9a137369de\") " pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.313831 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-config\") pod \"dnsmasq-dns-8b5c85b87-dqgvx\" (UID: \"17a547a9-a098-43b7-a153-ad9a137369de\") " pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.313854 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-dqgvx\" (UID: \"17a547a9-a098-43b7-a153-ad9a137369de\") " pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.313874 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-dqgvx\" (UID: \"17a547a9-a098-43b7-a153-ad9a137369de\") " pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.314633 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-dqgvx\" (UID: \"17a547a9-a098-43b7-a153-ad9a137369de\") " pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.314658 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-dqgvx\" (UID: \"17a547a9-a098-43b7-a153-ad9a137369de\") " pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.314901 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-config\") pod \"dnsmasq-dns-8b5c85b87-dqgvx\" (UID: \"17a547a9-a098-43b7-a153-ad9a137369de\") " pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.314911 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-dqgvx\" (UID: \"17a547a9-a098-43b7-a153-ad9a137369de\") " pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.315214 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-dqgvx\" (UID: \"17a547a9-a098-43b7-a153-ad9a137369de\") " pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.335719 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzhkt\" (UniqueName: \"kubernetes.io/projected/17a547a9-a098-43b7-a153-ad9a137369de-kube-api-access-xzhkt\") pod \"dnsmasq-dns-8b5c85b87-dqgvx\" (UID: \"17a547a9-a098-43b7-a153-ad9a137369de\") " pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.476919 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.987208 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.990363 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.992882 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.992956 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.994071 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-vd4fn" Jan 27 15:27:03 crc kubenswrapper[4772]: I0127 15:27:03.998542 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.025044 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42e34d70-0be1-400d-b214-62ba7d9e2e09-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.025134 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wphqd\" (UniqueName: \"kubernetes.io/projected/42e34d70-0be1-400d-b214-62ba7d9e2e09-kube-api-access-wphqd\") pod \"glance-default-external-api-0\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.025186 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.025227 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42e34d70-0be1-400d-b214-62ba7d9e2e09-scripts\") pod \"glance-default-external-api-0\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.025252 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/42e34d70-0be1-400d-b214-62ba7d9e2e09-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.025337 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42e34d70-0be1-400d-b214-62ba7d9e2e09-logs\") pod \"glance-default-external-api-0\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.025388 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42e34d70-0be1-400d-b214-62ba7d9e2e09-config-data\") pod \"glance-default-external-api-0\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.126513 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42e34d70-0be1-400d-b214-62ba7d9e2e09-logs\") pod \"glance-default-external-api-0\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.126582 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42e34d70-0be1-400d-b214-62ba7d9e2e09-config-data\") pod \"glance-default-external-api-0\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.126641 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42e34d70-0be1-400d-b214-62ba7d9e2e09-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.126700 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wphqd\" (UniqueName: \"kubernetes.io/projected/42e34d70-0be1-400d-b214-62ba7d9e2e09-kube-api-access-wphqd\") pod \"glance-default-external-api-0\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.126729 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.126768 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42e34d70-0be1-400d-b214-62ba7d9e2e09-scripts\") pod \"glance-default-external-api-0\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.126791 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/42e34d70-0be1-400d-b214-62ba7d9e2e09-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.127263 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42e34d70-0be1-400d-b214-62ba7d9e2e09-logs\") pod \"glance-default-external-api-0\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.127320 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/42e34d70-0be1-400d-b214-62ba7d9e2e09-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.127683 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.133230 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42e34d70-0be1-400d-b214-62ba7d9e2e09-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.141810 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42e34d70-0be1-400d-b214-62ba7d9e2e09-scripts\") pod \"glance-default-external-api-0\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.143033 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42e34d70-0be1-400d-b214-62ba7d9e2e09-config-data\") pod \"glance-default-external-api-0\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.146015 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wphqd\" (UniqueName: \"kubernetes.io/projected/42e34d70-0be1-400d-b214-62ba7d9e2e09-kube-api-access-wphqd\") pod \"glance-default-external-api-0\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.168413 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.315519 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.401453 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.403935 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.407004 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.415362 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.432700 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.432759 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.432780 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.432805 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-logs\") pod \"glance-default-internal-api-0\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.432879 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.432900 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.432919 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rkbs\" (UniqueName: \"kubernetes.io/projected/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-kube-api-access-6rkbs\") pod \"glance-default-internal-api-0\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.534283 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.534338 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.534372 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.534423 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-logs\") pod \"glance-default-internal-api-0\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.534513 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.534523 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.535026 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.535047 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rkbs\" (UniqueName: \"kubernetes.io/projected/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-kube-api-access-6rkbs\") pod \"glance-default-internal-api-0\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.535442 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-logs\") pod \"glance-default-internal-api-0\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.535464 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.541716 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.544571 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.544571 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.550827 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rkbs\" (UniqueName: \"kubernetes.io/projected/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-kube-api-access-6rkbs\") pod \"glance-default-internal-api-0\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.558863 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:04 crc kubenswrapper[4772]: I0127 15:27:04.732146 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 27 15:27:05 crc kubenswrapper[4772]: I0127 15:27:05.557405 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 15:27:05 crc kubenswrapper[4772]: I0127 15:27:05.637395 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 15:27:09 crc kubenswrapper[4772]: I0127 15:27:09.325746 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-lll86" podUID="b6aa637d-4418-4fa4-8a26-249446d2fb3f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.120:5353: i/o timeout" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.192996 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rl8kf" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.200024 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-lll86" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.298021 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-fernet-keys\") pod \"0851ad59-841c-4133-a043-13d2cfdb0803\" (UID: \"0851ad59-841c-4133-a043-13d2cfdb0803\") " Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.298081 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-combined-ca-bundle\") pod \"0851ad59-841c-4133-a043-13d2cfdb0803\" (UID: \"0851ad59-841c-4133-a043-13d2cfdb0803\") " Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.298242 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-config-data\") pod \"0851ad59-841c-4133-a043-13d2cfdb0803\" (UID: \"0851ad59-841c-4133-a043-13d2cfdb0803\") " Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.298283 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtsjh\" (UniqueName: \"kubernetes.io/projected/b6aa637d-4418-4fa4-8a26-249446d2fb3f-kube-api-access-gtsjh\") pod \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\" (UID: \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\") " Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.298335 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwktg\" (UniqueName: \"kubernetes.io/projected/0851ad59-841c-4133-a043-13d2cfdb0803-kube-api-access-fwktg\") pod \"0851ad59-841c-4133-a043-13d2cfdb0803\" (UID: \"0851ad59-841c-4133-a043-13d2cfdb0803\") " Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.298363 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-ovsdbserver-nb\") pod \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\" (UID: \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\") " Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.298403 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-credential-keys\") pod \"0851ad59-841c-4133-a043-13d2cfdb0803\" (UID: \"0851ad59-841c-4133-a043-13d2cfdb0803\") " Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.298438 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-dns-swift-storage-0\") pod \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\" (UID: \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\") " Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.298464 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-ovsdbserver-sb\") pod \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\" (UID: \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\") " Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.298527 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-dns-svc\") pod \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\" (UID: \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\") " Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.298575 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-config\") pod \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\" (UID: \"b6aa637d-4418-4fa4-8a26-249446d2fb3f\") " Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.298600 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-scripts\") pod \"0851ad59-841c-4133-a043-13d2cfdb0803\" (UID: \"0851ad59-841c-4133-a043-13d2cfdb0803\") " Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.304618 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6aa637d-4418-4fa4-8a26-249446d2fb3f-kube-api-access-gtsjh" (OuterVolumeSpecName: "kube-api-access-gtsjh") pod "b6aa637d-4418-4fa4-8a26-249446d2fb3f" (UID: "b6aa637d-4418-4fa4-8a26-249446d2fb3f"). InnerVolumeSpecName "kube-api-access-gtsjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.305912 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "0851ad59-841c-4133-a043-13d2cfdb0803" (UID: "0851ad59-841c-4133-a043-13d2cfdb0803"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.306794 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-scripts" (OuterVolumeSpecName: "scripts") pod "0851ad59-841c-4133-a043-13d2cfdb0803" (UID: "0851ad59-841c-4133-a043-13d2cfdb0803"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.307925 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0851ad59-841c-4133-a043-13d2cfdb0803-kube-api-access-fwktg" (OuterVolumeSpecName: "kube-api-access-fwktg") pod "0851ad59-841c-4133-a043-13d2cfdb0803" (UID: "0851ad59-841c-4133-a043-13d2cfdb0803"). InnerVolumeSpecName "kube-api-access-fwktg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.309450 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "0851ad59-841c-4133-a043-13d2cfdb0803" (UID: "0851ad59-841c-4133-a043-13d2cfdb0803"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.336610 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0851ad59-841c-4133-a043-13d2cfdb0803" (UID: "0851ad59-841c-4133-a043-13d2cfdb0803"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.340435 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-config-data" (OuterVolumeSpecName: "config-data") pod "0851ad59-841c-4133-a043-13d2cfdb0803" (UID: "0851ad59-841c-4133-a043-13d2cfdb0803"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.350444 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b6aa637d-4418-4fa4-8a26-249446d2fb3f" (UID: "b6aa637d-4418-4fa4-8a26-249446d2fb3f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.353026 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-config" (OuterVolumeSpecName: "config") pod "b6aa637d-4418-4fa4-8a26-249446d2fb3f" (UID: "b6aa637d-4418-4fa4-8a26-249446d2fb3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.355265 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b6aa637d-4418-4fa4-8a26-249446d2fb3f" (UID: "b6aa637d-4418-4fa4-8a26-249446d2fb3f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.357334 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b6aa637d-4418-4fa4-8a26-249446d2fb3f" (UID: "b6aa637d-4418-4fa4-8a26-249446d2fb3f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.361342 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b6aa637d-4418-4fa4-8a26-249446d2fb3f" (UID: "b6aa637d-4418-4fa4-8a26-249446d2fb3f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.400582 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.400612 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwktg\" (UniqueName: \"kubernetes.io/projected/0851ad59-841c-4133-a043-13d2cfdb0803-kube-api-access-fwktg\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.400662 4772 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.400675 4772 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.400686 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.400699 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.400709 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6aa637d-4418-4fa4-8a26-249446d2fb3f-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.400719 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.400727 4772 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.400735 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.400745 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0851ad59-841c-4133-a043-13d2cfdb0803-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.400754 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtsjh\" (UniqueName: \"kubernetes.io/projected/b6aa637d-4418-4fa4-8a26-249446d2fb3f-kube-api-access-gtsjh\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:13 crc kubenswrapper[4772]: E0127 15:27:13.670892 4772 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Jan 27 15:27:13 crc kubenswrapper[4772]: E0127 15:27:13.671110 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8cvd2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-pmk27_openstack(5a423229-06be-4934-9715-58105e1af686): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 27 15:27:13 crc kubenswrapper[4772]: E0127 15:27:13.672281 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-pmk27" podUID="5a423229-06be-4934-9715-58105e1af686" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.815019 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rl8kf" event={"ID":"0851ad59-841c-4133-a043-13d2cfdb0803","Type":"ContainerDied","Data":"c20becc5003b571ca45e8d820a72a46ddfed0eee505f84347fd06aa34646e7c4"} Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.815097 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rl8kf" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.815154 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c20becc5003b571ca45e8d820a72a46ddfed0eee505f84347fd06aa34646e7c4" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.818335 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-lll86" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.818344 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-lll86" event={"ID":"b6aa637d-4418-4fa4-8a26-249446d2fb3f","Type":"ContainerDied","Data":"353300bf1914ec8c1fafaa4dfe7633842f95697653e6f9ec7954d70422c9cfbd"} Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.818442 4772 scope.go:117] "RemoveContainer" containerID="ded8f7e741d736bdfe8cef79d54407ecbfa8926bb6d56e27836f39ea6ec4c8ef" Jan 27 15:27:13 crc kubenswrapper[4772]: E0127 15:27:13.819580 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-pmk27" podUID="5a423229-06be-4934-9715-58105e1af686" Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.876817 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-lll86"] Jan 27 15:27:13 crc kubenswrapper[4772]: I0127 15:27:13.884542 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-lll86"] Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.281613 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-rl8kf"] Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.291277 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-rl8kf"] Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.326812 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-lll86" podUID="b6aa637d-4418-4fa4-8a26-249446d2fb3f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.120:5353: i/o timeout" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.388145 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-fl4nt"] Jan 27 15:27:14 crc kubenswrapper[4772]: E0127 15:27:14.388567 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6aa637d-4418-4fa4-8a26-249446d2fb3f" containerName="dnsmasq-dns" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.388588 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6aa637d-4418-4fa4-8a26-249446d2fb3f" containerName="dnsmasq-dns" Jan 27 15:27:14 crc kubenswrapper[4772]: E0127 15:27:14.388612 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0851ad59-841c-4133-a043-13d2cfdb0803" containerName="keystone-bootstrap" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.388621 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0851ad59-841c-4133-a043-13d2cfdb0803" containerName="keystone-bootstrap" Jan 27 15:27:14 crc kubenswrapper[4772]: E0127 15:27:14.388634 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6aa637d-4418-4fa4-8a26-249446d2fb3f" containerName="init" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.388644 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6aa637d-4418-4fa4-8a26-249446d2fb3f" containerName="init" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.388889 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6aa637d-4418-4fa4-8a26-249446d2fb3f" containerName="dnsmasq-dns" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.388911 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="0851ad59-841c-4133-a043-13d2cfdb0803" containerName="keystone-bootstrap" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.389573 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fl4nt" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.392077 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-bdjsw" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.392313 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.392493 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.392898 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.393078 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.398252 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fl4nt"] Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.534871 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-config-data\") pod \"keystone-bootstrap-fl4nt\" (UID: \"8322baad-60c1-4d0b-96e3-51038f2e447a\") " pod="openstack/keystone-bootstrap-fl4nt" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.535329 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-968vm\" (UniqueName: \"kubernetes.io/projected/8322baad-60c1-4d0b-96e3-51038f2e447a-kube-api-access-968vm\") pod \"keystone-bootstrap-fl4nt\" (UID: \"8322baad-60c1-4d0b-96e3-51038f2e447a\") " pod="openstack/keystone-bootstrap-fl4nt" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.535408 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-fernet-keys\") pod \"keystone-bootstrap-fl4nt\" (UID: \"8322baad-60c1-4d0b-96e3-51038f2e447a\") " pod="openstack/keystone-bootstrap-fl4nt" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.535438 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-scripts\") pod \"keystone-bootstrap-fl4nt\" (UID: \"8322baad-60c1-4d0b-96e3-51038f2e447a\") " pod="openstack/keystone-bootstrap-fl4nt" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.535464 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-combined-ca-bundle\") pod \"keystone-bootstrap-fl4nt\" (UID: \"8322baad-60c1-4d0b-96e3-51038f2e447a\") " pod="openstack/keystone-bootstrap-fl4nt" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.535517 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-credential-keys\") pod \"keystone-bootstrap-fl4nt\" (UID: \"8322baad-60c1-4d0b-96e3-51038f2e447a\") " pod="openstack/keystone-bootstrap-fl4nt" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.637706 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-config-data\") pod \"keystone-bootstrap-fl4nt\" (UID: \"8322baad-60c1-4d0b-96e3-51038f2e447a\") " pod="openstack/keystone-bootstrap-fl4nt" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.637816 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-968vm\" (UniqueName: \"kubernetes.io/projected/8322baad-60c1-4d0b-96e3-51038f2e447a-kube-api-access-968vm\") pod \"keystone-bootstrap-fl4nt\" (UID: \"8322baad-60c1-4d0b-96e3-51038f2e447a\") " pod="openstack/keystone-bootstrap-fl4nt" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.637900 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-fernet-keys\") pod \"keystone-bootstrap-fl4nt\" (UID: \"8322baad-60c1-4d0b-96e3-51038f2e447a\") " pod="openstack/keystone-bootstrap-fl4nt" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.637929 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-scripts\") pod \"keystone-bootstrap-fl4nt\" (UID: \"8322baad-60c1-4d0b-96e3-51038f2e447a\") " pod="openstack/keystone-bootstrap-fl4nt" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.637954 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-combined-ca-bundle\") pod \"keystone-bootstrap-fl4nt\" (UID: \"8322baad-60c1-4d0b-96e3-51038f2e447a\") " pod="openstack/keystone-bootstrap-fl4nt" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.638014 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-credential-keys\") pod \"keystone-bootstrap-fl4nt\" (UID: \"8322baad-60c1-4d0b-96e3-51038f2e447a\") " pod="openstack/keystone-bootstrap-fl4nt" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.641946 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-config-data\") pod \"keystone-bootstrap-fl4nt\" (UID: \"8322baad-60c1-4d0b-96e3-51038f2e447a\") " pod="openstack/keystone-bootstrap-fl4nt" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.642141 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-fernet-keys\") pod \"keystone-bootstrap-fl4nt\" (UID: \"8322baad-60c1-4d0b-96e3-51038f2e447a\") " pod="openstack/keystone-bootstrap-fl4nt" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.642835 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-combined-ca-bundle\") pod \"keystone-bootstrap-fl4nt\" (UID: \"8322baad-60c1-4d0b-96e3-51038f2e447a\") " pod="openstack/keystone-bootstrap-fl4nt" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.643113 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-credential-keys\") pod \"keystone-bootstrap-fl4nt\" (UID: \"8322baad-60c1-4d0b-96e3-51038f2e447a\") " pod="openstack/keystone-bootstrap-fl4nt" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.652308 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-scripts\") pod \"keystone-bootstrap-fl4nt\" (UID: \"8322baad-60c1-4d0b-96e3-51038f2e447a\") " pod="openstack/keystone-bootstrap-fl4nt" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.655857 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-968vm\" (UniqueName: \"kubernetes.io/projected/8322baad-60c1-4d0b-96e3-51038f2e447a-kube-api-access-968vm\") pod \"keystone-bootstrap-fl4nt\" (UID: \"8322baad-60c1-4d0b-96e3-51038f2e447a\") " pod="openstack/keystone-bootstrap-fl4nt" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.673596 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0851ad59-841c-4133-a043-13d2cfdb0803" path="/var/lib/kubelet/pods/0851ad59-841c-4133-a043-13d2cfdb0803/volumes" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.674339 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6aa637d-4418-4fa4-8a26-249446d2fb3f" path="/var/lib/kubelet/pods/b6aa637d-4418-4fa4-8a26-249446d2fb3f/volumes" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.749079 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fl4nt" Jan 27 15:27:14 crc kubenswrapper[4772]: I0127 15:27:14.883051 4772 scope.go:117] "RemoveContainer" containerID="26a0610819b472b19e1babe3f9b5893ac7bd92b0c9047d536f0dadb42db99a12" Jan 27 15:27:14 crc kubenswrapper[4772]: E0127 15:27:14.937626 4772 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Jan 27 15:27:14 crc kubenswrapper[4772]: E0127 15:27:14.937873 4772 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d4tmg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-8l85z_openstack(9ae05919-68bf-43d1-abd9-9908ec287bd0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 27 15:27:14 crc kubenswrapper[4772]: E0127 15:27:14.943014 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-8l85z" podUID="9ae05919-68bf-43d1-abd9-9908ec287bd0" Jan 27 15:27:15 crc kubenswrapper[4772]: I0127 15:27:15.367950 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-dqgvx"] Jan 27 15:27:15 crc kubenswrapper[4772]: W0127 15:27:15.374800 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17a547a9_a098_43b7_a153_ad9a137369de.slice/crio-31e83f6ba26ca249b5435d61c8786bdc24b0777adb10cbb234cdaacbda3e0db7 WatchSource:0}: Error finding container 31e83f6ba26ca249b5435d61c8786bdc24b0777adb10cbb234cdaacbda3e0db7: Status 404 returned error can't find the container with id 31e83f6ba26ca249b5435d61c8786bdc24b0777adb10cbb234cdaacbda3e0db7 Jan 27 15:27:15 crc kubenswrapper[4772]: I0127 15:27:15.501618 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fl4nt"] Jan 27 15:27:15 crc kubenswrapper[4772]: W0127 15:27:15.505752 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8322baad_60c1_4d0b_96e3_51038f2e447a.slice/crio-00f64e139d84dd0a1b89ff44770726f3c1e45680dfa15727e0c2287a16f89b5d WatchSource:0}: Error finding container 00f64e139d84dd0a1b89ff44770726f3c1e45680dfa15727e0c2287a16f89b5d: Status 404 returned error can't find the container with id 00f64e139d84dd0a1b89ff44770726f3c1e45680dfa15727e0c2287a16f89b5d Jan 27 15:27:15 crc kubenswrapper[4772]: I0127 15:27:15.851966 4772 generic.go:334] "Generic (PLEG): container finished" podID="17a547a9-a098-43b7-a153-ad9a137369de" containerID="dbf85c4247c8ba3d5d079a5efed1ba22279b244d404fd6795f678db691f0b023" exitCode=0 Jan 27 15:27:15 crc kubenswrapper[4772]: I0127 15:27:15.852566 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" event={"ID":"17a547a9-a098-43b7-a153-ad9a137369de","Type":"ContainerDied","Data":"dbf85c4247c8ba3d5d079a5efed1ba22279b244d404fd6795f678db691f0b023"} Jan 27 15:27:15 crc kubenswrapper[4772]: I0127 15:27:15.852607 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" event={"ID":"17a547a9-a098-43b7-a153-ad9a137369de","Type":"ContainerStarted","Data":"31e83f6ba26ca249b5435d61c8786bdc24b0777adb10cbb234cdaacbda3e0db7"} Jan 27 15:27:15 crc kubenswrapper[4772]: I0127 15:27:15.858714 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 15:27:15 crc kubenswrapper[4772]: I0127 15:27:15.861147 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de415c6e-4424-49c4-bc9d-076a5b13ab4e","Type":"ContainerStarted","Data":"f76b5eae8b9d1fd746edffe9a9f5a02ca0ad4ea09665e63c5dbeacff4753fa40"} Jan 27 15:27:15 crc kubenswrapper[4772]: W0127 15:27:15.869505 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42e34d70_0be1_400d_b214_62ba7d9e2e09.slice/crio-38fe91c20bb8dce720ba73a97f9c737c6e74e3b73cf336c9f1c1013c6d14e07f WatchSource:0}: Error finding container 38fe91c20bb8dce720ba73a97f9c737c6e74e3b73cf336c9f1c1013c6d14e07f: Status 404 returned error can't find the container with id 38fe91c20bb8dce720ba73a97f9c737c6e74e3b73cf336c9f1c1013c6d14e07f Jan 27 15:27:15 crc kubenswrapper[4772]: I0127 15:27:15.872768 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zf2tx" event={"ID":"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6","Type":"ContainerStarted","Data":"d2b29cba9bcd684a9fa3005c73cbd809102e0bb6c21ef6ed5d53662bb4cdcdaa"} Jan 27 15:27:15 crc kubenswrapper[4772]: I0127 15:27:15.887361 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fl4nt" event={"ID":"8322baad-60c1-4d0b-96e3-51038f2e447a","Type":"ContainerStarted","Data":"5de6bd74908b324e47419d9f37b784b689e01e1c833ca0e1c7d7483a1e19037c"} Jan 27 15:27:15 crc kubenswrapper[4772]: I0127 15:27:15.887400 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fl4nt" event={"ID":"8322baad-60c1-4d0b-96e3-51038f2e447a","Type":"ContainerStarted","Data":"00f64e139d84dd0a1b89ff44770726f3c1e45680dfa15727e0c2287a16f89b5d"} Jan 27 15:27:15 crc kubenswrapper[4772]: E0127 15:27:15.891542 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-8l85z" podUID="9ae05919-68bf-43d1-abd9-9908ec287bd0" Jan 27 15:27:15 crc kubenswrapper[4772]: I0127 15:27:15.908000 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-zf2tx" podStartSLOduration=2.441509563 podStartE2EDuration="27.907978432s" podCreationTimestamp="2026-01-27 15:26:48 +0000 UTC" firstStartedPulling="2026-01-27 15:26:49.382637438 +0000 UTC m=+1195.363246536" lastFinishedPulling="2026-01-27 15:27:14.849106307 +0000 UTC m=+1220.829715405" observedRunningTime="2026-01-27 15:27:15.904495631 +0000 UTC m=+1221.885104729" watchObservedRunningTime="2026-01-27 15:27:15.907978432 +0000 UTC m=+1221.888587530" Jan 27 15:27:15 crc kubenswrapper[4772]: I0127 15:27:15.951848 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-fl4nt" podStartSLOduration=1.9518243050000001 podStartE2EDuration="1.951824305s" podCreationTimestamp="2026-01-27 15:27:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:27:15.945598765 +0000 UTC m=+1221.926207873" watchObservedRunningTime="2026-01-27 15:27:15.951824305 +0000 UTC m=+1221.932433403" Jan 27 15:27:16 crc kubenswrapper[4772]: I0127 15:27:16.752189 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 15:27:16 crc kubenswrapper[4772]: W0127 15:27:16.765645 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc3206f0_7c01_44c9_9d6a_c586a9b25db8.slice/crio-9244a9db558588c4d27b4c9d5dab7473287ca2114db14b6dfae8075b2b13cc13 WatchSource:0}: Error finding container 9244a9db558588c4d27b4c9d5dab7473287ca2114db14b6dfae8075b2b13cc13: Status 404 returned error can't find the container with id 9244a9db558588c4d27b4c9d5dab7473287ca2114db14b6dfae8075b2b13cc13 Jan 27 15:27:16 crc kubenswrapper[4772]: I0127 15:27:16.928751 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de415c6e-4424-49c4-bc9d-076a5b13ab4e","Type":"ContainerStarted","Data":"ed51d0aa4ae1c7166bbf0464f2b405f79a0faa50f99c4244c9717d1a1fd81db2"} Jan 27 15:27:16 crc kubenswrapper[4772]: I0127 15:27:16.931423 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" event={"ID":"17a547a9-a098-43b7-a153-ad9a137369de","Type":"ContainerStarted","Data":"9b585b108c0dc492c1e4b07fece64b278518a1b12e04b9e57c9b27d9183ca9ac"} Jan 27 15:27:16 crc kubenswrapper[4772]: I0127 15:27:16.931614 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" Jan 27 15:27:16 crc kubenswrapper[4772]: I0127 15:27:16.933911 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bc3206f0-7c01-44c9-9d6a-c586a9b25db8","Type":"ContainerStarted","Data":"9244a9db558588c4d27b4c9d5dab7473287ca2114db14b6dfae8075b2b13cc13"} Jan 27 15:27:16 crc kubenswrapper[4772]: I0127 15:27:16.944717 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"42e34d70-0be1-400d-b214-62ba7d9e2e09","Type":"ContainerStarted","Data":"d9b6567e565b2918f93dacf3cb905248aa46c864651dde1d9ea774cbf20bdb64"} Jan 27 15:27:16 crc kubenswrapper[4772]: I0127 15:27:16.944799 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"42e34d70-0be1-400d-b214-62ba7d9e2e09","Type":"ContainerStarted","Data":"38fe91c20bb8dce720ba73a97f9c737c6e74e3b73cf336c9f1c1013c6d14e07f"} Jan 27 15:27:16 crc kubenswrapper[4772]: I0127 15:27:16.951150 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" podStartSLOduration=13.951131055 podStartE2EDuration="13.951131055s" podCreationTimestamp="2026-01-27 15:27:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:27:16.950492416 +0000 UTC m=+1222.931101524" watchObservedRunningTime="2026-01-27 15:27:16.951131055 +0000 UTC m=+1222.931740153" Jan 27 15:27:17 crc kubenswrapper[4772]: I0127 15:27:17.963692 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bc3206f0-7c01-44c9-9d6a-c586a9b25db8","Type":"ContainerStarted","Data":"dfaebe2fb36461e7df7eef945e33c09ed5b7d7616129107b7ad1d7131c28fb7e"} Jan 27 15:27:17 crc kubenswrapper[4772]: I0127 15:27:17.967049 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="42e34d70-0be1-400d-b214-62ba7d9e2e09" containerName="glance-log" containerID="cri-o://d9b6567e565b2918f93dacf3cb905248aa46c864651dde1d9ea774cbf20bdb64" gracePeriod=30 Jan 27 15:27:17 crc kubenswrapper[4772]: I0127 15:27:17.967209 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"42e34d70-0be1-400d-b214-62ba7d9e2e09","Type":"ContainerStarted","Data":"612151cd11eb1c096a4f60cba2e500a7912da9863b27317fc35d7db2a913ec63"} Jan 27 15:27:17 crc kubenswrapper[4772]: I0127 15:27:17.967387 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="42e34d70-0be1-400d-b214-62ba7d9e2e09" containerName="glance-httpd" containerID="cri-o://612151cd11eb1c096a4f60cba2e500a7912da9863b27317fc35d7db2a913ec63" gracePeriod=30 Jan 27 15:27:17 crc kubenswrapper[4772]: I0127 15:27:17.999908 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=15.999882739 podStartE2EDuration="15.999882739s" podCreationTimestamp="2026-01-27 15:27:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:27:17.993227517 +0000 UTC m=+1223.973836615" watchObservedRunningTime="2026-01-27 15:27:17.999882739 +0000 UTC m=+1223.980491837" Jan 27 15:27:18 crc kubenswrapper[4772]: I0127 15:27:18.982822 4772 generic.go:334] "Generic (PLEG): container finished" podID="42e34d70-0be1-400d-b214-62ba7d9e2e09" containerID="d9b6567e565b2918f93dacf3cb905248aa46c864651dde1d9ea774cbf20bdb64" exitCode=143 Jan 27 15:27:18 crc kubenswrapper[4772]: I0127 15:27:18.983005 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"42e34d70-0be1-400d-b214-62ba7d9e2e09","Type":"ContainerDied","Data":"d9b6567e565b2918f93dacf3cb905248aa46c864651dde1d9ea774cbf20bdb64"} Jan 27 15:27:19 crc kubenswrapper[4772]: I0127 15:27:19.984859 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.004979 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bc3206f0-7c01-44c9-9d6a-c586a9b25db8","Type":"ContainerStarted","Data":"e5923297bfaf90562635dde0a7a065cb6e5b62f035580ac799c013859809f144"} Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.005162 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="bc3206f0-7c01-44c9-9d6a-c586a9b25db8" containerName="glance-log" containerID="cri-o://dfaebe2fb36461e7df7eef945e33c09ed5b7d7616129107b7ad1d7131c28fb7e" gracePeriod=30 Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.005447 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="bc3206f0-7c01-44c9-9d6a-c586a9b25db8" containerName="glance-httpd" containerID="cri-o://e5923297bfaf90562635dde0a7a065cb6e5b62f035580ac799c013859809f144" gracePeriod=30 Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.018740 4772 generic.go:334] "Generic (PLEG): container finished" podID="42e34d70-0be1-400d-b214-62ba7d9e2e09" containerID="612151cd11eb1c096a4f60cba2e500a7912da9863b27317fc35d7db2a913ec63" exitCode=0 Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.018789 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"42e34d70-0be1-400d-b214-62ba7d9e2e09","Type":"ContainerDied","Data":"612151cd11eb1c096a4f60cba2e500a7912da9863b27317fc35d7db2a913ec63"} Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.018820 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"42e34d70-0be1-400d-b214-62ba7d9e2e09","Type":"ContainerDied","Data":"38fe91c20bb8dce720ba73a97f9c737c6e74e3b73cf336c9f1c1013c6d14e07f"} Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.018839 4772 scope.go:117] "RemoveContainer" containerID="612151cd11eb1c096a4f60cba2e500a7912da9863b27317fc35d7db2a913ec63" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.019016 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.043531 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=17.043510515 podStartE2EDuration="17.043510515s" podCreationTimestamp="2026-01-27 15:27:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:27:20.038255613 +0000 UTC m=+1226.018864731" watchObservedRunningTime="2026-01-27 15:27:20.043510515 +0000 UTC m=+1226.024119613" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.065414 4772 scope.go:117] "RemoveContainer" containerID="d9b6567e565b2918f93dacf3cb905248aa46c864651dde1d9ea774cbf20bdb64" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.081098 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wphqd\" (UniqueName: \"kubernetes.io/projected/42e34d70-0be1-400d-b214-62ba7d9e2e09-kube-api-access-wphqd\") pod \"42e34d70-0be1-400d-b214-62ba7d9e2e09\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.081236 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42e34d70-0be1-400d-b214-62ba7d9e2e09-combined-ca-bundle\") pod \"42e34d70-0be1-400d-b214-62ba7d9e2e09\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.081367 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"42e34d70-0be1-400d-b214-62ba7d9e2e09\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.081430 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42e34d70-0be1-400d-b214-62ba7d9e2e09-logs\") pod \"42e34d70-0be1-400d-b214-62ba7d9e2e09\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.081468 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42e34d70-0be1-400d-b214-62ba7d9e2e09-scripts\") pod \"42e34d70-0be1-400d-b214-62ba7d9e2e09\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.081500 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42e34d70-0be1-400d-b214-62ba7d9e2e09-config-data\") pod \"42e34d70-0be1-400d-b214-62ba7d9e2e09\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.081548 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/42e34d70-0be1-400d-b214-62ba7d9e2e09-httpd-run\") pod \"42e34d70-0be1-400d-b214-62ba7d9e2e09\" (UID: \"42e34d70-0be1-400d-b214-62ba7d9e2e09\") " Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.082090 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42e34d70-0be1-400d-b214-62ba7d9e2e09-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "42e34d70-0be1-400d-b214-62ba7d9e2e09" (UID: "42e34d70-0be1-400d-b214-62ba7d9e2e09"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.082494 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42e34d70-0be1-400d-b214-62ba7d9e2e09-logs" (OuterVolumeSpecName: "logs") pod "42e34d70-0be1-400d-b214-62ba7d9e2e09" (UID: "42e34d70-0be1-400d-b214-62ba7d9e2e09"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.087867 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "42e34d70-0be1-400d-b214-62ba7d9e2e09" (UID: "42e34d70-0be1-400d-b214-62ba7d9e2e09"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.088044 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42e34d70-0be1-400d-b214-62ba7d9e2e09-scripts" (OuterVolumeSpecName: "scripts") pod "42e34d70-0be1-400d-b214-62ba7d9e2e09" (UID: "42e34d70-0be1-400d-b214-62ba7d9e2e09"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.088119 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42e34d70-0be1-400d-b214-62ba7d9e2e09-kube-api-access-wphqd" (OuterVolumeSpecName: "kube-api-access-wphqd") pod "42e34d70-0be1-400d-b214-62ba7d9e2e09" (UID: "42e34d70-0be1-400d-b214-62ba7d9e2e09"). InnerVolumeSpecName "kube-api-access-wphqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.107140 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42e34d70-0be1-400d-b214-62ba7d9e2e09-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "42e34d70-0be1-400d-b214-62ba7d9e2e09" (UID: "42e34d70-0be1-400d-b214-62ba7d9e2e09"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.127647 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42e34d70-0be1-400d-b214-62ba7d9e2e09-config-data" (OuterVolumeSpecName: "config-data") pod "42e34d70-0be1-400d-b214-62ba7d9e2e09" (UID: "42e34d70-0be1-400d-b214-62ba7d9e2e09"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.183735 4772 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.183765 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42e34d70-0be1-400d-b214-62ba7d9e2e09-logs\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.183773 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42e34d70-0be1-400d-b214-62ba7d9e2e09-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.183781 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42e34d70-0be1-400d-b214-62ba7d9e2e09-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.183790 4772 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/42e34d70-0be1-400d-b214-62ba7d9e2e09-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.183800 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wphqd\" (UniqueName: \"kubernetes.io/projected/42e34d70-0be1-400d-b214-62ba7d9e2e09-kube-api-access-wphqd\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.183810 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42e34d70-0be1-400d-b214-62ba7d9e2e09-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.202061 4772 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.226326 4772 scope.go:117] "RemoveContainer" containerID="612151cd11eb1c096a4f60cba2e500a7912da9863b27317fc35d7db2a913ec63" Jan 27 15:27:20 crc kubenswrapper[4772]: E0127 15:27:20.226821 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"612151cd11eb1c096a4f60cba2e500a7912da9863b27317fc35d7db2a913ec63\": container with ID starting with 612151cd11eb1c096a4f60cba2e500a7912da9863b27317fc35d7db2a913ec63 not found: ID does not exist" containerID="612151cd11eb1c096a4f60cba2e500a7912da9863b27317fc35d7db2a913ec63" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.226897 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"612151cd11eb1c096a4f60cba2e500a7912da9863b27317fc35d7db2a913ec63"} err="failed to get container status \"612151cd11eb1c096a4f60cba2e500a7912da9863b27317fc35d7db2a913ec63\": rpc error: code = NotFound desc = could not find container \"612151cd11eb1c096a4f60cba2e500a7912da9863b27317fc35d7db2a913ec63\": container with ID starting with 612151cd11eb1c096a4f60cba2e500a7912da9863b27317fc35d7db2a913ec63 not found: ID does not exist" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.227235 4772 scope.go:117] "RemoveContainer" containerID="d9b6567e565b2918f93dacf3cb905248aa46c864651dde1d9ea774cbf20bdb64" Jan 27 15:27:20 crc kubenswrapper[4772]: E0127 15:27:20.228038 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9b6567e565b2918f93dacf3cb905248aa46c864651dde1d9ea774cbf20bdb64\": container with ID starting with d9b6567e565b2918f93dacf3cb905248aa46c864651dde1d9ea774cbf20bdb64 not found: ID does not exist" containerID="d9b6567e565b2918f93dacf3cb905248aa46c864651dde1d9ea774cbf20bdb64" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.228084 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9b6567e565b2918f93dacf3cb905248aa46c864651dde1d9ea774cbf20bdb64"} err="failed to get container status \"d9b6567e565b2918f93dacf3cb905248aa46c864651dde1d9ea774cbf20bdb64\": rpc error: code = NotFound desc = could not find container \"d9b6567e565b2918f93dacf3cb905248aa46c864651dde1d9ea774cbf20bdb64\": container with ID starting with d9b6567e565b2918f93dacf3cb905248aa46c864651dde1d9ea774cbf20bdb64 not found: ID does not exist" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.285122 4772 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.366703 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.394785 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.407656 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 15:27:20 crc kubenswrapper[4772]: E0127 15:27:20.408376 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42e34d70-0be1-400d-b214-62ba7d9e2e09" containerName="glance-httpd" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.408397 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="42e34d70-0be1-400d-b214-62ba7d9e2e09" containerName="glance-httpd" Jan 27 15:27:20 crc kubenswrapper[4772]: E0127 15:27:20.408411 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42e34d70-0be1-400d-b214-62ba7d9e2e09" containerName="glance-log" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.408418 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="42e34d70-0be1-400d-b214-62ba7d9e2e09" containerName="glance-log" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.408636 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="42e34d70-0be1-400d-b214-62ba7d9e2e09" containerName="glance-httpd" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.408670 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="42e34d70-0be1-400d-b214-62ba7d9e2e09" containerName="glance-log" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.411701 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.418429 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.420075 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.420265 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.499070 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c94a7cfa-28e2-4d52-85a1-d5586f162227-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.499222 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q49rw\" (UniqueName: \"kubernetes.io/projected/c94a7cfa-28e2-4d52-85a1-d5586f162227-kube-api-access-q49rw\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.499270 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c94a7cfa-28e2-4d52-85a1-d5586f162227-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.499298 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c94a7cfa-28e2-4d52-85a1-d5586f162227-logs\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.499366 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c94a7cfa-28e2-4d52-85a1-d5586f162227-scripts\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.499393 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c94a7cfa-28e2-4d52-85a1-d5586f162227-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.499432 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c94a7cfa-28e2-4d52-85a1-d5586f162227-config-data\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.499638 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.569639 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.612062 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c94a7cfa-28e2-4d52-85a1-d5586f162227-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.612154 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c94a7cfa-28e2-4d52-85a1-d5586f162227-logs\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.612262 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c94a7cfa-28e2-4d52-85a1-d5586f162227-scripts\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.612292 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c94a7cfa-28e2-4d52-85a1-d5586f162227-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.612331 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c94a7cfa-28e2-4d52-85a1-d5586f162227-config-data\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.612434 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.612660 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c94a7cfa-28e2-4d52-85a1-d5586f162227-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.612737 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q49rw\" (UniqueName: \"kubernetes.io/projected/c94a7cfa-28e2-4d52-85a1-d5586f162227-kube-api-access-q49rw\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.612785 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c94a7cfa-28e2-4d52-85a1-d5586f162227-logs\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.613442 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.613668 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c94a7cfa-28e2-4d52-85a1-d5586f162227-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.618295 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c94a7cfa-28e2-4d52-85a1-d5586f162227-scripts\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.621263 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c94a7cfa-28e2-4d52-85a1-d5586f162227-config-data\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.622897 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c94a7cfa-28e2-4d52-85a1-d5586f162227-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.623664 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c94a7cfa-28e2-4d52-85a1-d5586f162227-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.635109 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q49rw\" (UniqueName: \"kubernetes.io/projected/c94a7cfa-28e2-4d52-85a1-d5586f162227-kube-api-access-q49rw\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.650612 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.673075 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42e34d70-0be1-400d-b214-62ba7d9e2e09" path="/var/lib/kubelet/pods/42e34d70-0be1-400d-b214-62ba7d9e2e09/volumes" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.713989 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rkbs\" (UniqueName: \"kubernetes.io/projected/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-kube-api-access-6rkbs\") pod \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.714091 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-config-data\") pod \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.714563 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-scripts\") pod \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.714607 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-httpd-run\") pod \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.714646 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.714674 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-logs\") pod \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.714722 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-combined-ca-bundle\") pod \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\" (UID: \"bc3206f0-7c01-44c9-9d6a-c586a9b25db8\") " Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.715396 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "bc3206f0-7c01-44c9-9d6a-c586a9b25db8" (UID: "bc3206f0-7c01-44c9-9d6a-c586a9b25db8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.715526 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-logs" (OuterVolumeSpecName: "logs") pod "bc3206f0-7c01-44c9-9d6a-c586a9b25db8" (UID: "bc3206f0-7c01-44c9-9d6a-c586a9b25db8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.719082 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-kube-api-access-6rkbs" (OuterVolumeSpecName: "kube-api-access-6rkbs") pod "bc3206f0-7c01-44c9-9d6a-c586a9b25db8" (UID: "bc3206f0-7c01-44c9-9d6a-c586a9b25db8"). InnerVolumeSpecName "kube-api-access-6rkbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.719410 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-scripts" (OuterVolumeSpecName: "scripts") pod "bc3206f0-7c01-44c9-9d6a-c586a9b25db8" (UID: "bc3206f0-7c01-44c9-9d6a-c586a9b25db8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.719509 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "bc3206f0-7c01-44c9-9d6a-c586a9b25db8" (UID: "bc3206f0-7c01-44c9-9d6a-c586a9b25db8"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.737705 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc3206f0-7c01-44c9-9d6a-c586a9b25db8" (UID: "bc3206f0-7c01-44c9-9d6a-c586a9b25db8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.762856 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-config-data" (OuterVolumeSpecName: "config-data") pod "bc3206f0-7c01-44c9-9d6a-c586a9b25db8" (UID: "bc3206f0-7c01-44c9-9d6a-c586a9b25db8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.765967 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.816731 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.816767 4772 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.816802 4772 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.816812 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-logs\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.816820 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.816831 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rkbs\" (UniqueName: \"kubernetes.io/projected/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-kube-api-access-6rkbs\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.816839 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc3206f0-7c01-44c9-9d6a-c586a9b25db8-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.833694 4772 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Jan 27 15:27:20 crc kubenswrapper[4772]: I0127 15:27:20.918628 4772 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.031353 4772 generic.go:334] "Generic (PLEG): container finished" podID="bc3206f0-7c01-44c9-9d6a-c586a9b25db8" containerID="e5923297bfaf90562635dde0a7a065cb6e5b62f035580ac799c013859809f144" exitCode=143 Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.031388 4772 generic.go:334] "Generic (PLEG): container finished" podID="bc3206f0-7c01-44c9-9d6a-c586a9b25db8" containerID="dfaebe2fb36461e7df7eef945e33c09ed5b7d7616129107b7ad1d7131c28fb7e" exitCode=143 Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.031444 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bc3206f0-7c01-44c9-9d6a-c586a9b25db8","Type":"ContainerDied","Data":"e5923297bfaf90562635dde0a7a065cb6e5b62f035580ac799c013859809f144"} Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.031480 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bc3206f0-7c01-44c9-9d6a-c586a9b25db8","Type":"ContainerDied","Data":"dfaebe2fb36461e7df7eef945e33c09ed5b7d7616129107b7ad1d7131c28fb7e"} Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.031487 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.031509 4772 scope.go:117] "RemoveContainer" containerID="e5923297bfaf90562635dde0a7a065cb6e5b62f035580ac799c013859809f144" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.031497 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bc3206f0-7c01-44c9-9d6a-c586a9b25db8","Type":"ContainerDied","Data":"9244a9db558588c4d27b4c9d5dab7473287ca2114db14b6dfae8075b2b13cc13"} Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.041147 4772 generic.go:334] "Generic (PLEG): container finished" podID="8322baad-60c1-4d0b-96e3-51038f2e447a" containerID="5de6bd74908b324e47419d9f37b784b689e01e1c833ca0e1c7d7483a1e19037c" exitCode=0 Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.041234 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fl4nt" event={"ID":"8322baad-60c1-4d0b-96e3-51038f2e447a","Type":"ContainerDied","Data":"5de6bd74908b324e47419d9f37b784b689e01e1c833ca0e1c7d7483a1e19037c"} Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.080206 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.082973 4772 scope.go:117] "RemoveContainer" containerID="dfaebe2fb36461e7df7eef945e33c09ed5b7d7616129107b7ad1d7131c28fb7e" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.090542 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.107740 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 15:27:21 crc kubenswrapper[4772]: E0127 15:27:21.108358 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc3206f0-7c01-44c9-9d6a-c586a9b25db8" containerName="glance-log" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.108373 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc3206f0-7c01-44c9-9d6a-c586a9b25db8" containerName="glance-log" Jan 27 15:27:21 crc kubenswrapper[4772]: E0127 15:27:21.108400 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc3206f0-7c01-44c9-9d6a-c586a9b25db8" containerName="glance-httpd" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.108406 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc3206f0-7c01-44c9-9d6a-c586a9b25db8" containerName="glance-httpd" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.108553 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc3206f0-7c01-44c9-9d6a-c586a9b25db8" containerName="glance-httpd" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.108568 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc3206f0-7c01-44c9-9d6a-c586a9b25db8" containerName="glance-log" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.108961 4772 scope.go:117] "RemoveContainer" containerID="e5923297bfaf90562635dde0a7a065cb6e5b62f035580ac799c013859809f144" Jan 27 15:27:21 crc kubenswrapper[4772]: E0127 15:27:21.110548 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5923297bfaf90562635dde0a7a065cb6e5b62f035580ac799c013859809f144\": container with ID starting with e5923297bfaf90562635dde0a7a065cb6e5b62f035580ac799c013859809f144 not found: ID does not exist" containerID="e5923297bfaf90562635dde0a7a065cb6e5b62f035580ac799c013859809f144" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.110591 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5923297bfaf90562635dde0a7a065cb6e5b62f035580ac799c013859809f144"} err="failed to get container status \"e5923297bfaf90562635dde0a7a065cb6e5b62f035580ac799c013859809f144\": rpc error: code = NotFound desc = could not find container \"e5923297bfaf90562635dde0a7a065cb6e5b62f035580ac799c013859809f144\": container with ID starting with e5923297bfaf90562635dde0a7a065cb6e5b62f035580ac799c013859809f144 not found: ID does not exist" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.110618 4772 scope.go:117] "RemoveContainer" containerID="dfaebe2fb36461e7df7eef945e33c09ed5b7d7616129107b7ad1d7131c28fb7e" Jan 27 15:27:21 crc kubenswrapper[4772]: E0127 15:27:21.111698 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfaebe2fb36461e7df7eef945e33c09ed5b7d7616129107b7ad1d7131c28fb7e\": container with ID starting with dfaebe2fb36461e7df7eef945e33c09ed5b7d7616129107b7ad1d7131c28fb7e not found: ID does not exist" containerID="dfaebe2fb36461e7df7eef945e33c09ed5b7d7616129107b7ad1d7131c28fb7e" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.111723 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfaebe2fb36461e7df7eef945e33c09ed5b7d7616129107b7ad1d7131c28fb7e"} err="failed to get container status \"dfaebe2fb36461e7df7eef945e33c09ed5b7d7616129107b7ad1d7131c28fb7e\": rpc error: code = NotFound desc = could not find container \"dfaebe2fb36461e7df7eef945e33c09ed5b7d7616129107b7ad1d7131c28fb7e\": container with ID starting with dfaebe2fb36461e7df7eef945e33c09ed5b7d7616129107b7ad1d7131c28fb7e not found: ID does not exist" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.111735 4772 scope.go:117] "RemoveContainer" containerID="e5923297bfaf90562635dde0a7a065cb6e5b62f035580ac799c013859809f144" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.112507 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5923297bfaf90562635dde0a7a065cb6e5b62f035580ac799c013859809f144"} err="failed to get container status \"e5923297bfaf90562635dde0a7a065cb6e5b62f035580ac799c013859809f144\": rpc error: code = NotFound desc = could not find container \"e5923297bfaf90562635dde0a7a065cb6e5b62f035580ac799c013859809f144\": container with ID starting with e5923297bfaf90562635dde0a7a065cb6e5b62f035580ac799c013859809f144 not found: ID does not exist" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.112532 4772 scope.go:117] "RemoveContainer" containerID="dfaebe2fb36461e7df7eef945e33c09ed5b7d7616129107b7ad1d7131c28fb7e" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.112768 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfaebe2fb36461e7df7eef945e33c09ed5b7d7616129107b7ad1d7131c28fb7e"} err="failed to get container status \"dfaebe2fb36461e7df7eef945e33c09ed5b7d7616129107b7ad1d7131c28fb7e\": rpc error: code = NotFound desc = could not find container \"dfaebe2fb36461e7df7eef945e33c09ed5b7d7616129107b7ad1d7131c28fb7e\": container with ID starting with dfaebe2fb36461e7df7eef945e33c09ed5b7d7616129107b7ad1d7131c28fb7e not found: ID does not exist" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.114675 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.121340 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.121444 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.134260 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.142396 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.224990 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc9t4\" (UniqueName: \"kubernetes.io/projected/41f85a83-f245-40ff-b994-50cab01b2530-kube-api-access-pc9t4\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.225119 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41f85a83-f245-40ff-b994-50cab01b2530-config-data\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.225203 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41f85a83-f245-40ff-b994-50cab01b2530-scripts\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.225251 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/41f85a83-f245-40ff-b994-50cab01b2530-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.225274 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41f85a83-f245-40ff-b994-50cab01b2530-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.225429 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41f85a83-f245-40ff-b994-50cab01b2530-logs\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.225527 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41f85a83-f245-40ff-b994-50cab01b2530-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.225610 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.327139 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc9t4\" (UniqueName: \"kubernetes.io/projected/41f85a83-f245-40ff-b994-50cab01b2530-kube-api-access-pc9t4\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.327224 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41f85a83-f245-40ff-b994-50cab01b2530-config-data\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.327283 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41f85a83-f245-40ff-b994-50cab01b2530-scripts\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.327339 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/41f85a83-f245-40ff-b994-50cab01b2530-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.327366 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41f85a83-f245-40ff-b994-50cab01b2530-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.327394 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41f85a83-f245-40ff-b994-50cab01b2530-logs\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.327429 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41f85a83-f245-40ff-b994-50cab01b2530-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.327462 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.328008 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.328067 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41f85a83-f245-40ff-b994-50cab01b2530-logs\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.328966 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/41f85a83-f245-40ff-b994-50cab01b2530-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.334529 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41f85a83-f245-40ff-b994-50cab01b2530-scripts\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.334529 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41f85a83-f245-40ff-b994-50cab01b2530-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.334878 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41f85a83-f245-40ff-b994-50cab01b2530-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.335248 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41f85a83-f245-40ff-b994-50cab01b2530-config-data\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.347064 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc9t4\" (UniqueName: \"kubernetes.io/projected/41f85a83-f245-40ff-b994-50cab01b2530-kube-api-access-pc9t4\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.353985 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:27:21 crc kubenswrapper[4772]: I0127 15:27:21.485614 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 27 15:27:22 crc kubenswrapper[4772]: I0127 15:27:22.056231 4772 generic.go:334] "Generic (PLEG): container finished" podID="f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6" containerID="d2b29cba9bcd684a9fa3005c73cbd809102e0bb6c21ef6ed5d53662bb4cdcdaa" exitCode=0 Jan 27 15:27:22 crc kubenswrapper[4772]: I0127 15:27:22.056301 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zf2tx" event={"ID":"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6","Type":"ContainerDied","Data":"d2b29cba9bcd684a9fa3005c73cbd809102e0bb6c21ef6ed5d53662bb4cdcdaa"} Jan 27 15:27:22 crc kubenswrapper[4772]: I0127 15:27:22.065038 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c94a7cfa-28e2-4d52-85a1-d5586f162227","Type":"ContainerStarted","Data":"c52299828ac41e83b1686de53ba3808d1e810b20370ec9d5bc6e9bbc6b64bbed"} Jan 27 15:27:22 crc kubenswrapper[4772]: I0127 15:27:22.065098 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c94a7cfa-28e2-4d52-85a1-d5586f162227","Type":"ContainerStarted","Data":"65117f0b87347a480b318a709dc150116a10a8d323bd2553e117803b3054a685"} Jan 27 15:27:22 crc kubenswrapper[4772]: I0127 15:27:22.696431 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fl4nt" Jan 27 15:27:22 crc kubenswrapper[4772]: I0127 15:27:22.700677 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc3206f0-7c01-44c9-9d6a-c586a9b25db8" path="/var/lib/kubelet/pods/bc3206f0-7c01-44c9-9d6a-c586a9b25db8/volumes" Jan 27 15:27:22 crc kubenswrapper[4772]: I0127 15:27:22.785805 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-credential-keys\") pod \"8322baad-60c1-4d0b-96e3-51038f2e447a\" (UID: \"8322baad-60c1-4d0b-96e3-51038f2e447a\") " Jan 27 15:27:22 crc kubenswrapper[4772]: I0127 15:27:22.785901 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-scripts\") pod \"8322baad-60c1-4d0b-96e3-51038f2e447a\" (UID: \"8322baad-60c1-4d0b-96e3-51038f2e447a\") " Jan 27 15:27:22 crc kubenswrapper[4772]: I0127 15:27:22.786067 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-config-data\") pod \"8322baad-60c1-4d0b-96e3-51038f2e447a\" (UID: \"8322baad-60c1-4d0b-96e3-51038f2e447a\") " Jan 27 15:27:22 crc kubenswrapper[4772]: I0127 15:27:22.786125 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-968vm\" (UniqueName: \"kubernetes.io/projected/8322baad-60c1-4d0b-96e3-51038f2e447a-kube-api-access-968vm\") pod \"8322baad-60c1-4d0b-96e3-51038f2e447a\" (UID: \"8322baad-60c1-4d0b-96e3-51038f2e447a\") " Jan 27 15:27:22 crc kubenswrapper[4772]: I0127 15:27:22.786210 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-fernet-keys\") pod \"8322baad-60c1-4d0b-96e3-51038f2e447a\" (UID: \"8322baad-60c1-4d0b-96e3-51038f2e447a\") " Jan 27 15:27:22 crc kubenswrapper[4772]: I0127 15:27:22.786250 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-combined-ca-bundle\") pod \"8322baad-60c1-4d0b-96e3-51038f2e447a\" (UID: \"8322baad-60c1-4d0b-96e3-51038f2e447a\") " Jan 27 15:27:22 crc kubenswrapper[4772]: I0127 15:27:22.791207 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8322baad-60c1-4d0b-96e3-51038f2e447a-kube-api-access-968vm" (OuterVolumeSpecName: "kube-api-access-968vm") pod "8322baad-60c1-4d0b-96e3-51038f2e447a" (UID: "8322baad-60c1-4d0b-96e3-51038f2e447a"). InnerVolumeSpecName "kube-api-access-968vm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:27:22 crc kubenswrapper[4772]: I0127 15:27:22.798717 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "8322baad-60c1-4d0b-96e3-51038f2e447a" (UID: "8322baad-60c1-4d0b-96e3-51038f2e447a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:22 crc kubenswrapper[4772]: I0127 15:27:22.804921 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-scripts" (OuterVolumeSpecName: "scripts") pod "8322baad-60c1-4d0b-96e3-51038f2e447a" (UID: "8322baad-60c1-4d0b-96e3-51038f2e447a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:22 crc kubenswrapper[4772]: I0127 15:27:22.810788 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8322baad-60c1-4d0b-96e3-51038f2e447a" (UID: "8322baad-60c1-4d0b-96e3-51038f2e447a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:22 crc kubenswrapper[4772]: I0127 15:27:22.819187 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-config-data" (OuterVolumeSpecName: "config-data") pod "8322baad-60c1-4d0b-96e3-51038f2e447a" (UID: "8322baad-60c1-4d0b-96e3-51038f2e447a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:22 crc kubenswrapper[4772]: I0127 15:27:22.823311 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8322baad-60c1-4d0b-96e3-51038f2e447a" (UID: "8322baad-60c1-4d0b-96e3-51038f2e447a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:22 crc kubenswrapper[4772]: I0127 15:27:22.887680 4772 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:22 crc kubenswrapper[4772]: I0127 15:27:22.887716 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:22 crc kubenswrapper[4772]: I0127 15:27:22.887729 4772 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:22 crc kubenswrapper[4772]: I0127 15:27:22.887738 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:22 crc kubenswrapper[4772]: I0127 15:27:22.887746 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8322baad-60c1-4d0b-96e3-51038f2e447a-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:22 crc kubenswrapper[4772]: I0127 15:27:22.887754 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-968vm\" (UniqueName: \"kubernetes.io/projected/8322baad-60c1-4d0b-96e3-51038f2e447a-kube-api-access-968vm\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.089270 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fl4nt" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.090436 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fl4nt" event={"ID":"8322baad-60c1-4d0b-96e3-51038f2e447a","Type":"ContainerDied","Data":"00f64e139d84dd0a1b89ff44770726f3c1e45680dfa15727e0c2287a16f89b5d"} Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.090484 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00f64e139d84dd0a1b89ff44770726f3c1e45680dfa15727e0c2287a16f89b5d" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.183899 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-677fb7d6fc-djjsx"] Jan 27 15:27:23 crc kubenswrapper[4772]: E0127 15:27:23.184350 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8322baad-60c1-4d0b-96e3-51038f2e447a" containerName="keystone-bootstrap" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.184363 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="8322baad-60c1-4d0b-96e3-51038f2e447a" containerName="keystone-bootstrap" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.184508 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="8322baad-60c1-4d0b-96e3-51038f2e447a" containerName="keystone-bootstrap" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.185074 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.189302 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.189397 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.189842 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.193398 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.193616 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-bdjsw" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.193772 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.198911 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-677fb7d6fc-djjsx"] Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.297909 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-config-data\") pod \"keystone-677fb7d6fc-djjsx\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.298054 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-internal-tls-certs\") pod \"keystone-677fb7d6fc-djjsx\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.298162 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-combined-ca-bundle\") pod \"keystone-677fb7d6fc-djjsx\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.298397 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-public-tls-certs\") pod \"keystone-677fb7d6fc-djjsx\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.298465 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-scripts\") pod \"keystone-677fb7d6fc-djjsx\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.298638 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksdpm\" (UniqueName: \"kubernetes.io/projected/6e790127-8223-4b0c-8a5d-21e1bb15fa30-kube-api-access-ksdpm\") pod \"keystone-677fb7d6fc-djjsx\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.298720 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-credential-keys\") pod \"keystone-677fb7d6fc-djjsx\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.298775 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-fernet-keys\") pod \"keystone-677fb7d6fc-djjsx\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.400561 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-combined-ca-bundle\") pod \"keystone-677fb7d6fc-djjsx\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.400659 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-public-tls-certs\") pod \"keystone-677fb7d6fc-djjsx\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.400695 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-scripts\") pod \"keystone-677fb7d6fc-djjsx\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.400780 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksdpm\" (UniqueName: \"kubernetes.io/projected/6e790127-8223-4b0c-8a5d-21e1bb15fa30-kube-api-access-ksdpm\") pod \"keystone-677fb7d6fc-djjsx\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.400841 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-credential-keys\") pod \"keystone-677fb7d6fc-djjsx\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.400884 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-fernet-keys\") pod \"keystone-677fb7d6fc-djjsx\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.400936 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-config-data\") pod \"keystone-677fb7d6fc-djjsx\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.400976 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-internal-tls-certs\") pod \"keystone-677fb7d6fc-djjsx\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.405320 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-internal-tls-certs\") pod \"keystone-677fb7d6fc-djjsx\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.405573 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-combined-ca-bundle\") pod \"keystone-677fb7d6fc-djjsx\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.407090 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-public-tls-certs\") pod \"keystone-677fb7d6fc-djjsx\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.408286 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-credential-keys\") pod \"keystone-677fb7d6fc-djjsx\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.410016 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-scripts\") pod \"keystone-677fb7d6fc-djjsx\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.410741 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-fernet-keys\") pod \"keystone-677fb7d6fc-djjsx\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.418292 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-config-data\") pod \"keystone-677fb7d6fc-djjsx\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.425446 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksdpm\" (UniqueName: \"kubernetes.io/projected/6e790127-8223-4b0c-8a5d-21e1bb15fa30-kube-api-access-ksdpm\") pod \"keystone-677fb7d6fc-djjsx\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.480502 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.550253 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.572993 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-8d9vp"] Jan 27 15:27:23 crc kubenswrapper[4772]: I0127 15:27:23.575391 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" podUID="e329efba-60e3-49c7-81ff-b073be77e34b" containerName="dnsmasq-dns" containerID="cri-o://0bc2eb78a83f1e9ddf6e0c975669640d497ff3e50951b0eaadaee82dc03caffd" gracePeriod=10 Jan 27 15:27:24 crc kubenswrapper[4772]: I0127 15:27:24.104355 4772 generic.go:334] "Generic (PLEG): container finished" podID="e329efba-60e3-49c7-81ff-b073be77e34b" containerID="0bc2eb78a83f1e9ddf6e0c975669640d497ff3e50951b0eaadaee82dc03caffd" exitCode=0 Jan 27 15:27:24 crc kubenswrapper[4772]: I0127 15:27:24.104531 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" event={"ID":"e329efba-60e3-49c7-81ff-b073be77e34b","Type":"ContainerDied","Data":"0bc2eb78a83f1e9ddf6e0c975669640d497ff3e50951b0eaadaee82dc03caffd"} Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.084659 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zf2tx" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.154395 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zf2tx" event={"ID":"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6","Type":"ContainerDied","Data":"7dfb28db1e03cbbc36a413590b93de83567a6c9fa02be76267be1180098e9795"} Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.154461 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7dfb28db1e03cbbc36a413590b93de83567a6c9fa02be76267be1180098e9795" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.154555 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zf2tx" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.184871 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-scripts\") pod \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\" (UID: \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\") " Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.184915 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-combined-ca-bundle\") pod \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\" (UID: \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\") " Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.184954 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-logs\") pod \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\" (UID: \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\") " Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.185000 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-config-data\") pod \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\" (UID: \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\") " Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.185097 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8x5b\" (UniqueName: \"kubernetes.io/projected/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-kube-api-access-d8x5b\") pod \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\" (UID: \"f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6\") " Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.186214 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-logs" (OuterVolumeSpecName: "logs") pod "f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6" (UID: "f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.205444 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-scripts" (OuterVolumeSpecName: "scripts") pod "f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6" (UID: "f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.205497 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-kube-api-access-d8x5b" (OuterVolumeSpecName: "kube-api-access-d8x5b") pod "f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6" (UID: "f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6"). InnerVolumeSpecName "kube-api-access-d8x5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.226573 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-config-data" (OuterVolumeSpecName: "config-data") pod "f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6" (UID: "f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.233121 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6" (UID: "f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.287633 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.287664 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.287676 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-logs\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.287686 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.287694 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8x5b\" (UniqueName: \"kubernetes.io/projected/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6-kube-api-access-d8x5b\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.326396 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.490073 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-ovsdbserver-sb\") pod \"e329efba-60e3-49c7-81ff-b073be77e34b\" (UID: \"e329efba-60e3-49c7-81ff-b073be77e34b\") " Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.490283 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-dns-svc\") pod \"e329efba-60e3-49c7-81ff-b073be77e34b\" (UID: \"e329efba-60e3-49c7-81ff-b073be77e34b\") " Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.490671 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-config\") pod \"e329efba-60e3-49c7-81ff-b073be77e34b\" (UID: \"e329efba-60e3-49c7-81ff-b073be77e34b\") " Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.490713 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmlv8\" (UniqueName: \"kubernetes.io/projected/e329efba-60e3-49c7-81ff-b073be77e34b-kube-api-access-zmlv8\") pod \"e329efba-60e3-49c7-81ff-b073be77e34b\" (UID: \"e329efba-60e3-49c7-81ff-b073be77e34b\") " Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.490741 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-ovsdbserver-nb\") pod \"e329efba-60e3-49c7-81ff-b073be77e34b\" (UID: \"e329efba-60e3-49c7-81ff-b073be77e34b\") " Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.491063 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-dns-swift-storage-0\") pod \"e329efba-60e3-49c7-81ff-b073be77e34b\" (UID: \"e329efba-60e3-49c7-81ff-b073be77e34b\") " Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.512197 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e329efba-60e3-49c7-81ff-b073be77e34b-kube-api-access-zmlv8" (OuterVolumeSpecName: "kube-api-access-zmlv8") pod "e329efba-60e3-49c7-81ff-b073be77e34b" (UID: "e329efba-60e3-49c7-81ff-b073be77e34b"). InnerVolumeSpecName "kube-api-access-zmlv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.527867 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-677fb7d6fc-djjsx"] Jan 27 15:27:27 crc kubenswrapper[4772]: W0127 15:27:27.533725 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e790127_8223_4b0c_8a5d_21e1bb15fa30.slice/crio-8b30129bf5b3504ae600edeaafe66f62d6f0c11b788461d423310f03199da3c5 WatchSource:0}: Error finding container 8b30129bf5b3504ae600edeaafe66f62d6f0c11b788461d423310f03199da3c5: Status 404 returned error can't find the container with id 8b30129bf5b3504ae600edeaafe66f62d6f0c11b788461d423310f03199da3c5 Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.566458 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e329efba-60e3-49c7-81ff-b073be77e34b" (UID: "e329efba-60e3-49c7-81ff-b073be77e34b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.581258 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e329efba-60e3-49c7-81ff-b073be77e34b" (UID: "e329efba-60e3-49c7-81ff-b073be77e34b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.581272 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e329efba-60e3-49c7-81ff-b073be77e34b" (UID: "e329efba-60e3-49c7-81ff-b073be77e34b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.582113 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e329efba-60e3-49c7-81ff-b073be77e34b" (UID: "e329efba-60e3-49c7-81ff-b073be77e34b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.593305 4772 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.593338 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.593348 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.593356 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmlv8\" (UniqueName: \"kubernetes.io/projected/e329efba-60e3-49c7-81ff-b073be77e34b-kube-api-access-zmlv8\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.593367 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.599448 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-config" (OuterVolumeSpecName: "config") pod "e329efba-60e3-49c7-81ff-b073be77e34b" (UID: "e329efba-60e3-49c7-81ff-b073be77e34b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.601381 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 15:27:27 crc kubenswrapper[4772]: W0127 15:27:27.604214 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41f85a83_f245_40ff_b994_50cab01b2530.slice/crio-70507e561102278ed4f801ac168676eb09960026059c30c45c5fe4950449c589 WatchSource:0}: Error finding container 70507e561102278ed4f801ac168676eb09960026059c30c45c5fe4950449c589: Status 404 returned error can't find the container with id 70507e561102278ed4f801ac168676eb09960026059c30c45c5fe4950449c589 Jan 27 15:27:27 crc kubenswrapper[4772]: I0127 15:27:27.694875 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e329efba-60e3-49c7-81ff-b073be77e34b-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.182082 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.182085 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-8d9vp" event={"ID":"e329efba-60e3-49c7-81ff-b073be77e34b","Type":"ContainerDied","Data":"dc924b979634b0cd0c7264ffb70a5a244bf22da4a19a82562f283a10b69f4841"} Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.182317 4772 scope.go:117] "RemoveContainer" containerID="0bc2eb78a83f1e9ddf6e0c975669640d497ff3e50951b0eaadaee82dc03caffd" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.185226 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"41f85a83-f245-40ff-b994-50cab01b2530","Type":"ContainerStarted","Data":"70507e561102278ed4f801ac168676eb09960026059c30c45c5fe4950449c589"} Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.187373 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c94a7cfa-28e2-4d52-85a1-d5586f162227","Type":"ContainerStarted","Data":"9775d2c5b4eda3cae695814a686a4a82d4426bf3d7d28a73dffa9b807c4c16b8"} Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.190400 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-677fb7d6fc-djjsx" event={"ID":"6e790127-8223-4b0c-8a5d-21e1bb15fa30","Type":"ContainerStarted","Data":"468321c234874e808e21c356adbece5162a3e84011f0215b573e541258fb76bc"} Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.190469 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-677fb7d6fc-djjsx" event={"ID":"6e790127-8223-4b0c-8a5d-21e1bb15fa30","Type":"ContainerStarted","Data":"8b30129bf5b3504ae600edeaafe66f62d6f0c11b788461d423310f03199da3c5"} Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.190636 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.232896 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-677fb7d6fc-djjsx" podStartSLOduration=5.232864742 podStartE2EDuration="5.232864742s" podCreationTimestamp="2026-01-27 15:27:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:27:28.224391365 +0000 UTC m=+1234.205000483" watchObservedRunningTime="2026-01-27 15:27:28.232864742 +0000 UTC m=+1234.213473840" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.251896 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-597699949b-q6msx"] Jan 27 15:27:28 crc kubenswrapper[4772]: E0127 15:27:28.252351 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6" containerName="placement-db-sync" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.252376 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6" containerName="placement-db-sync" Jan 27 15:27:28 crc kubenswrapper[4772]: E0127 15:27:28.252391 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e329efba-60e3-49c7-81ff-b073be77e34b" containerName="init" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.252397 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e329efba-60e3-49c7-81ff-b073be77e34b" containerName="init" Jan 27 15:27:28 crc kubenswrapper[4772]: E0127 15:27:28.252419 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e329efba-60e3-49c7-81ff-b073be77e34b" containerName="dnsmasq-dns" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.252424 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e329efba-60e3-49c7-81ff-b073be77e34b" containerName="dnsmasq-dns" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.252650 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6" containerName="placement-db-sync" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.252677 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="e329efba-60e3-49c7-81ff-b073be77e34b" containerName="dnsmasq-dns" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.253813 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.255919 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.255950 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.258043 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-4tg2g" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.258352 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.259238 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.259213538000001 podStartE2EDuration="8.259213538s" podCreationTimestamp="2026-01-27 15:27:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:27:28.253162022 +0000 UTC m=+1234.233771130" watchObservedRunningTime="2026-01-27 15:27:28.259213538 +0000 UTC m=+1234.239822646" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.260216 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.285924 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-597699949b-q6msx"] Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.304310 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-8d9vp"] Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.314488 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-8d9vp"] Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.409106 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-internal-tls-certs\") pod \"placement-597699949b-q6msx\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.409235 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-public-tls-certs\") pod \"placement-597699949b-q6msx\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.409267 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-scripts\") pod \"placement-597699949b-q6msx\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.409294 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-combined-ca-bundle\") pod \"placement-597699949b-q6msx\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.409322 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fknks\" (UniqueName: \"kubernetes.io/projected/4205dfea-7dc7-496a-9745-fc5e3d0a418a-kube-api-access-fknks\") pod \"placement-597699949b-q6msx\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.409573 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4205dfea-7dc7-496a-9745-fc5e3d0a418a-logs\") pod \"placement-597699949b-q6msx\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.409831 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-config-data\") pod \"placement-597699949b-q6msx\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.514650 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-config-data\") pod \"placement-597699949b-q6msx\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.514738 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-internal-tls-certs\") pod \"placement-597699949b-q6msx\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.514779 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-public-tls-certs\") pod \"placement-597699949b-q6msx\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.514805 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-scripts\") pod \"placement-597699949b-q6msx\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.514834 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-combined-ca-bundle\") pod \"placement-597699949b-q6msx\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.514856 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fknks\" (UniqueName: \"kubernetes.io/projected/4205dfea-7dc7-496a-9745-fc5e3d0a418a-kube-api-access-fknks\") pod \"placement-597699949b-q6msx\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.514939 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4205dfea-7dc7-496a-9745-fc5e3d0a418a-logs\") pod \"placement-597699949b-q6msx\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.517944 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4205dfea-7dc7-496a-9745-fc5e3d0a418a-logs\") pod \"placement-597699949b-q6msx\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.519001 4772 scope.go:117] "RemoveContainer" containerID="10196ecd014a671d2bb0c35a007cf89f6cc32f81e1a8290e8f8bb5f8f7575614" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.522395 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-public-tls-certs\") pod \"placement-597699949b-q6msx\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.522915 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-scripts\") pod \"placement-597699949b-q6msx\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.524310 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-internal-tls-certs\") pod \"placement-597699949b-q6msx\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.527982 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-config-data\") pod \"placement-597699949b-q6msx\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.528875 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-combined-ca-bundle\") pod \"placement-597699949b-q6msx\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.532118 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fknks\" (UniqueName: \"kubernetes.io/projected/4205dfea-7dc7-496a-9745-fc5e3d0a418a-kube-api-access-fknks\") pod \"placement-597699949b-q6msx\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.593617 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:28 crc kubenswrapper[4772]: I0127 15:27:28.683475 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e329efba-60e3-49c7-81ff-b073be77e34b" path="/var/lib/kubelet/pods/e329efba-60e3-49c7-81ff-b073be77e34b/volumes" Jan 27 15:27:29 crc kubenswrapper[4772]: I0127 15:27:29.201899 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"41f85a83-f245-40ff-b994-50cab01b2530","Type":"ContainerStarted","Data":"58d128e4a7f44cc529be47e9f224989cce3b8a08dc4e4f4d37d49e38c0c7b8d2"} Jan 27 15:27:29 crc kubenswrapper[4772]: I0127 15:27:29.204372 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de415c6e-4424-49c4-bc9d-076a5b13ab4e","Type":"ContainerStarted","Data":"b87da5e7b978350e6830e0f65fce50644eee1e1665a4ebcd45d4d0010f0f31d7"} Jan 27 15:27:29 crc kubenswrapper[4772]: I0127 15:27:29.303090 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-597699949b-q6msx"] Jan 27 15:27:29 crc kubenswrapper[4772]: W0127 15:27:29.353274 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4205dfea_7dc7_496a_9745_fc5e3d0a418a.slice/crio-77708c49aaa66488bf09da947ac24b469a4cd3c49071689cbd09cfa6aa9b79b5 WatchSource:0}: Error finding container 77708c49aaa66488bf09da947ac24b469a4cd3c49071689cbd09cfa6aa9b79b5: Status 404 returned error can't find the container with id 77708c49aaa66488bf09da947ac24b469a4cd3c49071689cbd09cfa6aa9b79b5 Jan 27 15:27:30 crc kubenswrapper[4772]: I0127 15:27:30.213964 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8l85z" event={"ID":"9ae05919-68bf-43d1-abd9-9908ec287bd0","Type":"ContainerStarted","Data":"f581dd644d182efa5f740dc0b5a2f4adfb865bef3f027972802161889179f1d4"} Jan 27 15:27:30 crc kubenswrapper[4772]: I0127 15:27:30.227208 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-597699949b-q6msx" event={"ID":"4205dfea-7dc7-496a-9745-fc5e3d0a418a","Type":"ContainerStarted","Data":"ad26ca4835a223df0b0aa3065e02d9e54b67030d2b6d0436f1f1a0dd7bf06415"} Jan 27 15:27:30 crc kubenswrapper[4772]: I0127 15:27:30.227251 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-597699949b-q6msx" event={"ID":"4205dfea-7dc7-496a-9745-fc5e3d0a418a","Type":"ContainerStarted","Data":"f10ed54f4ea68e56be83b8d8387a9768612b5c035b1fc42928132066af5bd689"} Jan 27 15:27:30 crc kubenswrapper[4772]: I0127 15:27:30.227264 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-597699949b-q6msx" event={"ID":"4205dfea-7dc7-496a-9745-fc5e3d0a418a","Type":"ContainerStarted","Data":"77708c49aaa66488bf09da947ac24b469a4cd3c49071689cbd09cfa6aa9b79b5"} Jan 27 15:27:30 crc kubenswrapper[4772]: I0127 15:27:30.227825 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:30 crc kubenswrapper[4772]: I0127 15:27:30.227854 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-597699949b-q6msx" Jan 27 15:27:30 crc kubenswrapper[4772]: I0127 15:27:30.233710 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pmk27" event={"ID":"5a423229-06be-4934-9715-58105e1af686","Type":"ContainerStarted","Data":"6ba95c7bf22c812cf8d7d855d86c702f5f7f90db05ec7fc2281ddec549f7d67b"} Jan 27 15:27:30 crc kubenswrapper[4772]: I0127 15:27:30.236435 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"41f85a83-f245-40ff-b994-50cab01b2530","Type":"ContainerStarted","Data":"f1accbd1db4a8c2dce7512a2eb2abaa265e29ed373b0fc121d29515c5bba0e55"} Jan 27 15:27:30 crc kubenswrapper[4772]: I0127 15:27:30.251096 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-8l85z" podStartSLOduration=2.19845139 podStartE2EDuration="42.251073127s" podCreationTimestamp="2026-01-27 15:26:48 +0000 UTC" firstStartedPulling="2026-01-27 15:26:49.310600793 +0000 UTC m=+1195.291209891" lastFinishedPulling="2026-01-27 15:27:29.36322253 +0000 UTC m=+1235.343831628" observedRunningTime="2026-01-27 15:27:30.24258258 +0000 UTC m=+1236.223191678" watchObservedRunningTime="2026-01-27 15:27:30.251073127 +0000 UTC m=+1236.231682225" Jan 27 15:27:30 crc kubenswrapper[4772]: I0127 15:27:30.277072 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-597699949b-q6msx" podStartSLOduration=2.277054472 podStartE2EDuration="2.277054472s" podCreationTimestamp="2026-01-27 15:27:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:27:30.269883874 +0000 UTC m=+1236.250492982" watchObservedRunningTime="2026-01-27 15:27:30.277054472 +0000 UTC m=+1236.257663570" Jan 27 15:27:30 crc kubenswrapper[4772]: I0127 15:27:30.302619 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-pmk27" podStartSLOduration=2.699131315 podStartE2EDuration="43.302595605s" podCreationTimestamp="2026-01-27 15:26:47 +0000 UTC" firstStartedPulling="2026-01-27 15:26:49.217989205 +0000 UTC m=+1195.198598303" lastFinishedPulling="2026-01-27 15:27:29.821453495 +0000 UTC m=+1235.802062593" observedRunningTime="2026-01-27 15:27:30.292606624 +0000 UTC m=+1236.273215732" watchObservedRunningTime="2026-01-27 15:27:30.302595605 +0000 UTC m=+1236.283204703" Jan 27 15:27:30 crc kubenswrapper[4772]: I0127 15:27:30.313528 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=9.313517882 podStartE2EDuration="9.313517882s" podCreationTimestamp="2026-01-27 15:27:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:27:30.311067241 +0000 UTC m=+1236.291676349" watchObservedRunningTime="2026-01-27 15:27:30.313517882 +0000 UTC m=+1236.294126980" Jan 27 15:27:30 crc kubenswrapper[4772]: I0127 15:27:30.767040 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 27 15:27:30 crc kubenswrapper[4772]: I0127 15:27:30.767101 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 27 15:27:30 crc kubenswrapper[4772]: I0127 15:27:30.804250 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 27 15:27:30 crc kubenswrapper[4772]: I0127 15:27:30.819293 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 27 15:27:31 crc kubenswrapper[4772]: I0127 15:27:31.256911 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 27 15:27:31 crc kubenswrapper[4772]: I0127 15:27:31.256952 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 27 15:27:31 crc kubenswrapper[4772]: I0127 15:27:31.485830 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 27 15:27:31 crc kubenswrapper[4772]: I0127 15:27:31.486070 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 27 15:27:31 crc kubenswrapper[4772]: I0127 15:27:31.522874 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 27 15:27:31 crc kubenswrapper[4772]: I0127 15:27:31.537126 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 27 15:27:32 crc kubenswrapper[4772]: I0127 15:27:32.264550 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 27 15:27:32 crc kubenswrapper[4772]: I0127 15:27:32.264879 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 27 15:27:33 crc kubenswrapper[4772]: I0127 15:27:33.211950 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 27 15:27:34 crc kubenswrapper[4772]: I0127 15:27:34.166558 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 27 15:27:39 crc kubenswrapper[4772]: I0127 15:27:39.333053 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de415c6e-4424-49c4-bc9d-076a5b13ab4e","Type":"ContainerStarted","Data":"03f8da2d80772e659c36db9a1b10a6be24dc704eb86ce89c04a5a14351b7726d"} Jan 27 15:27:39 crc kubenswrapper[4772]: I0127 15:27:39.333545 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de415c6e-4424-49c4-bc9d-076a5b13ab4e" containerName="ceilometer-central-agent" containerID="cri-o://f76b5eae8b9d1fd746edffe9a9f5a02ca0ad4ea09665e63c5dbeacff4753fa40" gracePeriod=30 Jan 27 15:27:39 crc kubenswrapper[4772]: I0127 15:27:39.333891 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 27 15:27:39 crc kubenswrapper[4772]: I0127 15:27:39.334054 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de415c6e-4424-49c4-bc9d-076a5b13ab4e" containerName="proxy-httpd" containerID="cri-o://03f8da2d80772e659c36db9a1b10a6be24dc704eb86ce89c04a5a14351b7726d" gracePeriod=30 Jan 27 15:27:39 crc kubenswrapper[4772]: I0127 15:27:39.334274 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de415c6e-4424-49c4-bc9d-076a5b13ab4e" containerName="sg-core" containerID="cri-o://b87da5e7b978350e6830e0f65fce50644eee1e1665a4ebcd45d4d0010f0f31d7" gracePeriod=30 Jan 27 15:27:39 crc kubenswrapper[4772]: I0127 15:27:39.334377 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de415c6e-4424-49c4-bc9d-076a5b13ab4e" containerName="ceilometer-notification-agent" containerID="cri-o://ed51d0aa4ae1c7166bbf0464f2b405f79a0faa50f99c4244c9717d1a1fd81db2" gracePeriod=30 Jan 27 15:27:39 crc kubenswrapper[4772]: I0127 15:27:39.371826 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.163481096 podStartE2EDuration="52.371798099s" podCreationTimestamp="2026-01-27 15:26:47 +0000 UTC" firstStartedPulling="2026-01-27 15:26:49.076431467 +0000 UTC m=+1195.057040565" lastFinishedPulling="2026-01-27 15:27:38.28474847 +0000 UTC m=+1244.265357568" observedRunningTime="2026-01-27 15:27:39.363124797 +0000 UTC m=+1245.343733985" watchObservedRunningTime="2026-01-27 15:27:39.371798099 +0000 UTC m=+1245.352407217" Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.343985 4772 generic.go:334] "Generic (PLEG): container finished" podID="de415c6e-4424-49c4-bc9d-076a5b13ab4e" containerID="03f8da2d80772e659c36db9a1b10a6be24dc704eb86ce89c04a5a14351b7726d" exitCode=0 Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.344469 4772 generic.go:334] "Generic (PLEG): container finished" podID="de415c6e-4424-49c4-bc9d-076a5b13ab4e" containerID="b87da5e7b978350e6830e0f65fce50644eee1e1665a4ebcd45d4d0010f0f31d7" exitCode=2 Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.344478 4772 generic.go:334] "Generic (PLEG): container finished" podID="de415c6e-4424-49c4-bc9d-076a5b13ab4e" containerID="ed51d0aa4ae1c7166bbf0464f2b405f79a0faa50f99c4244c9717d1a1fd81db2" exitCode=0 Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.344486 4772 generic.go:334] "Generic (PLEG): container finished" podID="de415c6e-4424-49c4-bc9d-076a5b13ab4e" containerID="f76b5eae8b9d1fd746edffe9a9f5a02ca0ad4ea09665e63c5dbeacff4753fa40" exitCode=0 Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.344079 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de415c6e-4424-49c4-bc9d-076a5b13ab4e","Type":"ContainerDied","Data":"03f8da2d80772e659c36db9a1b10a6be24dc704eb86ce89c04a5a14351b7726d"} Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.344518 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de415c6e-4424-49c4-bc9d-076a5b13ab4e","Type":"ContainerDied","Data":"b87da5e7b978350e6830e0f65fce50644eee1e1665a4ebcd45d4d0010f0f31d7"} Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.344532 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de415c6e-4424-49c4-bc9d-076a5b13ab4e","Type":"ContainerDied","Data":"ed51d0aa4ae1c7166bbf0464f2b405f79a0faa50f99c4244c9717d1a1fd81db2"} Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.344541 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de415c6e-4424-49c4-bc9d-076a5b13ab4e","Type":"ContainerDied","Data":"f76b5eae8b9d1fd746edffe9a9f5a02ca0ad4ea09665e63c5dbeacff4753fa40"} Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.344550 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de415c6e-4424-49c4-bc9d-076a5b13ab4e","Type":"ContainerDied","Data":"a0254acb416eb806ca40cead3274ef3b55185c0cdbabec25da60a2a08040318a"} Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.344559 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0254acb416eb806ca40cead3274ef3b55185c0cdbabec25da60a2a08040318a" Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.362482 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.461586 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de415c6e-4424-49c4-bc9d-076a5b13ab4e-combined-ca-bundle\") pod \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.461668 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de415c6e-4424-49c4-bc9d-076a5b13ab4e-config-data\") pod \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.461796 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de415c6e-4424-49c4-bc9d-076a5b13ab4e-scripts\") pod \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.461839 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de415c6e-4424-49c4-bc9d-076a5b13ab4e-sg-core-conf-yaml\") pod \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.461887 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de415c6e-4424-49c4-bc9d-076a5b13ab4e-run-httpd\") pod \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.461916 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvxkw\" (UniqueName: \"kubernetes.io/projected/de415c6e-4424-49c4-bc9d-076a5b13ab4e-kube-api-access-fvxkw\") pod \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.461978 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de415c6e-4424-49c4-bc9d-076a5b13ab4e-log-httpd\") pod \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\" (UID: \"de415c6e-4424-49c4-bc9d-076a5b13ab4e\") " Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.462371 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de415c6e-4424-49c4-bc9d-076a5b13ab4e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "de415c6e-4424-49c4-bc9d-076a5b13ab4e" (UID: "de415c6e-4424-49c4-bc9d-076a5b13ab4e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.462485 4772 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de415c6e-4424-49c4-bc9d-076a5b13ab4e-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.463209 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de415c6e-4424-49c4-bc9d-076a5b13ab4e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "de415c6e-4424-49c4-bc9d-076a5b13ab4e" (UID: "de415c6e-4424-49c4-bc9d-076a5b13ab4e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.467458 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de415c6e-4424-49c4-bc9d-076a5b13ab4e-kube-api-access-fvxkw" (OuterVolumeSpecName: "kube-api-access-fvxkw") pod "de415c6e-4424-49c4-bc9d-076a5b13ab4e" (UID: "de415c6e-4424-49c4-bc9d-076a5b13ab4e"). InnerVolumeSpecName "kube-api-access-fvxkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.468034 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de415c6e-4424-49c4-bc9d-076a5b13ab4e-scripts" (OuterVolumeSpecName: "scripts") pod "de415c6e-4424-49c4-bc9d-076a5b13ab4e" (UID: "de415c6e-4424-49c4-bc9d-076a5b13ab4e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.493715 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de415c6e-4424-49c4-bc9d-076a5b13ab4e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "de415c6e-4424-49c4-bc9d-076a5b13ab4e" (UID: "de415c6e-4424-49c4-bc9d-076a5b13ab4e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.525470 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de415c6e-4424-49c4-bc9d-076a5b13ab4e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de415c6e-4424-49c4-bc9d-076a5b13ab4e" (UID: "de415c6e-4424-49c4-bc9d-076a5b13ab4e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.547092 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de415c6e-4424-49c4-bc9d-076a5b13ab4e-config-data" (OuterVolumeSpecName: "config-data") pod "de415c6e-4424-49c4-bc9d-076a5b13ab4e" (UID: "de415c6e-4424-49c4-bc9d-076a5b13ab4e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.564289 4772 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de415c6e-4424-49c4-bc9d-076a5b13ab4e-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.564369 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de415c6e-4424-49c4-bc9d-076a5b13ab4e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.564380 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de415c6e-4424-49c4-bc9d-076a5b13ab4e-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.564389 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de415c6e-4424-49c4-bc9d-076a5b13ab4e-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.564400 4772 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de415c6e-4424-49c4-bc9d-076a5b13ab4e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:40 crc kubenswrapper[4772]: I0127 15:27:40.564408 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvxkw\" (UniqueName: \"kubernetes.io/projected/de415c6e-4424-49c4-bc9d-076a5b13ab4e-kube-api-access-fvxkw\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.352524 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.376939 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.383825 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.394975 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:27:41 crc kubenswrapper[4772]: E0127 15:27:41.395380 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de415c6e-4424-49c4-bc9d-076a5b13ab4e" containerName="ceilometer-central-agent" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.395407 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="de415c6e-4424-49c4-bc9d-076a5b13ab4e" containerName="ceilometer-central-agent" Jan 27 15:27:41 crc kubenswrapper[4772]: E0127 15:27:41.395421 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de415c6e-4424-49c4-bc9d-076a5b13ab4e" containerName="sg-core" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.395430 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="de415c6e-4424-49c4-bc9d-076a5b13ab4e" containerName="sg-core" Jan 27 15:27:41 crc kubenswrapper[4772]: E0127 15:27:41.395459 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de415c6e-4424-49c4-bc9d-076a5b13ab4e" containerName="proxy-httpd" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.395467 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="de415c6e-4424-49c4-bc9d-076a5b13ab4e" containerName="proxy-httpd" Jan 27 15:27:41 crc kubenswrapper[4772]: E0127 15:27:41.395489 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de415c6e-4424-49c4-bc9d-076a5b13ab4e" containerName="ceilometer-notification-agent" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.395498 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="de415c6e-4424-49c4-bc9d-076a5b13ab4e" containerName="ceilometer-notification-agent" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.395722 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="de415c6e-4424-49c4-bc9d-076a5b13ab4e" containerName="sg-core" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.395769 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="de415c6e-4424-49c4-bc9d-076a5b13ab4e" containerName="ceilometer-central-agent" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.395779 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="de415c6e-4424-49c4-bc9d-076a5b13ab4e" containerName="proxy-httpd" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.395813 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="de415c6e-4424-49c4-bc9d-076a5b13ab4e" containerName="ceilometer-notification-agent" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.397669 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.399668 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.400276 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.410151 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.477517 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8c7b9cc-5427-4c3a-92d7-cec9760975df-scripts\") pod \"ceilometer-0\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " pod="openstack/ceilometer-0" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.477951 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d8c7b9cc-5427-4c3a-92d7-cec9760975df-log-httpd\") pod \"ceilometer-0\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " pod="openstack/ceilometer-0" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.477994 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8c7b9cc-5427-4c3a-92d7-cec9760975df-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " pod="openstack/ceilometer-0" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.478019 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnnbp\" (UniqueName: \"kubernetes.io/projected/d8c7b9cc-5427-4c3a-92d7-cec9760975df-kube-api-access-tnnbp\") pod \"ceilometer-0\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " pod="openstack/ceilometer-0" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.478121 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d8c7b9cc-5427-4c3a-92d7-cec9760975df-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " pod="openstack/ceilometer-0" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.478318 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8c7b9cc-5427-4c3a-92d7-cec9760975df-config-data\") pod \"ceilometer-0\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " pod="openstack/ceilometer-0" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.478400 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d8c7b9cc-5427-4c3a-92d7-cec9760975df-run-httpd\") pod \"ceilometer-0\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " pod="openstack/ceilometer-0" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.488256 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:27:41 crc kubenswrapper[4772]: E0127 15:27:41.489005 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data kube-api-access-tnnbp log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="d8c7b9cc-5427-4c3a-92d7-cec9760975df" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.580343 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8c7b9cc-5427-4c3a-92d7-cec9760975df-config-data\") pod \"ceilometer-0\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " pod="openstack/ceilometer-0" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.580405 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d8c7b9cc-5427-4c3a-92d7-cec9760975df-run-httpd\") pod \"ceilometer-0\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " pod="openstack/ceilometer-0" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.580444 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8c7b9cc-5427-4c3a-92d7-cec9760975df-scripts\") pod \"ceilometer-0\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " pod="openstack/ceilometer-0" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.580467 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d8c7b9cc-5427-4c3a-92d7-cec9760975df-log-httpd\") pod \"ceilometer-0\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " pod="openstack/ceilometer-0" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.580497 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnnbp\" (UniqueName: \"kubernetes.io/projected/d8c7b9cc-5427-4c3a-92d7-cec9760975df-kube-api-access-tnnbp\") pod \"ceilometer-0\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " pod="openstack/ceilometer-0" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.580515 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8c7b9cc-5427-4c3a-92d7-cec9760975df-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " pod="openstack/ceilometer-0" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.580549 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d8c7b9cc-5427-4c3a-92d7-cec9760975df-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " pod="openstack/ceilometer-0" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.581101 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d8c7b9cc-5427-4c3a-92d7-cec9760975df-log-httpd\") pod \"ceilometer-0\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " pod="openstack/ceilometer-0" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.581355 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d8c7b9cc-5427-4c3a-92d7-cec9760975df-run-httpd\") pod \"ceilometer-0\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " pod="openstack/ceilometer-0" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.585787 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d8c7b9cc-5427-4c3a-92d7-cec9760975df-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " pod="openstack/ceilometer-0" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.587073 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8c7b9cc-5427-4c3a-92d7-cec9760975df-config-data\") pod \"ceilometer-0\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " pod="openstack/ceilometer-0" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.588018 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8c7b9cc-5427-4c3a-92d7-cec9760975df-scripts\") pod \"ceilometer-0\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " pod="openstack/ceilometer-0" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.589134 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8c7b9cc-5427-4c3a-92d7-cec9760975df-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " pod="openstack/ceilometer-0" Jan 27 15:27:41 crc kubenswrapper[4772]: I0127 15:27:41.597202 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnnbp\" (UniqueName: \"kubernetes.io/projected/d8c7b9cc-5427-4c3a-92d7-cec9760975df-kube-api-access-tnnbp\") pod \"ceilometer-0\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " pod="openstack/ceilometer-0" Jan 27 15:27:42 crc kubenswrapper[4772]: I0127 15:27:42.361202 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:27:42 crc kubenswrapper[4772]: I0127 15:27:42.371445 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:27:42 crc kubenswrapper[4772]: I0127 15:27:42.495279 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d8c7b9cc-5427-4c3a-92d7-cec9760975df-sg-core-conf-yaml\") pod \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " Jan 27 15:27:42 crc kubenswrapper[4772]: I0127 15:27:42.495410 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8c7b9cc-5427-4c3a-92d7-cec9760975df-scripts\") pod \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " Jan 27 15:27:42 crc kubenswrapper[4772]: I0127 15:27:42.495447 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8c7b9cc-5427-4c3a-92d7-cec9760975df-config-data\") pod \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " Jan 27 15:27:42 crc kubenswrapper[4772]: I0127 15:27:42.495497 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d8c7b9cc-5427-4c3a-92d7-cec9760975df-log-httpd\") pod \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " Jan 27 15:27:42 crc kubenswrapper[4772]: I0127 15:27:42.495568 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnnbp\" (UniqueName: \"kubernetes.io/projected/d8c7b9cc-5427-4c3a-92d7-cec9760975df-kube-api-access-tnnbp\") pod \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " Jan 27 15:27:42 crc kubenswrapper[4772]: I0127 15:27:42.495619 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8c7b9cc-5427-4c3a-92d7-cec9760975df-combined-ca-bundle\") pod \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " Jan 27 15:27:42 crc kubenswrapper[4772]: I0127 15:27:42.495638 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d8c7b9cc-5427-4c3a-92d7-cec9760975df-run-httpd\") pod \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\" (UID: \"d8c7b9cc-5427-4c3a-92d7-cec9760975df\") " Jan 27 15:27:42 crc kubenswrapper[4772]: I0127 15:27:42.495980 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8c7b9cc-5427-4c3a-92d7-cec9760975df-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d8c7b9cc-5427-4c3a-92d7-cec9760975df" (UID: "d8c7b9cc-5427-4c3a-92d7-cec9760975df"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:27:42 crc kubenswrapper[4772]: I0127 15:27:42.496088 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8c7b9cc-5427-4c3a-92d7-cec9760975df-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d8c7b9cc-5427-4c3a-92d7-cec9760975df" (UID: "d8c7b9cc-5427-4c3a-92d7-cec9760975df"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:27:42 crc kubenswrapper[4772]: I0127 15:27:42.496306 4772 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d8c7b9cc-5427-4c3a-92d7-cec9760975df-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:42 crc kubenswrapper[4772]: I0127 15:27:42.496321 4772 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d8c7b9cc-5427-4c3a-92d7-cec9760975df-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:42 crc kubenswrapper[4772]: I0127 15:27:42.499357 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8c7b9cc-5427-4c3a-92d7-cec9760975df-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d8c7b9cc-5427-4c3a-92d7-cec9760975df" (UID: "d8c7b9cc-5427-4c3a-92d7-cec9760975df"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:42 crc kubenswrapper[4772]: I0127 15:27:42.499469 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8c7b9cc-5427-4c3a-92d7-cec9760975df-kube-api-access-tnnbp" (OuterVolumeSpecName: "kube-api-access-tnnbp") pod "d8c7b9cc-5427-4c3a-92d7-cec9760975df" (UID: "d8c7b9cc-5427-4c3a-92d7-cec9760975df"). InnerVolumeSpecName "kube-api-access-tnnbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:27:42 crc kubenswrapper[4772]: I0127 15:27:42.499861 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8c7b9cc-5427-4c3a-92d7-cec9760975df-config-data" (OuterVolumeSpecName: "config-data") pod "d8c7b9cc-5427-4c3a-92d7-cec9760975df" (UID: "d8c7b9cc-5427-4c3a-92d7-cec9760975df"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:42 crc kubenswrapper[4772]: I0127 15:27:42.500955 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8c7b9cc-5427-4c3a-92d7-cec9760975df-scripts" (OuterVolumeSpecName: "scripts") pod "d8c7b9cc-5427-4c3a-92d7-cec9760975df" (UID: "d8c7b9cc-5427-4c3a-92d7-cec9760975df"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:42 crc kubenswrapper[4772]: I0127 15:27:42.502667 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8c7b9cc-5427-4c3a-92d7-cec9760975df-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d8c7b9cc-5427-4c3a-92d7-cec9760975df" (UID: "d8c7b9cc-5427-4c3a-92d7-cec9760975df"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:42 crc kubenswrapper[4772]: I0127 15:27:42.597292 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnnbp\" (UniqueName: \"kubernetes.io/projected/d8c7b9cc-5427-4c3a-92d7-cec9760975df-kube-api-access-tnnbp\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:42 crc kubenswrapper[4772]: I0127 15:27:42.597321 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8c7b9cc-5427-4c3a-92d7-cec9760975df-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:42 crc kubenswrapper[4772]: I0127 15:27:42.597330 4772 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d8c7b9cc-5427-4c3a-92d7-cec9760975df-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:42 crc kubenswrapper[4772]: I0127 15:27:42.597338 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8c7b9cc-5427-4c3a-92d7-cec9760975df-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:42 crc kubenswrapper[4772]: I0127 15:27:42.597346 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8c7b9cc-5427-4c3a-92d7-cec9760975df-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:42 crc kubenswrapper[4772]: I0127 15:27:42.676347 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de415c6e-4424-49c4-bc9d-076a5b13ab4e" path="/var/lib/kubelet/pods/de415c6e-4424-49c4-bc9d-076a5b13ab4e/volumes" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.372656 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.421544 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.432682 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.442279 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.445246 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.447290 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.451333 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.460357 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.511964 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8583377-67ef-4cca-83bb-08d7523ab0a8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " pod="openstack/ceilometer-0" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.512025 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8583377-67ef-4cca-83bb-08d7523ab0a8-config-data\") pod \"ceilometer-0\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " pod="openstack/ceilometer-0" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.512081 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jjsh\" (UniqueName: \"kubernetes.io/projected/f8583377-67ef-4cca-83bb-08d7523ab0a8-kube-api-access-9jjsh\") pod \"ceilometer-0\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " pod="openstack/ceilometer-0" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.512114 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f8583377-67ef-4cca-83bb-08d7523ab0a8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " pod="openstack/ceilometer-0" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.512143 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8583377-67ef-4cca-83bb-08d7523ab0a8-scripts\") pod \"ceilometer-0\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " pod="openstack/ceilometer-0" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.512273 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8583377-67ef-4cca-83bb-08d7523ab0a8-run-httpd\") pod \"ceilometer-0\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " pod="openstack/ceilometer-0" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.512382 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8583377-67ef-4cca-83bb-08d7523ab0a8-log-httpd\") pod \"ceilometer-0\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " pod="openstack/ceilometer-0" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.613680 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8583377-67ef-4cca-83bb-08d7523ab0a8-run-httpd\") pod \"ceilometer-0\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " pod="openstack/ceilometer-0" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.613804 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8583377-67ef-4cca-83bb-08d7523ab0a8-log-httpd\") pod \"ceilometer-0\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " pod="openstack/ceilometer-0" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.613886 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8583377-67ef-4cca-83bb-08d7523ab0a8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " pod="openstack/ceilometer-0" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.613934 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8583377-67ef-4cca-83bb-08d7523ab0a8-config-data\") pod \"ceilometer-0\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " pod="openstack/ceilometer-0" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.614018 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jjsh\" (UniqueName: \"kubernetes.io/projected/f8583377-67ef-4cca-83bb-08d7523ab0a8-kube-api-access-9jjsh\") pod \"ceilometer-0\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " pod="openstack/ceilometer-0" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.614084 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f8583377-67ef-4cca-83bb-08d7523ab0a8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " pod="openstack/ceilometer-0" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.614161 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8583377-67ef-4cca-83bb-08d7523ab0a8-scripts\") pod \"ceilometer-0\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " pod="openstack/ceilometer-0" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.614324 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8583377-67ef-4cca-83bb-08d7523ab0a8-run-httpd\") pod \"ceilometer-0\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " pod="openstack/ceilometer-0" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.614844 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8583377-67ef-4cca-83bb-08d7523ab0a8-log-httpd\") pod \"ceilometer-0\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " pod="openstack/ceilometer-0" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.620115 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8583377-67ef-4cca-83bb-08d7523ab0a8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " pod="openstack/ceilometer-0" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.620208 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8583377-67ef-4cca-83bb-08d7523ab0a8-config-data\") pod \"ceilometer-0\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " pod="openstack/ceilometer-0" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.626003 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8583377-67ef-4cca-83bb-08d7523ab0a8-scripts\") pod \"ceilometer-0\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " pod="openstack/ceilometer-0" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.628207 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f8583377-67ef-4cca-83bb-08d7523ab0a8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " pod="openstack/ceilometer-0" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.640354 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jjsh\" (UniqueName: \"kubernetes.io/projected/f8583377-67ef-4cca-83bb-08d7523ab0a8-kube-api-access-9jjsh\") pod \"ceilometer-0\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " pod="openstack/ceilometer-0" Jan 27 15:27:43 crc kubenswrapper[4772]: I0127 15:27:43.777553 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:27:44 crc kubenswrapper[4772]: I0127 15:27:44.251349 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:27:44 crc kubenswrapper[4772]: I0127 15:27:44.382526 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8583377-67ef-4cca-83bb-08d7523ab0a8","Type":"ContainerStarted","Data":"561a737e7b03fa9b4f9fd3fd05ad062a1b5523a6f22e1965b09d595f10adafd3"} Jan 27 15:27:44 crc kubenswrapper[4772]: I0127 15:27:44.675520 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8c7b9cc-5427-4c3a-92d7-cec9760975df" path="/var/lib/kubelet/pods/d8c7b9cc-5427-4c3a-92d7-cec9760975df/volumes" Jan 27 15:27:45 crc kubenswrapper[4772]: I0127 15:27:45.393064 4772 generic.go:334] "Generic (PLEG): container finished" podID="5a423229-06be-4934-9715-58105e1af686" containerID="6ba95c7bf22c812cf8d7d855d86c702f5f7f90db05ec7fc2281ddec549f7d67b" exitCode=0 Jan 27 15:27:45 crc kubenswrapper[4772]: I0127 15:27:45.393109 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pmk27" event={"ID":"5a423229-06be-4934-9715-58105e1af686","Type":"ContainerDied","Data":"6ba95c7bf22c812cf8d7d855d86c702f5f7f90db05ec7fc2281ddec549f7d67b"} Jan 27 15:27:45 crc kubenswrapper[4772]: I0127 15:27:45.395117 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8583377-67ef-4cca-83bb-08d7523ab0a8","Type":"ContainerStarted","Data":"941b08834b6f5b8dafbc182c67d3e458a94c7299ea32b8afd698f876b68ea015"} Jan 27 15:27:46 crc kubenswrapper[4772]: I0127 15:27:46.409858 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8583377-67ef-4cca-83bb-08d7523ab0a8","Type":"ContainerStarted","Data":"99f37d09a547a41878834ffbee7e0a0b90552016b42313fae983e9915266d761"} Jan 27 15:27:46 crc kubenswrapper[4772]: I0127 15:27:46.746584 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pmk27" Jan 27 15:27:46 crc kubenswrapper[4772]: I0127 15:27:46.773589 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a423229-06be-4934-9715-58105e1af686-combined-ca-bundle\") pod \"5a423229-06be-4934-9715-58105e1af686\" (UID: \"5a423229-06be-4934-9715-58105e1af686\") " Jan 27 15:27:46 crc kubenswrapper[4772]: I0127 15:27:46.773654 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5a423229-06be-4934-9715-58105e1af686-db-sync-config-data\") pod \"5a423229-06be-4934-9715-58105e1af686\" (UID: \"5a423229-06be-4934-9715-58105e1af686\") " Jan 27 15:27:46 crc kubenswrapper[4772]: I0127 15:27:46.773691 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cvd2\" (UniqueName: \"kubernetes.io/projected/5a423229-06be-4934-9715-58105e1af686-kube-api-access-8cvd2\") pod \"5a423229-06be-4934-9715-58105e1af686\" (UID: \"5a423229-06be-4934-9715-58105e1af686\") " Jan 27 15:27:46 crc kubenswrapper[4772]: I0127 15:27:46.790834 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a423229-06be-4934-9715-58105e1af686-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "5a423229-06be-4934-9715-58105e1af686" (UID: "5a423229-06be-4934-9715-58105e1af686"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:46 crc kubenswrapper[4772]: I0127 15:27:46.793053 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a423229-06be-4934-9715-58105e1af686-kube-api-access-8cvd2" (OuterVolumeSpecName: "kube-api-access-8cvd2") pod "5a423229-06be-4934-9715-58105e1af686" (UID: "5a423229-06be-4934-9715-58105e1af686"). InnerVolumeSpecName "kube-api-access-8cvd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:27:46 crc kubenswrapper[4772]: I0127 15:27:46.807362 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a423229-06be-4934-9715-58105e1af686-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a423229-06be-4934-9715-58105e1af686" (UID: "5a423229-06be-4934-9715-58105e1af686"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:46 crc kubenswrapper[4772]: I0127 15:27:46.876349 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a423229-06be-4934-9715-58105e1af686-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:46 crc kubenswrapper[4772]: I0127 15:27:46.877020 4772 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5a423229-06be-4934-9715-58105e1af686-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:46 crc kubenswrapper[4772]: I0127 15:27:46.877056 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cvd2\" (UniqueName: \"kubernetes.io/projected/5a423229-06be-4934-9715-58105e1af686-kube-api-access-8cvd2\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.420025 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8583377-67ef-4cca-83bb-08d7523ab0a8","Type":"ContainerStarted","Data":"ea8686b5fbb3cb04fd3d0cb81bec48b421aa5e6e9be9af4a4ad0ccc951c6bce4"} Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.421787 4772 generic.go:334] "Generic (PLEG): container finished" podID="9ae05919-68bf-43d1-abd9-9908ec287bd0" containerID="f581dd644d182efa5f740dc0b5a2f4adfb865bef3f027972802161889179f1d4" exitCode=0 Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.421859 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8l85z" event={"ID":"9ae05919-68bf-43d1-abd9-9908ec287bd0","Type":"ContainerDied","Data":"f581dd644d182efa5f740dc0b5a2f4adfb865bef3f027972802161889179f1d4"} Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.424700 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pmk27" event={"ID":"5a423229-06be-4934-9715-58105e1af686","Type":"ContainerDied","Data":"604f080c5b545eb272e78d6599f0497ec22c32b54d41f8331dbefcd9a29b19de"} Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.424800 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="604f080c5b545eb272e78d6599f0497ec22c32b54d41f8331dbefcd9a29b19de" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.424756 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pmk27" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.682892 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-6748df9c8c-zk7zp"] Jan 27 15:27:47 crc kubenswrapper[4772]: E0127 15:27:47.683309 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a423229-06be-4934-9715-58105e1af686" containerName="barbican-db-sync" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.683326 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a423229-06be-4934-9715-58105e1af686" containerName="barbican-db-sync" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.683551 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a423229-06be-4934-9715-58105e1af686" containerName="barbican-db-sync" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.684531 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6748df9c8c-zk7zp" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.692971 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-556764fb84-r628x"] Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.694789 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-556764fb84-r628x" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.698416 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.698641 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-flljj" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.698812 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.699446 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.708056 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ce27714-673f-47de-acc3-b6902b534bdd-combined-ca-bundle\") pod \"barbican-keystone-listener-556764fb84-r628x\" (UID: \"4ce27714-673f-47de-acc3-b6902b534bdd\") " pod="openstack/barbican-keystone-listener-556764fb84-r628x" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.708135 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmmsl\" (UniqueName: \"kubernetes.io/projected/4ce27714-673f-47de-acc3-b6902b534bdd-kube-api-access-pmmsl\") pod \"barbican-keystone-listener-556764fb84-r628x\" (UID: \"4ce27714-673f-47de-acc3-b6902b534bdd\") " pod="openstack/barbican-keystone-listener-556764fb84-r628x" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.708198 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ce27714-673f-47de-acc3-b6902b534bdd-config-data\") pod \"barbican-keystone-listener-556764fb84-r628x\" (UID: \"4ce27714-673f-47de-acc3-b6902b534bdd\") " pod="openstack/barbican-keystone-listener-556764fb84-r628x" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.708248 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ce27714-673f-47de-acc3-b6902b534bdd-config-data-custom\") pod \"barbican-keystone-listener-556764fb84-r628x\" (UID: \"4ce27714-673f-47de-acc3-b6902b534bdd\") " pod="openstack/barbican-keystone-listener-556764fb84-r628x" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.708277 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ce27714-673f-47de-acc3-b6902b534bdd-logs\") pod \"barbican-keystone-listener-556764fb84-r628x\" (UID: \"4ce27714-673f-47de-acc3-b6902b534bdd\") " pod="openstack/barbican-keystone-listener-556764fb84-r628x" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.720538 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6748df9c8c-zk7zp"] Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.739089 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-556764fb84-r628x"] Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.788104 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-62kx4"] Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.789689 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.803742 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-62kx4"] Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.815362 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ce27714-673f-47de-acc3-b6902b534bdd-config-data\") pod \"barbican-keystone-listener-556764fb84-r628x\" (UID: \"4ce27714-673f-47de-acc3-b6902b534bdd\") " pod="openstack/barbican-keystone-listener-556764fb84-r628x" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.815422 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ce27714-673f-47de-acc3-b6902b534bdd-config-data-custom\") pod \"barbican-keystone-listener-556764fb84-r628x\" (UID: \"4ce27714-673f-47de-acc3-b6902b534bdd\") " pod="openstack/barbican-keystone-listener-556764fb84-r628x" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.815465 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ce27714-673f-47de-acc3-b6902b534bdd-logs\") pod \"barbican-keystone-listener-556764fb84-r628x\" (UID: \"4ce27714-673f-47de-acc3-b6902b534bdd\") " pod="openstack/barbican-keystone-listener-556764fb84-r628x" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.815493 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/710edaa6-ba83-4b1f-a49a-769ca1911c9b-combined-ca-bundle\") pod \"barbican-worker-6748df9c8c-zk7zp\" (UID: \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\") " pod="openstack/barbican-worker-6748df9c8c-zk7zp" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.815600 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ce27714-673f-47de-acc3-b6902b534bdd-combined-ca-bundle\") pod \"barbican-keystone-listener-556764fb84-r628x\" (UID: \"4ce27714-673f-47de-acc3-b6902b534bdd\") " pod="openstack/barbican-keystone-listener-556764fb84-r628x" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.815620 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/710edaa6-ba83-4b1f-a49a-769ca1911c9b-logs\") pod \"barbican-worker-6748df9c8c-zk7zp\" (UID: \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\") " pod="openstack/barbican-worker-6748df9c8c-zk7zp" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.815645 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/710edaa6-ba83-4b1f-a49a-769ca1911c9b-config-data\") pod \"barbican-worker-6748df9c8c-zk7zp\" (UID: \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\") " pod="openstack/barbican-worker-6748df9c8c-zk7zp" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.815662 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvs49\" (UniqueName: \"kubernetes.io/projected/710edaa6-ba83-4b1f-a49a-769ca1911c9b-kube-api-access-nvs49\") pod \"barbican-worker-6748df9c8c-zk7zp\" (UID: \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\") " pod="openstack/barbican-worker-6748df9c8c-zk7zp" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.815699 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/710edaa6-ba83-4b1f-a49a-769ca1911c9b-config-data-custom\") pod \"barbican-worker-6748df9c8c-zk7zp\" (UID: \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\") " pod="openstack/barbican-worker-6748df9c8c-zk7zp" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.815719 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmmsl\" (UniqueName: \"kubernetes.io/projected/4ce27714-673f-47de-acc3-b6902b534bdd-kube-api-access-pmmsl\") pod \"barbican-keystone-listener-556764fb84-r628x\" (UID: \"4ce27714-673f-47de-acc3-b6902b534bdd\") " pod="openstack/barbican-keystone-listener-556764fb84-r628x" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.816743 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ce27714-673f-47de-acc3-b6902b534bdd-logs\") pod \"barbican-keystone-listener-556764fb84-r628x\" (UID: \"4ce27714-673f-47de-acc3-b6902b534bdd\") " pod="openstack/barbican-keystone-listener-556764fb84-r628x" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.821366 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ce27714-673f-47de-acc3-b6902b534bdd-config-data\") pod \"barbican-keystone-listener-556764fb84-r628x\" (UID: \"4ce27714-673f-47de-acc3-b6902b534bdd\") " pod="openstack/barbican-keystone-listener-556764fb84-r628x" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.830033 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ce27714-673f-47de-acc3-b6902b534bdd-combined-ca-bundle\") pod \"barbican-keystone-listener-556764fb84-r628x\" (UID: \"4ce27714-673f-47de-acc3-b6902b534bdd\") " pod="openstack/barbican-keystone-listener-556764fb84-r628x" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.830705 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ce27714-673f-47de-acc3-b6902b534bdd-config-data-custom\") pod \"barbican-keystone-listener-556764fb84-r628x\" (UID: \"4ce27714-673f-47de-acc3-b6902b534bdd\") " pod="openstack/barbican-keystone-listener-556764fb84-r628x" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.856907 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmmsl\" (UniqueName: \"kubernetes.io/projected/4ce27714-673f-47de-acc3-b6902b534bdd-kube-api-access-pmmsl\") pod \"barbican-keystone-listener-556764fb84-r628x\" (UID: \"4ce27714-673f-47de-acc3-b6902b534bdd\") " pod="openstack/barbican-keystone-listener-556764fb84-r628x" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.920464 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/710edaa6-ba83-4b1f-a49a-769ca1911c9b-logs\") pod \"barbican-worker-6748df9c8c-zk7zp\" (UID: \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\") " pod="openstack/barbican-worker-6748df9c8c-zk7zp" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.920897 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/710edaa6-ba83-4b1f-a49a-769ca1911c9b-config-data\") pod \"barbican-worker-6748df9c8c-zk7zp\" (UID: \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\") " pod="openstack/barbican-worker-6748df9c8c-zk7zp" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.920920 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvs49\" (UniqueName: \"kubernetes.io/projected/710edaa6-ba83-4b1f-a49a-769ca1911c9b-kube-api-access-nvs49\") pod \"barbican-worker-6748df9c8c-zk7zp\" (UID: \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\") " pod="openstack/barbican-worker-6748df9c8c-zk7zp" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.920919 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/710edaa6-ba83-4b1f-a49a-769ca1911c9b-logs\") pod \"barbican-worker-6748df9c8c-zk7zp\" (UID: \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\") " pod="openstack/barbican-worker-6748df9c8c-zk7zp" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.921006 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/710edaa6-ba83-4b1f-a49a-769ca1911c9b-config-data-custom\") pod \"barbican-worker-6748df9c8c-zk7zp\" (UID: \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\") " pod="openstack/barbican-worker-6748df9c8c-zk7zp" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.921056 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-dns-swift-storage-0\") pod \"dnsmasq-dns-59d5ff467f-62kx4\" (UID: \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\") " pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.921084 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vxqs\" (UniqueName: \"kubernetes.io/projected/eca3ffef-1a57-4aee-9302-64b59ee0fc44-kube-api-access-4vxqs\") pod \"dnsmasq-dns-59d5ff467f-62kx4\" (UID: \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\") " pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.921218 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5ff467f-62kx4\" (UID: \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\") " pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.922731 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-dns-svc\") pod \"dnsmasq-dns-59d5ff467f-62kx4\" (UID: \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\") " pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.922788 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/710edaa6-ba83-4b1f-a49a-769ca1911c9b-combined-ca-bundle\") pod \"barbican-worker-6748df9c8c-zk7zp\" (UID: \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\") " pod="openstack/barbican-worker-6748df9c8c-zk7zp" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.922842 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-config\") pod \"dnsmasq-dns-59d5ff467f-62kx4\" (UID: \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\") " pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.922928 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5ff467f-62kx4\" (UID: \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\") " pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.926666 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/710edaa6-ba83-4b1f-a49a-769ca1911c9b-config-data-custom\") pod \"barbican-worker-6748df9c8c-zk7zp\" (UID: \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\") " pod="openstack/barbican-worker-6748df9c8c-zk7zp" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.926733 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/710edaa6-ba83-4b1f-a49a-769ca1911c9b-combined-ca-bundle\") pod \"barbican-worker-6748df9c8c-zk7zp\" (UID: \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\") " pod="openstack/barbican-worker-6748df9c8c-zk7zp" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.936750 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/710edaa6-ba83-4b1f-a49a-769ca1911c9b-config-data\") pod \"barbican-worker-6748df9c8c-zk7zp\" (UID: \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\") " pod="openstack/barbican-worker-6748df9c8c-zk7zp" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.953184 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvs49\" (UniqueName: \"kubernetes.io/projected/710edaa6-ba83-4b1f-a49a-769ca1911c9b-kube-api-access-nvs49\") pod \"barbican-worker-6748df9c8c-zk7zp\" (UID: \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\") " pod="openstack/barbican-worker-6748df9c8c-zk7zp" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.979588 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-cb9d976b-flrwl"] Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.981216 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-cb9d976b-flrwl" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.987142 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Jan 27 15:27:47 crc kubenswrapper[4772]: I0127 15:27:47.997436 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-cb9d976b-flrwl"] Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.025304 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-dns-swift-storage-0\") pod \"dnsmasq-dns-59d5ff467f-62kx4\" (UID: \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\") " pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.025357 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vxqs\" (UniqueName: \"kubernetes.io/projected/eca3ffef-1a57-4aee-9302-64b59ee0fc44-kube-api-access-4vxqs\") pod \"dnsmasq-dns-59d5ff467f-62kx4\" (UID: \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\") " pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.025444 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5ff467f-62kx4\" (UID: \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\") " pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.025506 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-dns-svc\") pod \"dnsmasq-dns-59d5ff467f-62kx4\" (UID: \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\") " pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.025546 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-config\") pod \"dnsmasq-dns-59d5ff467f-62kx4\" (UID: \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\") " pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.025597 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5ff467f-62kx4\" (UID: \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\") " pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.026661 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5ff467f-62kx4\" (UID: \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\") " pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.027306 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-dns-swift-storage-0\") pod \"dnsmasq-dns-59d5ff467f-62kx4\" (UID: \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\") " pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.028216 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5ff467f-62kx4\" (UID: \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\") " pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.028805 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-dns-svc\") pod \"dnsmasq-dns-59d5ff467f-62kx4\" (UID: \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\") " pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.029534 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-config\") pod \"dnsmasq-dns-59d5ff467f-62kx4\" (UID: \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\") " pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.029893 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6748df9c8c-zk7zp" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.045593 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-556764fb84-r628x" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.049980 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vxqs\" (UniqueName: \"kubernetes.io/projected/eca3ffef-1a57-4aee-9302-64b59ee0fc44-kube-api-access-4vxqs\") pod \"dnsmasq-dns-59d5ff467f-62kx4\" (UID: \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\") " pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.127564 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a02a1b6c-d438-42bf-a577-88bbbcca2a00-config-data\") pod \"barbican-api-cb9d976b-flrwl\" (UID: \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\") " pod="openstack/barbican-api-cb9d976b-flrwl" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.127607 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbjvx\" (UniqueName: \"kubernetes.io/projected/a02a1b6c-d438-42bf-a577-88bbbcca2a00-kube-api-access-rbjvx\") pod \"barbican-api-cb9d976b-flrwl\" (UID: \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\") " pod="openstack/barbican-api-cb9d976b-flrwl" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.128009 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a02a1b6c-d438-42bf-a577-88bbbcca2a00-config-data-custom\") pod \"barbican-api-cb9d976b-flrwl\" (UID: \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\") " pod="openstack/barbican-api-cb9d976b-flrwl" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.128118 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a02a1b6c-d438-42bf-a577-88bbbcca2a00-combined-ca-bundle\") pod \"barbican-api-cb9d976b-flrwl\" (UID: \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\") " pod="openstack/barbican-api-cb9d976b-flrwl" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.128275 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a02a1b6c-d438-42bf-a577-88bbbcca2a00-logs\") pod \"barbican-api-cb9d976b-flrwl\" (UID: \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\") " pod="openstack/barbican-api-cb9d976b-flrwl" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.217266 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.229672 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a02a1b6c-d438-42bf-a577-88bbbcca2a00-config-data-custom\") pod \"barbican-api-cb9d976b-flrwl\" (UID: \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\") " pod="openstack/barbican-api-cb9d976b-flrwl" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.229741 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a02a1b6c-d438-42bf-a577-88bbbcca2a00-combined-ca-bundle\") pod \"barbican-api-cb9d976b-flrwl\" (UID: \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\") " pod="openstack/barbican-api-cb9d976b-flrwl" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.229797 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a02a1b6c-d438-42bf-a577-88bbbcca2a00-logs\") pod \"barbican-api-cb9d976b-flrwl\" (UID: \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\") " pod="openstack/barbican-api-cb9d976b-flrwl" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.229857 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a02a1b6c-d438-42bf-a577-88bbbcca2a00-config-data\") pod \"barbican-api-cb9d976b-flrwl\" (UID: \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\") " pod="openstack/barbican-api-cb9d976b-flrwl" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.229888 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbjvx\" (UniqueName: \"kubernetes.io/projected/a02a1b6c-d438-42bf-a577-88bbbcca2a00-kube-api-access-rbjvx\") pod \"barbican-api-cb9d976b-flrwl\" (UID: \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\") " pod="openstack/barbican-api-cb9d976b-flrwl" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.233743 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a02a1b6c-d438-42bf-a577-88bbbcca2a00-logs\") pod \"barbican-api-cb9d976b-flrwl\" (UID: \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\") " pod="openstack/barbican-api-cb9d976b-flrwl" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.234865 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a02a1b6c-d438-42bf-a577-88bbbcca2a00-config-data-custom\") pod \"barbican-api-cb9d976b-flrwl\" (UID: \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\") " pod="openstack/barbican-api-cb9d976b-flrwl" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.237989 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a02a1b6c-d438-42bf-a577-88bbbcca2a00-config-data\") pod \"barbican-api-cb9d976b-flrwl\" (UID: \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\") " pod="openstack/barbican-api-cb9d976b-flrwl" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.243837 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a02a1b6c-d438-42bf-a577-88bbbcca2a00-combined-ca-bundle\") pod \"barbican-api-cb9d976b-flrwl\" (UID: \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\") " pod="openstack/barbican-api-cb9d976b-flrwl" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.250081 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbjvx\" (UniqueName: \"kubernetes.io/projected/a02a1b6c-d438-42bf-a577-88bbbcca2a00-kube-api-access-rbjvx\") pod \"barbican-api-cb9d976b-flrwl\" (UID: \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\") " pod="openstack/barbican-api-cb9d976b-flrwl" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.297944 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-cb9d976b-flrwl" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.465903 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8583377-67ef-4cca-83bb-08d7523ab0a8","Type":"ContainerStarted","Data":"851c7b3936a50d21408fbed0918adde539924e5915ec73fdcccd952a3392565b"} Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.466073 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.498245 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.003279171 podStartE2EDuration="5.498219276s" podCreationTimestamp="2026-01-27 15:27:43 +0000 UTC" firstStartedPulling="2026-01-27 15:27:44.258624969 +0000 UTC m=+1250.239234067" lastFinishedPulling="2026-01-27 15:27:47.753565074 +0000 UTC m=+1253.734174172" observedRunningTime="2026-01-27 15:27:48.490723209 +0000 UTC m=+1254.471332337" watchObservedRunningTime="2026-01-27 15:27:48.498219276 +0000 UTC m=+1254.478828384" Jan 27 15:27:48 crc kubenswrapper[4772]: W0127 15:27:48.622793 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod710edaa6_ba83_4b1f_a49a_769ca1911c9b.slice/crio-c83991847bf683630e70d44722d44695d9152a02d09f0a3d6fe39436ebbf262d WatchSource:0}: Error finding container c83991847bf683630e70d44722d44695d9152a02d09f0a3d6fe39436ebbf262d: Status 404 returned error can't find the container with id c83991847bf683630e70d44722d44695d9152a02d09f0a3d6fe39436ebbf262d Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.623106 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6748df9c8c-zk7zp"] Jan 27 15:27:48 crc kubenswrapper[4772]: W0127 15:27:48.690665 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ce27714_673f_47de_acc3_b6902b534bdd.slice/crio-51e9e5e71be46820f9c3d1564ff14b9e6df8988ed057a1326779c07f7fee3331 WatchSource:0}: Error finding container 51e9e5e71be46820f9c3d1564ff14b9e6df8988ed057a1326779c07f7fee3331: Status 404 returned error can't find the container with id 51e9e5e71be46820f9c3d1564ff14b9e6df8988ed057a1326779c07f7fee3331 Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.694419 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-556764fb84-r628x"] Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.770789 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-62kx4"] Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.893259 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8l85z" Jan 27 15:27:48 crc kubenswrapper[4772]: I0127 15:27:48.953940 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-cb9d976b-flrwl"] Jan 27 15:27:48 crc kubenswrapper[4772]: W0127 15:27:48.964126 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda02a1b6c_d438_42bf_a577_88bbbcca2a00.slice/crio-37bb9f56f01e20e1b9f2066e0e10a19c1b316c11b490381320b663b46a9cc874 WatchSource:0}: Error finding container 37bb9f56f01e20e1b9f2066e0e10a19c1b316c11b490381320b663b46a9cc874: Status 404 returned error can't find the container with id 37bb9f56f01e20e1b9f2066e0e10a19c1b316c11b490381320b663b46a9cc874 Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.051328 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ae05919-68bf-43d1-abd9-9908ec287bd0-scripts\") pod \"9ae05919-68bf-43d1-abd9-9908ec287bd0\" (UID: \"9ae05919-68bf-43d1-abd9-9908ec287bd0\") " Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.051539 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4tmg\" (UniqueName: \"kubernetes.io/projected/9ae05919-68bf-43d1-abd9-9908ec287bd0-kube-api-access-d4tmg\") pod \"9ae05919-68bf-43d1-abd9-9908ec287bd0\" (UID: \"9ae05919-68bf-43d1-abd9-9908ec287bd0\") " Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.051814 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ae05919-68bf-43d1-abd9-9908ec287bd0-config-data\") pod \"9ae05919-68bf-43d1-abd9-9908ec287bd0\" (UID: \"9ae05919-68bf-43d1-abd9-9908ec287bd0\") " Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.051913 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ae05919-68bf-43d1-abd9-9908ec287bd0-combined-ca-bundle\") pod \"9ae05919-68bf-43d1-abd9-9908ec287bd0\" (UID: \"9ae05919-68bf-43d1-abd9-9908ec287bd0\") " Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.051959 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9ae05919-68bf-43d1-abd9-9908ec287bd0-etc-machine-id\") pod \"9ae05919-68bf-43d1-abd9-9908ec287bd0\" (UID: \"9ae05919-68bf-43d1-abd9-9908ec287bd0\") " Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.052013 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9ae05919-68bf-43d1-abd9-9908ec287bd0-db-sync-config-data\") pod \"9ae05919-68bf-43d1-abd9-9908ec287bd0\" (UID: \"9ae05919-68bf-43d1-abd9-9908ec287bd0\") " Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.052058 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ae05919-68bf-43d1-abd9-9908ec287bd0-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9ae05919-68bf-43d1-abd9-9908ec287bd0" (UID: "9ae05919-68bf-43d1-abd9-9908ec287bd0"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.052763 4772 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9ae05919-68bf-43d1-abd9-9908ec287bd0-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.057095 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ae05919-68bf-43d1-abd9-9908ec287bd0-scripts" (OuterVolumeSpecName: "scripts") pod "9ae05919-68bf-43d1-abd9-9908ec287bd0" (UID: "9ae05919-68bf-43d1-abd9-9908ec287bd0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.057285 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ae05919-68bf-43d1-abd9-9908ec287bd0-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "9ae05919-68bf-43d1-abd9-9908ec287bd0" (UID: "9ae05919-68bf-43d1-abd9-9908ec287bd0"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.059018 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ae05919-68bf-43d1-abd9-9908ec287bd0-kube-api-access-d4tmg" (OuterVolumeSpecName: "kube-api-access-d4tmg") pod "9ae05919-68bf-43d1-abd9-9908ec287bd0" (UID: "9ae05919-68bf-43d1-abd9-9908ec287bd0"). InnerVolumeSpecName "kube-api-access-d4tmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.083754 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ae05919-68bf-43d1-abd9-9908ec287bd0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9ae05919-68bf-43d1-abd9-9908ec287bd0" (UID: "9ae05919-68bf-43d1-abd9-9908ec287bd0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.110442 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ae05919-68bf-43d1-abd9-9908ec287bd0-config-data" (OuterVolumeSpecName: "config-data") pod "9ae05919-68bf-43d1-abd9-9908ec287bd0" (UID: "9ae05919-68bf-43d1-abd9-9908ec287bd0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.153906 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ae05919-68bf-43d1-abd9-9908ec287bd0-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.153943 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ae05919-68bf-43d1-abd9-9908ec287bd0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.153959 4772 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9ae05919-68bf-43d1-abd9-9908ec287bd0-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.153999 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ae05919-68bf-43d1-abd9-9908ec287bd0-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.154011 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4tmg\" (UniqueName: \"kubernetes.io/projected/9ae05919-68bf-43d1-abd9-9908ec287bd0-kube-api-access-d4tmg\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.476213 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6748df9c8c-zk7zp" event={"ID":"710edaa6-ba83-4b1f-a49a-769ca1911c9b","Type":"ContainerStarted","Data":"c83991847bf683630e70d44722d44695d9152a02d09f0a3d6fe39436ebbf262d"} Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.478431 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-556764fb84-r628x" event={"ID":"4ce27714-673f-47de-acc3-b6902b534bdd","Type":"ContainerStarted","Data":"51e9e5e71be46820f9c3d1564ff14b9e6df8988ed057a1326779c07f7fee3331"} Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.480459 4772 generic.go:334] "Generic (PLEG): container finished" podID="eca3ffef-1a57-4aee-9302-64b59ee0fc44" containerID="7edcc63bf9420fd4c188dbb2b09e3dd62f71b1aeec4c1159720516f01b6da375" exitCode=0 Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.480563 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" event={"ID":"eca3ffef-1a57-4aee-9302-64b59ee0fc44","Type":"ContainerDied","Data":"7edcc63bf9420fd4c188dbb2b09e3dd62f71b1aeec4c1159720516f01b6da375"} Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.480623 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" event={"ID":"eca3ffef-1a57-4aee-9302-64b59ee0fc44","Type":"ContainerStarted","Data":"89fbde952ff7b7a7ea9f206fc34ae7d5aa1cfb885454833c03af5e3e355d5fb3"} Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.482105 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8l85z" event={"ID":"9ae05919-68bf-43d1-abd9-9908ec287bd0","Type":"ContainerDied","Data":"6c7bfeb67dfdf4e440bd40114d111aab9077e461f93cb6bdda5f337cad29c97d"} Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.482131 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c7bfeb67dfdf4e440bd40114d111aab9077e461f93cb6bdda5f337cad29c97d" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.482246 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8l85z" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.484520 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cb9d976b-flrwl" event={"ID":"a02a1b6c-d438-42bf-a577-88bbbcca2a00","Type":"ContainerStarted","Data":"37bb9f56f01e20e1b9f2066e0e10a19c1b316c11b490381320b663b46a9cc874"} Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.694002 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Jan 27 15:27:49 crc kubenswrapper[4772]: E0127 15:27:49.694489 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ae05919-68bf-43d1-abd9-9908ec287bd0" containerName="cinder-db-sync" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.694506 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ae05919-68bf-43d1-abd9-9908ec287bd0" containerName="cinder-db-sync" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.694682 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ae05919-68bf-43d1-abd9-9908ec287bd0" containerName="cinder-db-sync" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.696582 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.703691 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.703840 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.704052 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.704184 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-8nhs4" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.715682 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.768206 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b7d9f3f-e366-421e-b00d-9c453da1adca-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7b7d9f3f-e366-421e-b00d-9c453da1adca\") " pod="openstack/cinder-scheduler-0" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.774244 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b7d9f3f-e366-421e-b00d-9c453da1adca-scripts\") pod \"cinder-scheduler-0\" (UID: \"7b7d9f3f-e366-421e-b00d-9c453da1adca\") " pod="openstack/cinder-scheduler-0" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.774647 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b7d9f3f-e366-421e-b00d-9c453da1adca-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7b7d9f3f-e366-421e-b00d-9c453da1adca\") " pod="openstack/cinder-scheduler-0" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.774707 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2tvv\" (UniqueName: \"kubernetes.io/projected/7b7d9f3f-e366-421e-b00d-9c453da1adca-kube-api-access-g2tvv\") pod \"cinder-scheduler-0\" (UID: \"7b7d9f3f-e366-421e-b00d-9c453da1adca\") " pod="openstack/cinder-scheduler-0" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.774761 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b7d9f3f-e366-421e-b00d-9c453da1adca-config-data\") pod \"cinder-scheduler-0\" (UID: \"7b7d9f3f-e366-421e-b00d-9c453da1adca\") " pod="openstack/cinder-scheduler-0" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.774849 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b7d9f3f-e366-421e-b00d-9c453da1adca-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7b7d9f3f-e366-421e-b00d-9c453da1adca\") " pod="openstack/cinder-scheduler-0" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.778012 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-62kx4"] Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.823515 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-69c986f6d7-gszgg"] Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.825075 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.881041 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b7d9f3f-e366-421e-b00d-9c453da1adca-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7b7d9f3f-e366-421e-b00d-9c453da1adca\") " pod="openstack/cinder-scheduler-0" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.881417 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2tvv\" (UniqueName: \"kubernetes.io/projected/7b7d9f3f-e366-421e-b00d-9c453da1adca-kube-api-access-g2tvv\") pod \"cinder-scheduler-0\" (UID: \"7b7d9f3f-e366-421e-b00d-9c453da1adca\") " pod="openstack/cinder-scheduler-0" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.881456 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b7d9f3f-e366-421e-b00d-9c453da1adca-config-data\") pod \"cinder-scheduler-0\" (UID: \"7b7d9f3f-e366-421e-b00d-9c453da1adca\") " pod="openstack/cinder-scheduler-0" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.881509 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b7d9f3f-e366-421e-b00d-9c453da1adca-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7b7d9f3f-e366-421e-b00d-9c453da1adca\") " pod="openstack/cinder-scheduler-0" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.881554 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b7d9f3f-e366-421e-b00d-9c453da1adca-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7b7d9f3f-e366-421e-b00d-9c453da1adca\") " pod="openstack/cinder-scheduler-0" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.881591 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b7d9f3f-e366-421e-b00d-9c453da1adca-scripts\") pod \"cinder-scheduler-0\" (UID: \"7b7d9f3f-e366-421e-b00d-9c453da1adca\") " pod="openstack/cinder-scheduler-0" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.882611 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b7d9f3f-e366-421e-b00d-9c453da1adca-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7b7d9f3f-e366-421e-b00d-9c453da1adca\") " pod="openstack/cinder-scheduler-0" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.892081 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69c986f6d7-gszgg"] Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.899486 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b7d9f3f-e366-421e-b00d-9c453da1adca-scripts\") pod \"cinder-scheduler-0\" (UID: \"7b7d9f3f-e366-421e-b00d-9c453da1adca\") " pod="openstack/cinder-scheduler-0" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.900732 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b7d9f3f-e366-421e-b00d-9c453da1adca-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7b7d9f3f-e366-421e-b00d-9c453da1adca\") " pod="openstack/cinder-scheduler-0" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.901394 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b7d9f3f-e366-421e-b00d-9c453da1adca-config-data\") pod \"cinder-scheduler-0\" (UID: \"7b7d9f3f-e366-421e-b00d-9c453da1adca\") " pod="openstack/cinder-scheduler-0" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.901471 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b7d9f3f-e366-421e-b00d-9c453da1adca-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7b7d9f3f-e366-421e-b00d-9c453da1adca\") " pod="openstack/cinder-scheduler-0" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.909803 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2tvv\" (UniqueName: \"kubernetes.io/projected/7b7d9f3f-e366-421e-b00d-9c453da1adca-kube-api-access-g2tvv\") pod \"cinder-scheduler-0\" (UID: \"7b7d9f3f-e366-421e-b00d-9c453da1adca\") " pod="openstack/cinder-scheduler-0" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.971284 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.972961 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 27 15:27:49 crc kubenswrapper[4772]: I0127 15:27:49.998890 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:49.999393 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpwmg\" (UniqueName: \"kubernetes.io/projected/ce081402-0ada-4fbf-8b22-eb88a50e804b-kube-api-access-hpwmg\") pod \"dnsmasq-dns-69c986f6d7-gszgg\" (UID: \"ce081402-0ada-4fbf-8b22-eb88a50e804b\") " pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:49.999485 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-dns-svc\") pod \"dnsmasq-dns-69c986f6d7-gszgg\" (UID: \"ce081402-0ada-4fbf-8b22-eb88a50e804b\") " pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:49.999514 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-config\") pod \"dnsmasq-dns-69c986f6d7-gszgg\" (UID: \"ce081402-0ada-4fbf-8b22-eb88a50e804b\") " pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:49.999593 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-ovsdbserver-nb\") pod \"dnsmasq-dns-69c986f6d7-gszgg\" (UID: \"ce081402-0ada-4fbf-8b22-eb88a50e804b\") " pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:49.999648 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-dns-swift-storage-0\") pod \"dnsmasq-dns-69c986f6d7-gszgg\" (UID: \"ce081402-0ada-4fbf-8b22-eb88a50e804b\") " pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:49.999731 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-ovsdbserver-sb\") pod \"dnsmasq-dns-69c986f6d7-gszgg\" (UID: \"ce081402-0ada-4fbf-8b22-eb88a50e804b\") " pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.028800 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.051653 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.101798 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-dns-svc\") pod \"dnsmasq-dns-69c986f6d7-gszgg\" (UID: \"ce081402-0ada-4fbf-8b22-eb88a50e804b\") " pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.101859 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-config\") pod \"dnsmasq-dns-69c986f6d7-gszgg\" (UID: \"ce081402-0ada-4fbf-8b22-eb88a50e804b\") " pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.101899 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l2r7\" (UniqueName: \"kubernetes.io/projected/da515cae-40ac-41af-aef5-9cef9f3b366e-kube-api-access-4l2r7\") pod \"cinder-api-0\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " pod="openstack/cinder-api-0" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.101923 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da515cae-40ac-41af-aef5-9cef9f3b366e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " pod="openstack/cinder-api-0" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.101961 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da515cae-40ac-41af-aef5-9cef9f3b366e-config-data-custom\") pod \"cinder-api-0\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " pod="openstack/cinder-api-0" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.101997 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-ovsdbserver-nb\") pod \"dnsmasq-dns-69c986f6d7-gszgg\" (UID: \"ce081402-0ada-4fbf-8b22-eb88a50e804b\") " pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.102045 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da515cae-40ac-41af-aef5-9cef9f3b366e-config-data\") pod \"cinder-api-0\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " pod="openstack/cinder-api-0" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.102069 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-dns-swift-storage-0\") pod \"dnsmasq-dns-69c986f6d7-gszgg\" (UID: \"ce081402-0ada-4fbf-8b22-eb88a50e804b\") " pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.102105 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da515cae-40ac-41af-aef5-9cef9f3b366e-scripts\") pod \"cinder-api-0\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " pod="openstack/cinder-api-0" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.102240 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-ovsdbserver-sb\") pod \"dnsmasq-dns-69c986f6d7-gszgg\" (UID: \"ce081402-0ada-4fbf-8b22-eb88a50e804b\") " pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.102277 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpwmg\" (UniqueName: \"kubernetes.io/projected/ce081402-0ada-4fbf-8b22-eb88a50e804b-kube-api-access-hpwmg\") pod \"dnsmasq-dns-69c986f6d7-gszgg\" (UID: \"ce081402-0ada-4fbf-8b22-eb88a50e804b\") " pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.102325 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da515cae-40ac-41af-aef5-9cef9f3b366e-logs\") pod \"cinder-api-0\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " pod="openstack/cinder-api-0" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.102351 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da515cae-40ac-41af-aef5-9cef9f3b366e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " pod="openstack/cinder-api-0" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.103733 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-config\") pod \"dnsmasq-dns-69c986f6d7-gszgg\" (UID: \"ce081402-0ada-4fbf-8b22-eb88a50e804b\") " pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.103803 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-dns-svc\") pod \"dnsmasq-dns-69c986f6d7-gszgg\" (UID: \"ce081402-0ada-4fbf-8b22-eb88a50e804b\") " pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.103979 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-ovsdbserver-nb\") pod \"dnsmasq-dns-69c986f6d7-gszgg\" (UID: \"ce081402-0ada-4fbf-8b22-eb88a50e804b\") " pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.105313 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-dns-swift-storage-0\") pod \"dnsmasq-dns-69c986f6d7-gszgg\" (UID: \"ce081402-0ada-4fbf-8b22-eb88a50e804b\") " pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.105462 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-ovsdbserver-sb\") pod \"dnsmasq-dns-69c986f6d7-gszgg\" (UID: \"ce081402-0ada-4fbf-8b22-eb88a50e804b\") " pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.136590 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpwmg\" (UniqueName: \"kubernetes.io/projected/ce081402-0ada-4fbf-8b22-eb88a50e804b-kube-api-access-hpwmg\") pod \"dnsmasq-dns-69c986f6d7-gszgg\" (UID: \"ce081402-0ada-4fbf-8b22-eb88a50e804b\") " pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.176155 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.203819 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l2r7\" (UniqueName: \"kubernetes.io/projected/da515cae-40ac-41af-aef5-9cef9f3b366e-kube-api-access-4l2r7\") pod \"cinder-api-0\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " pod="openstack/cinder-api-0" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.204012 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da515cae-40ac-41af-aef5-9cef9f3b366e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " pod="openstack/cinder-api-0" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.204047 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da515cae-40ac-41af-aef5-9cef9f3b366e-config-data-custom\") pod \"cinder-api-0\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " pod="openstack/cinder-api-0" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.204087 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da515cae-40ac-41af-aef5-9cef9f3b366e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " pod="openstack/cinder-api-0" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.204116 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da515cae-40ac-41af-aef5-9cef9f3b366e-config-data\") pod \"cinder-api-0\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " pod="openstack/cinder-api-0" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.204155 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da515cae-40ac-41af-aef5-9cef9f3b366e-scripts\") pod \"cinder-api-0\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " pod="openstack/cinder-api-0" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.204267 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da515cae-40ac-41af-aef5-9cef9f3b366e-logs\") pod \"cinder-api-0\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " pod="openstack/cinder-api-0" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.204294 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da515cae-40ac-41af-aef5-9cef9f3b366e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " pod="openstack/cinder-api-0" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.208697 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da515cae-40ac-41af-aef5-9cef9f3b366e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " pod="openstack/cinder-api-0" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.209439 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da515cae-40ac-41af-aef5-9cef9f3b366e-logs\") pod \"cinder-api-0\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " pod="openstack/cinder-api-0" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.210853 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da515cae-40ac-41af-aef5-9cef9f3b366e-scripts\") pod \"cinder-api-0\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " pod="openstack/cinder-api-0" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.212496 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da515cae-40ac-41af-aef5-9cef9f3b366e-config-data\") pod \"cinder-api-0\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " pod="openstack/cinder-api-0" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.222653 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l2r7\" (UniqueName: \"kubernetes.io/projected/da515cae-40ac-41af-aef5-9cef9f3b366e-kube-api-access-4l2r7\") pod \"cinder-api-0\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " pod="openstack/cinder-api-0" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.224681 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da515cae-40ac-41af-aef5-9cef9f3b366e-config-data-custom\") pod \"cinder-api-0\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " pod="openstack/cinder-api-0" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.332696 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.455093 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.512425 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" event={"ID":"eca3ffef-1a57-4aee-9302-64b59ee0fc44","Type":"ContainerStarted","Data":"3c25ec3ecdb9afdacbc10ba3b1a990e70919f2f5f3c0a3614045131a2f37d22d"} Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.512522 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" podUID="eca3ffef-1a57-4aee-9302-64b59ee0fc44" containerName="dnsmasq-dns" containerID="cri-o://3c25ec3ecdb9afdacbc10ba3b1a990e70919f2f5f3c0a3614045131a2f37d22d" gracePeriod=10 Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.512801 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.517970 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7b7d9f3f-e366-421e-b00d-9c453da1adca","Type":"ContainerStarted","Data":"60d07ef91bf8ed929ad3a7649d54a1c26d24e5b004c0a1e93cdd8cb08880d285"} Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.520232 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cb9d976b-flrwl" event={"ID":"a02a1b6c-d438-42bf-a577-88bbbcca2a00","Type":"ContainerStarted","Data":"51682243156b2ae0619aa3431049d4b3ccf7b64abb3cf0a6f8d8cabce6c5c3f3"} Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.520263 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cb9d976b-flrwl" event={"ID":"a02a1b6c-d438-42bf-a577-88bbbcca2a00","Type":"ContainerStarted","Data":"6263f4825b44d3954903ca3307e160ecf8edb8b8f916573aa23ea8b2efbc78ab"} Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.520460 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-cb9d976b-flrwl" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.520566 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-cb9d976b-flrwl" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.565720 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" podStartSLOduration=3.565698475 podStartE2EDuration="3.565698475s" podCreationTimestamp="2026-01-27 15:27:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:27:50.550240136 +0000 UTC m=+1256.530849244" watchObservedRunningTime="2026-01-27 15:27:50.565698475 +0000 UTC m=+1256.546307573" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.577041 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-cb9d976b-flrwl" podStartSLOduration=3.5770209250000002 podStartE2EDuration="3.577020925s" podCreationTimestamp="2026-01-27 15:27:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:27:50.572659858 +0000 UTC m=+1256.553268986" watchObservedRunningTime="2026-01-27 15:27:50.577020925 +0000 UTC m=+1256.557630023" Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.725266 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69c986f6d7-gszgg"] Jan 27 15:27:50 crc kubenswrapper[4772]: I0127 15:27:50.936923 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 27 15:27:51 crc kubenswrapper[4772]: W0127 15:27:51.256822 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda515cae_40ac_41af_aef5_9cef9f3b366e.slice/crio-005ceeb7f240212b9c37236e7c8187a3f42d3b14297b7bdfc9d9274e9fec1f9d WatchSource:0}: Error finding container 005ceeb7f240212b9c37236e7c8187a3f42d3b14297b7bdfc9d9274e9fec1f9d: Status 404 returned error can't find the container with id 005ceeb7f240212b9c37236e7c8187a3f42d3b14297b7bdfc9d9274e9fec1f9d Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.325868 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.434717 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-ovsdbserver-nb\") pod \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\" (UID: \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\") " Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.435995 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vxqs\" (UniqueName: \"kubernetes.io/projected/eca3ffef-1a57-4aee-9302-64b59ee0fc44-kube-api-access-4vxqs\") pod \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\" (UID: \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\") " Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.436028 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-dns-swift-storage-0\") pod \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\" (UID: \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\") " Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.436114 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-ovsdbserver-sb\") pod \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\" (UID: \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\") " Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.436156 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-config\") pod \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\" (UID: \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\") " Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.436195 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-dns-svc\") pod \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\" (UID: \"eca3ffef-1a57-4aee-9302-64b59ee0fc44\") " Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.445040 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eca3ffef-1a57-4aee-9302-64b59ee0fc44-kube-api-access-4vxqs" (OuterVolumeSpecName: "kube-api-access-4vxqs") pod "eca3ffef-1a57-4aee-9302-64b59ee0fc44" (UID: "eca3ffef-1a57-4aee-9302-64b59ee0fc44"). InnerVolumeSpecName "kube-api-access-4vxqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.493348 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "eca3ffef-1a57-4aee-9302-64b59ee0fc44" (UID: "eca3ffef-1a57-4aee-9302-64b59ee0fc44"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.494528 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eca3ffef-1a57-4aee-9302-64b59ee0fc44" (UID: "eca3ffef-1a57-4aee-9302-64b59ee0fc44"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.505754 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "eca3ffef-1a57-4aee-9302-64b59ee0fc44" (UID: "eca3ffef-1a57-4aee-9302-64b59ee0fc44"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.508736 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "eca3ffef-1a57-4aee-9302-64b59ee0fc44" (UID: "eca3ffef-1a57-4aee-9302-64b59ee0fc44"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.536370 4772 generic.go:334] "Generic (PLEG): container finished" podID="eca3ffef-1a57-4aee-9302-64b59ee0fc44" containerID="3c25ec3ecdb9afdacbc10ba3b1a990e70919f2f5f3c0a3614045131a2f37d22d" exitCode=0 Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.536433 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" event={"ID":"eca3ffef-1a57-4aee-9302-64b59ee0fc44","Type":"ContainerDied","Data":"3c25ec3ecdb9afdacbc10ba3b1a990e70919f2f5f3c0a3614045131a2f37d22d"} Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.536459 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" event={"ID":"eca3ffef-1a57-4aee-9302-64b59ee0fc44","Type":"ContainerDied","Data":"89fbde952ff7b7a7ea9f206fc34ae7d5aa1cfb885454833c03af5e3e355d5fb3"} Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.536475 4772 scope.go:117] "RemoveContainer" containerID="3c25ec3ecdb9afdacbc10ba3b1a990e70919f2f5f3c0a3614045131a2f37d22d" Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.536588 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-62kx4" Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.537896 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.537917 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vxqs\" (UniqueName: \"kubernetes.io/projected/eca3ffef-1a57-4aee-9302-64b59ee0fc44-kube-api-access-4vxqs\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.537927 4772 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.537938 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.537947 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.540789 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-config" (OuterVolumeSpecName: "config") pod "eca3ffef-1a57-4aee-9302-64b59ee0fc44" (UID: "eca3ffef-1a57-4aee-9302-64b59ee0fc44"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.547349 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"da515cae-40ac-41af-aef5-9cef9f3b366e","Type":"ContainerStarted","Data":"005ceeb7f240212b9c37236e7c8187a3f42d3b14297b7bdfc9d9274e9fec1f9d"} Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.554227 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" event={"ID":"ce081402-0ada-4fbf-8b22-eb88a50e804b","Type":"ContainerStarted","Data":"59a55a087cb1a139692664b7e8a0e67c4b8ddc2ffa8baef3cc11190a48375152"} Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.639154 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eca3ffef-1a57-4aee-9302-64b59ee0fc44-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.885150 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-62kx4"] Jan 27 15:27:51 crc kubenswrapper[4772]: I0127 15:27:51.901891 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-62kx4"] Jan 27 15:27:52 crc kubenswrapper[4772]: I0127 15:27:52.062397 4772 scope.go:117] "RemoveContainer" containerID="7edcc63bf9420fd4c188dbb2b09e3dd62f71b1aeec4c1159720516f01b6da375" Jan 27 15:27:52 crc kubenswrapper[4772]: I0127 15:27:52.108001 4772 scope.go:117] "RemoveContainer" containerID="3c25ec3ecdb9afdacbc10ba3b1a990e70919f2f5f3c0a3614045131a2f37d22d" Jan 27 15:27:52 crc kubenswrapper[4772]: E0127 15:27:52.114993 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c25ec3ecdb9afdacbc10ba3b1a990e70919f2f5f3c0a3614045131a2f37d22d\": container with ID starting with 3c25ec3ecdb9afdacbc10ba3b1a990e70919f2f5f3c0a3614045131a2f37d22d not found: ID does not exist" containerID="3c25ec3ecdb9afdacbc10ba3b1a990e70919f2f5f3c0a3614045131a2f37d22d" Jan 27 15:27:52 crc kubenswrapper[4772]: I0127 15:27:52.115052 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c25ec3ecdb9afdacbc10ba3b1a990e70919f2f5f3c0a3614045131a2f37d22d"} err="failed to get container status \"3c25ec3ecdb9afdacbc10ba3b1a990e70919f2f5f3c0a3614045131a2f37d22d\": rpc error: code = NotFound desc = could not find container \"3c25ec3ecdb9afdacbc10ba3b1a990e70919f2f5f3c0a3614045131a2f37d22d\": container with ID starting with 3c25ec3ecdb9afdacbc10ba3b1a990e70919f2f5f3c0a3614045131a2f37d22d not found: ID does not exist" Jan 27 15:27:52 crc kubenswrapper[4772]: I0127 15:27:52.115087 4772 scope.go:117] "RemoveContainer" containerID="7edcc63bf9420fd4c188dbb2b09e3dd62f71b1aeec4c1159720516f01b6da375" Jan 27 15:27:52 crc kubenswrapper[4772]: E0127 15:27:52.115592 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7edcc63bf9420fd4c188dbb2b09e3dd62f71b1aeec4c1159720516f01b6da375\": container with ID starting with 7edcc63bf9420fd4c188dbb2b09e3dd62f71b1aeec4c1159720516f01b6da375 not found: ID does not exist" containerID="7edcc63bf9420fd4c188dbb2b09e3dd62f71b1aeec4c1159720516f01b6da375" Jan 27 15:27:52 crc kubenswrapper[4772]: I0127 15:27:52.115616 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7edcc63bf9420fd4c188dbb2b09e3dd62f71b1aeec4c1159720516f01b6da375"} err="failed to get container status \"7edcc63bf9420fd4c188dbb2b09e3dd62f71b1aeec4c1159720516f01b6da375\": rpc error: code = NotFound desc = could not find container \"7edcc63bf9420fd4c188dbb2b09e3dd62f71b1aeec4c1159720516f01b6da375\": container with ID starting with 7edcc63bf9420fd4c188dbb2b09e3dd62f71b1aeec4c1159720516f01b6da375 not found: ID does not exist" Jan 27 15:27:52 crc kubenswrapper[4772]: I0127 15:27:52.570955 4772 generic.go:334] "Generic (PLEG): container finished" podID="ce081402-0ada-4fbf-8b22-eb88a50e804b" containerID="dc64f95e60b6b7231bfc72557652d02f45abe891bf9e4367174098a9820da0e9" exitCode=0 Jan 27 15:27:52 crc kubenswrapper[4772]: I0127 15:27:52.571270 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" event={"ID":"ce081402-0ada-4fbf-8b22-eb88a50e804b","Type":"ContainerDied","Data":"dc64f95e60b6b7231bfc72557652d02f45abe891bf9e4367174098a9820da0e9"} Jan 27 15:27:52 crc kubenswrapper[4772]: I0127 15:27:52.698010 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eca3ffef-1a57-4aee-9302-64b59ee0fc44" path="/var/lib/kubelet/pods/eca3ffef-1a57-4aee-9302-64b59ee0fc44/volumes" Jan 27 15:27:52 crc kubenswrapper[4772]: I0127 15:27:52.772224 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Jan 27 15:27:53 crc kubenswrapper[4772]: I0127 15:27:53.674930 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6748df9c8c-zk7zp" event={"ID":"710edaa6-ba83-4b1f-a49a-769ca1911c9b","Type":"ContainerStarted","Data":"a063d80b4cd5f0199157f5e139c54f744514f0203001f47e4bce93805443a546"} Jan 27 15:27:53 crc kubenswrapper[4772]: I0127 15:27:53.675206 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6748df9c8c-zk7zp" event={"ID":"710edaa6-ba83-4b1f-a49a-769ca1911c9b","Type":"ContainerStarted","Data":"b1ca77abcb5dfa41040a6625bbdc220ae80143a0714b3ff9a856057794a6d02a"} Jan 27 15:27:53 crc kubenswrapper[4772]: I0127 15:27:53.699652 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-6748df9c8c-zk7zp" podStartSLOduration=3.107491642 podStartE2EDuration="6.699638544s" podCreationTimestamp="2026-01-27 15:27:47 +0000 UTC" firstStartedPulling="2026-01-27 15:27:48.641318958 +0000 UTC m=+1254.621928066" lastFinishedPulling="2026-01-27 15:27:52.23346587 +0000 UTC m=+1258.214074968" observedRunningTime="2026-01-27 15:27:53.697718438 +0000 UTC m=+1259.678327536" watchObservedRunningTime="2026-01-27 15:27:53.699638544 +0000 UTC m=+1259.680247642" Jan 27 15:27:53 crc kubenswrapper[4772]: I0127 15:27:53.700690 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-556764fb84-r628x" event={"ID":"4ce27714-673f-47de-acc3-b6902b534bdd","Type":"ContainerStarted","Data":"aa76ea75f91196a6ccffd5d7e7d149b5efe900bfae2e86e19fa1ec88171321cc"} Jan 27 15:27:53 crc kubenswrapper[4772]: I0127 15:27:53.700741 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-556764fb84-r628x" event={"ID":"4ce27714-673f-47de-acc3-b6902b534bdd","Type":"ContainerStarted","Data":"f068099f2f85afe1f1db1c1b4191de3b3198e413724471d516ae5586de30eb8f"} Jan 27 15:27:53 crc kubenswrapper[4772]: I0127 15:27:53.718353 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7b7d9f3f-e366-421e-b00d-9c453da1adca","Type":"ContainerStarted","Data":"55338fb54abd9ad2a096debb2356f749682191abe6b851127e0e95fcec09a654"} Jan 27 15:27:53 crc kubenswrapper[4772]: I0127 15:27:53.740086 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-556764fb84-r628x" podStartSLOduration=3.302596144 podStartE2EDuration="6.740063209s" podCreationTimestamp="2026-01-27 15:27:47 +0000 UTC" firstStartedPulling="2026-01-27 15:27:48.69472078 +0000 UTC m=+1254.675329878" lastFinishedPulling="2026-01-27 15:27:52.132187845 +0000 UTC m=+1258.112796943" observedRunningTime="2026-01-27 15:27:53.725451555 +0000 UTC m=+1259.706060653" watchObservedRunningTime="2026-01-27 15:27:53.740063209 +0000 UTC m=+1259.720672307" Jan 27 15:27:53 crc kubenswrapper[4772]: I0127 15:27:53.765260 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"da515cae-40ac-41af-aef5-9cef9f3b366e","Type":"ContainerStarted","Data":"cd2c8c5778f70b0c56d0f79379ec5079a2ceb539c1d125108e64ede41e68dc9b"} Jan 27 15:27:53 crc kubenswrapper[4772]: I0127 15:27:53.775537 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" event={"ID":"ce081402-0ada-4fbf-8b22-eb88a50e804b","Type":"ContainerStarted","Data":"672b3b869d3cac682f2871f1cccebaabe0e2baa030c87285ea2dc87b60951bbf"} Jan 27 15:27:53 crc kubenswrapper[4772]: I0127 15:27:53.775852 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" Jan 27 15:27:53 crc kubenswrapper[4772]: I0127 15:27:53.802104 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" podStartSLOduration=4.802081813 podStartE2EDuration="4.802081813s" podCreationTimestamp="2026-01-27 15:27:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:27:53.799205539 +0000 UTC m=+1259.779814637" watchObservedRunningTime="2026-01-27 15:27:53.802081813 +0000 UTC m=+1259.782690911" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.353678 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.679802 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-659485ddbb-5bnzg"] Jan 27 15:27:54 crc kubenswrapper[4772]: E0127 15:27:54.680132 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eca3ffef-1a57-4aee-9302-64b59ee0fc44" containerName="dnsmasq-dns" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.680147 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="eca3ffef-1a57-4aee-9302-64b59ee0fc44" containerName="dnsmasq-dns" Jan 27 15:27:54 crc kubenswrapper[4772]: E0127 15:27:54.680186 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eca3ffef-1a57-4aee-9302-64b59ee0fc44" containerName="init" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.680195 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="eca3ffef-1a57-4aee-9302-64b59ee0fc44" containerName="init" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.680411 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="eca3ffef-1a57-4aee-9302-64b59ee0fc44" containerName="dnsmasq-dns" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.681388 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.684031 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.687672 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.692528 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-659485ddbb-5bnzg"] Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.759511 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.791308 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7b7d9f3f-e366-421e-b00d-9c453da1adca","Type":"ContainerStarted","Data":"7232aaea46edf2977ebe24eeb1188331c23b40d3efeaaca7dfc75e3658d209b6"} Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.809629 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"da515cae-40ac-41af-aef5-9cef9f3b366e","Type":"ContainerStarted","Data":"537d1bcf9f874816a464389d56071423783d326469277b895f554d80f412795e"} Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.809638 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="da515cae-40ac-41af-aef5-9cef9f3b366e" containerName="cinder-api-log" containerID="cri-o://cd2c8c5778f70b0c56d0f79379ec5079a2ceb539c1d125108e64ede41e68dc9b" gracePeriod=30 Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.809683 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.809771 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="da515cae-40ac-41af-aef5-9cef9f3b366e" containerName="cinder-api" containerID="cri-o://537d1bcf9f874816a464389d56071423783d326469277b895f554d80f412795e" gracePeriod=30 Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.815756 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.873166652 podStartE2EDuration="5.815742358s" podCreationTimestamp="2026-01-27 15:27:49 +0000 UTC" firstStartedPulling="2026-01-27 15:27:50.463871474 +0000 UTC m=+1256.444480572" lastFinishedPulling="2026-01-27 15:27:52.40644718 +0000 UTC m=+1258.387056278" observedRunningTime="2026-01-27 15:27:54.81444356 +0000 UTC m=+1260.795052658" watchObservedRunningTime="2026-01-27 15:27:54.815742358 +0000 UTC m=+1260.796351456" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.826520 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-public-tls-certs\") pod \"barbican-api-659485ddbb-5bnzg\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.826576 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/766c2a26-46ea-41b2-ba0c-2101ec9477d5-logs\") pod \"barbican-api-659485ddbb-5bnzg\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.826732 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-combined-ca-bundle\") pod \"barbican-api-659485ddbb-5bnzg\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.826761 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-internal-tls-certs\") pod \"barbican-api-659485ddbb-5bnzg\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.826816 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-config-data-custom\") pod \"barbican-api-659485ddbb-5bnzg\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.826879 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-config-data\") pod \"barbican-api-659485ddbb-5bnzg\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.826913 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqskw\" (UniqueName: \"kubernetes.io/projected/766c2a26-46ea-41b2-ba0c-2101ec9477d5-kube-api-access-sqskw\") pod \"barbican-api-659485ddbb-5bnzg\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.864075 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.8640541729999995 podStartE2EDuration="5.864054173s" podCreationTimestamp="2026-01-27 15:27:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:27:54.847566254 +0000 UTC m=+1260.828175372" watchObservedRunningTime="2026-01-27 15:27:54.864054173 +0000 UTC m=+1260.844663271" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.929261 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-combined-ca-bundle\") pod \"barbican-api-659485ddbb-5bnzg\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.929330 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-internal-tls-certs\") pod \"barbican-api-659485ddbb-5bnzg\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.929396 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-config-data-custom\") pod \"barbican-api-659485ddbb-5bnzg\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.929475 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-config-data\") pod \"barbican-api-659485ddbb-5bnzg\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.929507 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqskw\" (UniqueName: \"kubernetes.io/projected/766c2a26-46ea-41b2-ba0c-2101ec9477d5-kube-api-access-sqskw\") pod \"barbican-api-659485ddbb-5bnzg\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.929579 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-public-tls-certs\") pod \"barbican-api-659485ddbb-5bnzg\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.929597 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/766c2a26-46ea-41b2-ba0c-2101ec9477d5-logs\") pod \"barbican-api-659485ddbb-5bnzg\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.931300 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/766c2a26-46ea-41b2-ba0c-2101ec9477d5-logs\") pod \"barbican-api-659485ddbb-5bnzg\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.943125 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-combined-ca-bundle\") pod \"barbican-api-659485ddbb-5bnzg\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.944225 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-config-data\") pod \"barbican-api-659485ddbb-5bnzg\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.944722 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-public-tls-certs\") pod \"barbican-api-659485ddbb-5bnzg\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.944859 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-internal-tls-certs\") pod \"barbican-api-659485ddbb-5bnzg\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.951280 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-config-data-custom\") pod \"barbican-api-659485ddbb-5bnzg\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:54 crc kubenswrapper[4772]: I0127 15:27:54.981413 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqskw\" (UniqueName: \"kubernetes.io/projected/766c2a26-46ea-41b2-ba0c-2101ec9477d5-kube-api-access-sqskw\") pod \"barbican-api-659485ddbb-5bnzg\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.036255 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.052764 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.729862 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-659485ddbb-5bnzg"] Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.825001 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-659485ddbb-5bnzg" event={"ID":"766c2a26-46ea-41b2-ba0c-2101ec9477d5","Type":"ContainerStarted","Data":"181aa9237802812b703a88787d1d6892177f6147a0214d407241520c82b45857"} Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.829930 4772 generic.go:334] "Generic (PLEG): container finished" podID="b0625578-3b48-44c7-9082-174fce3a7e74" containerID="d2de8b3a1c27ebd01b5c3393c6dcb85d202fe549eef0c41d0f9f318c3b15d219" exitCode=0 Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.829989 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-v689b" event={"ID":"b0625578-3b48-44c7-9082-174fce3a7e74","Type":"ContainerDied","Data":"d2de8b3a1c27ebd01b5c3393c6dcb85d202fe549eef0c41d0f9f318c3b15d219"} Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.833099 4772 generic.go:334] "Generic (PLEG): container finished" podID="da515cae-40ac-41af-aef5-9cef9f3b366e" containerID="537d1bcf9f874816a464389d56071423783d326469277b895f554d80f412795e" exitCode=0 Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.833319 4772 generic.go:334] "Generic (PLEG): container finished" podID="da515cae-40ac-41af-aef5-9cef9f3b366e" containerID="cd2c8c5778f70b0c56d0f79379ec5079a2ceb539c1d125108e64ede41e68dc9b" exitCode=143 Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.833347 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"da515cae-40ac-41af-aef5-9cef9f3b366e","Type":"ContainerDied","Data":"537d1bcf9f874816a464389d56071423783d326469277b895f554d80f412795e"} Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.834429 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"da515cae-40ac-41af-aef5-9cef9f3b366e","Type":"ContainerDied","Data":"cd2c8c5778f70b0c56d0f79379ec5079a2ceb539c1d125108e64ede41e68dc9b"} Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.841094 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.973937 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da515cae-40ac-41af-aef5-9cef9f3b366e-config-data\") pod \"da515cae-40ac-41af-aef5-9cef9f3b366e\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.974253 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da515cae-40ac-41af-aef5-9cef9f3b366e-combined-ca-bundle\") pod \"da515cae-40ac-41af-aef5-9cef9f3b366e\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.974397 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da515cae-40ac-41af-aef5-9cef9f3b366e-etc-machine-id\") pod \"da515cae-40ac-41af-aef5-9cef9f3b366e\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.974549 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4l2r7\" (UniqueName: \"kubernetes.io/projected/da515cae-40ac-41af-aef5-9cef9f3b366e-kube-api-access-4l2r7\") pod \"da515cae-40ac-41af-aef5-9cef9f3b366e\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.974655 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da515cae-40ac-41af-aef5-9cef9f3b366e-logs\") pod \"da515cae-40ac-41af-aef5-9cef9f3b366e\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.974790 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da515cae-40ac-41af-aef5-9cef9f3b366e-config-data-custom\") pod \"da515cae-40ac-41af-aef5-9cef9f3b366e\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.975050 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da515cae-40ac-41af-aef5-9cef9f3b366e-scripts\") pod \"da515cae-40ac-41af-aef5-9cef9f3b366e\" (UID: \"da515cae-40ac-41af-aef5-9cef9f3b366e\") " Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.975270 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da515cae-40ac-41af-aef5-9cef9f3b366e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "da515cae-40ac-41af-aef5-9cef9f3b366e" (UID: "da515cae-40ac-41af-aef5-9cef9f3b366e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.975533 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da515cae-40ac-41af-aef5-9cef9f3b366e-logs" (OuterVolumeSpecName: "logs") pod "da515cae-40ac-41af-aef5-9cef9f3b366e" (UID: "da515cae-40ac-41af-aef5-9cef9f3b366e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.975996 4772 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/da515cae-40ac-41af-aef5-9cef9f3b366e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.976092 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da515cae-40ac-41af-aef5-9cef9f3b366e-logs\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.980618 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da515cae-40ac-41af-aef5-9cef9f3b366e-scripts" (OuterVolumeSpecName: "scripts") pod "da515cae-40ac-41af-aef5-9cef9f3b366e" (UID: "da515cae-40ac-41af-aef5-9cef9f3b366e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.985690 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da515cae-40ac-41af-aef5-9cef9f3b366e-kube-api-access-4l2r7" (OuterVolumeSpecName: "kube-api-access-4l2r7") pod "da515cae-40ac-41af-aef5-9cef9f3b366e" (UID: "da515cae-40ac-41af-aef5-9cef9f3b366e"). InnerVolumeSpecName "kube-api-access-4l2r7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:27:55 crc kubenswrapper[4772]: I0127 15:27:55.993663 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da515cae-40ac-41af-aef5-9cef9f3b366e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "da515cae-40ac-41af-aef5-9cef9f3b366e" (UID: "da515cae-40ac-41af-aef5-9cef9f3b366e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.063796 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da515cae-40ac-41af-aef5-9cef9f3b366e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "da515cae-40ac-41af-aef5-9cef9f3b366e" (UID: "da515cae-40ac-41af-aef5-9cef9f3b366e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.077862 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da515cae-40ac-41af-aef5-9cef9f3b366e-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.077897 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da515cae-40ac-41af-aef5-9cef9f3b366e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.077910 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4l2r7\" (UniqueName: \"kubernetes.io/projected/da515cae-40ac-41af-aef5-9cef9f3b366e-kube-api-access-4l2r7\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.077920 4772 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da515cae-40ac-41af-aef5-9cef9f3b366e-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.102293 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da515cae-40ac-41af-aef5-9cef9f3b366e-config-data" (OuterVolumeSpecName: "config-data") pod "da515cae-40ac-41af-aef5-9cef9f3b366e" (UID: "da515cae-40ac-41af-aef5-9cef9f3b366e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.179703 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da515cae-40ac-41af-aef5-9cef9f3b366e-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.187945 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.845597 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.845579 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"da515cae-40ac-41af-aef5-9cef9f3b366e","Type":"ContainerDied","Data":"005ceeb7f240212b9c37236e7c8187a3f42d3b14297b7bdfc9d9274e9fec1f9d"} Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.846072 4772 scope.go:117] "RemoveContainer" containerID="537d1bcf9f874816a464389d56071423783d326469277b895f554d80f412795e" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.849351 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-659485ddbb-5bnzg" event={"ID":"766c2a26-46ea-41b2-ba0c-2101ec9477d5","Type":"ContainerStarted","Data":"ffbe05081a83d720881627c45e3d405aaf574d1db8fa63481da9c229023c0874"} Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.849405 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-659485ddbb-5bnzg" event={"ID":"766c2a26-46ea-41b2-ba0c-2101ec9477d5","Type":"ContainerStarted","Data":"23481794981b6875427087c492230b72f248918903ae28fa47bb73190cfa8573"} Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.878701 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-659485ddbb-5bnzg" podStartSLOduration=2.878681384 podStartE2EDuration="2.878681384s" podCreationTimestamp="2026-01-27 15:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:27:56.870958299 +0000 UTC m=+1262.851567397" watchObservedRunningTime="2026-01-27 15:27:56.878681384 +0000 UTC m=+1262.859290482" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.895244 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.898009 4772 scope.go:117] "RemoveContainer" containerID="cd2c8c5778f70b0c56d0f79379ec5079a2ceb539c1d125108e64ede41e68dc9b" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.904073 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.920230 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Jan 27 15:27:56 crc kubenswrapper[4772]: E0127 15:27:56.920613 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da515cae-40ac-41af-aef5-9cef9f3b366e" containerName="cinder-api" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.920632 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="da515cae-40ac-41af-aef5-9cef9f3b366e" containerName="cinder-api" Jan 27 15:27:56 crc kubenswrapper[4772]: E0127 15:27:56.920653 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da515cae-40ac-41af-aef5-9cef9f3b366e" containerName="cinder-api-log" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.920660 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="da515cae-40ac-41af-aef5-9cef9f3b366e" containerName="cinder-api-log" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.920821 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="da515cae-40ac-41af-aef5-9cef9f3b366e" containerName="cinder-api" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.920841 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="da515cae-40ac-41af-aef5-9cef9f3b366e" containerName="cinder-api-log" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.923252 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.929181 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.929476 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.929817 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.933536 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.994068 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-public-tls-certs\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.994184 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-config-data-custom\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.994223 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-scripts\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.994244 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.994314 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vclj\" (UniqueName: \"kubernetes.io/projected/be772158-a71c-448d-8972-014f0d3a9ab8-kube-api-access-2vclj\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.994359 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.994490 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be772158-a71c-448d-8972-014f0d3a9ab8-logs\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.994528 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-config-data\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:56 crc kubenswrapper[4772]: I0127 15:27:56.994604 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/be772158-a71c-448d-8972-014f0d3a9ab8-etc-machine-id\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.097224 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-public-tls-certs\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.097302 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-config-data-custom\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.097353 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-scripts\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.097374 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.097402 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vclj\" (UniqueName: \"kubernetes.io/projected/be772158-a71c-448d-8972-014f0d3a9ab8-kube-api-access-2vclj\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.097449 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.097531 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be772158-a71c-448d-8972-014f0d3a9ab8-logs\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.097554 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-config-data\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.097611 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/be772158-a71c-448d-8972-014f0d3a9ab8-etc-machine-id\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.097683 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/be772158-a71c-448d-8972-014f0d3a9ab8-etc-machine-id\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.101737 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be772158-a71c-448d-8972-014f0d3a9ab8-logs\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.102938 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-public-tls-certs\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.103045 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.103799 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.105989 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-config-data\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.111209 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-config-data-custom\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.115246 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-scripts\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.133504 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vclj\" (UniqueName: \"kubernetes.io/projected/be772158-a71c-448d-8972-014f0d3a9ab8-kube-api-access-2vclj\") pod \"cinder-api-0\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " pod="openstack/cinder-api-0" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.271664 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.316212 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-v689b" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.409028 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0625578-3b48-44c7-9082-174fce3a7e74-combined-ca-bundle\") pod \"b0625578-3b48-44c7-9082-174fce3a7e74\" (UID: \"b0625578-3b48-44c7-9082-174fce3a7e74\") " Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.409099 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b0625578-3b48-44c7-9082-174fce3a7e74-config\") pod \"b0625578-3b48-44c7-9082-174fce3a7e74\" (UID: \"b0625578-3b48-44c7-9082-174fce3a7e74\") " Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.409235 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jn9b\" (UniqueName: \"kubernetes.io/projected/b0625578-3b48-44c7-9082-174fce3a7e74-kube-api-access-9jn9b\") pod \"b0625578-3b48-44c7-9082-174fce3a7e74\" (UID: \"b0625578-3b48-44c7-9082-174fce3a7e74\") " Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.415348 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0625578-3b48-44c7-9082-174fce3a7e74-kube-api-access-9jn9b" (OuterVolumeSpecName: "kube-api-access-9jn9b") pod "b0625578-3b48-44c7-9082-174fce3a7e74" (UID: "b0625578-3b48-44c7-9082-174fce3a7e74"). InnerVolumeSpecName "kube-api-access-9jn9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.434488 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0625578-3b48-44c7-9082-174fce3a7e74-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b0625578-3b48-44c7-9082-174fce3a7e74" (UID: "b0625578-3b48-44c7-9082-174fce3a7e74"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.436800 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0625578-3b48-44c7-9082-174fce3a7e74-config" (OuterVolumeSpecName: "config") pod "b0625578-3b48-44c7-9082-174fce3a7e74" (UID: "b0625578-3b48-44c7-9082-174fce3a7e74"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.512650 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jn9b\" (UniqueName: \"kubernetes.io/projected/b0625578-3b48-44c7-9082-174fce3a7e74-kube-api-access-9jn9b\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.512694 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0625578-3b48-44c7-9082-174fce3a7e74-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.512732 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/b0625578-3b48-44c7-9082-174fce3a7e74-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.857345 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Jan 27 15:27:57 crc kubenswrapper[4772]: E0127 15:27:57.859052 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0625578-3b48-44c7-9082-174fce3a7e74" containerName="neutron-db-sync" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.859073 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0625578-3b48-44c7-9082-174fce3a7e74" containerName="neutron-db-sync" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.859339 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0625578-3b48-44c7-9082-174fce3a7e74" containerName="neutron-db-sync" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.860123 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.873567 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.876579 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.876651 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-7d9v8" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.876771 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.884905 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-v689b" event={"ID":"b0625578-3b48-44c7-9082-174fce3a7e74","Type":"ContainerDied","Data":"b3123ce803c91e7738d4af911f91769cd0703aad347549f4989b2ccc532f36ea"} Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.884963 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3123ce803c91e7738d4af911f91769cd0703aad347549f4989b2ccc532f36ea" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.884963 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-v689b" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.893840 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:57 crc kubenswrapper[4772]: I0127 15:27:57.893894 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.037800 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0edf6707-14dd-4986-8d64-0e48a31d6a39-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0edf6707-14dd-4986-8d64-0e48a31d6a39\") " pod="openstack/openstackclient" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.037879 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0edf6707-14dd-4986-8d64-0e48a31d6a39-openstack-config\") pod \"openstackclient\" (UID: \"0edf6707-14dd-4986-8d64-0e48a31d6a39\") " pod="openstack/openstackclient" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.037919 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0edf6707-14dd-4986-8d64-0e48a31d6a39-openstack-config-secret\") pod \"openstackclient\" (UID: \"0edf6707-14dd-4986-8d64-0e48a31d6a39\") " pod="openstack/openstackclient" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.038113 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkhnj\" (UniqueName: \"kubernetes.io/projected/0edf6707-14dd-4986-8d64-0e48a31d6a39-kube-api-access-xkhnj\") pod \"openstackclient\" (UID: \"0edf6707-14dd-4986-8d64-0e48a31d6a39\") " pod="openstack/openstackclient" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.082095 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69c986f6d7-gszgg"] Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.087825 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" podUID="ce081402-0ada-4fbf-8b22-eb88a50e804b" containerName="dnsmasq-dns" containerID="cri-o://672b3b869d3cac682f2871f1cccebaabe0e2baa030c87285ea2dc87b60951bbf" gracePeriod=10 Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.091387 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.121583 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-2849v"] Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.123236 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-2849v" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.144973 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkhnj\" (UniqueName: \"kubernetes.io/projected/0edf6707-14dd-4986-8d64-0e48a31d6a39-kube-api-access-xkhnj\") pod \"openstackclient\" (UID: \"0edf6707-14dd-4986-8d64-0e48a31d6a39\") " pod="openstack/openstackclient" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.145065 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0edf6707-14dd-4986-8d64-0e48a31d6a39-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0edf6707-14dd-4986-8d64-0e48a31d6a39\") " pod="openstack/openstackclient" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.145114 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0edf6707-14dd-4986-8d64-0e48a31d6a39-openstack-config\") pod \"openstackclient\" (UID: \"0edf6707-14dd-4986-8d64-0e48a31d6a39\") " pod="openstack/openstackclient" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.145138 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0edf6707-14dd-4986-8d64-0e48a31d6a39-openstack-config-secret\") pod \"openstackclient\" (UID: \"0edf6707-14dd-4986-8d64-0e48a31d6a39\") " pod="openstack/openstackclient" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.147898 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0edf6707-14dd-4986-8d64-0e48a31d6a39-openstack-config\") pod \"openstackclient\" (UID: \"0edf6707-14dd-4986-8d64-0e48a31d6a39\") " pod="openstack/openstackclient" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.155186 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0edf6707-14dd-4986-8d64-0e48a31d6a39-openstack-config-secret\") pod \"openstackclient\" (UID: \"0edf6707-14dd-4986-8d64-0e48a31d6a39\") " pod="openstack/openstackclient" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.157035 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0edf6707-14dd-4986-8d64-0e48a31d6a39-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0edf6707-14dd-4986-8d64-0e48a31d6a39\") " pod="openstack/openstackclient" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.159242 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-2849v"] Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.179814 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkhnj\" (UniqueName: \"kubernetes.io/projected/0edf6707-14dd-4986-8d64-0e48a31d6a39-kube-api-access-xkhnj\") pod \"openstackclient\" (UID: \"0edf6707-14dd-4986-8d64-0e48a31d6a39\") " pod="openstack/openstackclient" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.194983 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.250146 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-dns-svc\") pod \"dnsmasq-dns-5784cf869f-2849v\" (UID: \"4ab060da-8587-413a-a410-ee0e9cec40c6\") " pod="openstack/dnsmasq-dns-5784cf869f-2849v" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.250241 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-2849v\" (UID: \"4ab060da-8587-413a-a410-ee0e9cec40c6\") " pod="openstack/dnsmasq-dns-5784cf869f-2849v" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.250280 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-config\") pod \"dnsmasq-dns-5784cf869f-2849v\" (UID: \"4ab060da-8587-413a-a410-ee0e9cec40c6\") " pod="openstack/dnsmasq-dns-5784cf869f-2849v" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.250366 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s774r\" (UniqueName: \"kubernetes.io/projected/4ab060da-8587-413a-a410-ee0e9cec40c6-kube-api-access-s774r\") pod \"dnsmasq-dns-5784cf869f-2849v\" (UID: \"4ab060da-8587-413a-a410-ee0e9cec40c6\") " pod="openstack/dnsmasq-dns-5784cf869f-2849v" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.250397 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-2849v\" (UID: \"4ab060da-8587-413a-a410-ee0e9cec40c6\") " pod="openstack/dnsmasq-dns-5784cf869f-2849v" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.250423 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-2849v\" (UID: \"4ab060da-8587-413a-a410-ee0e9cec40c6\") " pod="openstack/dnsmasq-dns-5784cf869f-2849v" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.298094 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-66bf894476-wz7b5"] Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.309314 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66bf894476-wz7b5" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.313275 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.313505 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.313652 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.316500 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-rjm9r" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.320776 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-66bf894476-wz7b5"] Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.355307 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s774r\" (UniqueName: \"kubernetes.io/projected/4ab060da-8587-413a-a410-ee0e9cec40c6-kube-api-access-s774r\") pod \"dnsmasq-dns-5784cf869f-2849v\" (UID: \"4ab060da-8587-413a-a410-ee0e9cec40c6\") " pod="openstack/dnsmasq-dns-5784cf869f-2849v" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.355368 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-2849v\" (UID: \"4ab060da-8587-413a-a410-ee0e9cec40c6\") " pod="openstack/dnsmasq-dns-5784cf869f-2849v" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.355405 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-2849v\" (UID: \"4ab060da-8587-413a-a410-ee0e9cec40c6\") " pod="openstack/dnsmasq-dns-5784cf869f-2849v" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.355442 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-dns-svc\") pod \"dnsmasq-dns-5784cf869f-2849v\" (UID: \"4ab060da-8587-413a-a410-ee0e9cec40c6\") " pod="openstack/dnsmasq-dns-5784cf869f-2849v" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.355481 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-2849v\" (UID: \"4ab060da-8587-413a-a410-ee0e9cec40c6\") " pod="openstack/dnsmasq-dns-5784cf869f-2849v" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.355521 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-config\") pod \"dnsmasq-dns-5784cf869f-2849v\" (UID: \"4ab060da-8587-413a-a410-ee0e9cec40c6\") " pod="openstack/dnsmasq-dns-5784cf869f-2849v" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.356589 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-config\") pod \"dnsmasq-dns-5784cf869f-2849v\" (UID: \"4ab060da-8587-413a-a410-ee0e9cec40c6\") " pod="openstack/dnsmasq-dns-5784cf869f-2849v" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.356872 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-2849v\" (UID: \"4ab060da-8587-413a-a410-ee0e9cec40c6\") " pod="openstack/dnsmasq-dns-5784cf869f-2849v" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.357156 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-2849v\" (UID: \"4ab060da-8587-413a-a410-ee0e9cec40c6\") " pod="openstack/dnsmasq-dns-5784cf869f-2849v" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.357254 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-dns-svc\") pod \"dnsmasq-dns-5784cf869f-2849v\" (UID: \"4ab060da-8587-413a-a410-ee0e9cec40c6\") " pod="openstack/dnsmasq-dns-5784cf869f-2849v" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.357796 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-2849v\" (UID: \"4ab060da-8587-413a-a410-ee0e9cec40c6\") " pod="openstack/dnsmasq-dns-5784cf869f-2849v" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.391105 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s774r\" (UniqueName: \"kubernetes.io/projected/4ab060da-8587-413a-a410-ee0e9cec40c6-kube-api-access-s774r\") pod \"dnsmasq-dns-5784cf869f-2849v\" (UID: \"4ab060da-8587-413a-a410-ee0e9cec40c6\") " pod="openstack/dnsmasq-dns-5784cf869f-2849v" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.457381 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb8pb\" (UniqueName: \"kubernetes.io/projected/e7385520-8ffb-40e5-802e-ff0db348c5c1-kube-api-access-bb8pb\") pod \"neutron-66bf894476-wz7b5\" (UID: \"e7385520-8ffb-40e5-802e-ff0db348c5c1\") " pod="openstack/neutron-66bf894476-wz7b5" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.457464 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7385520-8ffb-40e5-802e-ff0db348c5c1-combined-ca-bundle\") pod \"neutron-66bf894476-wz7b5\" (UID: \"e7385520-8ffb-40e5-802e-ff0db348c5c1\") " pod="openstack/neutron-66bf894476-wz7b5" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.457528 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e7385520-8ffb-40e5-802e-ff0db348c5c1-config\") pod \"neutron-66bf894476-wz7b5\" (UID: \"e7385520-8ffb-40e5-802e-ff0db348c5c1\") " pod="openstack/neutron-66bf894476-wz7b5" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.457798 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7385520-8ffb-40e5-802e-ff0db348c5c1-ovndb-tls-certs\") pod \"neutron-66bf894476-wz7b5\" (UID: \"e7385520-8ffb-40e5-802e-ff0db348c5c1\") " pod="openstack/neutron-66bf894476-wz7b5" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.457884 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e7385520-8ffb-40e5-802e-ff0db348c5c1-httpd-config\") pod \"neutron-66bf894476-wz7b5\" (UID: \"e7385520-8ffb-40e5-802e-ff0db348c5c1\") " pod="openstack/neutron-66bf894476-wz7b5" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.560276 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb8pb\" (UniqueName: \"kubernetes.io/projected/e7385520-8ffb-40e5-802e-ff0db348c5c1-kube-api-access-bb8pb\") pod \"neutron-66bf894476-wz7b5\" (UID: \"e7385520-8ffb-40e5-802e-ff0db348c5c1\") " pod="openstack/neutron-66bf894476-wz7b5" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.560379 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7385520-8ffb-40e5-802e-ff0db348c5c1-combined-ca-bundle\") pod \"neutron-66bf894476-wz7b5\" (UID: \"e7385520-8ffb-40e5-802e-ff0db348c5c1\") " pod="openstack/neutron-66bf894476-wz7b5" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.560431 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e7385520-8ffb-40e5-802e-ff0db348c5c1-config\") pod \"neutron-66bf894476-wz7b5\" (UID: \"e7385520-8ffb-40e5-802e-ff0db348c5c1\") " pod="openstack/neutron-66bf894476-wz7b5" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.560597 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7385520-8ffb-40e5-802e-ff0db348c5c1-ovndb-tls-certs\") pod \"neutron-66bf894476-wz7b5\" (UID: \"e7385520-8ffb-40e5-802e-ff0db348c5c1\") " pod="openstack/neutron-66bf894476-wz7b5" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.560643 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e7385520-8ffb-40e5-802e-ff0db348c5c1-httpd-config\") pod \"neutron-66bf894476-wz7b5\" (UID: \"e7385520-8ffb-40e5-802e-ff0db348c5c1\") " pod="openstack/neutron-66bf894476-wz7b5" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.565218 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7385520-8ffb-40e5-802e-ff0db348c5c1-ovndb-tls-certs\") pod \"neutron-66bf894476-wz7b5\" (UID: \"e7385520-8ffb-40e5-802e-ff0db348c5c1\") " pod="openstack/neutron-66bf894476-wz7b5" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.566487 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7385520-8ffb-40e5-802e-ff0db348c5c1-combined-ca-bundle\") pod \"neutron-66bf894476-wz7b5\" (UID: \"e7385520-8ffb-40e5-802e-ff0db348c5c1\") " pod="openstack/neutron-66bf894476-wz7b5" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.567216 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e7385520-8ffb-40e5-802e-ff0db348c5c1-config\") pod \"neutron-66bf894476-wz7b5\" (UID: \"e7385520-8ffb-40e5-802e-ff0db348c5c1\") " pod="openstack/neutron-66bf894476-wz7b5" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.579011 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-2849v" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.581878 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e7385520-8ffb-40e5-802e-ff0db348c5c1-httpd-config\") pod \"neutron-66bf894476-wz7b5\" (UID: \"e7385520-8ffb-40e5-802e-ff0db348c5c1\") " pod="openstack/neutron-66bf894476-wz7b5" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.583900 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb8pb\" (UniqueName: \"kubernetes.io/projected/e7385520-8ffb-40e5-802e-ff0db348c5c1-kube-api-access-bb8pb\") pod \"neutron-66bf894476-wz7b5\" (UID: \"e7385520-8ffb-40e5-802e-ff0db348c5c1\") " pod="openstack/neutron-66bf894476-wz7b5" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.639245 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66bf894476-wz7b5" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.681966 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da515cae-40ac-41af-aef5-9cef9f3b366e" path="/var/lib/kubelet/pods/da515cae-40ac-41af-aef5-9cef9f3b366e/volumes" Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.969839 4772 generic.go:334] "Generic (PLEG): container finished" podID="ce081402-0ada-4fbf-8b22-eb88a50e804b" containerID="672b3b869d3cac682f2871f1cccebaabe0e2baa030c87285ea2dc87b60951bbf" exitCode=0 Jan 27 15:27:58 crc kubenswrapper[4772]: I0127 15:27:58.971692 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" event={"ID":"ce081402-0ada-4fbf-8b22-eb88a50e804b","Type":"ContainerDied","Data":"672b3b869d3cac682f2871f1cccebaabe0e2baa030c87285ea2dc87b60951bbf"} Jan 27 15:27:59 crc kubenswrapper[4772]: I0127 15:27:59.477889 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" Jan 27 15:27:59 crc kubenswrapper[4772]: I0127 15:27:59.569939 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Jan 27 15:27:59 crc kubenswrapper[4772]: I0127 15:27:59.590401 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-config\") pod \"ce081402-0ada-4fbf-8b22-eb88a50e804b\" (UID: \"ce081402-0ada-4fbf-8b22-eb88a50e804b\") " Jan 27 15:27:59 crc kubenswrapper[4772]: I0127 15:27:59.590503 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpwmg\" (UniqueName: \"kubernetes.io/projected/ce081402-0ada-4fbf-8b22-eb88a50e804b-kube-api-access-hpwmg\") pod \"ce081402-0ada-4fbf-8b22-eb88a50e804b\" (UID: \"ce081402-0ada-4fbf-8b22-eb88a50e804b\") " Jan 27 15:27:59 crc kubenswrapper[4772]: I0127 15:27:59.590646 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-ovsdbserver-sb\") pod \"ce081402-0ada-4fbf-8b22-eb88a50e804b\" (UID: \"ce081402-0ada-4fbf-8b22-eb88a50e804b\") " Jan 27 15:27:59 crc kubenswrapper[4772]: I0127 15:27:59.590752 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-ovsdbserver-nb\") pod \"ce081402-0ada-4fbf-8b22-eb88a50e804b\" (UID: \"ce081402-0ada-4fbf-8b22-eb88a50e804b\") " Jan 27 15:27:59 crc kubenswrapper[4772]: I0127 15:27:59.590787 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-dns-swift-storage-0\") pod \"ce081402-0ada-4fbf-8b22-eb88a50e804b\" (UID: \"ce081402-0ada-4fbf-8b22-eb88a50e804b\") " Jan 27 15:27:59 crc kubenswrapper[4772]: I0127 15:27:59.590811 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-dns-svc\") pod \"ce081402-0ada-4fbf-8b22-eb88a50e804b\" (UID: \"ce081402-0ada-4fbf-8b22-eb88a50e804b\") " Jan 27 15:27:59 crc kubenswrapper[4772]: I0127 15:27:59.597873 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 27 15:27:59 crc kubenswrapper[4772]: I0127 15:27:59.626649 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce081402-0ada-4fbf-8b22-eb88a50e804b-kube-api-access-hpwmg" (OuterVolumeSpecName: "kube-api-access-hpwmg") pod "ce081402-0ada-4fbf-8b22-eb88a50e804b" (UID: "ce081402-0ada-4fbf-8b22-eb88a50e804b"). InnerVolumeSpecName "kube-api-access-hpwmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:27:59 crc kubenswrapper[4772]: W0127 15:27:59.654309 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe772158_a71c_448d_8972_014f0d3a9ab8.slice/crio-2e68d940e0eebbc1216da3357187ae70827b7d508fb0a26f0e91d9593aac8852 WatchSource:0}: Error finding container 2e68d940e0eebbc1216da3357187ae70827b7d508fb0a26f0e91d9593aac8852: Status 404 returned error can't find the container with id 2e68d940e0eebbc1216da3357187ae70827b7d508fb0a26f0e91d9593aac8852 Jan 27 15:27:59 crc kubenswrapper[4772]: I0127 15:27:59.692556 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpwmg\" (UniqueName: \"kubernetes.io/projected/ce081402-0ada-4fbf-8b22-eb88a50e804b-kube-api-access-hpwmg\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:59 crc kubenswrapper[4772]: I0127 15:27:59.806349 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-2849v"] Jan 27 15:27:59 crc kubenswrapper[4772]: I0127 15:27:59.807457 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ce081402-0ada-4fbf-8b22-eb88a50e804b" (UID: "ce081402-0ada-4fbf-8b22-eb88a50e804b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:27:59 crc kubenswrapper[4772]: I0127 15:27:59.813773 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ce081402-0ada-4fbf-8b22-eb88a50e804b" (UID: "ce081402-0ada-4fbf-8b22-eb88a50e804b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:27:59 crc kubenswrapper[4772]: I0127 15:27:59.841931 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-66bf894476-wz7b5"] Jan 27 15:27:59 crc kubenswrapper[4772]: I0127 15:27:59.862604 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-config" (OuterVolumeSpecName: "config") pod "ce081402-0ada-4fbf-8b22-eb88a50e804b" (UID: "ce081402-0ada-4fbf-8b22-eb88a50e804b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:27:59 crc kubenswrapper[4772]: I0127 15:27:59.892932 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ce081402-0ada-4fbf-8b22-eb88a50e804b" (UID: "ce081402-0ada-4fbf-8b22-eb88a50e804b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:27:59 crc kubenswrapper[4772]: I0127 15:27:59.898719 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:59 crc kubenswrapper[4772]: I0127 15:27:59.898762 4772 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:59 crc kubenswrapper[4772]: I0127 15:27:59.898781 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:59 crc kubenswrapper[4772]: I0127 15:27:59.898796 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:27:59 crc kubenswrapper[4772]: I0127 15:27:59.902492 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ce081402-0ada-4fbf-8b22-eb88a50e804b" (UID: "ce081402-0ada-4fbf-8b22-eb88a50e804b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:27:59 crc kubenswrapper[4772]: I0127 15:27:59.994642 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"be772158-a71c-448d-8972-014f0d3a9ab8","Type":"ContainerStarted","Data":"2e68d940e0eebbc1216da3357187ae70827b7d508fb0a26f0e91d9593aac8852"} Jan 27 15:28:00 crc kubenswrapper[4772]: I0127 15:28:00.001071 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce081402-0ada-4fbf-8b22-eb88a50e804b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:00 crc kubenswrapper[4772]: I0127 15:28:00.008587 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" event={"ID":"ce081402-0ada-4fbf-8b22-eb88a50e804b","Type":"ContainerDied","Data":"59a55a087cb1a139692664b7e8a0e67c4b8ddc2ffa8baef3cc11190a48375152"} Jan 27 15:28:00 crc kubenswrapper[4772]: I0127 15:28:00.008601 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69c986f6d7-gszgg" Jan 27 15:28:00 crc kubenswrapper[4772]: I0127 15:28:00.008642 4772 scope.go:117] "RemoveContainer" containerID="672b3b869d3cac682f2871f1cccebaabe0e2baa030c87285ea2dc87b60951bbf" Jan 27 15:28:00 crc kubenswrapper[4772]: I0127 15:28:00.031647 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66bf894476-wz7b5" event={"ID":"e7385520-8ffb-40e5-802e-ff0db348c5c1","Type":"ContainerStarted","Data":"d30df4d73e5cfb24af9149a2561d9917bae2965d7778272568f0bfb1966f855f"} Jan 27 15:28:00 crc kubenswrapper[4772]: I0127 15:28:00.064564 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69c986f6d7-gszgg"] Jan 27 15:28:00 crc kubenswrapper[4772]: I0127 15:28:00.064623 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-2849v" event={"ID":"4ab060da-8587-413a-a410-ee0e9cec40c6","Type":"ContainerStarted","Data":"894acb53cf18e87f7e2c3c3b36e872c3494f3f5477487f3a865082faee222b95"} Jan 27 15:28:00 crc kubenswrapper[4772]: I0127 15:28:00.075768 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-69c986f6d7-gszgg"] Jan 27 15:28:00 crc kubenswrapper[4772]: I0127 15:28:00.083454 4772 scope.go:117] "RemoveContainer" containerID="dc64f95e60b6b7231bfc72557652d02f45abe891bf9e4367174098a9820da0e9" Jan 27 15:28:00 crc kubenswrapper[4772]: I0127 15:28:00.084291 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"0edf6707-14dd-4986-8d64-0e48a31d6a39","Type":"ContainerStarted","Data":"5c52f5cf3b82427db2a187bbd0708a64e4f14f826b96324500c229ad2e72a4cf"} Jan 27 15:28:00 crc kubenswrapper[4772]: I0127 15:28:00.385961 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Jan 27 15:28:00 crc kubenswrapper[4772]: I0127 15:28:00.607442 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-cb9d976b-flrwl" Jan 27 15:28:00 crc kubenswrapper[4772]: I0127 15:28:00.616843 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 27 15:28:00 crc kubenswrapper[4772]: I0127 15:28:00.688706 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce081402-0ada-4fbf-8b22-eb88a50e804b" path="/var/lib/kubelet/pods/ce081402-0ada-4fbf-8b22-eb88a50e804b/volumes" Jan 27 15:28:00 crc kubenswrapper[4772]: I0127 15:28:00.689360 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-597699949b-q6msx" Jan 27 15:28:00 crc kubenswrapper[4772]: I0127 15:28:00.940733 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-cb9d976b-flrwl" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.154198 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66bf894476-wz7b5" event={"ID":"e7385520-8ffb-40e5-802e-ff0db348c5c1","Type":"ContainerStarted","Data":"b5a8f7019a8ae14ffdea4c25f43d7ff45e4469316acbf03b2364b347f5933e7c"} Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.154234 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66bf894476-wz7b5" event={"ID":"e7385520-8ffb-40e5-802e-ff0db348c5c1","Type":"ContainerStarted","Data":"8859f4bb50887ba9951c0e2249a3e56deff79409c3a080683519e71c92360a6d"} Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.154531 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-66bf894476-wz7b5" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.161844 4772 generic.go:334] "Generic (PLEG): container finished" podID="4ab060da-8587-413a-a410-ee0e9cec40c6" containerID="d9accb6fdf89d9e80604718e7ef1a89b03857412edcadace6f8fed5be8e5dfab" exitCode=0 Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.161947 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-2849v" event={"ID":"4ab060da-8587-413a-a410-ee0e9cec40c6","Type":"ContainerDied","Data":"d9accb6fdf89d9e80604718e7ef1a89b03857412edcadace6f8fed5be8e5dfab"} Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.175767 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"be772158-a71c-448d-8972-014f0d3a9ab8","Type":"ContainerStarted","Data":"26cc6d1f580535edc969fb0f7d0d2e7d716fa8450f944ca1657554f90801529b"} Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.175936 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="7b7d9f3f-e366-421e-b00d-9c453da1adca" containerName="cinder-scheduler" containerID="cri-o://55338fb54abd9ad2a096debb2356f749682191abe6b851127e0e95fcec09a654" gracePeriod=30 Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.176132 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="7b7d9f3f-e366-421e-b00d-9c453da1adca" containerName="probe" containerID="cri-o://7232aaea46edf2977ebe24eeb1188331c23b40d3efeaaca7dfc75e3658d209b6" gracePeriod=30 Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.187318 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-66bf894476-wz7b5" podStartSLOduration=3.18729179 podStartE2EDuration="3.18729179s" podCreationTimestamp="2026-01-27 15:27:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:28:01.176434204 +0000 UTC m=+1267.157043312" watchObservedRunningTime="2026-01-27 15:28:01.18729179 +0000 UTC m=+1267.167900888" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.252063 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-597699949b-q6msx" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.708809 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-647c88bb6f-wzf82"] Jan 27 15:28:01 crc kubenswrapper[4772]: E0127 15:28:01.721828 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce081402-0ada-4fbf-8b22-eb88a50e804b" containerName="init" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.721874 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce081402-0ada-4fbf-8b22-eb88a50e804b" containerName="init" Jan 27 15:28:01 crc kubenswrapper[4772]: E0127 15:28:01.721887 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce081402-0ada-4fbf-8b22-eb88a50e804b" containerName="dnsmasq-dns" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.721895 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce081402-0ada-4fbf-8b22-eb88a50e804b" containerName="dnsmasq-dns" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.722161 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce081402-0ada-4fbf-8b22-eb88a50e804b" containerName="dnsmasq-dns" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.723369 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.727205 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.727385 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.733885 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-647c88bb6f-wzf82"] Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.876345 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-httpd-config\") pod \"neutron-647c88bb6f-wzf82\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.876417 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-internal-tls-certs\") pod \"neutron-647c88bb6f-wzf82\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.876483 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-public-tls-certs\") pod \"neutron-647c88bb6f-wzf82\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.876538 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-combined-ca-bundle\") pod \"neutron-647c88bb6f-wzf82\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.876561 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-ovndb-tls-certs\") pod \"neutron-647c88bb6f-wzf82\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.876604 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brxbx\" (UniqueName: \"kubernetes.io/projected/6cf131c4-a5bd-452b-8598-42312c3a0270-kube-api-access-brxbx\") pod \"neutron-647c88bb6f-wzf82\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.876675 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-config\") pod \"neutron-647c88bb6f-wzf82\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.978124 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-config\") pod \"neutron-647c88bb6f-wzf82\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.978538 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-httpd-config\") pod \"neutron-647c88bb6f-wzf82\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.978587 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-internal-tls-certs\") pod \"neutron-647c88bb6f-wzf82\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.978630 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-public-tls-certs\") pod \"neutron-647c88bb6f-wzf82\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.978710 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-combined-ca-bundle\") pod \"neutron-647c88bb6f-wzf82\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.978815 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-ovndb-tls-certs\") pod \"neutron-647c88bb6f-wzf82\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.978938 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brxbx\" (UniqueName: \"kubernetes.io/projected/6cf131c4-a5bd-452b-8598-42312c3a0270-kube-api-access-brxbx\") pod \"neutron-647c88bb6f-wzf82\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.986017 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-config\") pod \"neutron-647c88bb6f-wzf82\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.986879 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-ovndb-tls-certs\") pod \"neutron-647c88bb6f-wzf82\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.987952 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-internal-tls-certs\") pod \"neutron-647c88bb6f-wzf82\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.988645 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-httpd-config\") pod \"neutron-647c88bb6f-wzf82\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.988884 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-public-tls-certs\") pod \"neutron-647c88bb6f-wzf82\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.992004 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-combined-ca-bundle\") pod \"neutron-647c88bb6f-wzf82\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:01 crc kubenswrapper[4772]: I0127 15:28:01.994821 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brxbx\" (UniqueName: \"kubernetes.io/projected/6cf131c4-a5bd-452b-8598-42312c3a0270-kube-api-access-brxbx\") pod \"neutron-647c88bb6f-wzf82\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:02 crc kubenswrapper[4772]: I0127 15:28:02.078946 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:02 crc kubenswrapper[4772]: I0127 15:28:02.216382 4772 generic.go:334] "Generic (PLEG): container finished" podID="7b7d9f3f-e366-421e-b00d-9c453da1adca" containerID="7232aaea46edf2977ebe24eeb1188331c23b40d3efeaaca7dfc75e3658d209b6" exitCode=0 Jan 27 15:28:02 crc kubenswrapper[4772]: I0127 15:28:02.216467 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7b7d9f3f-e366-421e-b00d-9c453da1adca","Type":"ContainerDied","Data":"7232aaea46edf2977ebe24eeb1188331c23b40d3efeaaca7dfc75e3658d209b6"} Jan 27 15:28:02 crc kubenswrapper[4772]: I0127 15:28:02.251227 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-2849v" event={"ID":"4ab060da-8587-413a-a410-ee0e9cec40c6","Type":"ContainerStarted","Data":"b3b0580b2d9a989010c2055ae938a024c281531976b862414fc303ffddcf01e5"} Jan 27 15:28:02 crc kubenswrapper[4772]: I0127 15:28:02.251307 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-2849v" Jan 27 15:28:02 crc kubenswrapper[4772]: I0127 15:28:02.311745 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-2849v" podStartSLOduration=4.311722635 podStartE2EDuration="4.311722635s" podCreationTimestamp="2026-01-27 15:27:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:28:02.301499628 +0000 UTC m=+1268.282108746" watchObservedRunningTime="2026-01-27 15:28:02.311722635 +0000 UTC m=+1268.292331733" Jan 27 15:28:02 crc kubenswrapper[4772]: I0127 15:28:02.830060 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-647c88bb6f-wzf82"] Jan 27 15:28:03 crc kubenswrapper[4772]: I0127 15:28:03.021265 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:28:03 crc kubenswrapper[4772]: I0127 15:28:03.279348 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"be772158-a71c-448d-8972-014f0d3a9ab8","Type":"ContainerStarted","Data":"c47159ab0aee5087f5a44073988d2ad8d6aaaa0e47ba7702dc2a03eab229b375"} Jan 27 15:28:03 crc kubenswrapper[4772]: I0127 15:28:03.279614 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Jan 27 15:28:03 crc kubenswrapper[4772]: I0127 15:28:03.295046 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-647c88bb6f-wzf82" event={"ID":"6cf131c4-a5bd-452b-8598-42312c3a0270","Type":"ContainerStarted","Data":"72824ad39b806a2254b462f1a46f766a404dd5dd1e5172059745c7930bd54b37"} Jan 27 15:28:03 crc kubenswrapper[4772]: I0127 15:28:03.295107 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-647c88bb6f-wzf82" event={"ID":"6cf131c4-a5bd-452b-8598-42312c3a0270","Type":"ContainerStarted","Data":"eb3fc136e47d75ea92171b2a25f1728b294a61ff0f248fa056a324eadfc98f00"} Jan 27 15:28:03 crc kubenswrapper[4772]: I0127 15:28:03.321044 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=7.321024384 podStartE2EDuration="7.321024384s" podCreationTimestamp="2026-01-27 15:27:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:28:03.303089343 +0000 UTC m=+1269.283698441" watchObservedRunningTime="2026-01-27 15:28:03.321024384 +0000 UTC m=+1269.301633482" Jan 27 15:28:04 crc kubenswrapper[4772]: I0127 15:28:04.314946 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-647c88bb6f-wzf82" event={"ID":"6cf131c4-a5bd-452b-8598-42312c3a0270","Type":"ContainerStarted","Data":"b96f34157cbed4eef2143feeb0fd51ea8ff8193f4fa6d28ad6a9487061aba8b2"} Jan 27 15:28:04 crc kubenswrapper[4772]: I0127 15:28:04.315101 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:04 crc kubenswrapper[4772]: I0127 15:28:04.319571 4772 generic.go:334] "Generic (PLEG): container finished" podID="7b7d9f3f-e366-421e-b00d-9c453da1adca" containerID="55338fb54abd9ad2a096debb2356f749682191abe6b851127e0e95fcec09a654" exitCode=0 Jan 27 15:28:04 crc kubenswrapper[4772]: I0127 15:28:04.320745 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7b7d9f3f-e366-421e-b00d-9c453da1adca","Type":"ContainerDied","Data":"55338fb54abd9ad2a096debb2356f749682191abe6b851127e0e95fcec09a654"} Jan 27 15:28:04 crc kubenswrapper[4772]: I0127 15:28:04.337658 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-647c88bb6f-wzf82" podStartSLOduration=3.337639085 podStartE2EDuration="3.337639085s" podCreationTimestamp="2026-01-27 15:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:28:04.333508305 +0000 UTC m=+1270.314117403" watchObservedRunningTime="2026-01-27 15:28:04.337639085 +0000 UTC m=+1270.318248183" Jan 27 15:28:04 crc kubenswrapper[4772]: I0127 15:28:04.724513 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 27 15:28:04 crc kubenswrapper[4772]: I0127 15:28:04.756361 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2tvv\" (UniqueName: \"kubernetes.io/projected/7b7d9f3f-e366-421e-b00d-9c453da1adca-kube-api-access-g2tvv\") pod \"7b7d9f3f-e366-421e-b00d-9c453da1adca\" (UID: \"7b7d9f3f-e366-421e-b00d-9c453da1adca\") " Jan 27 15:28:04 crc kubenswrapper[4772]: I0127 15:28:04.756439 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b7d9f3f-e366-421e-b00d-9c453da1adca-scripts\") pod \"7b7d9f3f-e366-421e-b00d-9c453da1adca\" (UID: \"7b7d9f3f-e366-421e-b00d-9c453da1adca\") " Jan 27 15:28:04 crc kubenswrapper[4772]: I0127 15:28:04.756463 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b7d9f3f-e366-421e-b00d-9c453da1adca-etc-machine-id\") pod \"7b7d9f3f-e366-421e-b00d-9c453da1adca\" (UID: \"7b7d9f3f-e366-421e-b00d-9c453da1adca\") " Jan 27 15:28:04 crc kubenswrapper[4772]: I0127 15:28:04.756493 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b7d9f3f-e366-421e-b00d-9c453da1adca-config-data\") pod \"7b7d9f3f-e366-421e-b00d-9c453da1adca\" (UID: \"7b7d9f3f-e366-421e-b00d-9c453da1adca\") " Jan 27 15:28:04 crc kubenswrapper[4772]: I0127 15:28:04.756597 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b7d9f3f-e366-421e-b00d-9c453da1adca-combined-ca-bundle\") pod \"7b7d9f3f-e366-421e-b00d-9c453da1adca\" (UID: \"7b7d9f3f-e366-421e-b00d-9c453da1adca\") " Jan 27 15:28:04 crc kubenswrapper[4772]: I0127 15:28:04.756669 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b7d9f3f-e366-421e-b00d-9c453da1adca-config-data-custom\") pod \"7b7d9f3f-e366-421e-b00d-9c453da1adca\" (UID: \"7b7d9f3f-e366-421e-b00d-9c453da1adca\") " Jan 27 15:28:04 crc kubenswrapper[4772]: I0127 15:28:04.758613 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b7d9f3f-e366-421e-b00d-9c453da1adca-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7b7d9f3f-e366-421e-b00d-9c453da1adca" (UID: "7b7d9f3f-e366-421e-b00d-9c453da1adca"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:28:04 crc kubenswrapper[4772]: I0127 15:28:04.774141 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b7d9f3f-e366-421e-b00d-9c453da1adca-kube-api-access-g2tvv" (OuterVolumeSpecName: "kube-api-access-g2tvv") pod "7b7d9f3f-e366-421e-b00d-9c453da1adca" (UID: "7b7d9f3f-e366-421e-b00d-9c453da1adca"). InnerVolumeSpecName "kube-api-access-g2tvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:28:04 crc kubenswrapper[4772]: I0127 15:28:04.784849 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b7d9f3f-e366-421e-b00d-9c453da1adca-scripts" (OuterVolumeSpecName: "scripts") pod "7b7d9f3f-e366-421e-b00d-9c453da1adca" (UID: "7b7d9f3f-e366-421e-b00d-9c453da1adca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:04 crc kubenswrapper[4772]: I0127 15:28:04.792121 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b7d9f3f-e366-421e-b00d-9c453da1adca-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7b7d9f3f-e366-421e-b00d-9c453da1adca" (UID: "7b7d9f3f-e366-421e-b00d-9c453da1adca"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:04 crc kubenswrapper[4772]: I0127 15:28:04.861506 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2tvv\" (UniqueName: \"kubernetes.io/projected/7b7d9f3f-e366-421e-b00d-9c453da1adca-kube-api-access-g2tvv\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:04 crc kubenswrapper[4772]: I0127 15:28:04.861803 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b7d9f3f-e366-421e-b00d-9c453da1adca-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:04 crc kubenswrapper[4772]: I0127 15:28:04.861928 4772 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b7d9f3f-e366-421e-b00d-9c453da1adca-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:04 crc kubenswrapper[4772]: I0127 15:28:04.862026 4772 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b7d9f3f-e366-421e-b00d-9c453da1adca-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:04 crc kubenswrapper[4772]: I0127 15:28:04.897893 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b7d9f3f-e366-421e-b00d-9c453da1adca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7b7d9f3f-e366-421e-b00d-9c453da1adca" (UID: "7b7d9f3f-e366-421e-b00d-9c453da1adca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:04 crc kubenswrapper[4772]: I0127 15:28:04.921553 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b7d9f3f-e366-421e-b00d-9c453da1adca-config-data" (OuterVolumeSpecName: "config-data") pod "7b7d9f3f-e366-421e-b00d-9c453da1adca" (UID: "7b7d9f3f-e366-421e-b00d-9c453da1adca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:04 crc kubenswrapper[4772]: I0127 15:28:04.963916 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b7d9f3f-e366-421e-b00d-9c453da1adca-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:04 crc kubenswrapper[4772]: I0127 15:28:04.964229 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b7d9f3f-e366-421e-b00d-9c453da1adca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.337556 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.338674 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7b7d9f3f-e366-421e-b00d-9c453da1adca","Type":"ContainerDied","Data":"60d07ef91bf8ed929ad3a7649d54a1c26d24e5b004c0a1e93cdd8cb08880d285"} Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.338739 4772 scope.go:117] "RemoveContainer" containerID="7232aaea46edf2977ebe24eeb1188331c23b40d3efeaaca7dfc75e3658d209b6" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.374529 4772 scope.go:117] "RemoveContainer" containerID="55338fb54abd9ad2a096debb2356f749682191abe6b851127e0e95fcec09a654" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.375910 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.383431 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.465254 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Jan 27 15:28:05 crc kubenswrapper[4772]: E0127 15:28:05.465988 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b7d9f3f-e366-421e-b00d-9c453da1adca" containerName="cinder-scheduler" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.466011 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b7d9f3f-e366-421e-b00d-9c453da1adca" containerName="cinder-scheduler" Jan 27 15:28:05 crc kubenswrapper[4772]: E0127 15:28:05.466059 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b7d9f3f-e366-421e-b00d-9c453da1adca" containerName="probe" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.466070 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b7d9f3f-e366-421e-b00d-9c453da1adca" containerName="probe" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.467576 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b7d9f3f-e366-421e-b00d-9c453da1adca" containerName="probe" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.467615 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b7d9f3f-e366-421e-b00d-9c453da1adca" containerName="cinder-scheduler" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.476501 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.478884 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-scripts\") pod \"cinder-scheduler-0\" (UID: \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\") " pod="openstack/cinder-scheduler-0" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.479005 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\") " pod="openstack/cinder-scheduler-0" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.479042 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrz6x\" (UniqueName: \"kubernetes.io/projected/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-kube-api-access-rrz6x\") pod \"cinder-scheduler-0\" (UID: \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\") " pod="openstack/cinder-scheduler-0" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.479112 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\") " pod="openstack/cinder-scheduler-0" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.479151 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-config-data\") pod \"cinder-scheduler-0\" (UID: \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\") " pod="openstack/cinder-scheduler-0" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.479198 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\") " pod="openstack/cinder-scheduler-0" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.484974 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.489714 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.560906 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.579903 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\") " pod="openstack/cinder-scheduler-0" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.579951 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrz6x\" (UniqueName: \"kubernetes.io/projected/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-kube-api-access-rrz6x\") pod \"cinder-scheduler-0\" (UID: \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\") " pod="openstack/cinder-scheduler-0" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.580007 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\") " pod="openstack/cinder-scheduler-0" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.580045 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-config-data\") pod \"cinder-scheduler-0\" (UID: \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\") " pod="openstack/cinder-scheduler-0" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.580068 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\") " pod="openstack/cinder-scheduler-0" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.580104 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-scripts\") pod \"cinder-scheduler-0\" (UID: \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\") " pod="openstack/cinder-scheduler-0" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.580432 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\") " pod="openstack/cinder-scheduler-0" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.597038 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-scripts\") pod \"cinder-scheduler-0\" (UID: \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\") " pod="openstack/cinder-scheduler-0" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.599715 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\") " pod="openstack/cinder-scheduler-0" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.613717 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\") " pod="openstack/cinder-scheduler-0" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.628893 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrz6x\" (UniqueName: \"kubernetes.io/projected/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-kube-api-access-rrz6x\") pod \"cinder-scheduler-0\" (UID: \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\") " pod="openstack/cinder-scheduler-0" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.636906 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-config-data\") pod \"cinder-scheduler-0\" (UID: \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\") " pod="openstack/cinder-scheduler-0" Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.647698 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-cb9d976b-flrwl"] Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.647969 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-cb9d976b-flrwl" podUID="a02a1b6c-d438-42bf-a577-88bbbcca2a00" containerName="barbican-api-log" containerID="cri-o://6263f4825b44d3954903ca3307e160ecf8edb8b8f916573aa23ea8b2efbc78ab" gracePeriod=30 Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.648464 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-cb9d976b-flrwl" podUID="a02a1b6c-d438-42bf-a577-88bbbcca2a00" containerName="barbican-api" containerID="cri-o://51682243156b2ae0619aa3431049d4b3ccf7b64abb3cf0a6f8d8cabce6c5c3f3" gracePeriod=30 Jan 27 15:28:05 crc kubenswrapper[4772]: I0127 15:28:05.824136 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 27 15:28:06 crc kubenswrapper[4772]: I0127 15:28:06.321150 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 27 15:28:06 crc kubenswrapper[4772]: W0127 15:28:06.343196 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod683f458e_44e9_49ea_a66b_4ac91a3f2bc1.slice/crio-bb200c044803c6c5491d60dc192f271f4cdf0adcf18a5f0f12ab40acb77fdf72 WatchSource:0}: Error finding container bb200c044803c6c5491d60dc192f271f4cdf0adcf18a5f0f12ab40acb77fdf72: Status 404 returned error can't find the container with id bb200c044803c6c5491d60dc192f271f4cdf0adcf18a5f0f12ab40acb77fdf72 Jan 27 15:28:06 crc kubenswrapper[4772]: I0127 15:28:06.377048 4772 generic.go:334] "Generic (PLEG): container finished" podID="a02a1b6c-d438-42bf-a577-88bbbcca2a00" containerID="6263f4825b44d3954903ca3307e160ecf8edb8b8f916573aa23ea8b2efbc78ab" exitCode=143 Jan 27 15:28:06 crc kubenswrapper[4772]: I0127 15:28:06.377285 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cb9d976b-flrwl" event={"ID":"a02a1b6c-d438-42bf-a577-88bbbcca2a00","Type":"ContainerDied","Data":"6263f4825b44d3954903ca3307e160ecf8edb8b8f916573aa23ea8b2efbc78ab"} Jan 27 15:28:06 crc kubenswrapper[4772]: I0127 15:28:06.545649 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:06 crc kubenswrapper[4772]: I0127 15:28:06.546551 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f8583377-67ef-4cca-83bb-08d7523ab0a8" containerName="ceilometer-central-agent" containerID="cri-o://941b08834b6f5b8dafbc182c67d3e458a94c7299ea32b8afd698f876b68ea015" gracePeriod=30 Jan 27 15:28:06 crc kubenswrapper[4772]: I0127 15:28:06.547675 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f8583377-67ef-4cca-83bb-08d7523ab0a8" containerName="proxy-httpd" containerID="cri-o://851c7b3936a50d21408fbed0918adde539924e5915ec73fdcccd952a3392565b" gracePeriod=30 Jan 27 15:28:06 crc kubenswrapper[4772]: I0127 15:28:06.547712 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f8583377-67ef-4cca-83bb-08d7523ab0a8" containerName="ceilometer-notification-agent" containerID="cri-o://99f37d09a547a41878834ffbee7e0a0b90552016b42313fae983e9915266d761" gracePeriod=30 Jan 27 15:28:06 crc kubenswrapper[4772]: I0127 15:28:06.547792 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f8583377-67ef-4cca-83bb-08d7523ab0a8" containerName="sg-core" containerID="cri-o://ea8686b5fbb3cb04fd3d0cb81bec48b421aa5e6e9be9af4a4ad0ccc951c6bce4" gracePeriod=30 Jan 27 15:28:06 crc kubenswrapper[4772]: I0127 15:28:06.567549 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="f8583377-67ef-4cca-83bb-08d7523ab0a8" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.154:3000/\": EOF" Jan 27 15:28:06 crc kubenswrapper[4772]: I0127 15:28:06.684323 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b7d9f3f-e366-421e-b00d-9c453da1adca" path="/var/lib/kubelet/pods/7b7d9f3f-e366-421e-b00d-9c453da1adca/volumes" Jan 27 15:28:06 crc kubenswrapper[4772]: I0127 15:28:06.762830 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-d86f6cfbc-cwfmc"] Jan 27 15:28:06 crc kubenswrapper[4772]: I0127 15:28:06.769809 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:06 crc kubenswrapper[4772]: I0127 15:28:06.777964 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Jan 27 15:28:06 crc kubenswrapper[4772]: I0127 15:28:06.778460 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Jan 27 15:28:06 crc kubenswrapper[4772]: I0127 15:28:06.778315 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Jan 27 15:28:06 crc kubenswrapper[4772]: I0127 15:28:06.791334 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-d86f6cfbc-cwfmc"] Jan 27 15:28:06 crc kubenswrapper[4772]: I0127 15:28:06.917243 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c16a29a0-7238-4a5e-b892-8f5195a1f486-public-tls-certs\") pod \"swift-proxy-d86f6cfbc-cwfmc\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:06 crc kubenswrapper[4772]: I0127 15:28:06.917300 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c16a29a0-7238-4a5e-b892-8f5195a1f486-run-httpd\") pod \"swift-proxy-d86f6cfbc-cwfmc\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:06 crc kubenswrapper[4772]: I0127 15:28:06.917378 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c16a29a0-7238-4a5e-b892-8f5195a1f486-combined-ca-bundle\") pod \"swift-proxy-d86f6cfbc-cwfmc\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:06 crc kubenswrapper[4772]: I0127 15:28:06.917399 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c16a29a0-7238-4a5e-b892-8f5195a1f486-config-data\") pod \"swift-proxy-d86f6cfbc-cwfmc\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:06 crc kubenswrapper[4772]: I0127 15:28:06.917419 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c16a29a0-7238-4a5e-b892-8f5195a1f486-internal-tls-certs\") pod \"swift-proxy-d86f6cfbc-cwfmc\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:06 crc kubenswrapper[4772]: I0127 15:28:06.917446 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfzf2\" (UniqueName: \"kubernetes.io/projected/c16a29a0-7238-4a5e-b892-8f5195a1f486-kube-api-access-pfzf2\") pod \"swift-proxy-d86f6cfbc-cwfmc\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:06 crc kubenswrapper[4772]: I0127 15:28:06.917497 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c16a29a0-7238-4a5e-b892-8f5195a1f486-etc-swift\") pod \"swift-proxy-d86f6cfbc-cwfmc\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:06 crc kubenswrapper[4772]: I0127 15:28:06.917515 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c16a29a0-7238-4a5e-b892-8f5195a1f486-log-httpd\") pod \"swift-proxy-d86f6cfbc-cwfmc\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.019529 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfzf2\" (UniqueName: \"kubernetes.io/projected/c16a29a0-7238-4a5e-b892-8f5195a1f486-kube-api-access-pfzf2\") pod \"swift-proxy-d86f6cfbc-cwfmc\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.019611 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c16a29a0-7238-4a5e-b892-8f5195a1f486-etc-swift\") pod \"swift-proxy-d86f6cfbc-cwfmc\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.019644 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c16a29a0-7238-4a5e-b892-8f5195a1f486-log-httpd\") pod \"swift-proxy-d86f6cfbc-cwfmc\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.019723 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c16a29a0-7238-4a5e-b892-8f5195a1f486-public-tls-certs\") pod \"swift-proxy-d86f6cfbc-cwfmc\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.019749 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c16a29a0-7238-4a5e-b892-8f5195a1f486-run-httpd\") pod \"swift-proxy-d86f6cfbc-cwfmc\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.019817 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c16a29a0-7238-4a5e-b892-8f5195a1f486-combined-ca-bundle\") pod \"swift-proxy-d86f6cfbc-cwfmc\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.019890 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c16a29a0-7238-4a5e-b892-8f5195a1f486-config-data\") pod \"swift-proxy-d86f6cfbc-cwfmc\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.019949 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c16a29a0-7238-4a5e-b892-8f5195a1f486-internal-tls-certs\") pod \"swift-proxy-d86f6cfbc-cwfmc\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.023681 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c16a29a0-7238-4a5e-b892-8f5195a1f486-log-httpd\") pod \"swift-proxy-d86f6cfbc-cwfmc\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.023752 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c16a29a0-7238-4a5e-b892-8f5195a1f486-run-httpd\") pod \"swift-proxy-d86f6cfbc-cwfmc\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.026152 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c16a29a0-7238-4a5e-b892-8f5195a1f486-internal-tls-certs\") pod \"swift-proxy-d86f6cfbc-cwfmc\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.026423 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c16a29a0-7238-4a5e-b892-8f5195a1f486-etc-swift\") pod \"swift-proxy-d86f6cfbc-cwfmc\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.026745 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c16a29a0-7238-4a5e-b892-8f5195a1f486-public-tls-certs\") pod \"swift-proxy-d86f6cfbc-cwfmc\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.028044 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c16a29a0-7238-4a5e-b892-8f5195a1f486-combined-ca-bundle\") pod \"swift-proxy-d86f6cfbc-cwfmc\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.028961 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c16a29a0-7238-4a5e-b892-8f5195a1f486-config-data\") pod \"swift-proxy-d86f6cfbc-cwfmc\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.042191 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfzf2\" (UniqueName: \"kubernetes.io/projected/c16a29a0-7238-4a5e-b892-8f5195a1f486-kube-api-access-pfzf2\") pod \"swift-proxy-d86f6cfbc-cwfmc\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.111156 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.415554 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"683f458e-44e9-49ea-a66b-4ac91a3f2bc1","Type":"ContainerStarted","Data":"3e806373a2604b5465de7a3913d6865c82f0689bac61f26c430950d7d4efb948"} Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.416286 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"683f458e-44e9-49ea-a66b-4ac91a3f2bc1","Type":"ContainerStarted","Data":"bb200c044803c6c5491d60dc192f271f4cdf0adcf18a5f0f12ab40acb77fdf72"} Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.432693 4772 generic.go:334] "Generic (PLEG): container finished" podID="f8583377-67ef-4cca-83bb-08d7523ab0a8" containerID="851c7b3936a50d21408fbed0918adde539924e5915ec73fdcccd952a3392565b" exitCode=0 Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.432737 4772 generic.go:334] "Generic (PLEG): container finished" podID="f8583377-67ef-4cca-83bb-08d7523ab0a8" containerID="ea8686b5fbb3cb04fd3d0cb81bec48b421aa5e6e9be9af4a4ad0ccc951c6bce4" exitCode=2 Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.432745 4772 generic.go:334] "Generic (PLEG): container finished" podID="f8583377-67ef-4cca-83bb-08d7523ab0a8" containerID="99f37d09a547a41878834ffbee7e0a0b90552016b42313fae983e9915266d761" exitCode=0 Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.432752 4772 generic.go:334] "Generic (PLEG): container finished" podID="f8583377-67ef-4cca-83bb-08d7523ab0a8" containerID="941b08834b6f5b8dafbc182c67d3e458a94c7299ea32b8afd698f876b68ea015" exitCode=0 Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.432771 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8583377-67ef-4cca-83bb-08d7523ab0a8","Type":"ContainerDied","Data":"851c7b3936a50d21408fbed0918adde539924e5915ec73fdcccd952a3392565b"} Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.432814 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8583377-67ef-4cca-83bb-08d7523ab0a8","Type":"ContainerDied","Data":"ea8686b5fbb3cb04fd3d0cb81bec48b421aa5e6e9be9af4a4ad0ccc951c6bce4"} Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.432825 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8583377-67ef-4cca-83bb-08d7523ab0a8","Type":"ContainerDied","Data":"99f37d09a547a41878834ffbee7e0a0b90552016b42313fae983e9915266d761"} Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.432833 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8583377-67ef-4cca-83bb-08d7523ab0a8","Type":"ContainerDied","Data":"941b08834b6f5b8dafbc182c67d3e458a94c7299ea32b8afd698f876b68ea015"} Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.726656 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.769099 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-d86f6cfbc-cwfmc"] Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.847473 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f8583377-67ef-4cca-83bb-08d7523ab0a8-sg-core-conf-yaml\") pod \"f8583377-67ef-4cca-83bb-08d7523ab0a8\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.847524 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8583377-67ef-4cca-83bb-08d7523ab0a8-run-httpd\") pod \"f8583377-67ef-4cca-83bb-08d7523ab0a8\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.847565 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8583377-67ef-4cca-83bb-08d7523ab0a8-scripts\") pod \"f8583377-67ef-4cca-83bb-08d7523ab0a8\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.847584 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8583377-67ef-4cca-83bb-08d7523ab0a8-config-data\") pod \"f8583377-67ef-4cca-83bb-08d7523ab0a8\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.847649 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8583377-67ef-4cca-83bb-08d7523ab0a8-combined-ca-bundle\") pod \"f8583377-67ef-4cca-83bb-08d7523ab0a8\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.847700 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8583377-67ef-4cca-83bb-08d7523ab0a8-log-httpd\") pod \"f8583377-67ef-4cca-83bb-08d7523ab0a8\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.847755 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jjsh\" (UniqueName: \"kubernetes.io/projected/f8583377-67ef-4cca-83bb-08d7523ab0a8-kube-api-access-9jjsh\") pod \"f8583377-67ef-4cca-83bb-08d7523ab0a8\" (UID: \"f8583377-67ef-4cca-83bb-08d7523ab0a8\") " Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.850502 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8583377-67ef-4cca-83bb-08d7523ab0a8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f8583377-67ef-4cca-83bb-08d7523ab0a8" (UID: "f8583377-67ef-4cca-83bb-08d7523ab0a8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.852923 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8583377-67ef-4cca-83bb-08d7523ab0a8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f8583377-67ef-4cca-83bb-08d7523ab0a8" (UID: "f8583377-67ef-4cca-83bb-08d7523ab0a8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.857381 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8583377-67ef-4cca-83bb-08d7523ab0a8-kube-api-access-9jjsh" (OuterVolumeSpecName: "kube-api-access-9jjsh") pod "f8583377-67ef-4cca-83bb-08d7523ab0a8" (UID: "f8583377-67ef-4cca-83bb-08d7523ab0a8"). InnerVolumeSpecName "kube-api-access-9jjsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.858185 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8583377-67ef-4cca-83bb-08d7523ab0a8-scripts" (OuterVolumeSpecName: "scripts") pod "f8583377-67ef-4cca-83bb-08d7523ab0a8" (UID: "f8583377-67ef-4cca-83bb-08d7523ab0a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.922665 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8583377-67ef-4cca-83bb-08d7523ab0a8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f8583377-67ef-4cca-83bb-08d7523ab0a8" (UID: "f8583377-67ef-4cca-83bb-08d7523ab0a8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.949823 4772 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f8583377-67ef-4cca-83bb-08d7523ab0a8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.949864 4772 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8583377-67ef-4cca-83bb-08d7523ab0a8-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.949876 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8583377-67ef-4cca-83bb-08d7523ab0a8-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.949887 4772 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f8583377-67ef-4cca-83bb-08d7523ab0a8-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:07 crc kubenswrapper[4772]: I0127 15:28:07.949897 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jjsh\" (UniqueName: \"kubernetes.io/projected/f8583377-67ef-4cca-83bb-08d7523ab0a8-kube-api-access-9jjsh\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.001467 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8583377-67ef-4cca-83bb-08d7523ab0a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8583377-67ef-4cca-83bb-08d7523ab0a8" (UID: "f8583377-67ef-4cca-83bb-08d7523ab0a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.029277 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8583377-67ef-4cca-83bb-08d7523ab0a8-config-data" (OuterVolumeSpecName: "config-data") pod "f8583377-67ef-4cca-83bb-08d7523ab0a8" (UID: "f8583377-67ef-4cca-83bb-08d7523ab0a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.051761 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8583377-67ef-4cca-83bb-08d7523ab0a8-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.052003 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8583377-67ef-4cca-83bb-08d7523ab0a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.480499 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-d86f6cfbc-cwfmc" event={"ID":"c16a29a0-7238-4a5e-b892-8f5195a1f486","Type":"ContainerStarted","Data":"47a1d8c4913044388b407e6a5c05783d2d3731216d7862873425d28265a5fe05"} Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.480771 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-d86f6cfbc-cwfmc" event={"ID":"c16a29a0-7238-4a5e-b892-8f5195a1f486","Type":"ContainerStarted","Data":"a476d84a3741734575b073569a645d9d973c5cdbb39812aa454a7257859db22b"} Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.480782 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-d86f6cfbc-cwfmc" event={"ID":"c16a29a0-7238-4a5e-b892-8f5195a1f486","Type":"ContainerStarted","Data":"92e9170b2797b87fe5816f61d1944a7f0cca88f2e0e21f7420f27a5ed25d4005"} Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.501503 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f8583377-67ef-4cca-83bb-08d7523ab0a8","Type":"ContainerDied","Data":"561a737e7b03fa9b4f9fd3fd05ad062a1b5523a6f22e1965b09d595f10adafd3"} Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.501565 4772 scope.go:117] "RemoveContainer" containerID="851c7b3936a50d21408fbed0918adde539924e5915ec73fdcccd952a3392565b" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.501741 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.542147 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"683f458e-44e9-49ea-a66b-4ac91a3f2bc1","Type":"ContainerStarted","Data":"112ddc6068b3694383f83c1ffece42788a7623920d1c02ff9f46202f7c8c0d7e"} Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.550079 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.597129 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-2849v" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.608995 4772 scope.go:117] "RemoveContainer" containerID="ea8686b5fbb3cb04fd3d0cb81bec48b421aa5e6e9be9af4a4ad0ccc951c6bce4" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.642175 4772 scope.go:117] "RemoveContainer" containerID="99f37d09a547a41878834ffbee7e0a0b90552016b42313fae983e9915266d761" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.650409 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.692233 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8583377-67ef-4cca-83bb-08d7523ab0a8" path="/var/lib/kubelet/pods/f8583377-67ef-4cca-83bb-08d7523ab0a8/volumes" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.692992 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:08 crc kubenswrapper[4772]: E0127 15:28:08.693348 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8583377-67ef-4cca-83bb-08d7523ab0a8" containerName="sg-core" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.693361 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8583377-67ef-4cca-83bb-08d7523ab0a8" containerName="sg-core" Jan 27 15:28:08 crc kubenswrapper[4772]: E0127 15:28:08.693375 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8583377-67ef-4cca-83bb-08d7523ab0a8" containerName="ceilometer-notification-agent" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.693381 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8583377-67ef-4cca-83bb-08d7523ab0a8" containerName="ceilometer-notification-agent" Jan 27 15:28:08 crc kubenswrapper[4772]: E0127 15:28:08.693399 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8583377-67ef-4cca-83bb-08d7523ab0a8" containerName="proxy-httpd" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.693404 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8583377-67ef-4cca-83bb-08d7523ab0a8" containerName="proxy-httpd" Jan 27 15:28:08 crc kubenswrapper[4772]: E0127 15:28:08.693421 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8583377-67ef-4cca-83bb-08d7523ab0a8" containerName="ceilometer-central-agent" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.693427 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8583377-67ef-4cca-83bb-08d7523ab0a8" containerName="ceilometer-central-agent" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.693596 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8583377-67ef-4cca-83bb-08d7523ab0a8" containerName="sg-core" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.693607 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8583377-67ef-4cca-83bb-08d7523ab0a8" containerName="ceilometer-central-agent" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.693621 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8583377-67ef-4cca-83bb-08d7523ab0a8" containerName="proxy-httpd" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.693629 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8583377-67ef-4cca-83bb-08d7523ab0a8" containerName="ceilometer-notification-agent" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.695857 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.697832 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.698916 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.698269 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.698257854 podStartE2EDuration="3.698257854s" podCreationTimestamp="2026-01-27 15:28:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:28:08.574045412 +0000 UTC m=+1274.554654510" watchObservedRunningTime="2026-01-27 15:28:08.698257854 +0000 UTC m=+1274.678866952" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.722198 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.752706 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-dqgvx"] Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.752925 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" podUID="17a547a9-a098-43b7-a153-ad9a137369de" containerName="dnsmasq-dns" containerID="cri-o://9b585b108c0dc492c1e4b07fece64b278518a1b12e04b9e57c9b27d9183ca9ac" gracePeriod=10 Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.757756 4772 scope.go:117] "RemoveContainer" containerID="941b08834b6f5b8dafbc182c67d3e458a94c7299ea32b8afd698f876b68ea015" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.776209 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhrl9\" (UniqueName: \"kubernetes.io/projected/65e74d64-83f8-4964-8950-bf76816dd5fc-kube-api-access-nhrl9\") pod \"ceilometer-0\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " pod="openstack/ceilometer-0" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.776562 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65e74d64-83f8-4964-8950-bf76816dd5fc-config-data\") pod \"ceilometer-0\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " pod="openstack/ceilometer-0" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.776585 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65e74d64-83f8-4964-8950-bf76816dd5fc-scripts\") pod \"ceilometer-0\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " pod="openstack/ceilometer-0" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.776611 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65e74d64-83f8-4964-8950-bf76816dd5fc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " pod="openstack/ceilometer-0" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.776636 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65e74d64-83f8-4964-8950-bf76816dd5fc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " pod="openstack/ceilometer-0" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.776681 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65e74d64-83f8-4964-8950-bf76816dd5fc-run-httpd\") pod \"ceilometer-0\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " pod="openstack/ceilometer-0" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.776738 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65e74d64-83f8-4964-8950-bf76816dd5fc-log-httpd\") pod \"ceilometer-0\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " pod="openstack/ceilometer-0" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.862782 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-cb9d976b-flrwl" podUID="a02a1b6c-d438-42bf-a577-88bbbcca2a00" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:49270->10.217.0.158:9311: read: connection reset by peer" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.863142 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-cb9d976b-flrwl" podUID="a02a1b6c-d438-42bf-a577-88bbbcca2a00" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:49258->10.217.0.158:9311: read: connection reset by peer" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.878114 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhrl9\" (UniqueName: \"kubernetes.io/projected/65e74d64-83f8-4964-8950-bf76816dd5fc-kube-api-access-nhrl9\") pod \"ceilometer-0\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " pod="openstack/ceilometer-0" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.878729 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65e74d64-83f8-4964-8950-bf76816dd5fc-config-data\") pod \"ceilometer-0\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " pod="openstack/ceilometer-0" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.879561 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65e74d64-83f8-4964-8950-bf76816dd5fc-scripts\") pod \"ceilometer-0\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " pod="openstack/ceilometer-0" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.880378 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65e74d64-83f8-4964-8950-bf76816dd5fc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " pod="openstack/ceilometer-0" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.880580 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65e74d64-83f8-4964-8950-bf76816dd5fc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " pod="openstack/ceilometer-0" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.880770 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65e74d64-83f8-4964-8950-bf76816dd5fc-run-httpd\") pod \"ceilometer-0\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " pod="openstack/ceilometer-0" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.880941 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65e74d64-83f8-4964-8950-bf76816dd5fc-log-httpd\") pod \"ceilometer-0\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " pod="openstack/ceilometer-0" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.881518 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65e74d64-83f8-4964-8950-bf76816dd5fc-log-httpd\") pod \"ceilometer-0\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " pod="openstack/ceilometer-0" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.883051 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65e74d64-83f8-4964-8950-bf76816dd5fc-run-httpd\") pod \"ceilometer-0\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " pod="openstack/ceilometer-0" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.884809 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65e74d64-83f8-4964-8950-bf76816dd5fc-config-data\") pod \"ceilometer-0\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " pod="openstack/ceilometer-0" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.887804 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65e74d64-83f8-4964-8950-bf76816dd5fc-scripts\") pod \"ceilometer-0\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " pod="openstack/ceilometer-0" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.889203 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65e74d64-83f8-4964-8950-bf76816dd5fc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " pod="openstack/ceilometer-0" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.893023 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65e74d64-83f8-4964-8950-bf76816dd5fc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " pod="openstack/ceilometer-0" Jan 27 15:28:08 crc kubenswrapper[4772]: I0127 15:28:08.904252 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhrl9\" (UniqueName: \"kubernetes.io/projected/65e74d64-83f8-4964-8950-bf76816dd5fc-kube-api-access-nhrl9\") pod \"ceilometer-0\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " pod="openstack/ceilometer-0" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.031988 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.466041 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.583367 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-cb9d976b-flrwl" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.589210 4772 generic.go:334] "Generic (PLEG): container finished" podID="17a547a9-a098-43b7-a153-ad9a137369de" containerID="9b585b108c0dc492c1e4b07fece64b278518a1b12e04b9e57c9b27d9183ca9ac" exitCode=0 Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.589265 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" event={"ID":"17a547a9-a098-43b7-a153-ad9a137369de","Type":"ContainerDied","Data":"9b585b108c0dc492c1e4b07fece64b278518a1b12e04b9e57c9b27d9183ca9ac"} Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.589292 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" event={"ID":"17a547a9-a098-43b7-a153-ad9a137369de","Type":"ContainerDied","Data":"31e83f6ba26ca249b5435d61c8786bdc24b0777adb10cbb234cdaacbda3e0db7"} Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.589308 4772 scope.go:117] "RemoveContainer" containerID="9b585b108c0dc492c1e4b07fece64b278518a1b12e04b9e57c9b27d9183ca9ac" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.589415 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-dqgvx" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.604013 4772 generic.go:334] "Generic (PLEG): container finished" podID="a02a1b6c-d438-42bf-a577-88bbbcca2a00" containerID="51682243156b2ae0619aa3431049d4b3ccf7b64abb3cf0a6f8d8cabce6c5c3f3" exitCode=0 Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.604940 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cb9d976b-flrwl" event={"ID":"a02a1b6c-d438-42bf-a577-88bbbcca2a00","Type":"ContainerDied","Data":"51682243156b2ae0619aa3431049d4b3ccf7b64abb3cf0a6f8d8cabce6c5c3f3"} Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.604997 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cb9d976b-flrwl" event={"ID":"a02a1b6c-d438-42bf-a577-88bbbcca2a00","Type":"ContainerDied","Data":"37bb9f56f01e20e1b9f2066e0e10a19c1b316c11b490381320b663b46a9cc874"} Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.605180 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-cb9d976b-flrwl" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.605426 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.605448 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.607127 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-ovsdbserver-sb\") pod \"17a547a9-a098-43b7-a153-ad9a137369de\" (UID: \"17a547a9-a098-43b7-a153-ad9a137369de\") " Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.607253 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-ovsdbserver-nb\") pod \"17a547a9-a098-43b7-a153-ad9a137369de\" (UID: \"17a547a9-a098-43b7-a153-ad9a137369de\") " Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.607299 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-config\") pod \"17a547a9-a098-43b7-a153-ad9a137369de\" (UID: \"17a547a9-a098-43b7-a153-ad9a137369de\") " Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.607325 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-dns-svc\") pod \"17a547a9-a098-43b7-a153-ad9a137369de\" (UID: \"17a547a9-a098-43b7-a153-ad9a137369de\") " Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.607384 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-dns-swift-storage-0\") pod \"17a547a9-a098-43b7-a153-ad9a137369de\" (UID: \"17a547a9-a098-43b7-a153-ad9a137369de\") " Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.607436 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzhkt\" (UniqueName: \"kubernetes.io/projected/17a547a9-a098-43b7-a153-ad9a137369de-kube-api-access-xzhkt\") pod \"17a547a9-a098-43b7-a153-ad9a137369de\" (UID: \"17a547a9-a098-43b7-a153-ad9a137369de\") " Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.631580 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17a547a9-a098-43b7-a153-ad9a137369de-kube-api-access-xzhkt" (OuterVolumeSpecName: "kube-api-access-xzhkt") pod "17a547a9-a098-43b7-a153-ad9a137369de" (UID: "17a547a9-a098-43b7-a153-ad9a137369de"). InnerVolumeSpecName "kube-api-access-xzhkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.682266 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-d86f6cfbc-cwfmc" podStartSLOduration=3.682240335 podStartE2EDuration="3.682240335s" podCreationTimestamp="2026-01-27 15:28:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:28:09.644089806 +0000 UTC m=+1275.624698924" watchObservedRunningTime="2026-01-27 15:28:09.682240335 +0000 UTC m=+1275.662849443" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.709987 4772 scope.go:117] "RemoveContainer" containerID="dbf85c4247c8ba3d5d079a5efed1ba22279b244d404fd6795f678db691f0b023" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.712261 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a02a1b6c-d438-42bf-a577-88bbbcca2a00-config-data\") pod \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\" (UID: \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\") " Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.712307 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a02a1b6c-d438-42bf-a577-88bbbcca2a00-config-data-custom\") pod \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\" (UID: \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\") " Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.712389 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a02a1b6c-d438-42bf-a577-88bbbcca2a00-combined-ca-bundle\") pod \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\" (UID: \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\") " Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.712461 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbjvx\" (UniqueName: \"kubernetes.io/projected/a02a1b6c-d438-42bf-a577-88bbbcca2a00-kube-api-access-rbjvx\") pod \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\" (UID: \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\") " Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.712546 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a02a1b6c-d438-42bf-a577-88bbbcca2a00-logs\") pod \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\" (UID: \"a02a1b6c-d438-42bf-a577-88bbbcca2a00\") " Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.713208 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzhkt\" (UniqueName: \"kubernetes.io/projected/17a547a9-a098-43b7-a153-ad9a137369de-kube-api-access-xzhkt\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.732344 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "17a547a9-a098-43b7-a153-ad9a137369de" (UID: "17a547a9-a098-43b7-a153-ad9a137369de"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.737891 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a02a1b6c-d438-42bf-a577-88bbbcca2a00-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a02a1b6c-d438-42bf-a577-88bbbcca2a00" (UID: "a02a1b6c-d438-42bf-a577-88bbbcca2a00"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.738539 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a02a1b6c-d438-42bf-a577-88bbbcca2a00-logs" (OuterVolumeSpecName: "logs") pod "a02a1b6c-d438-42bf-a577-88bbbcca2a00" (UID: "a02a1b6c-d438-42bf-a577-88bbbcca2a00"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.750970 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a02a1b6c-d438-42bf-a577-88bbbcca2a00-kube-api-access-rbjvx" (OuterVolumeSpecName: "kube-api-access-rbjvx") pod "a02a1b6c-d438-42bf-a577-88bbbcca2a00" (UID: "a02a1b6c-d438-42bf-a577-88bbbcca2a00"). InnerVolumeSpecName "kube-api-access-rbjvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.758132 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "17a547a9-a098-43b7-a153-ad9a137369de" (UID: "17a547a9-a098-43b7-a153-ad9a137369de"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.759099 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.782525 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "17a547a9-a098-43b7-a153-ad9a137369de" (UID: "17a547a9-a098-43b7-a153-ad9a137369de"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.783311 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a02a1b6c-d438-42bf-a577-88bbbcca2a00-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a02a1b6c-d438-42bf-a577-88bbbcca2a00" (UID: "a02a1b6c-d438-42bf-a577-88bbbcca2a00"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.805383 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-config" (OuterVolumeSpecName: "config") pod "17a547a9-a098-43b7-a153-ad9a137369de" (UID: "17a547a9-a098-43b7-a153-ad9a137369de"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.824665 4772 scope.go:117] "RemoveContainer" containerID="9b585b108c0dc492c1e4b07fece64b278518a1b12e04b9e57c9b27d9183ca9ac" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.830823 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "17a547a9-a098-43b7-a153-ad9a137369de" (UID: "17a547a9-a098-43b7-a153-ad9a137369de"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.831954 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a02a1b6c-d438-42bf-a577-88bbbcca2a00-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.832068 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbjvx\" (UniqueName: \"kubernetes.io/projected/a02a1b6c-d438-42bf-a577-88bbbcca2a00-kube-api-access-rbjvx\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.832146 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.833025 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a02a1b6c-d438-42bf-a577-88bbbcca2a00-logs\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.833116 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.833208 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.833383 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.833461 4772 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17a547a9-a098-43b7-a153-ad9a137369de-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.833536 4772 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a02a1b6c-d438-42bf-a577-88bbbcca2a00-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:09 crc kubenswrapper[4772]: E0127 15:28:09.837937 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b585b108c0dc492c1e4b07fece64b278518a1b12e04b9e57c9b27d9183ca9ac\": container with ID starting with 9b585b108c0dc492c1e4b07fece64b278518a1b12e04b9e57c9b27d9183ca9ac not found: ID does not exist" containerID="9b585b108c0dc492c1e4b07fece64b278518a1b12e04b9e57c9b27d9183ca9ac" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.837994 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b585b108c0dc492c1e4b07fece64b278518a1b12e04b9e57c9b27d9183ca9ac"} err="failed to get container status \"9b585b108c0dc492c1e4b07fece64b278518a1b12e04b9e57c9b27d9183ca9ac\": rpc error: code = NotFound desc = could not find container \"9b585b108c0dc492c1e4b07fece64b278518a1b12e04b9e57c9b27d9183ca9ac\": container with ID starting with 9b585b108c0dc492c1e4b07fece64b278518a1b12e04b9e57c9b27d9183ca9ac not found: ID does not exist" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.838028 4772 scope.go:117] "RemoveContainer" containerID="dbf85c4247c8ba3d5d079a5efed1ba22279b244d404fd6795f678db691f0b023" Jan 27 15:28:09 crc kubenswrapper[4772]: E0127 15:28:09.838627 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbf85c4247c8ba3d5d079a5efed1ba22279b244d404fd6795f678db691f0b023\": container with ID starting with dbf85c4247c8ba3d5d079a5efed1ba22279b244d404fd6795f678db691f0b023 not found: ID does not exist" containerID="dbf85c4247c8ba3d5d079a5efed1ba22279b244d404fd6795f678db691f0b023" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.838651 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbf85c4247c8ba3d5d079a5efed1ba22279b244d404fd6795f678db691f0b023"} err="failed to get container status \"dbf85c4247c8ba3d5d079a5efed1ba22279b244d404fd6795f678db691f0b023\": rpc error: code = NotFound desc = could not find container \"dbf85c4247c8ba3d5d079a5efed1ba22279b244d404fd6795f678db691f0b023\": container with ID starting with dbf85c4247c8ba3d5d079a5efed1ba22279b244d404fd6795f678db691f0b023 not found: ID does not exist" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.838672 4772 scope.go:117] "RemoveContainer" containerID="51682243156b2ae0619aa3431049d4b3ccf7b64abb3cf0a6f8d8cabce6c5c3f3" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.875510 4772 scope.go:117] "RemoveContainer" containerID="6263f4825b44d3954903ca3307e160ecf8edb8b8f916573aa23ea8b2efbc78ab" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.900901 4772 scope.go:117] "RemoveContainer" containerID="51682243156b2ae0619aa3431049d4b3ccf7b64abb3cf0a6f8d8cabce6c5c3f3" Jan 27 15:28:09 crc kubenswrapper[4772]: E0127 15:28:09.901616 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51682243156b2ae0619aa3431049d4b3ccf7b64abb3cf0a6f8d8cabce6c5c3f3\": container with ID starting with 51682243156b2ae0619aa3431049d4b3ccf7b64abb3cf0a6f8d8cabce6c5c3f3 not found: ID does not exist" containerID="51682243156b2ae0619aa3431049d4b3ccf7b64abb3cf0a6f8d8cabce6c5c3f3" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.901788 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51682243156b2ae0619aa3431049d4b3ccf7b64abb3cf0a6f8d8cabce6c5c3f3"} err="failed to get container status \"51682243156b2ae0619aa3431049d4b3ccf7b64abb3cf0a6f8d8cabce6c5c3f3\": rpc error: code = NotFound desc = could not find container \"51682243156b2ae0619aa3431049d4b3ccf7b64abb3cf0a6f8d8cabce6c5c3f3\": container with ID starting with 51682243156b2ae0619aa3431049d4b3ccf7b64abb3cf0a6f8d8cabce6c5c3f3 not found: ID does not exist" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.902032 4772 scope.go:117] "RemoveContainer" containerID="6263f4825b44d3954903ca3307e160ecf8edb8b8f916573aa23ea8b2efbc78ab" Jan 27 15:28:09 crc kubenswrapper[4772]: E0127 15:28:09.902473 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6263f4825b44d3954903ca3307e160ecf8edb8b8f916573aa23ea8b2efbc78ab\": container with ID starting with 6263f4825b44d3954903ca3307e160ecf8edb8b8f916573aa23ea8b2efbc78ab not found: ID does not exist" containerID="6263f4825b44d3954903ca3307e160ecf8edb8b8f916573aa23ea8b2efbc78ab" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.902542 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6263f4825b44d3954903ca3307e160ecf8edb8b8f916573aa23ea8b2efbc78ab"} err="failed to get container status \"6263f4825b44d3954903ca3307e160ecf8edb8b8f916573aa23ea8b2efbc78ab\": rpc error: code = NotFound desc = could not find container \"6263f4825b44d3954903ca3307e160ecf8edb8b8f916573aa23ea8b2efbc78ab\": container with ID starting with 6263f4825b44d3954903ca3307e160ecf8edb8b8f916573aa23ea8b2efbc78ab not found: ID does not exist" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.916727 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a02a1b6c-d438-42bf-a577-88bbbcca2a00-config-data" (OuterVolumeSpecName: "config-data") pod "a02a1b6c-d438-42bf-a577-88bbbcca2a00" (UID: "a02a1b6c-d438-42bf-a577-88bbbcca2a00"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.932182 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-dqgvx"] Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.935484 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a02a1b6c-d438-42bf-a577-88bbbcca2a00-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:09 crc kubenswrapper[4772]: I0127 15:28:09.942234 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-dqgvx"] Jan 27 15:28:10 crc kubenswrapper[4772]: I0127 15:28:10.266297 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-cb9d976b-flrwl"] Jan 27 15:28:10 crc kubenswrapper[4772]: I0127 15:28:10.283507 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-cb9d976b-flrwl"] Jan 27 15:28:10 crc kubenswrapper[4772]: I0127 15:28:10.641758 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65e74d64-83f8-4964-8950-bf76816dd5fc","Type":"ContainerStarted","Data":"efec7f44770a0bed43e7dc53d45c9b414c83392d3729ddafeb636db96612decb"} Jan 27 15:28:10 crc kubenswrapper[4772]: I0127 15:28:10.675972 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17a547a9-a098-43b7-a153-ad9a137369de" path="/var/lib/kubelet/pods/17a547a9-a098-43b7-a153-ad9a137369de/volumes" Jan 27 15:28:10 crc kubenswrapper[4772]: I0127 15:28:10.676866 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a02a1b6c-d438-42bf-a577-88bbbcca2a00" path="/var/lib/kubelet/pods/a02a1b6c-d438-42bf-a577-88bbbcca2a00/volumes" Jan 27 15:28:10 crc kubenswrapper[4772]: I0127 15:28:10.762996 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Jan 27 15:28:10 crc kubenswrapper[4772]: I0127 15:28:10.824514 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Jan 27 15:28:11 crc kubenswrapper[4772]: I0127 15:28:11.657761 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65e74d64-83f8-4964-8950-bf76816dd5fc","Type":"ContainerStarted","Data":"96c51a2493cf3f280c96b0bf8545deb92d6624f539170b50f2d2a50880a71c96"} Jan 27 15:28:12 crc kubenswrapper[4772]: I0127 15:28:12.683617 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65e74d64-83f8-4964-8950-bf76816dd5fc","Type":"ContainerStarted","Data":"902d06fc4d38a39dd40f6481a4374dc4c1e7ef0957c3355454b84d0778d1bc2b"} Jan 27 15:28:13 crc kubenswrapper[4772]: I0127 15:28:13.738077 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 15:28:13 crc kubenswrapper[4772]: I0127 15:28:13.738686 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c94a7cfa-28e2-4d52-85a1-d5586f162227" containerName="glance-log" containerID="cri-o://c52299828ac41e83b1686de53ba3808d1e810b20370ec9d5bc6e9bbc6b64bbed" gracePeriod=30 Jan 27 15:28:13 crc kubenswrapper[4772]: I0127 15:28:13.739141 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c94a7cfa-28e2-4d52-85a1-d5586f162227" containerName="glance-httpd" containerID="cri-o://9775d2c5b4eda3cae695814a686a4a82d4426bf3d7d28a73dffa9b807c4c16b8" gracePeriod=30 Jan 27 15:28:14 crc kubenswrapper[4772]: I0127 15:28:14.697801 4772 generic.go:334] "Generic (PLEG): container finished" podID="c94a7cfa-28e2-4d52-85a1-d5586f162227" containerID="c52299828ac41e83b1686de53ba3808d1e810b20370ec9d5bc6e9bbc6b64bbed" exitCode=143 Jan 27 15:28:14 crc kubenswrapper[4772]: I0127 15:28:14.697891 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c94a7cfa-28e2-4d52-85a1-d5586f162227","Type":"ContainerDied","Data":"c52299828ac41e83b1686de53ba3808d1e810b20370ec9d5bc6e9bbc6b64bbed"} Jan 27 15:28:16 crc kubenswrapper[4772]: I0127 15:28:16.076181 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Jan 27 15:28:16 crc kubenswrapper[4772]: I0127 15:28:16.601806 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 15:28:16 crc kubenswrapper[4772]: I0127 15:28:16.602422 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="41f85a83-f245-40ff-b994-50cab01b2530" containerName="glance-log" containerID="cri-o://58d128e4a7f44cc529be47e9f224989cce3b8a08dc4e4f4d37d49e38c0c7b8d2" gracePeriod=30 Jan 27 15:28:16 crc kubenswrapper[4772]: I0127 15:28:16.602493 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="41f85a83-f245-40ff-b994-50cab01b2530" containerName="glance-httpd" containerID="cri-o://f1accbd1db4a8c2dce7512a2eb2abaa265e29ed373b0fc121d29515c5bba0e55" gracePeriod=30 Jan 27 15:28:17 crc kubenswrapper[4772]: I0127 15:28:17.125825 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:17 crc kubenswrapper[4772]: I0127 15:28:17.126731 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:28:17 crc kubenswrapper[4772]: I0127 15:28:17.731541 4772 generic.go:334] "Generic (PLEG): container finished" podID="c94a7cfa-28e2-4d52-85a1-d5586f162227" containerID="9775d2c5b4eda3cae695814a686a4a82d4426bf3d7d28a73dffa9b807c4c16b8" exitCode=0 Jan 27 15:28:17 crc kubenswrapper[4772]: I0127 15:28:17.731683 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c94a7cfa-28e2-4d52-85a1-d5586f162227","Type":"ContainerDied","Data":"9775d2c5b4eda3cae695814a686a4a82d4426bf3d7d28a73dffa9b807c4c16b8"} Jan 27 15:28:17 crc kubenswrapper[4772]: I0127 15:28:17.743026 4772 generic.go:334] "Generic (PLEG): container finished" podID="41f85a83-f245-40ff-b994-50cab01b2530" containerID="58d128e4a7f44cc529be47e9f224989cce3b8a08dc4e4f4d37d49e38c0c7b8d2" exitCode=143 Jan 27 15:28:17 crc kubenswrapper[4772]: I0127 15:28:17.743408 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"41f85a83-f245-40ff-b994-50cab01b2530","Type":"ContainerDied","Data":"58d128e4a7f44cc529be47e9f224989cce3b8a08dc4e4f4d37d49e38c0c7b8d2"} Jan 27 15:28:18 crc kubenswrapper[4772]: I0127 15:28:18.765806 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.653852 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-pszgr"] Jan 27 15:28:19 crc kubenswrapper[4772]: E0127 15:28:19.664383 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a02a1b6c-d438-42bf-a577-88bbbcca2a00" containerName="barbican-api" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.664430 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a02a1b6c-d438-42bf-a577-88bbbcca2a00" containerName="barbican-api" Jan 27 15:28:19 crc kubenswrapper[4772]: E0127 15:28:19.664471 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17a547a9-a098-43b7-a153-ad9a137369de" containerName="dnsmasq-dns" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.664480 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="17a547a9-a098-43b7-a153-ad9a137369de" containerName="dnsmasq-dns" Jan 27 15:28:19 crc kubenswrapper[4772]: E0127 15:28:19.664518 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a02a1b6c-d438-42bf-a577-88bbbcca2a00" containerName="barbican-api-log" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.664527 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a02a1b6c-d438-42bf-a577-88bbbcca2a00" containerName="barbican-api-log" Jan 27 15:28:19 crc kubenswrapper[4772]: E0127 15:28:19.664574 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17a547a9-a098-43b7-a153-ad9a137369de" containerName="init" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.664582 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="17a547a9-a098-43b7-a153-ad9a137369de" containerName="init" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.664943 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="a02a1b6c-d438-42bf-a577-88bbbcca2a00" containerName="barbican-api" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.664969 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="17a547a9-a098-43b7-a153-ad9a137369de" containerName="dnsmasq-dns" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.664979 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="a02a1b6c-d438-42bf-a577-88bbbcca2a00" containerName="barbican-api-log" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.665902 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pszgr" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.668618 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-pszgr"] Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.748059 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-gbrww"] Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.749715 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-gbrww" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.750674 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54bbbf38-088b-4e4d-8154-569667fcf9a9-operator-scripts\") pod \"nova-api-db-create-pszgr\" (UID: \"54bbbf38-088b-4e4d-8154-569667fcf9a9\") " pod="openstack/nova-api-db-create-pszgr" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.750821 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwt4l\" (UniqueName: \"kubernetes.io/projected/54bbbf38-088b-4e4d-8154-569667fcf9a9-kube-api-access-vwt4l\") pod \"nova-api-db-create-pszgr\" (UID: \"54bbbf38-088b-4e4d-8154-569667fcf9a9\") " pod="openstack/nova-api-db-create-pszgr" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.759812 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-gbrww"] Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.775351 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-5104-account-create-update-vp7x7"] Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.776480 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5104-account-create-update-vp7x7" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.784520 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.834272 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-5104-account-create-update-vp7x7"] Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.853397 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69f24c00-a64a-4e82-a125-c0ee3fe8fa8f-operator-scripts\") pod \"nova-api-5104-account-create-update-vp7x7\" (UID: \"69f24c00-a64a-4e82-a125-c0ee3fe8fa8f\") " pod="openstack/nova-api-5104-account-create-update-vp7x7" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.853475 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54bbbf38-088b-4e4d-8154-569667fcf9a9-operator-scripts\") pod \"nova-api-db-create-pszgr\" (UID: \"54bbbf38-088b-4e4d-8154-569667fcf9a9\") " pod="openstack/nova-api-db-create-pszgr" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.853583 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwt4l\" (UniqueName: \"kubernetes.io/projected/54bbbf38-088b-4e4d-8154-569667fcf9a9-kube-api-access-vwt4l\") pod \"nova-api-db-create-pszgr\" (UID: \"54bbbf38-088b-4e4d-8154-569667fcf9a9\") " pod="openstack/nova-api-db-create-pszgr" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.853613 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be888039-f158-4d05-9f7d-6d01b2478b08-operator-scripts\") pod \"nova-cell0-db-create-gbrww\" (UID: \"be888039-f158-4d05-9f7d-6d01b2478b08\") " pod="openstack/nova-cell0-db-create-gbrww" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.853695 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvqvj\" (UniqueName: \"kubernetes.io/projected/be888039-f158-4d05-9f7d-6d01b2478b08-kube-api-access-pvqvj\") pod \"nova-cell0-db-create-gbrww\" (UID: \"be888039-f158-4d05-9f7d-6d01b2478b08\") " pod="openstack/nova-cell0-db-create-gbrww" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.853749 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtnzx\" (UniqueName: \"kubernetes.io/projected/69f24c00-a64a-4e82-a125-c0ee3fe8fa8f-kube-api-access-dtnzx\") pod \"nova-api-5104-account-create-update-vp7x7\" (UID: \"69f24c00-a64a-4e82-a125-c0ee3fe8fa8f\") " pod="openstack/nova-api-5104-account-create-update-vp7x7" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.854497 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54bbbf38-088b-4e4d-8154-569667fcf9a9-operator-scripts\") pod \"nova-api-db-create-pszgr\" (UID: \"54bbbf38-088b-4e4d-8154-569667fcf9a9\") " pod="openstack/nova-api-db-create-pszgr" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.874560 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwt4l\" (UniqueName: \"kubernetes.io/projected/54bbbf38-088b-4e4d-8154-569667fcf9a9-kube-api-access-vwt4l\") pod \"nova-api-db-create-pszgr\" (UID: \"54bbbf38-088b-4e4d-8154-569667fcf9a9\") " pod="openstack/nova-api-db-create-pszgr" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.954860 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69f24c00-a64a-4e82-a125-c0ee3fe8fa8f-operator-scripts\") pod \"nova-api-5104-account-create-update-vp7x7\" (UID: \"69f24c00-a64a-4e82-a125-c0ee3fe8fa8f\") " pod="openstack/nova-api-5104-account-create-update-vp7x7" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.954980 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be888039-f158-4d05-9f7d-6d01b2478b08-operator-scripts\") pod \"nova-cell0-db-create-gbrww\" (UID: \"be888039-f158-4d05-9f7d-6d01b2478b08\") " pod="openstack/nova-cell0-db-create-gbrww" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.955062 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvqvj\" (UniqueName: \"kubernetes.io/projected/be888039-f158-4d05-9f7d-6d01b2478b08-kube-api-access-pvqvj\") pod \"nova-cell0-db-create-gbrww\" (UID: \"be888039-f158-4d05-9f7d-6d01b2478b08\") " pod="openstack/nova-cell0-db-create-gbrww" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.955119 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtnzx\" (UniqueName: \"kubernetes.io/projected/69f24c00-a64a-4e82-a125-c0ee3fe8fa8f-kube-api-access-dtnzx\") pod \"nova-api-5104-account-create-update-vp7x7\" (UID: \"69f24c00-a64a-4e82-a125-c0ee3fe8fa8f\") " pod="openstack/nova-api-5104-account-create-update-vp7x7" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.955808 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-6w7p7"] Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.955848 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69f24c00-a64a-4e82-a125-c0ee3fe8fa8f-operator-scripts\") pod \"nova-api-5104-account-create-update-vp7x7\" (UID: \"69f24c00-a64a-4e82-a125-c0ee3fe8fa8f\") " pod="openstack/nova-api-5104-account-create-update-vp7x7" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.956031 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be888039-f158-4d05-9f7d-6d01b2478b08-operator-scripts\") pod \"nova-cell0-db-create-gbrww\" (UID: \"be888039-f158-4d05-9f7d-6d01b2478b08\") " pod="openstack/nova-cell0-db-create-gbrww" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.957913 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-6w7p7" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.975668 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-6w7p7"] Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.985466 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtnzx\" (UniqueName: \"kubernetes.io/projected/69f24c00-a64a-4e82-a125-c0ee3fe8fa8f-kube-api-access-dtnzx\") pod \"nova-api-5104-account-create-update-vp7x7\" (UID: \"69f24c00-a64a-4e82-a125-c0ee3fe8fa8f\") " pod="openstack/nova-api-5104-account-create-update-vp7x7" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.998546 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pszgr" Jan 27 15:28:19 crc kubenswrapper[4772]: I0127 15:28:19.999006 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvqvj\" (UniqueName: \"kubernetes.io/projected/be888039-f158-4d05-9f7d-6d01b2478b08-kube-api-access-pvqvj\") pod \"nova-cell0-db-create-gbrww\" (UID: \"be888039-f158-4d05-9f7d-6d01b2478b08\") " pod="openstack/nova-cell0-db-create-gbrww" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.015870 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-6af8-account-create-update-ltwnh"] Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.017423 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6af8-account-create-update-ltwnh" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.021428 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.055861 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/564de425-5170-45df-9080-5b02579483ee-operator-scripts\") pod \"nova-cell0-6af8-account-create-update-ltwnh\" (UID: \"564de425-5170-45df-9080-5b02579483ee\") " pod="openstack/nova-cell0-6af8-account-create-update-ltwnh" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.055916 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7907cc16-7665-49d3-ad17-f9e6e0fc2f09-operator-scripts\") pod \"nova-cell1-db-create-6w7p7\" (UID: \"7907cc16-7665-49d3-ad17-f9e6e0fc2f09\") " pod="openstack/nova-cell1-db-create-6w7p7" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.056008 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znsll\" (UniqueName: \"kubernetes.io/projected/564de425-5170-45df-9080-5b02579483ee-kube-api-access-znsll\") pod \"nova-cell0-6af8-account-create-update-ltwnh\" (UID: \"564de425-5170-45df-9080-5b02579483ee\") " pod="openstack/nova-cell0-6af8-account-create-update-ltwnh" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.056117 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv946\" (UniqueName: \"kubernetes.io/projected/7907cc16-7665-49d3-ad17-f9e6e0fc2f09-kube-api-access-hv946\") pod \"nova-cell1-db-create-6w7p7\" (UID: \"7907cc16-7665-49d3-ad17-f9e6e0fc2f09\") " pod="openstack/nova-cell1-db-create-6w7p7" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.062817 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-6af8-account-create-update-ltwnh"] Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.070224 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-gbrww" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.102929 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5104-account-create-update-vp7x7" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.158819 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/564de425-5170-45df-9080-5b02579483ee-operator-scripts\") pod \"nova-cell0-6af8-account-create-update-ltwnh\" (UID: \"564de425-5170-45df-9080-5b02579483ee\") " pod="openstack/nova-cell0-6af8-account-create-update-ltwnh" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.158877 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7907cc16-7665-49d3-ad17-f9e6e0fc2f09-operator-scripts\") pod \"nova-cell1-db-create-6w7p7\" (UID: \"7907cc16-7665-49d3-ad17-f9e6e0fc2f09\") " pod="openstack/nova-cell1-db-create-6w7p7" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.158929 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znsll\" (UniqueName: \"kubernetes.io/projected/564de425-5170-45df-9080-5b02579483ee-kube-api-access-znsll\") pod \"nova-cell0-6af8-account-create-update-ltwnh\" (UID: \"564de425-5170-45df-9080-5b02579483ee\") " pod="openstack/nova-cell0-6af8-account-create-update-ltwnh" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.158996 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv946\" (UniqueName: \"kubernetes.io/projected/7907cc16-7665-49d3-ad17-f9e6e0fc2f09-kube-api-access-hv946\") pod \"nova-cell1-db-create-6w7p7\" (UID: \"7907cc16-7665-49d3-ad17-f9e6e0fc2f09\") " pod="openstack/nova-cell1-db-create-6w7p7" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.160640 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/564de425-5170-45df-9080-5b02579483ee-operator-scripts\") pod \"nova-cell0-6af8-account-create-update-ltwnh\" (UID: \"564de425-5170-45df-9080-5b02579483ee\") " pod="openstack/nova-cell0-6af8-account-create-update-ltwnh" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.161257 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7907cc16-7665-49d3-ad17-f9e6e0fc2f09-operator-scripts\") pod \"nova-cell1-db-create-6w7p7\" (UID: \"7907cc16-7665-49d3-ad17-f9e6e0fc2f09\") " pod="openstack/nova-cell1-db-create-6w7p7" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.162422 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-12a3-account-create-update-mdv84"] Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.163492 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-12a3-account-create-update-mdv84" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.168223 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.183230 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-12a3-account-create-update-mdv84"] Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.194445 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znsll\" (UniqueName: \"kubernetes.io/projected/564de425-5170-45df-9080-5b02579483ee-kube-api-access-znsll\") pod \"nova-cell0-6af8-account-create-update-ltwnh\" (UID: \"564de425-5170-45df-9080-5b02579483ee\") " pod="openstack/nova-cell0-6af8-account-create-update-ltwnh" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.199655 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv946\" (UniqueName: \"kubernetes.io/projected/7907cc16-7665-49d3-ad17-f9e6e0fc2f09-kube-api-access-hv946\") pod \"nova-cell1-db-create-6w7p7\" (UID: \"7907cc16-7665-49d3-ad17-f9e6e0fc2f09\") " pod="openstack/nova-cell1-db-create-6w7p7" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.262129 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-878xd\" (UniqueName: \"kubernetes.io/projected/08d7e14a-70d3-446e-8250-ca1047b5bc4b-kube-api-access-878xd\") pod \"nova-cell1-12a3-account-create-update-mdv84\" (UID: \"08d7e14a-70d3-446e-8250-ca1047b5bc4b\") " pod="openstack/nova-cell1-12a3-account-create-update-mdv84" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.262310 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08d7e14a-70d3-446e-8250-ca1047b5bc4b-operator-scripts\") pod \"nova-cell1-12a3-account-create-update-mdv84\" (UID: \"08d7e14a-70d3-446e-8250-ca1047b5bc4b\") " pod="openstack/nova-cell1-12a3-account-create-update-mdv84" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.279890 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-6w7p7" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.363898 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-878xd\" (UniqueName: \"kubernetes.io/projected/08d7e14a-70d3-446e-8250-ca1047b5bc4b-kube-api-access-878xd\") pod \"nova-cell1-12a3-account-create-update-mdv84\" (UID: \"08d7e14a-70d3-446e-8250-ca1047b5bc4b\") " pod="openstack/nova-cell1-12a3-account-create-update-mdv84" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.364424 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08d7e14a-70d3-446e-8250-ca1047b5bc4b-operator-scripts\") pod \"nova-cell1-12a3-account-create-update-mdv84\" (UID: \"08d7e14a-70d3-446e-8250-ca1047b5bc4b\") " pod="openstack/nova-cell1-12a3-account-create-update-mdv84" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.365288 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08d7e14a-70d3-446e-8250-ca1047b5bc4b-operator-scripts\") pod \"nova-cell1-12a3-account-create-update-mdv84\" (UID: \"08d7e14a-70d3-446e-8250-ca1047b5bc4b\") " pod="openstack/nova-cell1-12a3-account-create-update-mdv84" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.385864 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-878xd\" (UniqueName: \"kubernetes.io/projected/08d7e14a-70d3-446e-8250-ca1047b5bc4b-kube-api-access-878xd\") pod \"nova-cell1-12a3-account-create-update-mdv84\" (UID: \"08d7e14a-70d3-446e-8250-ca1047b5bc4b\") " pod="openstack/nova-cell1-12a3-account-create-update-mdv84" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.396707 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6af8-account-create-update-ltwnh" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.571942 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-12a3-account-create-update-mdv84" Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.879772 4772 generic.go:334] "Generic (PLEG): container finished" podID="41f85a83-f245-40ff-b994-50cab01b2530" containerID="f1accbd1db4a8c2dce7512a2eb2abaa265e29ed373b0fc121d29515c5bba0e55" exitCode=0 Jan 27 15:28:20 crc kubenswrapper[4772]: I0127 15:28:20.879880 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"41f85a83-f245-40ff-b994-50cab01b2530","Type":"ContainerDied","Data":"f1accbd1db4a8c2dce7512a2eb2abaa265e29ed373b0fc121d29515c5bba0e55"} Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.196720 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.198147 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q49rw\" (UniqueName: \"kubernetes.io/projected/c94a7cfa-28e2-4d52-85a1-d5586f162227-kube-api-access-q49rw\") pod \"c94a7cfa-28e2-4d52-85a1-d5586f162227\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.198231 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c94a7cfa-28e2-4d52-85a1-d5586f162227-combined-ca-bundle\") pod \"c94a7cfa-28e2-4d52-85a1-d5586f162227\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.198319 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"c94a7cfa-28e2-4d52-85a1-d5586f162227\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.200058 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c94a7cfa-28e2-4d52-85a1-d5586f162227-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c94a7cfa-28e2-4d52-85a1-d5586f162227" (UID: "c94a7cfa-28e2-4d52-85a1-d5586f162227"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.199051 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c94a7cfa-28e2-4d52-85a1-d5586f162227-httpd-run\") pod \"c94a7cfa-28e2-4d52-85a1-d5586f162227\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.201210 4772 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c94a7cfa-28e2-4d52-85a1-d5586f162227-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.209223 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c94a7cfa-28e2-4d52-85a1-d5586f162227-kube-api-access-q49rw" (OuterVolumeSpecName: "kube-api-access-q49rw") pod "c94a7cfa-28e2-4d52-85a1-d5586f162227" (UID: "c94a7cfa-28e2-4d52-85a1-d5586f162227"). InnerVolumeSpecName "kube-api-access-q49rw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.212067 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "c94a7cfa-28e2-4d52-85a1-d5586f162227" (UID: "c94a7cfa-28e2-4d52-85a1-d5586f162227"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.273992 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c94a7cfa-28e2-4d52-85a1-d5586f162227-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c94a7cfa-28e2-4d52-85a1-d5586f162227" (UID: "c94a7cfa-28e2-4d52-85a1-d5586f162227"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.303630 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c94a7cfa-28e2-4d52-85a1-d5586f162227-logs\") pod \"c94a7cfa-28e2-4d52-85a1-d5586f162227\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.303723 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c94a7cfa-28e2-4d52-85a1-d5586f162227-public-tls-certs\") pod \"c94a7cfa-28e2-4d52-85a1-d5586f162227\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.303907 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c94a7cfa-28e2-4d52-85a1-d5586f162227-config-data\") pod \"c94a7cfa-28e2-4d52-85a1-d5586f162227\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.303959 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c94a7cfa-28e2-4d52-85a1-d5586f162227-scripts\") pod \"c94a7cfa-28e2-4d52-85a1-d5586f162227\" (UID: \"c94a7cfa-28e2-4d52-85a1-d5586f162227\") " Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.304690 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q49rw\" (UniqueName: \"kubernetes.io/projected/c94a7cfa-28e2-4d52-85a1-d5586f162227-kube-api-access-q49rw\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.304712 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c94a7cfa-28e2-4d52-85a1-d5586f162227-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.304738 4772 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.306301 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c94a7cfa-28e2-4d52-85a1-d5586f162227-logs" (OuterVolumeSpecName: "logs") pod "c94a7cfa-28e2-4d52-85a1-d5586f162227" (UID: "c94a7cfa-28e2-4d52-85a1-d5586f162227"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.312411 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c94a7cfa-28e2-4d52-85a1-d5586f162227-scripts" (OuterVolumeSpecName: "scripts") pod "c94a7cfa-28e2-4d52-85a1-d5586f162227" (UID: "c94a7cfa-28e2-4d52-85a1-d5586f162227"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.405652 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c94a7cfa-28e2-4d52-85a1-d5586f162227-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.405682 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c94a7cfa-28e2-4d52-85a1-d5586f162227-logs\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.414682 4772 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.442374 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c94a7cfa-28e2-4d52-85a1-d5586f162227-config-data" (OuterVolumeSpecName: "config-data") pod "c94a7cfa-28e2-4d52-85a1-d5586f162227" (UID: "c94a7cfa-28e2-4d52-85a1-d5586f162227"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.446089 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c94a7cfa-28e2-4d52-85a1-d5586f162227-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c94a7cfa-28e2-4d52-85a1-d5586f162227" (UID: "c94a7cfa-28e2-4d52-85a1-d5586f162227"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.460255 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.506392 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"41f85a83-f245-40ff-b994-50cab01b2530\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.506473 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41f85a83-f245-40ff-b994-50cab01b2530-logs\") pod \"41f85a83-f245-40ff-b994-50cab01b2530\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.506501 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/41f85a83-f245-40ff-b994-50cab01b2530-httpd-run\") pod \"41f85a83-f245-40ff-b994-50cab01b2530\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.506557 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pc9t4\" (UniqueName: \"kubernetes.io/projected/41f85a83-f245-40ff-b994-50cab01b2530-kube-api-access-pc9t4\") pod \"41f85a83-f245-40ff-b994-50cab01b2530\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.506584 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41f85a83-f245-40ff-b994-50cab01b2530-config-data\") pod \"41f85a83-f245-40ff-b994-50cab01b2530\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.506614 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41f85a83-f245-40ff-b994-50cab01b2530-internal-tls-certs\") pod \"41f85a83-f245-40ff-b994-50cab01b2530\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.506682 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41f85a83-f245-40ff-b994-50cab01b2530-combined-ca-bundle\") pod \"41f85a83-f245-40ff-b994-50cab01b2530\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.506714 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41f85a83-f245-40ff-b994-50cab01b2530-scripts\") pod \"41f85a83-f245-40ff-b994-50cab01b2530\" (UID: \"41f85a83-f245-40ff-b994-50cab01b2530\") " Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.506969 4772 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c94a7cfa-28e2-4d52-85a1-d5586f162227-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.506994 4772 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.507008 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c94a7cfa-28e2-4d52-85a1-d5586f162227-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.508228 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41f85a83-f245-40ff-b994-50cab01b2530-logs" (OuterVolumeSpecName: "logs") pod "41f85a83-f245-40ff-b994-50cab01b2530" (UID: "41f85a83-f245-40ff-b994-50cab01b2530"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.508438 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41f85a83-f245-40ff-b994-50cab01b2530-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "41f85a83-f245-40ff-b994-50cab01b2530" (UID: "41f85a83-f245-40ff-b994-50cab01b2530"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.509958 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "41f85a83-f245-40ff-b994-50cab01b2530" (UID: "41f85a83-f245-40ff-b994-50cab01b2530"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.510413 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41f85a83-f245-40ff-b994-50cab01b2530-scripts" (OuterVolumeSpecName: "scripts") pod "41f85a83-f245-40ff-b994-50cab01b2530" (UID: "41f85a83-f245-40ff-b994-50cab01b2530"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.526709 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41f85a83-f245-40ff-b994-50cab01b2530-kube-api-access-pc9t4" (OuterVolumeSpecName: "kube-api-access-pc9t4") pod "41f85a83-f245-40ff-b994-50cab01b2530" (UID: "41f85a83-f245-40ff-b994-50cab01b2530"). InnerVolumeSpecName "kube-api-access-pc9t4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.554282 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41f85a83-f245-40ff-b994-50cab01b2530-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41f85a83-f245-40ff-b994-50cab01b2530" (UID: "41f85a83-f245-40ff-b994-50cab01b2530"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.574092 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41f85a83-f245-40ff-b994-50cab01b2530-config-data" (OuterVolumeSpecName: "config-data") pod "41f85a83-f245-40ff-b994-50cab01b2530" (UID: "41f85a83-f245-40ff-b994-50cab01b2530"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.578660 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41f85a83-f245-40ff-b994-50cab01b2530-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "41f85a83-f245-40ff-b994-50cab01b2530" (UID: "41f85a83-f245-40ff-b994-50cab01b2530"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.608137 4772 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.608185 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41f85a83-f245-40ff-b994-50cab01b2530-logs\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.608197 4772 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/41f85a83-f245-40ff-b994-50cab01b2530-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.608207 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pc9t4\" (UniqueName: \"kubernetes.io/projected/41f85a83-f245-40ff-b994-50cab01b2530-kube-api-access-pc9t4\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.608219 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41f85a83-f245-40ff-b994-50cab01b2530-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.608226 4772 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41f85a83-f245-40ff-b994-50cab01b2530-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.608234 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41f85a83-f245-40ff-b994-50cab01b2530-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.608242 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41f85a83-f245-40ff-b994-50cab01b2530-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.644782 4772 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.711917 4772 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.818969 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-5104-account-create-update-vp7x7"] Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.864274 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-6af8-account-create-update-ltwnh"] Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.877073 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-gbrww"] Jan 27 15:28:21 crc kubenswrapper[4772]: W0127 15:28:21.890290 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7907cc16_7665_49d3_ad17_f9e6e0fc2f09.slice/crio-5abfcf1bdd92bd3b6c78b0f416ccffb023fa1f4cdaf92b953a574b587cc4a6d0 WatchSource:0}: Error finding container 5abfcf1bdd92bd3b6c78b0f416ccffb023fa1f4cdaf92b953a574b587cc4a6d0: Status 404 returned error can't find the container with id 5abfcf1bdd92bd3b6c78b0f416ccffb023fa1f4cdaf92b953a574b587cc4a6d0 Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.928375 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-6w7p7"] Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.943494 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"41f85a83-f245-40ff-b994-50cab01b2530","Type":"ContainerDied","Data":"70507e561102278ed4f801ac168676eb09960026059c30c45c5fe4950449c589"} Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.943542 4772 scope.go:117] "RemoveContainer" containerID="f1accbd1db4a8c2dce7512a2eb2abaa265e29ed373b0fc121d29515c5bba0e55" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.943660 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.951507 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-pszgr"] Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.964105 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pszgr" event={"ID":"54bbbf38-088b-4e4d-8154-569667fcf9a9","Type":"ContainerStarted","Data":"303a25d43e818afc0b1deec34d71e454f848a7bcfa271b0aa6ddd404903c4f4b"} Jan 27 15:28:21 crc kubenswrapper[4772]: I0127 15:28:21.989429 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-12a3-account-create-update-mdv84"] Jan 27 15:28:22 crc kubenswrapper[4772]: W0127 15:28:22.053532 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08d7e14a_70d3_446e_8250_ca1047b5bc4b.slice/crio-4ae0384de6c053e3c0b5213f6dd65e86d1254cb54c25b58aa19f1adb13fef254 WatchSource:0}: Error finding container 4ae0384de6c053e3c0b5213f6dd65e86d1254cb54c25b58aa19f1adb13fef254: Status 404 returned error can't find the container with id 4ae0384de6c053e3c0b5213f6dd65e86d1254cb54c25b58aa19f1adb13fef254 Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.053695 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5104-account-create-update-vp7x7" event={"ID":"69f24c00-a64a-4e82-a125-c0ee3fe8fa8f","Type":"ContainerStarted","Data":"9536a6635c6de01fa9cbbeb2b3e4a3db2498f81bff722635d21115dedc7f8ce3"} Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.091381 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"0edf6707-14dd-4986-8d64-0e48a31d6a39","Type":"ContainerStarted","Data":"0c37dc673e475cc4ca1e8b831b0543b26650ceedc799dace964e07fb4c7c7ffa"} Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.125489 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.139375 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.139690 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.139743 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c94a7cfa-28e2-4d52-85a1-d5586f162227","Type":"ContainerDied","Data":"65117f0b87347a480b318a709dc150116a10a8d323bd2553e117803b3054a685"} Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.149340 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 15:28:22 crc kubenswrapper[4772]: E0127 15:28:22.150433 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41f85a83-f245-40ff-b994-50cab01b2530" containerName="glance-log" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.150550 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="41f85a83-f245-40ff-b994-50cab01b2530" containerName="glance-log" Jan 27 15:28:22 crc kubenswrapper[4772]: E0127 15:28:22.150620 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41f85a83-f245-40ff-b994-50cab01b2530" containerName="glance-httpd" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.150686 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="41f85a83-f245-40ff-b994-50cab01b2530" containerName="glance-httpd" Jan 27 15:28:22 crc kubenswrapper[4772]: E0127 15:28:22.150743 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c94a7cfa-28e2-4d52-85a1-d5586f162227" containerName="glance-log" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.150791 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="c94a7cfa-28e2-4d52-85a1-d5586f162227" containerName="glance-log" Jan 27 15:28:22 crc kubenswrapper[4772]: E0127 15:28:22.150862 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c94a7cfa-28e2-4d52-85a1-d5586f162227" containerName="glance-httpd" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.150920 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="c94a7cfa-28e2-4d52-85a1-d5586f162227" containerName="glance-httpd" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.151156 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="c94a7cfa-28e2-4d52-85a1-d5586f162227" containerName="glance-log" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.151309 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="41f85a83-f245-40ff-b994-50cab01b2530" containerName="glance-httpd" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.151472 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="41f85a83-f245-40ff-b994-50cab01b2530" containerName="glance-log" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.151553 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="c94a7cfa-28e2-4d52-85a1-d5586f162227" containerName="glance-httpd" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.153507 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.158752 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.158805 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.158903 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-vd4fn" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.158929 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.171119 4772 scope.go:117] "RemoveContainer" containerID="58d128e4a7f44cc529be47e9f224989cce3b8a08dc4e4f4d37d49e38c0c7b8d2" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.179879 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.201687 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=4.112610248 podStartE2EDuration="25.201667156s" podCreationTimestamp="2026-01-27 15:27:57 +0000 UTC" firstStartedPulling="2026-01-27 15:27:59.635104725 +0000 UTC m=+1265.615713823" lastFinishedPulling="2026-01-27 15:28:20.724161633 +0000 UTC m=+1286.704770731" observedRunningTime="2026-01-27 15:28:22.15638888 +0000 UTC m=+1288.136997978" watchObservedRunningTime="2026-01-27 15:28:22.201667156 +0000 UTC m=+1288.182276254" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.213006 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65e74d64-83f8-4964-8950-bf76816dd5fc","Type":"ContainerStarted","Data":"2d2ef8b64e19f03ac931485698899fb937f36aead7d1925135401934fbecd74a"} Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.222192 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-6af8-account-create-update-ltwnh" event={"ID":"564de425-5170-45df-9080-5b02579483ee","Type":"ContainerStarted","Data":"deaf013f2c9b59ae86d1e302aff4c0c47992ea99da1873382d3b8f4f077d78da"} Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.232640 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpm5h\" (UniqueName: \"kubernetes.io/projected/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-kube-api-access-rpm5h\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.232920 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-gbrww" event={"ID":"be888039-f158-4d05-9f7d-6d01b2478b08","Type":"ContainerStarted","Data":"56740a6f5142d64002a343f18ba92dbb512d9b8a106c7a40a6e41b349e43508e"} Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.237212 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-logs\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.237342 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.241213 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.241377 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.241414 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.241467 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.241509 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.283096 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.296317 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.319389 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.323031 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.335460 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.336874 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.337606 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.340157 4772 scope.go:117] "RemoveContainer" containerID="9775d2c5b4eda3cae695814a686a4a82d4426bf3d7d28a73dffa9b807c4c16b8" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.343813 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.343937 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.344087 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpm5h\" (UniqueName: \"kubernetes.io/projected/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-kube-api-access-rpm5h\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.344571 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-logs\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.344687 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.344861 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.344999 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.347053 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.344515 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.345092 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-logs\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.346953 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.360201 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.361011 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.376937 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.383329 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpm5h\" (UniqueName: \"kubernetes.io/projected/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-kube-api-access-rpm5h\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.388582 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.431955 4772 scope.go:117] "RemoveContainer" containerID="c52299828ac41e83b1686de53ba3808d1e810b20370ec9d5bc6e9bbc6b64bbed" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.438472 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.558866 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a02b617-28a7-4262-a110-f1c71763ad19-config-data\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.558936 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a02b617-28a7-4262-a110-f1c71763ad19-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.558957 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a02b617-28a7-4262-a110-f1c71763ad19-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.559053 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a02b617-28a7-4262-a110-f1c71763ad19-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.559096 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.559119 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a02b617-28a7-4262-a110-f1c71763ad19-scripts\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.559136 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a02b617-28a7-4262-a110-f1c71763ad19-logs\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.559208 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghzrk\" (UniqueName: \"kubernetes.io/projected/9a02b617-28a7-4262-a110-f1c71763ad19-kube-api-access-ghzrk\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.601703 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.660838 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a02b617-28a7-4262-a110-f1c71763ad19-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.661111 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a02b617-28a7-4262-a110-f1c71763ad19-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.661351 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a02b617-28a7-4262-a110-f1c71763ad19-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.661428 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.661456 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a02b617-28a7-4262-a110-f1c71763ad19-scripts\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.661481 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a02b617-28a7-4262-a110-f1c71763ad19-logs\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.661606 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghzrk\" (UniqueName: \"kubernetes.io/projected/9a02b617-28a7-4262-a110-f1c71763ad19-kube-api-access-ghzrk\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.661675 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a02b617-28a7-4262-a110-f1c71763ad19-config-data\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.661832 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.662413 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a02b617-28a7-4262-a110-f1c71763ad19-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.662646 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a02b617-28a7-4262-a110-f1c71763ad19-logs\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.681760 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a02b617-28a7-4262-a110-f1c71763ad19-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.691036 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41f85a83-f245-40ff-b994-50cab01b2530" path="/var/lib/kubelet/pods/41f85a83-f245-40ff-b994-50cab01b2530/volumes" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.692705 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c94a7cfa-28e2-4d52-85a1-d5586f162227" path="/var/lib/kubelet/pods/c94a7cfa-28e2-4d52-85a1-d5586f162227/volumes" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.707285 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a02b617-28a7-4262-a110-f1c71763ad19-scripts\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.707397 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a02b617-28a7-4262-a110-f1c71763ad19-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.708220 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a02b617-28a7-4262-a110-f1c71763ad19-config-data\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.708619 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghzrk\" (UniqueName: \"kubernetes.io/projected/9a02b617-28a7-4262-a110-f1c71763ad19-kube-api-access-ghzrk\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " pod="openstack/glance-default-external-api-0" Jan 27 15:28:22 crc kubenswrapper[4772]: I0127 15:28:22.835951 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " pod="openstack/glance-default-external-api-0" Jan 27 15:28:23 crc kubenswrapper[4772]: I0127 15:28:23.109195 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 27 15:28:23 crc kubenswrapper[4772]: I0127 15:28:23.244592 4772 generic.go:334] "Generic (PLEG): container finished" podID="54bbbf38-088b-4e4d-8154-569667fcf9a9" containerID="4da9288c82c7401f434d2a53ff336e0d653eb3932d204eafc2869a5860cee4bc" exitCode=0 Jan 27 15:28:23 crc kubenswrapper[4772]: I0127 15:28:23.244669 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pszgr" event={"ID":"54bbbf38-088b-4e4d-8154-569667fcf9a9","Type":"ContainerDied","Data":"4da9288c82c7401f434d2a53ff336e0d653eb3932d204eafc2869a5860cee4bc"} Jan 27 15:28:23 crc kubenswrapper[4772]: I0127 15:28:23.246089 4772 generic.go:334] "Generic (PLEG): container finished" podID="69f24c00-a64a-4e82-a125-c0ee3fe8fa8f" containerID="e995550ae720943eacfd405b30c920c20d450c9bc6c2389b27261b188859406e" exitCode=0 Jan 27 15:28:23 crc kubenswrapper[4772]: I0127 15:28:23.246176 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5104-account-create-update-vp7x7" event={"ID":"69f24c00-a64a-4e82-a125-c0ee3fe8fa8f","Type":"ContainerDied","Data":"e995550ae720943eacfd405b30c920c20d450c9bc6c2389b27261b188859406e"} Jan 27 15:28:23 crc kubenswrapper[4772]: I0127 15:28:23.249115 4772 generic.go:334] "Generic (PLEG): container finished" podID="7907cc16-7665-49d3-ad17-f9e6e0fc2f09" containerID="9853bf54eae9ce0f1c3b8ddee31101fe10bc44f0b0f41d495f936c0ac3cc7ec8" exitCode=0 Jan 27 15:28:23 crc kubenswrapper[4772]: I0127 15:28:23.249217 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-6w7p7" event={"ID":"7907cc16-7665-49d3-ad17-f9e6e0fc2f09","Type":"ContainerDied","Data":"9853bf54eae9ce0f1c3b8ddee31101fe10bc44f0b0f41d495f936c0ac3cc7ec8"} Jan 27 15:28:23 crc kubenswrapper[4772]: I0127 15:28:23.249240 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-6w7p7" event={"ID":"7907cc16-7665-49d3-ad17-f9e6e0fc2f09","Type":"ContainerStarted","Data":"5abfcf1bdd92bd3b6c78b0f416ccffb023fa1f4cdaf92b953a574b587cc4a6d0"} Jan 27 15:28:23 crc kubenswrapper[4772]: I0127 15:28:23.250362 4772 generic.go:334] "Generic (PLEG): container finished" podID="08d7e14a-70d3-446e-8250-ca1047b5bc4b" containerID="f28ff63f10f8899bc8cd8fd5a42bd4249a187e430ea97934ef6b489554310751" exitCode=0 Jan 27 15:28:23 crc kubenswrapper[4772]: I0127 15:28:23.250458 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-12a3-account-create-update-mdv84" event={"ID":"08d7e14a-70d3-446e-8250-ca1047b5bc4b","Type":"ContainerDied","Data":"f28ff63f10f8899bc8cd8fd5a42bd4249a187e430ea97934ef6b489554310751"} Jan 27 15:28:23 crc kubenswrapper[4772]: I0127 15:28:23.250489 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-12a3-account-create-update-mdv84" event={"ID":"08d7e14a-70d3-446e-8250-ca1047b5bc4b","Type":"ContainerStarted","Data":"4ae0384de6c053e3c0b5213f6dd65e86d1254cb54c25b58aa19f1adb13fef254"} Jan 27 15:28:23 crc kubenswrapper[4772]: I0127 15:28:23.252627 4772 generic.go:334] "Generic (PLEG): container finished" podID="564de425-5170-45df-9080-5b02579483ee" containerID="46996df047d6fd10b3034c52a93ce3634cebbfdcb4bf44854f66da5e6d342110" exitCode=0 Jan 27 15:28:23 crc kubenswrapper[4772]: I0127 15:28:23.252679 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-6af8-account-create-update-ltwnh" event={"ID":"564de425-5170-45df-9080-5b02579483ee","Type":"ContainerDied","Data":"46996df047d6fd10b3034c52a93ce3634cebbfdcb4bf44854f66da5e6d342110"} Jan 27 15:28:23 crc kubenswrapper[4772]: I0127 15:28:23.255778 4772 generic.go:334] "Generic (PLEG): container finished" podID="be888039-f158-4d05-9f7d-6d01b2478b08" containerID="e1b312b7631d415f567909a3003da4cdfd7208b6894d1397aa7da34098746b5a" exitCode=0 Jan 27 15:28:23 crc kubenswrapper[4772]: I0127 15:28:23.256614 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-gbrww" event={"ID":"be888039-f158-4d05-9f7d-6d01b2478b08","Type":"ContainerDied","Data":"e1b312b7631d415f567909a3003da4cdfd7208b6894d1397aa7da34098746b5a"} Jan 27 15:28:23 crc kubenswrapper[4772]: I0127 15:28:23.373387 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 15:28:23 crc kubenswrapper[4772]: I0127 15:28:23.719583 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 15:28:24 crc kubenswrapper[4772]: I0127 15:28:24.282117 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65e74d64-83f8-4964-8950-bf76816dd5fc","Type":"ContainerStarted","Data":"db4f9e747383adbadbac961f6c8e5009d6edc10dcb010e6f2eb4e5637b296b8e"} Jan 27 15:28:24 crc kubenswrapper[4772]: I0127 15:28:24.282276 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="65e74d64-83f8-4964-8950-bf76816dd5fc" containerName="ceilometer-central-agent" containerID="cri-o://96c51a2493cf3f280c96b0bf8545deb92d6624f539170b50f2d2a50880a71c96" gracePeriod=30 Jan 27 15:28:24 crc kubenswrapper[4772]: I0127 15:28:24.282343 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="65e74d64-83f8-4964-8950-bf76816dd5fc" containerName="sg-core" containerID="cri-o://2d2ef8b64e19f03ac931485698899fb937f36aead7d1925135401934fbecd74a" gracePeriod=30 Jan 27 15:28:24 crc kubenswrapper[4772]: I0127 15:28:24.282359 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="65e74d64-83f8-4964-8950-bf76816dd5fc" containerName="ceilometer-notification-agent" containerID="cri-o://902d06fc4d38a39dd40f6481a4374dc4c1e7ef0957c3355454b84d0778d1bc2b" gracePeriod=30 Jan 27 15:28:24 crc kubenswrapper[4772]: I0127 15:28:24.282327 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="65e74d64-83f8-4964-8950-bf76816dd5fc" containerName="proxy-httpd" containerID="cri-o://db4f9e747383adbadbac961f6c8e5009d6edc10dcb010e6f2eb4e5637b296b8e" gracePeriod=30 Jan 27 15:28:24 crc kubenswrapper[4772]: I0127 15:28:24.282769 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 27 15:28:24 crc kubenswrapper[4772]: I0127 15:28:24.290961 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9a02b617-28a7-4262-a110-f1c71763ad19","Type":"ContainerStarted","Data":"1536a68238e83bb2c89cfe9a0fce1841bc4d60d2a518fdc49dc1b005d27a6470"} Jan 27 15:28:24 crc kubenswrapper[4772]: I0127 15:28:24.297971 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343","Type":"ContainerStarted","Data":"3454f9899adaff309b52934e71697924735c1f269fb473444cba03b5baf4e1e5"} Jan 27 15:28:24 crc kubenswrapper[4772]: I0127 15:28:24.298059 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343","Type":"ContainerStarted","Data":"64eb2d8855af54c245dc9d145df3ac0064c424271a5cf4af6c9815a1aa8bc16e"} Jan 27 15:28:24 crc kubenswrapper[4772]: I0127 15:28:24.315213 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.425757034 podStartE2EDuration="16.315150772s" podCreationTimestamp="2026-01-27 15:28:08 +0000 UTC" firstStartedPulling="2026-01-27 15:28:09.811238676 +0000 UTC m=+1275.791847774" lastFinishedPulling="2026-01-27 15:28:23.700632414 +0000 UTC m=+1289.681241512" observedRunningTime="2026-01-27 15:28:24.313945607 +0000 UTC m=+1290.294554715" watchObservedRunningTime="2026-01-27 15:28:24.315150772 +0000 UTC m=+1290.295759880" Jan 27 15:28:24 crc kubenswrapper[4772]: I0127 15:28:24.609906 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pszgr" Jan 27 15:28:24 crc kubenswrapper[4772]: I0127 15:28:24.701305 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54bbbf38-088b-4e4d-8154-569667fcf9a9-operator-scripts\") pod \"54bbbf38-088b-4e4d-8154-569667fcf9a9\" (UID: \"54bbbf38-088b-4e4d-8154-569667fcf9a9\") " Jan 27 15:28:24 crc kubenswrapper[4772]: I0127 15:28:24.701491 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwt4l\" (UniqueName: \"kubernetes.io/projected/54bbbf38-088b-4e4d-8154-569667fcf9a9-kube-api-access-vwt4l\") pod \"54bbbf38-088b-4e4d-8154-569667fcf9a9\" (UID: \"54bbbf38-088b-4e4d-8154-569667fcf9a9\") " Jan 27 15:28:24 crc kubenswrapper[4772]: I0127 15:28:24.707285 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54bbbf38-088b-4e4d-8154-569667fcf9a9-kube-api-access-vwt4l" (OuterVolumeSpecName: "kube-api-access-vwt4l") pod "54bbbf38-088b-4e4d-8154-569667fcf9a9" (UID: "54bbbf38-088b-4e4d-8154-569667fcf9a9"). InnerVolumeSpecName "kube-api-access-vwt4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:28:24 crc kubenswrapper[4772]: I0127 15:28:24.714643 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54bbbf38-088b-4e4d-8154-569667fcf9a9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "54bbbf38-088b-4e4d-8154-569667fcf9a9" (UID: "54bbbf38-088b-4e4d-8154-569667fcf9a9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:28:24 crc kubenswrapper[4772]: I0127 15:28:24.804516 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54bbbf38-088b-4e4d-8154-569667fcf9a9-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:24 crc kubenswrapper[4772]: I0127 15:28:24.804866 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwt4l\" (UniqueName: \"kubernetes.io/projected/54bbbf38-088b-4e4d-8154-569667fcf9a9-kube-api-access-vwt4l\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.310950 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-6w7p7" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.350266 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-gbrww" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.354887 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-12a3-account-create-update-mdv84" event={"ID":"08d7e14a-70d3-446e-8250-ca1047b5bc4b","Type":"ContainerDied","Data":"4ae0384de6c053e3c0b5213f6dd65e86d1254cb54c25b58aa19f1adb13fef254"} Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.354918 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5104-account-create-update-vp7x7" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.354920 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ae0384de6c053e3c0b5213f6dd65e86d1254cb54c25b58aa19f1adb13fef254" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.355889 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-12a3-account-create-update-mdv84" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.364711 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6af8-account-create-update-ltwnh" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.397265 4772 generic.go:334] "Generic (PLEG): container finished" podID="65e74d64-83f8-4964-8950-bf76816dd5fc" containerID="db4f9e747383adbadbac961f6c8e5009d6edc10dcb010e6f2eb4e5637b296b8e" exitCode=0 Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.397311 4772 generic.go:334] "Generic (PLEG): container finished" podID="65e74d64-83f8-4964-8950-bf76816dd5fc" containerID="2d2ef8b64e19f03ac931485698899fb937f36aead7d1925135401934fbecd74a" exitCode=2 Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.397326 4772 generic.go:334] "Generic (PLEG): container finished" podID="65e74d64-83f8-4964-8950-bf76816dd5fc" containerID="96c51a2493cf3f280c96b0bf8545deb92d6624f539170b50f2d2a50880a71c96" exitCode=0 Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.397400 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65e74d64-83f8-4964-8950-bf76816dd5fc","Type":"ContainerDied","Data":"db4f9e747383adbadbac961f6c8e5009d6edc10dcb010e6f2eb4e5637b296b8e"} Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.397428 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65e74d64-83f8-4964-8950-bf76816dd5fc","Type":"ContainerDied","Data":"2d2ef8b64e19f03ac931485698899fb937f36aead7d1925135401934fbecd74a"} Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.397441 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65e74d64-83f8-4964-8950-bf76816dd5fc","Type":"ContainerDied","Data":"96c51a2493cf3f280c96b0bf8545deb92d6624f539170b50f2d2a50880a71c96"} Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.410319 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9a02b617-28a7-4262-a110-f1c71763ad19","Type":"ContainerStarted","Data":"d767e789b4befb7b8caac693075691222c00bb6ae1189417345706dad41621f9"} Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.440733 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-gbrww" event={"ID":"be888039-f158-4d05-9f7d-6d01b2478b08","Type":"ContainerDied","Data":"56740a6f5142d64002a343f18ba92dbb512d9b8a106c7a40a6e41b349e43508e"} Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.441270 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56740a6f5142d64002a343f18ba92dbb512d9b8a106c7a40a6e41b349e43508e" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.441410 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-gbrww" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.442955 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtnzx\" (UniqueName: \"kubernetes.io/projected/69f24c00-a64a-4e82-a125-c0ee3fe8fa8f-kube-api-access-dtnzx\") pod \"69f24c00-a64a-4e82-a125-c0ee3fe8fa8f\" (UID: \"69f24c00-a64a-4e82-a125-c0ee3fe8fa8f\") " Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.447496 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7907cc16-7665-49d3-ad17-f9e6e0fc2f09-operator-scripts\") pod \"7907cc16-7665-49d3-ad17-f9e6e0fc2f09\" (UID: \"7907cc16-7665-49d3-ad17-f9e6e0fc2f09\") " Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.447659 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be888039-f158-4d05-9f7d-6d01b2478b08-operator-scripts\") pod \"be888039-f158-4d05-9f7d-6d01b2478b08\" (UID: \"be888039-f158-4d05-9f7d-6d01b2478b08\") " Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.447732 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv946\" (UniqueName: \"kubernetes.io/projected/7907cc16-7665-49d3-ad17-f9e6e0fc2f09-kube-api-access-hv946\") pod \"7907cc16-7665-49d3-ad17-f9e6e0fc2f09\" (UID: \"7907cc16-7665-49d3-ad17-f9e6e0fc2f09\") " Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.447809 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvqvj\" (UniqueName: \"kubernetes.io/projected/be888039-f158-4d05-9f7d-6d01b2478b08-kube-api-access-pvqvj\") pod \"be888039-f158-4d05-9f7d-6d01b2478b08\" (UID: \"be888039-f158-4d05-9f7d-6d01b2478b08\") " Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.447842 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69f24c00-a64a-4e82-a125-c0ee3fe8fa8f-operator-scripts\") pod \"69f24c00-a64a-4e82-a125-c0ee3fe8fa8f\" (UID: \"69f24c00-a64a-4e82-a125-c0ee3fe8fa8f\") " Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.451891 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be888039-f158-4d05-9f7d-6d01b2478b08-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "be888039-f158-4d05-9f7d-6d01b2478b08" (UID: "be888039-f158-4d05-9f7d-6d01b2478b08"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.452021 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69f24c00-a64a-4e82-a125-c0ee3fe8fa8f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "69f24c00-a64a-4e82-a125-c0ee3fe8fa8f" (UID: "69f24c00-a64a-4e82-a125-c0ee3fe8fa8f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.452159 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7907cc16-7665-49d3-ad17-f9e6e0fc2f09-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7907cc16-7665-49d3-ad17-f9e6e0fc2f09" (UID: "7907cc16-7665-49d3-ad17-f9e6e0fc2f09"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.461228 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69f24c00-a64a-4e82-a125-c0ee3fe8fa8f-kube-api-access-dtnzx" (OuterVolumeSpecName: "kube-api-access-dtnzx") pod "69f24c00-a64a-4e82-a125-c0ee3fe8fa8f" (UID: "69f24c00-a64a-4e82-a125-c0ee3fe8fa8f"). InnerVolumeSpecName "kube-api-access-dtnzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.463447 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7907cc16-7665-49d3-ad17-f9e6e0fc2f09-kube-api-access-hv946" (OuterVolumeSpecName: "kube-api-access-hv946") pod "7907cc16-7665-49d3-ad17-f9e6e0fc2f09" (UID: "7907cc16-7665-49d3-ad17-f9e6e0fc2f09"). InnerVolumeSpecName "kube-api-access-hv946". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.470102 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be888039-f158-4d05-9f7d-6d01b2478b08-kube-api-access-pvqvj" (OuterVolumeSpecName: "kube-api-access-pvqvj") pod "be888039-f158-4d05-9f7d-6d01b2478b08" (UID: "be888039-f158-4d05-9f7d-6d01b2478b08"). InnerVolumeSpecName "kube-api-access-pvqvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.477612 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pszgr" event={"ID":"54bbbf38-088b-4e4d-8154-569667fcf9a9","Type":"ContainerDied","Data":"303a25d43e818afc0b1deec34d71e454f848a7bcfa271b0aa6ddd404903c4f4b"} Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.477663 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="303a25d43e818afc0b1deec34d71e454f848a7bcfa271b0aa6ddd404903c4f4b" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.477674 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pszgr" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.486742 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5104-account-create-update-vp7x7" event={"ID":"69f24c00-a64a-4e82-a125-c0ee3fe8fa8f","Type":"ContainerDied","Data":"9536a6635c6de01fa9cbbeb2b3e4a3db2498f81bff722635d21115dedc7f8ce3"} Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.486790 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9536a6635c6de01fa9cbbeb2b3e4a3db2498f81bff722635d21115dedc7f8ce3" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.486906 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5104-account-create-update-vp7x7" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.515454 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343","Type":"ContainerStarted","Data":"6481b50eed7f8997cc197c4b50a1b5d1b9aa395b3745aa30ff2d6ee451d23215"} Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.519514 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-6w7p7" event={"ID":"7907cc16-7665-49d3-ad17-f9e6e0fc2f09","Type":"ContainerDied","Data":"5abfcf1bdd92bd3b6c78b0f416ccffb023fa1f4cdaf92b953a574b587cc4a6d0"} Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.519561 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5abfcf1bdd92bd3b6c78b0f416ccffb023fa1f4cdaf92b953a574b587cc4a6d0" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.519614 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-6w7p7" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.552521 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08d7e14a-70d3-446e-8250-ca1047b5bc4b-operator-scripts\") pod \"08d7e14a-70d3-446e-8250-ca1047b5bc4b\" (UID: \"08d7e14a-70d3-446e-8250-ca1047b5bc4b\") " Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.552636 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znsll\" (UniqueName: \"kubernetes.io/projected/564de425-5170-45df-9080-5b02579483ee-kube-api-access-znsll\") pod \"564de425-5170-45df-9080-5b02579483ee\" (UID: \"564de425-5170-45df-9080-5b02579483ee\") " Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.552662 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-878xd\" (UniqueName: \"kubernetes.io/projected/08d7e14a-70d3-446e-8250-ca1047b5bc4b-kube-api-access-878xd\") pod \"08d7e14a-70d3-446e-8250-ca1047b5bc4b\" (UID: \"08d7e14a-70d3-446e-8250-ca1047b5bc4b\") " Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.552732 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/564de425-5170-45df-9080-5b02579483ee-operator-scripts\") pod \"564de425-5170-45df-9080-5b02579483ee\" (UID: \"564de425-5170-45df-9080-5b02579483ee\") " Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.553374 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtnzx\" (UniqueName: \"kubernetes.io/projected/69f24c00-a64a-4e82-a125-c0ee3fe8fa8f-kube-api-access-dtnzx\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.553400 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7907cc16-7665-49d3-ad17-f9e6e0fc2f09-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.553413 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be888039-f158-4d05-9f7d-6d01b2478b08-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.553424 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv946\" (UniqueName: \"kubernetes.io/projected/7907cc16-7665-49d3-ad17-f9e6e0fc2f09-kube-api-access-hv946\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.553438 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvqvj\" (UniqueName: \"kubernetes.io/projected/be888039-f158-4d05-9f7d-6d01b2478b08-kube-api-access-pvqvj\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.553451 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69f24c00-a64a-4e82-a125-c0ee3fe8fa8f-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.553960 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08d7e14a-70d3-446e-8250-ca1047b5bc4b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "08d7e14a-70d3-446e-8250-ca1047b5bc4b" (UID: "08d7e14a-70d3-446e-8250-ca1047b5bc4b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.555975 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/564de425-5170-45df-9080-5b02579483ee-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "564de425-5170-45df-9080-5b02579483ee" (UID: "564de425-5170-45df-9080-5b02579483ee"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.561854 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/564de425-5170-45df-9080-5b02579483ee-kube-api-access-znsll" (OuterVolumeSpecName: "kube-api-access-znsll") pod "564de425-5170-45df-9080-5b02579483ee" (UID: "564de425-5170-45df-9080-5b02579483ee"). InnerVolumeSpecName "kube-api-access-znsll". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.566427 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08d7e14a-70d3-446e-8250-ca1047b5bc4b-kube-api-access-878xd" (OuterVolumeSpecName: "kube-api-access-878xd") pod "08d7e14a-70d3-446e-8250-ca1047b5bc4b" (UID: "08d7e14a-70d3-446e-8250-ca1047b5bc4b"). InnerVolumeSpecName "kube-api-access-878xd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.651338 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.651313775 podStartE2EDuration="3.651313775s" podCreationTimestamp="2026-01-27 15:28:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:28:25.550504004 +0000 UTC m=+1291.531113122" watchObservedRunningTime="2026-01-27 15:28:25.651313775 +0000 UTC m=+1291.631922873" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.655404 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08d7e14a-70d3-446e-8250-ca1047b5bc4b-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.655442 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znsll\" (UniqueName: \"kubernetes.io/projected/564de425-5170-45df-9080-5b02579483ee-kube-api-access-znsll\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.655456 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-878xd\" (UniqueName: \"kubernetes.io/projected/08d7e14a-70d3-446e-8250-ca1047b5bc4b-kube-api-access-878xd\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:25 crc kubenswrapper[4772]: I0127 15:28:25.655467 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/564de425-5170-45df-9080-5b02579483ee-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.510139 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.529134 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9a02b617-28a7-4262-a110-f1c71763ad19","Type":"ContainerStarted","Data":"3114715e24bc63a93ce31ec7ec2cc2fdeaad0a6c7647de22f23d06ac45e3d864"} Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.532354 4772 generic.go:334] "Generic (PLEG): container finished" podID="65e74d64-83f8-4964-8950-bf76816dd5fc" containerID="902d06fc4d38a39dd40f6481a4374dc4c1e7ef0957c3355454b84d0778d1bc2b" exitCode=0 Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.532444 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65e74d64-83f8-4964-8950-bf76816dd5fc","Type":"ContainerDied","Data":"902d06fc4d38a39dd40f6481a4374dc4c1e7ef0957c3355454b84d0778d1bc2b"} Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.532459 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.532487 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65e74d64-83f8-4964-8950-bf76816dd5fc","Type":"ContainerDied","Data":"efec7f44770a0bed43e7dc53d45c9b414c83392d3729ddafeb636db96612decb"} Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.532531 4772 scope.go:117] "RemoveContainer" containerID="db4f9e747383adbadbac961f6c8e5009d6edc10dcb010e6f2eb4e5637b296b8e" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.537057 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-12a3-account-create-update-mdv84" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.540733 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-6af8-account-create-update-ltwnh" event={"ID":"564de425-5170-45df-9080-5b02579483ee","Type":"ContainerDied","Data":"deaf013f2c9b59ae86d1e302aff4c0c47992ea99da1873382d3b8f4f077d78da"} Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.540772 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="deaf013f2c9b59ae86d1e302aff4c0c47992ea99da1873382d3b8f4f077d78da" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.540861 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6af8-account-create-update-ltwnh" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.568295 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.568277939 podStartE2EDuration="4.568277939s" podCreationTimestamp="2026-01-27 15:28:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:28:26.558192145 +0000 UTC m=+1292.538801243" watchObservedRunningTime="2026-01-27 15:28:26.568277939 +0000 UTC m=+1292.548887037" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.611460 4772 scope.go:117] "RemoveContainer" containerID="2d2ef8b64e19f03ac931485698899fb937f36aead7d1925135401934fbecd74a" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.639630 4772 scope.go:117] "RemoveContainer" containerID="902d06fc4d38a39dd40f6481a4374dc4c1e7ef0957c3355454b84d0778d1bc2b" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.665404 4772 scope.go:117] "RemoveContainer" containerID="96c51a2493cf3f280c96b0bf8545deb92d6624f539170b50f2d2a50880a71c96" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.687338 4772 scope.go:117] "RemoveContainer" containerID="db4f9e747383adbadbac961f6c8e5009d6edc10dcb010e6f2eb4e5637b296b8e" Jan 27 15:28:26 crc kubenswrapper[4772]: E0127 15:28:26.688677 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db4f9e747383adbadbac961f6c8e5009d6edc10dcb010e6f2eb4e5637b296b8e\": container with ID starting with db4f9e747383adbadbac961f6c8e5009d6edc10dcb010e6f2eb4e5637b296b8e not found: ID does not exist" containerID="db4f9e747383adbadbac961f6c8e5009d6edc10dcb010e6f2eb4e5637b296b8e" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.688730 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db4f9e747383adbadbac961f6c8e5009d6edc10dcb010e6f2eb4e5637b296b8e"} err="failed to get container status \"db4f9e747383adbadbac961f6c8e5009d6edc10dcb010e6f2eb4e5637b296b8e\": rpc error: code = NotFound desc = could not find container \"db4f9e747383adbadbac961f6c8e5009d6edc10dcb010e6f2eb4e5637b296b8e\": container with ID starting with db4f9e747383adbadbac961f6c8e5009d6edc10dcb010e6f2eb4e5637b296b8e not found: ID does not exist" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.688761 4772 scope.go:117] "RemoveContainer" containerID="2d2ef8b64e19f03ac931485698899fb937f36aead7d1925135401934fbecd74a" Jan 27 15:28:26 crc kubenswrapper[4772]: E0127 15:28:26.689375 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d2ef8b64e19f03ac931485698899fb937f36aead7d1925135401934fbecd74a\": container with ID starting with 2d2ef8b64e19f03ac931485698899fb937f36aead7d1925135401934fbecd74a not found: ID does not exist" containerID="2d2ef8b64e19f03ac931485698899fb937f36aead7d1925135401934fbecd74a" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.689406 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d2ef8b64e19f03ac931485698899fb937f36aead7d1925135401934fbecd74a"} err="failed to get container status \"2d2ef8b64e19f03ac931485698899fb937f36aead7d1925135401934fbecd74a\": rpc error: code = NotFound desc = could not find container \"2d2ef8b64e19f03ac931485698899fb937f36aead7d1925135401934fbecd74a\": container with ID starting with 2d2ef8b64e19f03ac931485698899fb937f36aead7d1925135401934fbecd74a not found: ID does not exist" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.689428 4772 scope.go:117] "RemoveContainer" containerID="902d06fc4d38a39dd40f6481a4374dc4c1e7ef0957c3355454b84d0778d1bc2b" Jan 27 15:28:26 crc kubenswrapper[4772]: E0127 15:28:26.689661 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"902d06fc4d38a39dd40f6481a4374dc4c1e7ef0957c3355454b84d0778d1bc2b\": container with ID starting with 902d06fc4d38a39dd40f6481a4374dc4c1e7ef0957c3355454b84d0778d1bc2b not found: ID does not exist" containerID="902d06fc4d38a39dd40f6481a4374dc4c1e7ef0957c3355454b84d0778d1bc2b" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.689686 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"902d06fc4d38a39dd40f6481a4374dc4c1e7ef0957c3355454b84d0778d1bc2b"} err="failed to get container status \"902d06fc4d38a39dd40f6481a4374dc4c1e7ef0957c3355454b84d0778d1bc2b\": rpc error: code = NotFound desc = could not find container \"902d06fc4d38a39dd40f6481a4374dc4c1e7ef0957c3355454b84d0778d1bc2b\": container with ID starting with 902d06fc4d38a39dd40f6481a4374dc4c1e7ef0957c3355454b84d0778d1bc2b not found: ID does not exist" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.689699 4772 scope.go:117] "RemoveContainer" containerID="96c51a2493cf3f280c96b0bf8545deb92d6624f539170b50f2d2a50880a71c96" Jan 27 15:28:26 crc kubenswrapper[4772]: E0127 15:28:26.690733 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96c51a2493cf3f280c96b0bf8545deb92d6624f539170b50f2d2a50880a71c96\": container with ID starting with 96c51a2493cf3f280c96b0bf8545deb92d6624f539170b50f2d2a50880a71c96 not found: ID does not exist" containerID="96c51a2493cf3f280c96b0bf8545deb92d6624f539170b50f2d2a50880a71c96" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.690754 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96c51a2493cf3f280c96b0bf8545deb92d6624f539170b50f2d2a50880a71c96"} err="failed to get container status \"96c51a2493cf3f280c96b0bf8545deb92d6624f539170b50f2d2a50880a71c96\": rpc error: code = NotFound desc = could not find container \"96c51a2493cf3f280c96b0bf8545deb92d6624f539170b50f2d2a50880a71c96\": container with ID starting with 96c51a2493cf3f280c96b0bf8545deb92d6624f539170b50f2d2a50880a71c96 not found: ID does not exist" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.702461 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65e74d64-83f8-4964-8950-bf76816dd5fc-sg-core-conf-yaml\") pod \"65e74d64-83f8-4964-8950-bf76816dd5fc\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.702573 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65e74d64-83f8-4964-8950-bf76816dd5fc-config-data\") pod \"65e74d64-83f8-4964-8950-bf76816dd5fc\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.702640 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65e74d64-83f8-4964-8950-bf76816dd5fc-log-httpd\") pod \"65e74d64-83f8-4964-8950-bf76816dd5fc\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.702750 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65e74d64-83f8-4964-8950-bf76816dd5fc-run-httpd\") pod \"65e74d64-83f8-4964-8950-bf76816dd5fc\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.702772 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhrl9\" (UniqueName: \"kubernetes.io/projected/65e74d64-83f8-4964-8950-bf76816dd5fc-kube-api-access-nhrl9\") pod \"65e74d64-83f8-4964-8950-bf76816dd5fc\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.702820 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65e74d64-83f8-4964-8950-bf76816dd5fc-combined-ca-bundle\") pod \"65e74d64-83f8-4964-8950-bf76816dd5fc\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.702850 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65e74d64-83f8-4964-8950-bf76816dd5fc-scripts\") pod \"65e74d64-83f8-4964-8950-bf76816dd5fc\" (UID: \"65e74d64-83f8-4964-8950-bf76816dd5fc\") " Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.705516 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65e74d64-83f8-4964-8950-bf76816dd5fc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "65e74d64-83f8-4964-8950-bf76816dd5fc" (UID: "65e74d64-83f8-4964-8950-bf76816dd5fc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.706029 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65e74d64-83f8-4964-8950-bf76816dd5fc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "65e74d64-83f8-4964-8950-bf76816dd5fc" (UID: "65e74d64-83f8-4964-8950-bf76816dd5fc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.712253 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65e74d64-83f8-4964-8950-bf76816dd5fc-scripts" (OuterVolumeSpecName: "scripts") pod "65e74d64-83f8-4964-8950-bf76816dd5fc" (UID: "65e74d64-83f8-4964-8950-bf76816dd5fc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.727494 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65e74d64-83f8-4964-8950-bf76816dd5fc-kube-api-access-nhrl9" (OuterVolumeSpecName: "kube-api-access-nhrl9") pod "65e74d64-83f8-4964-8950-bf76816dd5fc" (UID: "65e74d64-83f8-4964-8950-bf76816dd5fc"). InnerVolumeSpecName "kube-api-access-nhrl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.756406 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65e74d64-83f8-4964-8950-bf76816dd5fc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "65e74d64-83f8-4964-8950-bf76816dd5fc" (UID: "65e74d64-83f8-4964-8950-bf76816dd5fc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.785357 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65e74d64-83f8-4964-8950-bf76816dd5fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65e74d64-83f8-4964-8950-bf76816dd5fc" (UID: "65e74d64-83f8-4964-8950-bf76816dd5fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.804960 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65e74d64-83f8-4964-8950-bf76816dd5fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.805115 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65e74d64-83f8-4964-8950-bf76816dd5fc-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.805237 4772 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65e74d64-83f8-4964-8950-bf76816dd5fc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.805322 4772 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65e74d64-83f8-4964-8950-bf76816dd5fc-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.805411 4772 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65e74d64-83f8-4964-8950-bf76816dd5fc-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.805483 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhrl9\" (UniqueName: \"kubernetes.io/projected/65e74d64-83f8-4964-8950-bf76816dd5fc-kube-api-access-nhrl9\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.821158 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65e74d64-83f8-4964-8950-bf76816dd5fc-config-data" (OuterVolumeSpecName: "config-data") pod "65e74d64-83f8-4964-8950-bf76816dd5fc" (UID: "65e74d64-83f8-4964-8950-bf76816dd5fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.866797 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.876577 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.895689 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:26 crc kubenswrapper[4772]: E0127 15:28:26.896057 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65e74d64-83f8-4964-8950-bf76816dd5fc" containerName="ceilometer-central-agent" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.896074 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="65e74d64-83f8-4964-8950-bf76816dd5fc" containerName="ceilometer-central-agent" Jan 27 15:28:26 crc kubenswrapper[4772]: E0127 15:28:26.896088 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="564de425-5170-45df-9080-5b02579483ee" containerName="mariadb-account-create-update" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.896094 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="564de425-5170-45df-9080-5b02579483ee" containerName="mariadb-account-create-update" Jan 27 15:28:26 crc kubenswrapper[4772]: E0127 15:28:26.896107 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69f24c00-a64a-4e82-a125-c0ee3fe8fa8f" containerName="mariadb-account-create-update" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.896113 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="69f24c00-a64a-4e82-a125-c0ee3fe8fa8f" containerName="mariadb-account-create-update" Jan 27 15:28:26 crc kubenswrapper[4772]: E0127 15:28:26.896121 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7907cc16-7665-49d3-ad17-f9e6e0fc2f09" containerName="mariadb-database-create" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.896127 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="7907cc16-7665-49d3-ad17-f9e6e0fc2f09" containerName="mariadb-database-create" Jan 27 15:28:26 crc kubenswrapper[4772]: E0127 15:28:26.896142 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65e74d64-83f8-4964-8950-bf76816dd5fc" containerName="proxy-httpd" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.896148 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="65e74d64-83f8-4964-8950-bf76816dd5fc" containerName="proxy-httpd" Jan 27 15:28:26 crc kubenswrapper[4772]: E0127 15:28:26.896157 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be888039-f158-4d05-9f7d-6d01b2478b08" containerName="mariadb-database-create" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.896200 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="be888039-f158-4d05-9f7d-6d01b2478b08" containerName="mariadb-database-create" Jan 27 15:28:26 crc kubenswrapper[4772]: E0127 15:28:26.896218 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54bbbf38-088b-4e4d-8154-569667fcf9a9" containerName="mariadb-database-create" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.896225 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="54bbbf38-088b-4e4d-8154-569667fcf9a9" containerName="mariadb-database-create" Jan 27 15:28:26 crc kubenswrapper[4772]: E0127 15:28:26.896240 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08d7e14a-70d3-446e-8250-ca1047b5bc4b" containerName="mariadb-account-create-update" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.896247 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="08d7e14a-70d3-446e-8250-ca1047b5bc4b" containerName="mariadb-account-create-update" Jan 27 15:28:26 crc kubenswrapper[4772]: E0127 15:28:26.896258 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65e74d64-83f8-4964-8950-bf76816dd5fc" containerName="sg-core" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.896265 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="65e74d64-83f8-4964-8950-bf76816dd5fc" containerName="sg-core" Jan 27 15:28:26 crc kubenswrapper[4772]: E0127 15:28:26.896281 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65e74d64-83f8-4964-8950-bf76816dd5fc" containerName="ceilometer-notification-agent" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.896288 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="65e74d64-83f8-4964-8950-bf76816dd5fc" containerName="ceilometer-notification-agent" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.896466 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="564de425-5170-45df-9080-5b02579483ee" containerName="mariadb-account-create-update" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.896485 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="be888039-f158-4d05-9f7d-6d01b2478b08" containerName="mariadb-database-create" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.896494 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="65e74d64-83f8-4964-8950-bf76816dd5fc" containerName="ceilometer-central-agent" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.896503 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="54bbbf38-088b-4e4d-8154-569667fcf9a9" containerName="mariadb-database-create" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.896513 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="65e74d64-83f8-4964-8950-bf76816dd5fc" containerName="proxy-httpd" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.896520 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="65e74d64-83f8-4964-8950-bf76816dd5fc" containerName="ceilometer-notification-agent" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.896529 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="69f24c00-a64a-4e82-a125-c0ee3fe8fa8f" containerName="mariadb-account-create-update" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.896540 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="7907cc16-7665-49d3-ad17-f9e6e0fc2f09" containerName="mariadb-database-create" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.896551 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="08d7e14a-70d3-446e-8250-ca1047b5bc4b" containerName="mariadb-account-create-update" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.896562 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="65e74d64-83f8-4964-8950-bf76816dd5fc" containerName="sg-core" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.898091 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.905493 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.905672 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.907649 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65e74d64-83f8-4964-8950-bf76816dd5fc-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:26 crc kubenswrapper[4772]: I0127 15:28:26.913793 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:27 crc kubenswrapper[4772]: I0127 15:28:27.008775 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de3b39d5-b15f-46c0-881a-e747e07e76a5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " pod="openstack/ceilometer-0" Jan 27 15:28:27 crc kubenswrapper[4772]: I0127 15:28:27.008824 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de3b39d5-b15f-46c0-881a-e747e07e76a5-run-httpd\") pod \"ceilometer-0\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " pod="openstack/ceilometer-0" Jan 27 15:28:27 crc kubenswrapper[4772]: I0127 15:28:27.008871 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de3b39d5-b15f-46c0-881a-e747e07e76a5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " pod="openstack/ceilometer-0" Jan 27 15:28:27 crc kubenswrapper[4772]: I0127 15:28:27.008985 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de3b39d5-b15f-46c0-881a-e747e07e76a5-scripts\") pod \"ceilometer-0\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " pod="openstack/ceilometer-0" Jan 27 15:28:27 crc kubenswrapper[4772]: I0127 15:28:27.009019 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de3b39d5-b15f-46c0-881a-e747e07e76a5-config-data\") pod \"ceilometer-0\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " pod="openstack/ceilometer-0" Jan 27 15:28:27 crc kubenswrapper[4772]: I0127 15:28:27.009159 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxklb\" (UniqueName: \"kubernetes.io/projected/de3b39d5-b15f-46c0-881a-e747e07e76a5-kube-api-access-xxklb\") pod \"ceilometer-0\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " pod="openstack/ceilometer-0" Jan 27 15:28:27 crc kubenswrapper[4772]: I0127 15:28:27.009490 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de3b39d5-b15f-46c0-881a-e747e07e76a5-log-httpd\") pod \"ceilometer-0\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " pod="openstack/ceilometer-0" Jan 27 15:28:27 crc kubenswrapper[4772]: I0127 15:28:27.110899 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de3b39d5-b15f-46c0-881a-e747e07e76a5-log-httpd\") pod \"ceilometer-0\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " pod="openstack/ceilometer-0" Jan 27 15:28:27 crc kubenswrapper[4772]: I0127 15:28:27.110946 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de3b39d5-b15f-46c0-881a-e747e07e76a5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " pod="openstack/ceilometer-0" Jan 27 15:28:27 crc kubenswrapper[4772]: I0127 15:28:27.110969 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de3b39d5-b15f-46c0-881a-e747e07e76a5-run-httpd\") pod \"ceilometer-0\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " pod="openstack/ceilometer-0" Jan 27 15:28:27 crc kubenswrapper[4772]: I0127 15:28:27.111006 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de3b39d5-b15f-46c0-881a-e747e07e76a5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " pod="openstack/ceilometer-0" Jan 27 15:28:27 crc kubenswrapper[4772]: I0127 15:28:27.111035 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de3b39d5-b15f-46c0-881a-e747e07e76a5-scripts\") pod \"ceilometer-0\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " pod="openstack/ceilometer-0" Jan 27 15:28:27 crc kubenswrapper[4772]: I0127 15:28:27.111483 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de3b39d5-b15f-46c0-881a-e747e07e76a5-log-httpd\") pod \"ceilometer-0\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " pod="openstack/ceilometer-0" Jan 27 15:28:27 crc kubenswrapper[4772]: I0127 15:28:27.111700 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de3b39d5-b15f-46c0-881a-e747e07e76a5-run-httpd\") pod \"ceilometer-0\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " pod="openstack/ceilometer-0" Jan 27 15:28:27 crc kubenswrapper[4772]: I0127 15:28:27.111752 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de3b39d5-b15f-46c0-881a-e747e07e76a5-config-data\") pod \"ceilometer-0\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " pod="openstack/ceilometer-0" Jan 27 15:28:27 crc kubenswrapper[4772]: I0127 15:28:27.111788 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxklb\" (UniqueName: \"kubernetes.io/projected/de3b39d5-b15f-46c0-881a-e747e07e76a5-kube-api-access-xxklb\") pod \"ceilometer-0\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " pod="openstack/ceilometer-0" Jan 27 15:28:27 crc kubenswrapper[4772]: I0127 15:28:27.117363 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de3b39d5-b15f-46c0-881a-e747e07e76a5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " pod="openstack/ceilometer-0" Jan 27 15:28:27 crc kubenswrapper[4772]: I0127 15:28:27.118650 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de3b39d5-b15f-46c0-881a-e747e07e76a5-scripts\") pod \"ceilometer-0\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " pod="openstack/ceilometer-0" Jan 27 15:28:27 crc kubenswrapper[4772]: I0127 15:28:27.119468 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de3b39d5-b15f-46c0-881a-e747e07e76a5-config-data\") pod \"ceilometer-0\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " pod="openstack/ceilometer-0" Jan 27 15:28:27 crc kubenswrapper[4772]: I0127 15:28:27.120039 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de3b39d5-b15f-46c0-881a-e747e07e76a5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " pod="openstack/ceilometer-0" Jan 27 15:28:27 crc kubenswrapper[4772]: I0127 15:28:27.138314 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxklb\" (UniqueName: \"kubernetes.io/projected/de3b39d5-b15f-46c0-881a-e747e07e76a5-kube-api-access-xxklb\") pod \"ceilometer-0\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " pod="openstack/ceilometer-0" Jan 27 15:28:27 crc kubenswrapper[4772]: I0127 15:28:27.258850 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:28:27 crc kubenswrapper[4772]: I0127 15:28:27.754978 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:27 crc kubenswrapper[4772]: W0127 15:28:27.757713 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde3b39d5_b15f_46c0_881a_e747e07e76a5.slice/crio-fe477ed2e6bff622d3bb919bb2a03b5372b4eb037875880ee816351cacaad478 WatchSource:0}: Error finding container fe477ed2e6bff622d3bb919bb2a03b5372b4eb037875880ee816351cacaad478: Status 404 returned error can't find the container with id fe477ed2e6bff622d3bb919bb2a03b5372b4eb037875880ee816351cacaad478 Jan 27 15:28:28 crc kubenswrapper[4772]: I0127 15:28:28.576031 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de3b39d5-b15f-46c0-881a-e747e07e76a5","Type":"ContainerStarted","Data":"4c6da56f01306accbad60e3ba02a91f4cc6ed8bb905bd9286671fd7f32153ed5"} Jan 27 15:28:28 crc kubenswrapper[4772]: I0127 15:28:28.576403 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de3b39d5-b15f-46c0-881a-e747e07e76a5","Type":"ContainerStarted","Data":"fe477ed2e6bff622d3bb919bb2a03b5372b4eb037875880ee816351cacaad478"} Jan 27 15:28:28 crc kubenswrapper[4772]: I0127 15:28:28.677225 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65e74d64-83f8-4964-8950-bf76816dd5fc" path="/var/lib/kubelet/pods/65e74d64-83f8-4964-8950-bf76816dd5fc/volumes" Jan 27 15:28:28 crc kubenswrapper[4772]: I0127 15:28:28.764713 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-66bf894476-wz7b5" Jan 27 15:28:29 crc kubenswrapper[4772]: I0127 15:28:29.589441 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de3b39d5-b15f-46c0-881a-e747e07e76a5","Type":"ContainerStarted","Data":"48911a4a107b6bf266b45bdb20df360ce0efcf35791daa4bc1413cb966d28fb0"} Jan 27 15:28:30 crc kubenswrapper[4772]: I0127 15:28:30.126966 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:30 crc kubenswrapper[4772]: I0127 15:28:30.394299 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-v9mqp"] Jan 27 15:28:30 crc kubenswrapper[4772]: I0127 15:28:30.395514 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-v9mqp" Jan 27 15:28:30 crc kubenswrapper[4772]: I0127 15:28:30.400930 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Jan 27 15:28:30 crc kubenswrapper[4772]: I0127 15:28:30.401324 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Jan 27 15:28:30 crc kubenswrapper[4772]: I0127 15:28:30.401463 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bq7vb" Jan 27 15:28:30 crc kubenswrapper[4772]: I0127 15:28:30.405144 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-v9mqp"] Jan 27 15:28:30 crc kubenswrapper[4772]: I0127 15:28:30.480519 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pz7jw\" (UniqueName: \"kubernetes.io/projected/fe34fbf1-61c4-46a9-9954-64ed431d2cb7-kube-api-access-pz7jw\") pod \"nova-cell0-conductor-db-sync-v9mqp\" (UID: \"fe34fbf1-61c4-46a9-9954-64ed431d2cb7\") " pod="openstack/nova-cell0-conductor-db-sync-v9mqp" Jan 27 15:28:30 crc kubenswrapper[4772]: I0127 15:28:30.480602 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe34fbf1-61c4-46a9-9954-64ed431d2cb7-scripts\") pod \"nova-cell0-conductor-db-sync-v9mqp\" (UID: \"fe34fbf1-61c4-46a9-9954-64ed431d2cb7\") " pod="openstack/nova-cell0-conductor-db-sync-v9mqp" Jan 27 15:28:30 crc kubenswrapper[4772]: I0127 15:28:30.480778 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe34fbf1-61c4-46a9-9954-64ed431d2cb7-config-data\") pod \"nova-cell0-conductor-db-sync-v9mqp\" (UID: \"fe34fbf1-61c4-46a9-9954-64ed431d2cb7\") " pod="openstack/nova-cell0-conductor-db-sync-v9mqp" Jan 27 15:28:30 crc kubenswrapper[4772]: I0127 15:28:30.480830 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe34fbf1-61c4-46a9-9954-64ed431d2cb7-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-v9mqp\" (UID: \"fe34fbf1-61c4-46a9-9954-64ed431d2cb7\") " pod="openstack/nova-cell0-conductor-db-sync-v9mqp" Jan 27 15:28:30 crc kubenswrapper[4772]: I0127 15:28:30.583043 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe34fbf1-61c4-46a9-9954-64ed431d2cb7-config-data\") pod \"nova-cell0-conductor-db-sync-v9mqp\" (UID: \"fe34fbf1-61c4-46a9-9954-64ed431d2cb7\") " pod="openstack/nova-cell0-conductor-db-sync-v9mqp" Jan 27 15:28:30 crc kubenswrapper[4772]: I0127 15:28:30.583414 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe34fbf1-61c4-46a9-9954-64ed431d2cb7-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-v9mqp\" (UID: \"fe34fbf1-61c4-46a9-9954-64ed431d2cb7\") " pod="openstack/nova-cell0-conductor-db-sync-v9mqp" Jan 27 15:28:30 crc kubenswrapper[4772]: I0127 15:28:30.583596 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz7jw\" (UniqueName: \"kubernetes.io/projected/fe34fbf1-61c4-46a9-9954-64ed431d2cb7-kube-api-access-pz7jw\") pod \"nova-cell0-conductor-db-sync-v9mqp\" (UID: \"fe34fbf1-61c4-46a9-9954-64ed431d2cb7\") " pod="openstack/nova-cell0-conductor-db-sync-v9mqp" Jan 27 15:28:30 crc kubenswrapper[4772]: I0127 15:28:30.584059 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe34fbf1-61c4-46a9-9954-64ed431d2cb7-scripts\") pod \"nova-cell0-conductor-db-sync-v9mqp\" (UID: \"fe34fbf1-61c4-46a9-9954-64ed431d2cb7\") " pod="openstack/nova-cell0-conductor-db-sync-v9mqp" Jan 27 15:28:30 crc kubenswrapper[4772]: I0127 15:28:30.588704 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe34fbf1-61c4-46a9-9954-64ed431d2cb7-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-v9mqp\" (UID: \"fe34fbf1-61c4-46a9-9954-64ed431d2cb7\") " pod="openstack/nova-cell0-conductor-db-sync-v9mqp" Jan 27 15:28:30 crc kubenswrapper[4772]: I0127 15:28:30.595691 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe34fbf1-61c4-46a9-9954-64ed431d2cb7-scripts\") pod \"nova-cell0-conductor-db-sync-v9mqp\" (UID: \"fe34fbf1-61c4-46a9-9954-64ed431d2cb7\") " pod="openstack/nova-cell0-conductor-db-sync-v9mqp" Jan 27 15:28:30 crc kubenswrapper[4772]: I0127 15:28:30.596316 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe34fbf1-61c4-46a9-9954-64ed431d2cb7-config-data\") pod \"nova-cell0-conductor-db-sync-v9mqp\" (UID: \"fe34fbf1-61c4-46a9-9954-64ed431d2cb7\") " pod="openstack/nova-cell0-conductor-db-sync-v9mqp" Jan 27 15:28:30 crc kubenswrapper[4772]: I0127 15:28:30.600678 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de3b39d5-b15f-46c0-881a-e747e07e76a5","Type":"ContainerStarted","Data":"9627fca4ce2bbd20c54de88fa2250d98bc1976636644d325a8225826fd2e9ef2"} Jan 27 15:28:30 crc kubenswrapper[4772]: I0127 15:28:30.603414 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pz7jw\" (UniqueName: \"kubernetes.io/projected/fe34fbf1-61c4-46a9-9954-64ed431d2cb7-kube-api-access-pz7jw\") pod \"nova-cell0-conductor-db-sync-v9mqp\" (UID: \"fe34fbf1-61c4-46a9-9954-64ed431d2cb7\") " pod="openstack/nova-cell0-conductor-db-sync-v9mqp" Jan 27 15:28:30 crc kubenswrapper[4772]: I0127 15:28:30.717492 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-v9mqp" Jan 27 15:28:31 crc kubenswrapper[4772]: I0127 15:28:31.277704 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-v9mqp"] Jan 27 15:28:31 crc kubenswrapper[4772]: W0127 15:28:31.279155 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe34fbf1_61c4_46a9_9954_64ed431d2cb7.slice/crio-e097abc13b5b9498583b3db2e2e89e5740a2e06f350bc375a47375cb723458af WatchSource:0}: Error finding container e097abc13b5b9498583b3db2e2e89e5740a2e06f350bc375a47375cb723458af: Status 404 returned error can't find the container with id e097abc13b5b9498583b3db2e2e89e5740a2e06f350bc375a47375cb723458af Jan 27 15:28:31 crc kubenswrapper[4772]: I0127 15:28:31.613908 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de3b39d5-b15f-46c0-881a-e747e07e76a5","Type":"ContainerStarted","Data":"dce84557790ce392eba68b822eea435ede1d05fd9a392c9bd393123a9c7bf467"} Jan 27 15:28:31 crc kubenswrapper[4772]: I0127 15:28:31.614017 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de3b39d5-b15f-46c0-881a-e747e07e76a5" containerName="ceilometer-central-agent" containerID="cri-o://4c6da56f01306accbad60e3ba02a91f4cc6ed8bb905bd9286671fd7f32153ed5" gracePeriod=30 Jan 27 15:28:31 crc kubenswrapper[4772]: I0127 15:28:31.614041 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 27 15:28:31 crc kubenswrapper[4772]: I0127 15:28:31.614079 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de3b39d5-b15f-46c0-881a-e747e07e76a5" containerName="proxy-httpd" containerID="cri-o://dce84557790ce392eba68b822eea435ede1d05fd9a392c9bd393123a9c7bf467" gracePeriod=30 Jan 27 15:28:31 crc kubenswrapper[4772]: I0127 15:28:31.614105 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de3b39d5-b15f-46c0-881a-e747e07e76a5" containerName="sg-core" containerID="cri-o://9627fca4ce2bbd20c54de88fa2250d98bc1976636644d325a8225826fd2e9ef2" gracePeriod=30 Jan 27 15:28:31 crc kubenswrapper[4772]: I0127 15:28:31.614115 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de3b39d5-b15f-46c0-881a-e747e07e76a5" containerName="ceilometer-notification-agent" containerID="cri-o://48911a4a107b6bf266b45bdb20df360ce0efcf35791daa4bc1413cb966d28fb0" gracePeriod=30 Jan 27 15:28:31 crc kubenswrapper[4772]: I0127 15:28:31.617335 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-v9mqp" event={"ID":"fe34fbf1-61c4-46a9-9954-64ed431d2cb7","Type":"ContainerStarted","Data":"e097abc13b5b9498583b3db2e2e89e5740a2e06f350bc375a47375cb723458af"} Jan 27 15:28:31 crc kubenswrapper[4772]: I0127 15:28:31.644885 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.392901135 podStartE2EDuration="5.644866735s" podCreationTimestamp="2026-01-27 15:28:26 +0000 UTC" firstStartedPulling="2026-01-27 15:28:27.760403653 +0000 UTC m=+1293.741012751" lastFinishedPulling="2026-01-27 15:28:31.012369253 +0000 UTC m=+1296.992978351" observedRunningTime="2026-01-27 15:28:31.636758109 +0000 UTC m=+1297.617367207" watchObservedRunningTime="2026-01-27 15:28:31.644866735 +0000 UTC m=+1297.625475833" Jan 27 15:28:32 crc kubenswrapper[4772]: I0127 15:28:32.104263 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:28:32 crc kubenswrapper[4772]: I0127 15:28:32.168645 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-66bf894476-wz7b5"] Jan 27 15:28:32 crc kubenswrapper[4772]: I0127 15:28:32.168868 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-66bf894476-wz7b5" podUID="e7385520-8ffb-40e5-802e-ff0db348c5c1" containerName="neutron-api" containerID="cri-o://8859f4bb50887ba9951c0e2249a3e56deff79409c3a080683519e71c92360a6d" gracePeriod=30 Jan 27 15:28:32 crc kubenswrapper[4772]: I0127 15:28:32.169406 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-66bf894476-wz7b5" podUID="e7385520-8ffb-40e5-802e-ff0db348c5c1" containerName="neutron-httpd" containerID="cri-o://b5a8f7019a8ae14ffdea4c25f43d7ff45e4469316acbf03b2364b347f5933e7c" gracePeriod=30 Jan 27 15:28:32 crc kubenswrapper[4772]: I0127 15:28:32.601781 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 27 15:28:32 crc kubenswrapper[4772]: I0127 15:28:32.601865 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 27 15:28:32 crc kubenswrapper[4772]: I0127 15:28:32.644145 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 27 15:28:32 crc kubenswrapper[4772]: I0127 15:28:32.645561 4772 generic.go:334] "Generic (PLEG): container finished" podID="e7385520-8ffb-40e5-802e-ff0db348c5c1" containerID="b5a8f7019a8ae14ffdea4c25f43d7ff45e4469316acbf03b2364b347f5933e7c" exitCode=0 Jan 27 15:28:32 crc kubenswrapper[4772]: I0127 15:28:32.646030 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66bf894476-wz7b5" event={"ID":"e7385520-8ffb-40e5-802e-ff0db348c5c1","Type":"ContainerDied","Data":"b5a8f7019a8ae14ffdea4c25f43d7ff45e4469316acbf03b2364b347f5933e7c"} Jan 27 15:28:32 crc kubenswrapper[4772]: I0127 15:28:32.653405 4772 generic.go:334] "Generic (PLEG): container finished" podID="de3b39d5-b15f-46c0-881a-e747e07e76a5" containerID="dce84557790ce392eba68b822eea435ede1d05fd9a392c9bd393123a9c7bf467" exitCode=0 Jan 27 15:28:32 crc kubenswrapper[4772]: I0127 15:28:32.653446 4772 generic.go:334] "Generic (PLEG): container finished" podID="de3b39d5-b15f-46c0-881a-e747e07e76a5" containerID="9627fca4ce2bbd20c54de88fa2250d98bc1976636644d325a8225826fd2e9ef2" exitCode=2 Jan 27 15:28:32 crc kubenswrapper[4772]: I0127 15:28:32.653457 4772 generic.go:334] "Generic (PLEG): container finished" podID="de3b39d5-b15f-46c0-881a-e747e07e76a5" containerID="48911a4a107b6bf266b45bdb20df360ce0efcf35791daa4bc1413cb966d28fb0" exitCode=0 Jan 27 15:28:32 crc kubenswrapper[4772]: I0127 15:28:32.653522 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de3b39d5-b15f-46c0-881a-e747e07e76a5","Type":"ContainerDied","Data":"dce84557790ce392eba68b822eea435ede1d05fd9a392c9bd393123a9c7bf467"} Jan 27 15:28:32 crc kubenswrapper[4772]: I0127 15:28:32.653559 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de3b39d5-b15f-46c0-881a-e747e07e76a5","Type":"ContainerDied","Data":"9627fca4ce2bbd20c54de88fa2250d98bc1976636644d325a8225826fd2e9ef2"} Jan 27 15:28:32 crc kubenswrapper[4772]: I0127 15:28:32.653574 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de3b39d5-b15f-46c0-881a-e747e07e76a5","Type":"ContainerDied","Data":"48911a4a107b6bf266b45bdb20df360ce0efcf35791daa4bc1413cb966d28fb0"} Jan 27 15:28:32 crc kubenswrapper[4772]: I0127 15:28:32.654731 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 27 15:28:32 crc kubenswrapper[4772]: I0127 15:28:32.656286 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 27 15:28:33 crc kubenswrapper[4772]: I0127 15:28:33.109620 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 27 15:28:33 crc kubenswrapper[4772]: I0127 15:28:33.109667 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 27 15:28:33 crc kubenswrapper[4772]: I0127 15:28:33.157101 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 27 15:28:33 crc kubenswrapper[4772]: I0127 15:28:33.159024 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 27 15:28:33 crc kubenswrapper[4772]: I0127 15:28:33.662060 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 27 15:28:33 crc kubenswrapper[4772]: I0127 15:28:33.662093 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 27 15:28:33 crc kubenswrapper[4772]: I0127 15:28:33.662103 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 27 15:28:34 crc kubenswrapper[4772]: I0127 15:28:34.674509 4772 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 27 15:28:34 crc kubenswrapper[4772]: I0127 15:28:34.986394 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 27 15:28:34 crc kubenswrapper[4772]: I0127 15:28:34.990061 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 27 15:28:36 crc kubenswrapper[4772]: I0127 15:28:36.209775 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 27 15:28:36 crc kubenswrapper[4772]: I0127 15:28:36.210109 4772 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 27 15:28:36 crc kubenswrapper[4772]: I0127 15:28:36.519755 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 27 15:28:36 crc kubenswrapper[4772]: I0127 15:28:36.704315 4772 generic.go:334] "Generic (PLEG): container finished" podID="e7385520-8ffb-40e5-802e-ff0db348c5c1" containerID="8859f4bb50887ba9951c0e2249a3e56deff79409c3a080683519e71c92360a6d" exitCode=0 Jan 27 15:28:36 crc kubenswrapper[4772]: I0127 15:28:36.704366 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66bf894476-wz7b5" event={"ID":"e7385520-8ffb-40e5-802e-ff0db348c5c1","Type":"ContainerDied","Data":"8859f4bb50887ba9951c0e2249a3e56deff79409c3a080683519e71c92360a6d"} Jan 27 15:28:36 crc kubenswrapper[4772]: I0127 15:28:36.708536 4772 generic.go:334] "Generic (PLEG): container finished" podID="de3b39d5-b15f-46c0-881a-e747e07e76a5" containerID="4c6da56f01306accbad60e3ba02a91f4cc6ed8bb905bd9286671fd7f32153ed5" exitCode=0 Jan 27 15:28:36 crc kubenswrapper[4772]: I0127 15:28:36.709472 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de3b39d5-b15f-46c0-881a-e747e07e76a5","Type":"ContainerDied","Data":"4c6da56f01306accbad60e3ba02a91f4cc6ed8bb905bd9286671fd7f32153ed5"} Jan 27 15:28:40 crc kubenswrapper[4772]: I0127 15:28:40.749003 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de3b39d5-b15f-46c0-881a-e747e07e76a5","Type":"ContainerDied","Data":"fe477ed2e6bff622d3bb919bb2a03b5372b4eb037875880ee816351cacaad478"} Jan 27 15:28:40 crc kubenswrapper[4772]: I0127 15:28:40.749547 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe477ed2e6bff622d3bb919bb2a03b5372b4eb037875880ee816351cacaad478" Jan 27 15:28:40 crc kubenswrapper[4772]: I0127 15:28:40.785544 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:28:40 crc kubenswrapper[4772]: I0127 15:28:40.893932 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de3b39d5-b15f-46c0-881a-e747e07e76a5-run-httpd\") pod \"de3b39d5-b15f-46c0-881a-e747e07e76a5\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " Jan 27 15:28:40 crc kubenswrapper[4772]: I0127 15:28:40.894232 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxklb\" (UniqueName: \"kubernetes.io/projected/de3b39d5-b15f-46c0-881a-e747e07e76a5-kube-api-access-xxklb\") pod \"de3b39d5-b15f-46c0-881a-e747e07e76a5\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " Jan 27 15:28:40 crc kubenswrapper[4772]: I0127 15:28:40.894268 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de3b39d5-b15f-46c0-881a-e747e07e76a5-sg-core-conf-yaml\") pod \"de3b39d5-b15f-46c0-881a-e747e07e76a5\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " Jan 27 15:28:40 crc kubenswrapper[4772]: I0127 15:28:40.894354 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de3b39d5-b15f-46c0-881a-e747e07e76a5-config-data\") pod \"de3b39d5-b15f-46c0-881a-e747e07e76a5\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " Jan 27 15:28:40 crc kubenswrapper[4772]: I0127 15:28:40.894441 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de3b39d5-b15f-46c0-881a-e747e07e76a5-log-httpd\") pod \"de3b39d5-b15f-46c0-881a-e747e07e76a5\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " Jan 27 15:28:40 crc kubenswrapper[4772]: I0127 15:28:40.894464 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de3b39d5-b15f-46c0-881a-e747e07e76a5-combined-ca-bundle\") pod \"de3b39d5-b15f-46c0-881a-e747e07e76a5\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " Jan 27 15:28:40 crc kubenswrapper[4772]: I0127 15:28:40.894535 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de3b39d5-b15f-46c0-881a-e747e07e76a5-scripts\") pod \"de3b39d5-b15f-46c0-881a-e747e07e76a5\" (UID: \"de3b39d5-b15f-46c0-881a-e747e07e76a5\") " Jan 27 15:28:40 crc kubenswrapper[4772]: I0127 15:28:40.896401 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de3b39d5-b15f-46c0-881a-e747e07e76a5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "de3b39d5-b15f-46c0-881a-e747e07e76a5" (UID: "de3b39d5-b15f-46c0-881a-e747e07e76a5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:28:40 crc kubenswrapper[4772]: I0127 15:28:40.896464 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de3b39d5-b15f-46c0-881a-e747e07e76a5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "de3b39d5-b15f-46c0-881a-e747e07e76a5" (UID: "de3b39d5-b15f-46c0-881a-e747e07e76a5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:28:40 crc kubenswrapper[4772]: I0127 15:28:40.903979 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de3b39d5-b15f-46c0-881a-e747e07e76a5-kube-api-access-xxklb" (OuterVolumeSpecName: "kube-api-access-xxklb") pod "de3b39d5-b15f-46c0-881a-e747e07e76a5" (UID: "de3b39d5-b15f-46c0-881a-e747e07e76a5"). InnerVolumeSpecName "kube-api-access-xxklb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:28:40 crc kubenswrapper[4772]: I0127 15:28:40.904750 4772 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de3b39d5-b15f-46c0-881a-e747e07e76a5-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:40 crc kubenswrapper[4772]: I0127 15:28:40.904786 4772 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de3b39d5-b15f-46c0-881a-e747e07e76a5-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:40 crc kubenswrapper[4772]: I0127 15:28:40.904801 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxklb\" (UniqueName: \"kubernetes.io/projected/de3b39d5-b15f-46c0-881a-e747e07e76a5-kube-api-access-xxklb\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:40 crc kubenswrapper[4772]: I0127 15:28:40.908055 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de3b39d5-b15f-46c0-881a-e747e07e76a5-scripts" (OuterVolumeSpecName: "scripts") pod "de3b39d5-b15f-46c0-881a-e747e07e76a5" (UID: "de3b39d5-b15f-46c0-881a-e747e07e76a5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:40 crc kubenswrapper[4772]: I0127 15:28:40.937144 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de3b39d5-b15f-46c0-881a-e747e07e76a5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "de3b39d5-b15f-46c0-881a-e747e07e76a5" (UID: "de3b39d5-b15f-46c0-881a-e747e07e76a5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:40 crc kubenswrapper[4772]: I0127 15:28:40.955728 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66bf894476-wz7b5" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.005253 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de3b39d5-b15f-46c0-881a-e747e07e76a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de3b39d5-b15f-46c0-881a-e747e07e76a5" (UID: "de3b39d5-b15f-46c0-881a-e747e07e76a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.006359 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de3b39d5-b15f-46c0-881a-e747e07e76a5-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.006378 4772 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de3b39d5-b15f-46c0-881a-e747e07e76a5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.006388 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de3b39d5-b15f-46c0-881a-e747e07e76a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.024466 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de3b39d5-b15f-46c0-881a-e747e07e76a5-config-data" (OuterVolumeSpecName: "config-data") pod "de3b39d5-b15f-46c0-881a-e747e07e76a5" (UID: "de3b39d5-b15f-46c0-881a-e747e07e76a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.107160 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e7385520-8ffb-40e5-802e-ff0db348c5c1-config\") pod \"e7385520-8ffb-40e5-802e-ff0db348c5c1\" (UID: \"e7385520-8ffb-40e5-802e-ff0db348c5c1\") " Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.107262 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bb8pb\" (UniqueName: \"kubernetes.io/projected/e7385520-8ffb-40e5-802e-ff0db348c5c1-kube-api-access-bb8pb\") pod \"e7385520-8ffb-40e5-802e-ff0db348c5c1\" (UID: \"e7385520-8ffb-40e5-802e-ff0db348c5c1\") " Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.107301 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7385520-8ffb-40e5-802e-ff0db348c5c1-combined-ca-bundle\") pod \"e7385520-8ffb-40e5-802e-ff0db348c5c1\" (UID: \"e7385520-8ffb-40e5-802e-ff0db348c5c1\") " Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.107342 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7385520-8ffb-40e5-802e-ff0db348c5c1-ovndb-tls-certs\") pod \"e7385520-8ffb-40e5-802e-ff0db348c5c1\" (UID: \"e7385520-8ffb-40e5-802e-ff0db348c5c1\") " Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.107442 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e7385520-8ffb-40e5-802e-ff0db348c5c1-httpd-config\") pod \"e7385520-8ffb-40e5-802e-ff0db348c5c1\" (UID: \"e7385520-8ffb-40e5-802e-ff0db348c5c1\") " Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.107834 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de3b39d5-b15f-46c0-881a-e747e07e76a5-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.111723 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7385520-8ffb-40e5-802e-ff0db348c5c1-kube-api-access-bb8pb" (OuterVolumeSpecName: "kube-api-access-bb8pb") pod "e7385520-8ffb-40e5-802e-ff0db348c5c1" (UID: "e7385520-8ffb-40e5-802e-ff0db348c5c1"). InnerVolumeSpecName "kube-api-access-bb8pb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.112741 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7385520-8ffb-40e5-802e-ff0db348c5c1-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "e7385520-8ffb-40e5-802e-ff0db348c5c1" (UID: "e7385520-8ffb-40e5-802e-ff0db348c5c1"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.160514 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7385520-8ffb-40e5-802e-ff0db348c5c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e7385520-8ffb-40e5-802e-ff0db348c5c1" (UID: "e7385520-8ffb-40e5-802e-ff0db348c5c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.175746 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7385520-8ffb-40e5-802e-ff0db348c5c1-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "e7385520-8ffb-40e5-802e-ff0db348c5c1" (UID: "e7385520-8ffb-40e5-802e-ff0db348c5c1"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.184038 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7385520-8ffb-40e5-802e-ff0db348c5c1-config" (OuterVolumeSpecName: "config") pod "e7385520-8ffb-40e5-802e-ff0db348c5c1" (UID: "e7385520-8ffb-40e5-802e-ff0db348c5c1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.209638 4772 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e7385520-8ffb-40e5-802e-ff0db348c5c1-httpd-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.209667 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e7385520-8ffb-40e5-802e-ff0db348c5c1-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.209678 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bb8pb\" (UniqueName: \"kubernetes.io/projected/e7385520-8ffb-40e5-802e-ff0db348c5c1-kube-api-access-bb8pb\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.209687 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7385520-8ffb-40e5-802e-ff0db348c5c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.209695 4772 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7385520-8ffb-40e5-802e-ff0db348c5c1-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.759491 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66bf894476-wz7b5" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.759489 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66bf894476-wz7b5" event={"ID":"e7385520-8ffb-40e5-802e-ff0db348c5c1","Type":"ContainerDied","Data":"d30df4d73e5cfb24af9149a2561d9917bae2965d7778272568f0bfb1966f855f"} Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.759895 4772 scope.go:117] "RemoveContainer" containerID="b5a8f7019a8ae14ffdea4c25f43d7ff45e4469316acbf03b2364b347f5933e7c" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.762512 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.764923 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-v9mqp" event={"ID":"fe34fbf1-61c4-46a9-9954-64ed431d2cb7","Type":"ContainerStarted","Data":"35964dfe2e497930630aeb0996d17bf7bbe0e9d5e7bfb1d7efca05167ac578fc"} Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.788504 4772 scope.go:117] "RemoveContainer" containerID="8859f4bb50887ba9951c0e2249a3e56deff79409c3a080683519e71c92360a6d" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.788936 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-v9mqp" podStartSLOduration=2.444191847 podStartE2EDuration="11.788913933s" podCreationTimestamp="2026-01-27 15:28:30 +0000 UTC" firstStartedPulling="2026-01-27 15:28:31.285684411 +0000 UTC m=+1297.266293509" lastFinishedPulling="2026-01-27 15:28:40.630406497 +0000 UTC m=+1306.611015595" observedRunningTime="2026-01-27 15:28:41.78019221 +0000 UTC m=+1307.760801298" watchObservedRunningTime="2026-01-27 15:28:41.788913933 +0000 UTC m=+1307.769523031" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.807254 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.817600 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.833033 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-66bf894476-wz7b5"] Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.853891 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-66bf894476-wz7b5"] Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.870614 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:41 crc kubenswrapper[4772]: E0127 15:28:41.871136 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7385520-8ffb-40e5-802e-ff0db348c5c1" containerName="neutron-httpd" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.871155 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7385520-8ffb-40e5-802e-ff0db348c5c1" containerName="neutron-httpd" Jan 27 15:28:41 crc kubenswrapper[4772]: E0127 15:28:41.871204 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de3b39d5-b15f-46c0-881a-e747e07e76a5" containerName="sg-core" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.871213 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="de3b39d5-b15f-46c0-881a-e747e07e76a5" containerName="sg-core" Jan 27 15:28:41 crc kubenswrapper[4772]: E0127 15:28:41.871240 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de3b39d5-b15f-46c0-881a-e747e07e76a5" containerName="proxy-httpd" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.871247 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="de3b39d5-b15f-46c0-881a-e747e07e76a5" containerName="proxy-httpd" Jan 27 15:28:41 crc kubenswrapper[4772]: E0127 15:28:41.871285 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de3b39d5-b15f-46c0-881a-e747e07e76a5" containerName="ceilometer-central-agent" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.871335 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="de3b39d5-b15f-46c0-881a-e747e07e76a5" containerName="ceilometer-central-agent" Jan 27 15:28:41 crc kubenswrapper[4772]: E0127 15:28:41.871358 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de3b39d5-b15f-46c0-881a-e747e07e76a5" containerName="ceilometer-notification-agent" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.871366 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="de3b39d5-b15f-46c0-881a-e747e07e76a5" containerName="ceilometer-notification-agent" Jan 27 15:28:41 crc kubenswrapper[4772]: E0127 15:28:41.871382 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7385520-8ffb-40e5-802e-ff0db348c5c1" containerName="neutron-api" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.871389 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7385520-8ffb-40e5-802e-ff0db348c5c1" containerName="neutron-api" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.871597 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7385520-8ffb-40e5-802e-ff0db348c5c1" containerName="neutron-api" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.871613 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="de3b39d5-b15f-46c0-881a-e747e07e76a5" containerName="proxy-httpd" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.871633 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7385520-8ffb-40e5-802e-ff0db348c5c1" containerName="neutron-httpd" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.871644 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="de3b39d5-b15f-46c0-881a-e747e07e76a5" containerName="ceilometer-notification-agent" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.871658 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="de3b39d5-b15f-46c0-881a-e747e07e76a5" containerName="ceilometer-central-agent" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.871676 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="de3b39d5-b15f-46c0-881a-e747e07e76a5" containerName="sg-core" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.874052 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.879078 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.881932 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 27 15:28:41 crc kubenswrapper[4772]: I0127 15:28:41.882196 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.046247 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " pod="openstack/ceilometer-0" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.046929 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-scripts\") pod \"ceilometer-0\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " pod="openstack/ceilometer-0" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.047080 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5526\" (UniqueName: \"kubernetes.io/projected/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-kube-api-access-p5526\") pod \"ceilometer-0\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " pod="openstack/ceilometer-0" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.047247 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-config-data\") pod \"ceilometer-0\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " pod="openstack/ceilometer-0" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.047359 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-run-httpd\") pod \"ceilometer-0\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " pod="openstack/ceilometer-0" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.047586 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " pod="openstack/ceilometer-0" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.047710 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-log-httpd\") pod \"ceilometer-0\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " pod="openstack/ceilometer-0" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.058627 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.058694 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.149111 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-scripts\") pod \"ceilometer-0\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " pod="openstack/ceilometer-0" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.149629 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5526\" (UniqueName: \"kubernetes.io/projected/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-kube-api-access-p5526\") pod \"ceilometer-0\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " pod="openstack/ceilometer-0" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.149687 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-config-data\") pod \"ceilometer-0\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " pod="openstack/ceilometer-0" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.149719 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-run-httpd\") pod \"ceilometer-0\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " pod="openstack/ceilometer-0" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.149784 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " pod="openstack/ceilometer-0" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.149826 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-log-httpd\") pod \"ceilometer-0\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " pod="openstack/ceilometer-0" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.149900 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " pod="openstack/ceilometer-0" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.150922 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-run-httpd\") pod \"ceilometer-0\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " pod="openstack/ceilometer-0" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.151644 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-log-httpd\") pod \"ceilometer-0\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " pod="openstack/ceilometer-0" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.156587 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " pod="openstack/ceilometer-0" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.157602 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-config-data\") pod \"ceilometer-0\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " pod="openstack/ceilometer-0" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.158724 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-scripts\") pod \"ceilometer-0\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " pod="openstack/ceilometer-0" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.160530 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " pod="openstack/ceilometer-0" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.176651 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5526\" (UniqueName: \"kubernetes.io/projected/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-kube-api-access-p5526\") pod \"ceilometer-0\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " pod="openstack/ceilometer-0" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.201619 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.685273 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de3b39d5-b15f-46c0-881a-e747e07e76a5" path="/var/lib/kubelet/pods/de3b39d5-b15f-46c0-881a-e747e07e76a5/volumes" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.687471 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7385520-8ffb-40e5-802e-ff0db348c5c1" path="/var/lib/kubelet/pods/e7385520-8ffb-40e5-802e-ff0db348c5c1/volumes" Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.758000 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.792215 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4ce266c-6d03-4c51-a8e7-2439eecdf67d","Type":"ContainerStarted","Data":"b742511aad6534f52532262eb0cfdd05571cd1513b70b4dc28a5263607304430"} Jan 27 15:28:42 crc kubenswrapper[4772]: I0127 15:28:42.966291 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:43 crc kubenswrapper[4772]: I0127 15:28:43.802355 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4ce266c-6d03-4c51-a8e7-2439eecdf67d","Type":"ContainerStarted","Data":"afcf06fa22d2533b1f1a226452ae6ceefe63b8b90f23a95e55e5536a352c31c5"} Jan 27 15:28:44 crc kubenswrapper[4772]: I0127 15:28:44.812517 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4ce266c-6d03-4c51-a8e7-2439eecdf67d","Type":"ContainerStarted","Data":"01b8e4f8a171a9643b0341141cead865a87cb972ae5d54e123bbcc5bbb627212"} Jan 27 15:28:45 crc kubenswrapper[4772]: I0127 15:28:45.849418 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4ce266c-6d03-4c51-a8e7-2439eecdf67d","Type":"ContainerStarted","Data":"aeaac2f858ccbdd7513c3cf040b3290daac6472fdc5c899f1438b9ebc94bf571"} Jan 27 15:28:47 crc kubenswrapper[4772]: I0127 15:28:47.870500 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4ce266c-6d03-4c51-a8e7-2439eecdf67d","Type":"ContainerStarted","Data":"a9e4d8d8cdfe57821f29d14d1ac46a5f3e7ed0b5e31d0de7bea6c91615349a23"} Jan 27 15:28:47 crc kubenswrapper[4772]: I0127 15:28:47.871282 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b4ce266c-6d03-4c51-a8e7-2439eecdf67d" containerName="ceilometer-central-agent" containerID="cri-o://afcf06fa22d2533b1f1a226452ae6ceefe63b8b90f23a95e55e5536a352c31c5" gracePeriod=30 Jan 27 15:28:47 crc kubenswrapper[4772]: I0127 15:28:47.871600 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 27 15:28:47 crc kubenswrapper[4772]: I0127 15:28:47.871948 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b4ce266c-6d03-4c51-a8e7-2439eecdf67d" containerName="proxy-httpd" containerID="cri-o://a9e4d8d8cdfe57821f29d14d1ac46a5f3e7ed0b5e31d0de7bea6c91615349a23" gracePeriod=30 Jan 27 15:28:47 crc kubenswrapper[4772]: I0127 15:28:47.872008 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b4ce266c-6d03-4c51-a8e7-2439eecdf67d" containerName="sg-core" containerID="cri-o://aeaac2f858ccbdd7513c3cf040b3290daac6472fdc5c899f1438b9ebc94bf571" gracePeriod=30 Jan 27 15:28:47 crc kubenswrapper[4772]: I0127 15:28:47.872051 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b4ce266c-6d03-4c51-a8e7-2439eecdf67d" containerName="ceilometer-notification-agent" containerID="cri-o://01b8e4f8a171a9643b0341141cead865a87cb972ae5d54e123bbcc5bbb627212" gracePeriod=30 Jan 27 15:28:47 crc kubenswrapper[4772]: I0127 15:28:47.906506 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.876167947 podStartE2EDuration="6.906479221s" podCreationTimestamp="2026-01-27 15:28:41 +0000 UTC" firstStartedPulling="2026-01-27 15:28:42.773758801 +0000 UTC m=+1308.754367909" lastFinishedPulling="2026-01-27 15:28:46.804070085 +0000 UTC m=+1312.784679183" observedRunningTime="2026-01-27 15:28:47.895963975 +0000 UTC m=+1313.876573083" watchObservedRunningTime="2026-01-27 15:28:47.906479221 +0000 UTC m=+1313.887088319" Jan 27 15:28:48 crc kubenswrapper[4772]: I0127 15:28:48.882920 4772 generic.go:334] "Generic (PLEG): container finished" podID="b4ce266c-6d03-4c51-a8e7-2439eecdf67d" containerID="a9e4d8d8cdfe57821f29d14d1ac46a5f3e7ed0b5e31d0de7bea6c91615349a23" exitCode=0 Jan 27 15:28:48 crc kubenswrapper[4772]: I0127 15:28:48.882961 4772 generic.go:334] "Generic (PLEG): container finished" podID="b4ce266c-6d03-4c51-a8e7-2439eecdf67d" containerID="aeaac2f858ccbdd7513c3cf040b3290daac6472fdc5c899f1438b9ebc94bf571" exitCode=2 Jan 27 15:28:48 crc kubenswrapper[4772]: I0127 15:28:48.882975 4772 generic.go:334] "Generic (PLEG): container finished" podID="b4ce266c-6d03-4c51-a8e7-2439eecdf67d" containerID="01b8e4f8a171a9643b0341141cead865a87cb972ae5d54e123bbcc5bbb627212" exitCode=0 Jan 27 15:28:48 crc kubenswrapper[4772]: I0127 15:28:48.882985 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4ce266c-6d03-4c51-a8e7-2439eecdf67d","Type":"ContainerDied","Data":"a9e4d8d8cdfe57821f29d14d1ac46a5f3e7ed0b5e31d0de7bea6c91615349a23"} Jan 27 15:28:48 crc kubenswrapper[4772]: I0127 15:28:48.883036 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4ce266c-6d03-4c51-a8e7-2439eecdf67d","Type":"ContainerDied","Data":"aeaac2f858ccbdd7513c3cf040b3290daac6472fdc5c899f1438b9ebc94bf571"} Jan 27 15:28:48 crc kubenswrapper[4772]: I0127 15:28:48.883048 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4ce266c-6d03-4c51-a8e7-2439eecdf67d","Type":"ContainerDied","Data":"01b8e4f8a171a9643b0341141cead865a87cb972ae5d54e123bbcc5bbb627212"} Jan 27 15:28:50 crc kubenswrapper[4772]: I0127 15:28:50.768067 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="c94a7cfa-28e2-4d52-85a1-d5586f162227" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.149:9292/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 27 15:28:50 crc kubenswrapper[4772]: I0127 15:28:50.768182 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="c94a7cfa-28e2-4d52-85a1-d5586f162227" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.149:9292/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 27 15:28:53 crc kubenswrapper[4772]: I0127 15:28:53.943399 4772 generic.go:334] "Generic (PLEG): container finished" podID="b4ce266c-6d03-4c51-a8e7-2439eecdf67d" containerID="afcf06fa22d2533b1f1a226452ae6ceefe63b8b90f23a95e55e5536a352c31c5" exitCode=0 Jan 27 15:28:53 crc kubenswrapper[4772]: I0127 15:28:53.943551 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4ce266c-6d03-4c51-a8e7-2439eecdf67d","Type":"ContainerDied","Data":"afcf06fa22d2533b1f1a226452ae6ceefe63b8b90f23a95e55e5536a352c31c5"} Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.236641 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.357506 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-run-httpd\") pod \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.357802 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-combined-ca-bundle\") pod \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.357826 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5526\" (UniqueName: \"kubernetes.io/projected/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-kube-api-access-p5526\") pod \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.357855 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-config-data\") pod \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.357956 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-sg-core-conf-yaml\") pod \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.358046 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-scripts\") pod \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.358083 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-log-httpd\") pod \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\" (UID: \"b4ce266c-6d03-4c51-a8e7-2439eecdf67d\") " Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.359020 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b4ce266c-6d03-4c51-a8e7-2439eecdf67d" (UID: "b4ce266c-6d03-4c51-a8e7-2439eecdf67d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.359272 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b4ce266c-6d03-4c51-a8e7-2439eecdf67d" (UID: "b4ce266c-6d03-4c51-a8e7-2439eecdf67d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.365343 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-kube-api-access-p5526" (OuterVolumeSpecName: "kube-api-access-p5526") pod "b4ce266c-6d03-4c51-a8e7-2439eecdf67d" (UID: "b4ce266c-6d03-4c51-a8e7-2439eecdf67d"). InnerVolumeSpecName "kube-api-access-p5526". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.365344 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-scripts" (OuterVolumeSpecName: "scripts") pod "b4ce266c-6d03-4c51-a8e7-2439eecdf67d" (UID: "b4ce266c-6d03-4c51-a8e7-2439eecdf67d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.388098 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b4ce266c-6d03-4c51-a8e7-2439eecdf67d" (UID: "b4ce266c-6d03-4c51-a8e7-2439eecdf67d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.440316 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b4ce266c-6d03-4c51-a8e7-2439eecdf67d" (UID: "b4ce266c-6d03-4c51-a8e7-2439eecdf67d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.460226 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.460337 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5526\" (UniqueName: \"kubernetes.io/projected/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-kube-api-access-p5526\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.460353 4772 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.460364 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.460375 4772 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.460389 4772 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.478358 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-config-data" (OuterVolumeSpecName: "config-data") pod "b4ce266c-6d03-4c51-a8e7-2439eecdf67d" (UID: "b4ce266c-6d03-4c51-a8e7-2439eecdf67d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.562266 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4ce266c-6d03-4c51-a8e7-2439eecdf67d-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.958547 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4ce266c-6d03-4c51-a8e7-2439eecdf67d","Type":"ContainerDied","Data":"b742511aad6534f52532262eb0cfdd05571cd1513b70b4dc28a5263607304430"} Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.958600 4772 scope.go:117] "RemoveContainer" containerID="a9e4d8d8cdfe57821f29d14d1ac46a5f3e7ed0b5e31d0de7bea6c91615349a23" Jan 27 15:28:54 crc kubenswrapper[4772]: I0127 15:28:54.958732 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:54.993094 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.015247 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.029468 4772 scope.go:117] "RemoveContainer" containerID="aeaac2f858ccbdd7513c3cf040b3290daac6472fdc5c899f1438b9ebc94bf571" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.042227 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:55 crc kubenswrapper[4772]: E0127 15:28:55.042754 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4ce266c-6d03-4c51-a8e7-2439eecdf67d" containerName="ceilometer-central-agent" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.042779 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4ce266c-6d03-4c51-a8e7-2439eecdf67d" containerName="ceilometer-central-agent" Jan 27 15:28:55 crc kubenswrapper[4772]: E0127 15:28:55.042806 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4ce266c-6d03-4c51-a8e7-2439eecdf67d" containerName="ceilometer-notification-agent" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.042814 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4ce266c-6d03-4c51-a8e7-2439eecdf67d" containerName="ceilometer-notification-agent" Jan 27 15:28:55 crc kubenswrapper[4772]: E0127 15:28:55.042827 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4ce266c-6d03-4c51-a8e7-2439eecdf67d" containerName="sg-core" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.042837 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4ce266c-6d03-4c51-a8e7-2439eecdf67d" containerName="sg-core" Jan 27 15:28:55 crc kubenswrapper[4772]: E0127 15:28:55.042858 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4ce266c-6d03-4c51-a8e7-2439eecdf67d" containerName="proxy-httpd" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.042865 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4ce266c-6d03-4c51-a8e7-2439eecdf67d" containerName="proxy-httpd" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.043060 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4ce266c-6d03-4c51-a8e7-2439eecdf67d" containerName="proxy-httpd" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.043078 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4ce266c-6d03-4c51-a8e7-2439eecdf67d" containerName="ceilometer-notification-agent" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.043108 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4ce266c-6d03-4c51-a8e7-2439eecdf67d" containerName="sg-core" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.043121 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4ce266c-6d03-4c51-a8e7-2439eecdf67d" containerName="ceilometer-central-agent" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.045121 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.050032 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.050864 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.070208 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.122219 4772 scope.go:117] "RemoveContainer" containerID="01b8e4f8a171a9643b0341141cead865a87cb972ae5d54e123bbcc5bbb627212" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.200086 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2869a695-9773-4816-90d1-34f45555b442-run-httpd\") pod \"ceilometer-0\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " pod="openstack/ceilometer-0" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.200159 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2869a695-9773-4816-90d1-34f45555b442-scripts\") pod \"ceilometer-0\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " pod="openstack/ceilometer-0" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.200216 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2869a695-9773-4816-90d1-34f45555b442-log-httpd\") pod \"ceilometer-0\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " pod="openstack/ceilometer-0" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.200274 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2869a695-9773-4816-90d1-34f45555b442-config-data\") pod \"ceilometer-0\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " pod="openstack/ceilometer-0" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.200345 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnzcc\" (UniqueName: \"kubernetes.io/projected/2869a695-9773-4816-90d1-34f45555b442-kube-api-access-tnzcc\") pod \"ceilometer-0\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " pod="openstack/ceilometer-0" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.200382 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2869a695-9773-4816-90d1-34f45555b442-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " pod="openstack/ceilometer-0" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.200428 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2869a695-9773-4816-90d1-34f45555b442-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " pod="openstack/ceilometer-0" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.212440 4772 scope.go:117] "RemoveContainer" containerID="afcf06fa22d2533b1f1a226452ae6ceefe63b8b90f23a95e55e5536a352c31c5" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.301514 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnzcc\" (UniqueName: \"kubernetes.io/projected/2869a695-9773-4816-90d1-34f45555b442-kube-api-access-tnzcc\") pod \"ceilometer-0\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " pod="openstack/ceilometer-0" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.301799 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2869a695-9773-4816-90d1-34f45555b442-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " pod="openstack/ceilometer-0" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.301841 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2869a695-9773-4816-90d1-34f45555b442-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " pod="openstack/ceilometer-0" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.301902 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2869a695-9773-4816-90d1-34f45555b442-run-httpd\") pod \"ceilometer-0\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " pod="openstack/ceilometer-0" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.301938 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2869a695-9773-4816-90d1-34f45555b442-scripts\") pod \"ceilometer-0\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " pod="openstack/ceilometer-0" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.301959 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2869a695-9773-4816-90d1-34f45555b442-log-httpd\") pod \"ceilometer-0\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " pod="openstack/ceilometer-0" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.302505 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2869a695-9773-4816-90d1-34f45555b442-log-httpd\") pod \"ceilometer-0\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " pod="openstack/ceilometer-0" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.302750 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2869a695-9773-4816-90d1-34f45555b442-run-httpd\") pod \"ceilometer-0\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " pod="openstack/ceilometer-0" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.303428 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2869a695-9773-4816-90d1-34f45555b442-config-data\") pod \"ceilometer-0\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " pod="openstack/ceilometer-0" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.306739 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2869a695-9773-4816-90d1-34f45555b442-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " pod="openstack/ceilometer-0" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.307350 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2869a695-9773-4816-90d1-34f45555b442-config-data\") pod \"ceilometer-0\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " pod="openstack/ceilometer-0" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.311855 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2869a695-9773-4816-90d1-34f45555b442-scripts\") pod \"ceilometer-0\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " pod="openstack/ceilometer-0" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.312423 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2869a695-9773-4816-90d1-34f45555b442-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " pod="openstack/ceilometer-0" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.318401 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnzcc\" (UniqueName: \"kubernetes.io/projected/2869a695-9773-4816-90d1-34f45555b442-kube-api-access-tnzcc\") pod \"ceilometer-0\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " pod="openstack/ceilometer-0" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.433449 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.878825 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.968223 4772 generic.go:334] "Generic (PLEG): container finished" podID="fe34fbf1-61c4-46a9-9954-64ed431d2cb7" containerID="35964dfe2e497930630aeb0996d17bf7bbe0e9d5e7bfb1d7efca05167ac578fc" exitCode=0 Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.968297 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-v9mqp" event={"ID":"fe34fbf1-61c4-46a9-9954-64ed431d2cb7","Type":"ContainerDied","Data":"35964dfe2e497930630aeb0996d17bf7bbe0e9d5e7bfb1d7efca05167ac578fc"} Jan 27 15:28:55 crc kubenswrapper[4772]: I0127 15:28:55.970781 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2869a695-9773-4816-90d1-34f45555b442","Type":"ContainerStarted","Data":"6e415f02d11a2c8f15d05a7114c5f8606abeaaf34e280ab6666002c1dea01ba4"} Jan 27 15:28:56 crc kubenswrapper[4772]: I0127 15:28:56.679311 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4ce266c-6d03-4c51-a8e7-2439eecdf67d" path="/var/lib/kubelet/pods/b4ce266c-6d03-4c51-a8e7-2439eecdf67d/volumes" Jan 27 15:28:56 crc kubenswrapper[4772]: I0127 15:28:56.982073 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2869a695-9773-4816-90d1-34f45555b442","Type":"ContainerStarted","Data":"ede1e3bccf115909c4db09d1e983ec3d2f0212adfce0f8676afa3715415fb310"} Jan 27 15:28:57 crc kubenswrapper[4772]: I0127 15:28:57.329305 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-v9mqp" Jan 27 15:28:57 crc kubenswrapper[4772]: I0127 15:28:57.435033 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe34fbf1-61c4-46a9-9954-64ed431d2cb7-config-data\") pod \"fe34fbf1-61c4-46a9-9954-64ed431d2cb7\" (UID: \"fe34fbf1-61c4-46a9-9954-64ed431d2cb7\") " Jan 27 15:28:57 crc kubenswrapper[4772]: I0127 15:28:57.435148 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe34fbf1-61c4-46a9-9954-64ed431d2cb7-combined-ca-bundle\") pod \"fe34fbf1-61c4-46a9-9954-64ed431d2cb7\" (UID: \"fe34fbf1-61c4-46a9-9954-64ed431d2cb7\") " Jan 27 15:28:57 crc kubenswrapper[4772]: I0127 15:28:57.435213 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe34fbf1-61c4-46a9-9954-64ed431d2cb7-scripts\") pod \"fe34fbf1-61c4-46a9-9954-64ed431d2cb7\" (UID: \"fe34fbf1-61c4-46a9-9954-64ed431d2cb7\") " Jan 27 15:28:57 crc kubenswrapper[4772]: I0127 15:28:57.435307 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pz7jw\" (UniqueName: \"kubernetes.io/projected/fe34fbf1-61c4-46a9-9954-64ed431d2cb7-kube-api-access-pz7jw\") pod \"fe34fbf1-61c4-46a9-9954-64ed431d2cb7\" (UID: \"fe34fbf1-61c4-46a9-9954-64ed431d2cb7\") " Jan 27 15:28:57 crc kubenswrapper[4772]: I0127 15:28:57.453359 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe34fbf1-61c4-46a9-9954-64ed431d2cb7-kube-api-access-pz7jw" (OuterVolumeSpecName: "kube-api-access-pz7jw") pod "fe34fbf1-61c4-46a9-9954-64ed431d2cb7" (UID: "fe34fbf1-61c4-46a9-9954-64ed431d2cb7"). InnerVolumeSpecName "kube-api-access-pz7jw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:28:57 crc kubenswrapper[4772]: I0127 15:28:57.455815 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe34fbf1-61c4-46a9-9954-64ed431d2cb7-scripts" (OuterVolumeSpecName: "scripts") pod "fe34fbf1-61c4-46a9-9954-64ed431d2cb7" (UID: "fe34fbf1-61c4-46a9-9954-64ed431d2cb7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:57 crc kubenswrapper[4772]: I0127 15:28:57.464618 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe34fbf1-61c4-46a9-9954-64ed431d2cb7-config-data" (OuterVolumeSpecName: "config-data") pod "fe34fbf1-61c4-46a9-9954-64ed431d2cb7" (UID: "fe34fbf1-61c4-46a9-9954-64ed431d2cb7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:57 crc kubenswrapper[4772]: I0127 15:28:57.466247 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe34fbf1-61c4-46a9-9954-64ed431d2cb7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe34fbf1-61c4-46a9-9954-64ed431d2cb7" (UID: "fe34fbf1-61c4-46a9-9954-64ed431d2cb7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:28:57 crc kubenswrapper[4772]: I0127 15:28:57.537507 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe34fbf1-61c4-46a9-9954-64ed431d2cb7-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:57 crc kubenswrapper[4772]: I0127 15:28:57.537558 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe34fbf1-61c4-46a9-9954-64ed431d2cb7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:57 crc kubenswrapper[4772]: I0127 15:28:57.537577 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe34fbf1-61c4-46a9-9954-64ed431d2cb7-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:57 crc kubenswrapper[4772]: I0127 15:28:57.537590 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pz7jw\" (UniqueName: \"kubernetes.io/projected/fe34fbf1-61c4-46a9-9954-64ed431d2cb7-kube-api-access-pz7jw\") on node \"crc\" DevicePath \"\"" Jan 27 15:28:57 crc kubenswrapper[4772]: I0127 15:28:57.994995 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2869a695-9773-4816-90d1-34f45555b442","Type":"ContainerStarted","Data":"98f0561ee38d8d108638c05ac609999f7f05dffc671417770e7d54c659d7c770"} Jan 27 15:28:57 crc kubenswrapper[4772]: I0127 15:28:57.996948 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-v9mqp" event={"ID":"fe34fbf1-61c4-46a9-9954-64ed431d2cb7","Type":"ContainerDied","Data":"e097abc13b5b9498583b3db2e2e89e5740a2e06f350bc375a47375cb723458af"} Jan 27 15:28:57 crc kubenswrapper[4772]: I0127 15:28:57.997001 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e097abc13b5b9498583b3db2e2e89e5740a2e06f350bc375a47375cb723458af" Jan 27 15:28:57 crc kubenswrapper[4772]: I0127 15:28:57.997014 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-v9mqp" Jan 27 15:28:58 crc kubenswrapper[4772]: I0127 15:28:58.099875 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 27 15:28:58 crc kubenswrapper[4772]: E0127 15:28:58.100271 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe34fbf1-61c4-46a9-9954-64ed431d2cb7" containerName="nova-cell0-conductor-db-sync" Jan 27 15:28:58 crc kubenswrapper[4772]: I0127 15:28:58.100291 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe34fbf1-61c4-46a9-9954-64ed431d2cb7" containerName="nova-cell0-conductor-db-sync" Jan 27 15:28:58 crc kubenswrapper[4772]: I0127 15:28:58.100531 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe34fbf1-61c4-46a9-9954-64ed431d2cb7" containerName="nova-cell0-conductor-db-sync" Jan 27 15:28:58 crc kubenswrapper[4772]: I0127 15:28:58.101180 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 27 15:28:58 crc kubenswrapper[4772]: I0127 15:28:58.105789 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Jan 27 15:28:58 crc kubenswrapper[4772]: I0127 15:28:58.105850 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bq7vb" Jan 27 15:28:58 crc kubenswrapper[4772]: I0127 15:28:58.112730 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 27 15:28:58 crc kubenswrapper[4772]: I0127 15:28:58.250000 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd9ac534-7732-417d-81a3-573fe821b26d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"bd9ac534-7732-417d-81a3-573fe821b26d\") " pod="openstack/nova-cell0-conductor-0" Jan 27 15:28:58 crc kubenswrapper[4772]: I0127 15:28:58.250060 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd9ac534-7732-417d-81a3-573fe821b26d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"bd9ac534-7732-417d-81a3-573fe821b26d\") " pod="openstack/nova-cell0-conductor-0" Jan 27 15:28:58 crc kubenswrapper[4772]: I0127 15:28:58.250091 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txnx2\" (UniqueName: \"kubernetes.io/projected/bd9ac534-7732-417d-81a3-573fe821b26d-kube-api-access-txnx2\") pod \"nova-cell0-conductor-0\" (UID: \"bd9ac534-7732-417d-81a3-573fe821b26d\") " pod="openstack/nova-cell0-conductor-0" Jan 27 15:28:58 crc kubenswrapper[4772]: I0127 15:28:58.352699 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd9ac534-7732-417d-81a3-573fe821b26d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"bd9ac534-7732-417d-81a3-573fe821b26d\") " pod="openstack/nova-cell0-conductor-0" Jan 27 15:28:58 crc kubenswrapper[4772]: I0127 15:28:58.352978 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd9ac534-7732-417d-81a3-573fe821b26d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"bd9ac534-7732-417d-81a3-573fe821b26d\") " pod="openstack/nova-cell0-conductor-0" Jan 27 15:28:58 crc kubenswrapper[4772]: I0127 15:28:58.352998 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txnx2\" (UniqueName: \"kubernetes.io/projected/bd9ac534-7732-417d-81a3-573fe821b26d-kube-api-access-txnx2\") pod \"nova-cell0-conductor-0\" (UID: \"bd9ac534-7732-417d-81a3-573fe821b26d\") " pod="openstack/nova-cell0-conductor-0" Jan 27 15:28:58 crc kubenswrapper[4772]: I0127 15:28:58.359009 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd9ac534-7732-417d-81a3-573fe821b26d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"bd9ac534-7732-417d-81a3-573fe821b26d\") " pod="openstack/nova-cell0-conductor-0" Jan 27 15:28:58 crc kubenswrapper[4772]: I0127 15:28:58.372269 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd9ac534-7732-417d-81a3-573fe821b26d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"bd9ac534-7732-417d-81a3-573fe821b26d\") " pod="openstack/nova-cell0-conductor-0" Jan 27 15:28:58 crc kubenswrapper[4772]: I0127 15:28:58.376679 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txnx2\" (UniqueName: \"kubernetes.io/projected/bd9ac534-7732-417d-81a3-573fe821b26d-kube-api-access-txnx2\") pod \"nova-cell0-conductor-0\" (UID: \"bd9ac534-7732-417d-81a3-573fe821b26d\") " pod="openstack/nova-cell0-conductor-0" Jan 27 15:28:58 crc kubenswrapper[4772]: I0127 15:28:58.464228 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 27 15:28:58 crc kubenswrapper[4772]: I0127 15:28:58.574090 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 27 15:28:58 crc kubenswrapper[4772]: W0127 15:28:58.974673 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd9ac534_7732_417d_81a3_573fe821b26d.slice/crio-1fa0fda721794873849780c939c36b0ae92fda6799dd0ff519efa3b4fb4008b6 WatchSource:0}: Error finding container 1fa0fda721794873849780c939c36b0ae92fda6799dd0ff519efa3b4fb4008b6: Status 404 returned error can't find the container with id 1fa0fda721794873849780c939c36b0ae92fda6799dd0ff519efa3b4fb4008b6 Jan 27 15:28:58 crc kubenswrapper[4772]: I0127 15:28:58.983154 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 27 15:28:59 crc kubenswrapper[4772]: I0127 15:28:59.014045 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:28:59 crc kubenswrapper[4772]: I0127 15:28:59.043788 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2869a695-9773-4816-90d1-34f45555b442","Type":"ContainerStarted","Data":"ab34d41bdb3da5c2ca1372f2b064ab426f9d0b9c15e4bb1bb989886eacb17cd9"} Jan 27 15:28:59 crc kubenswrapper[4772]: I0127 15:28:59.059807 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"bd9ac534-7732-417d-81a3-573fe821b26d","Type":"ContainerStarted","Data":"1fa0fda721794873849780c939c36b0ae92fda6799dd0ff519efa3b4fb4008b6"} Jan 27 15:29:00 crc kubenswrapper[4772]: I0127 15:29:00.071856 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"bd9ac534-7732-417d-81a3-573fe821b26d","Type":"ContainerStarted","Data":"d0ebea91d7cab43bb0777a4dc747e6d73a7eab735ebf1eea102733e045c246c7"} Jan 27 15:29:00 crc kubenswrapper[4772]: I0127 15:29:00.072352 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Jan 27 15:29:00 crc kubenswrapper[4772]: I0127 15:29:00.071978 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="bd9ac534-7732-417d-81a3-573fe821b26d" containerName="nova-cell0-conductor-conductor" containerID="cri-o://d0ebea91d7cab43bb0777a4dc747e6d73a7eab735ebf1eea102733e045c246c7" gracePeriod=30 Jan 27 15:29:00 crc kubenswrapper[4772]: I0127 15:29:00.094899 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.094878454 podStartE2EDuration="2.094878454s" podCreationTimestamp="2026-01-27 15:28:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:29:00.089212709 +0000 UTC m=+1326.069821817" watchObservedRunningTime="2026-01-27 15:29:00.094878454 +0000 UTC m=+1326.075487552" Jan 27 15:29:01 crc kubenswrapper[4772]: I0127 15:29:01.083115 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2869a695-9773-4816-90d1-34f45555b442","Type":"ContainerStarted","Data":"dd17ba31af2f6ba28c2a4b9d132f086ba22d76034e3e3228e522e489759ecad1"} Jan 27 15:29:01 crc kubenswrapper[4772]: I0127 15:29:01.083317 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2869a695-9773-4816-90d1-34f45555b442" containerName="ceilometer-central-agent" containerID="cri-o://ede1e3bccf115909c4db09d1e983ec3d2f0212adfce0f8676afa3715415fb310" gracePeriod=30 Jan 27 15:29:01 crc kubenswrapper[4772]: I0127 15:29:01.083370 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2869a695-9773-4816-90d1-34f45555b442" containerName="sg-core" containerID="cri-o://ab34d41bdb3da5c2ca1372f2b064ab426f9d0b9c15e4bb1bb989886eacb17cd9" gracePeriod=30 Jan 27 15:29:01 crc kubenswrapper[4772]: I0127 15:29:01.083396 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2869a695-9773-4816-90d1-34f45555b442" containerName="proxy-httpd" containerID="cri-o://dd17ba31af2f6ba28c2a4b9d132f086ba22d76034e3e3228e522e489759ecad1" gracePeriod=30 Jan 27 15:29:01 crc kubenswrapper[4772]: I0127 15:29:01.083409 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2869a695-9773-4816-90d1-34f45555b442" containerName="ceilometer-notification-agent" containerID="cri-o://98f0561ee38d8d108638c05ac609999f7f05dffc671417770e7d54c659d7c770" gracePeriod=30 Jan 27 15:29:01 crc kubenswrapper[4772]: I0127 15:29:01.084093 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 27 15:29:01 crc kubenswrapper[4772]: I0127 15:29:01.119666 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.400734446 podStartE2EDuration="7.119644702s" podCreationTimestamp="2026-01-27 15:28:54 +0000 UTC" firstStartedPulling="2026-01-27 15:28:55.880665514 +0000 UTC m=+1321.861274622" lastFinishedPulling="2026-01-27 15:29:00.59957578 +0000 UTC m=+1326.580184878" observedRunningTime="2026-01-27 15:29:01.107785678 +0000 UTC m=+1327.088394776" watchObservedRunningTime="2026-01-27 15:29:01.119644702 +0000 UTC m=+1327.100253810" Jan 27 15:29:02 crc kubenswrapper[4772]: I0127 15:29:02.094253 4772 generic.go:334] "Generic (PLEG): container finished" podID="2869a695-9773-4816-90d1-34f45555b442" containerID="dd17ba31af2f6ba28c2a4b9d132f086ba22d76034e3e3228e522e489759ecad1" exitCode=0 Jan 27 15:29:02 crc kubenswrapper[4772]: I0127 15:29:02.094480 4772 generic.go:334] "Generic (PLEG): container finished" podID="2869a695-9773-4816-90d1-34f45555b442" containerID="ab34d41bdb3da5c2ca1372f2b064ab426f9d0b9c15e4bb1bb989886eacb17cd9" exitCode=2 Jan 27 15:29:02 crc kubenswrapper[4772]: I0127 15:29:02.094489 4772 generic.go:334] "Generic (PLEG): container finished" podID="2869a695-9773-4816-90d1-34f45555b442" containerID="98f0561ee38d8d108638c05ac609999f7f05dffc671417770e7d54c659d7c770" exitCode=0 Jan 27 15:29:02 crc kubenswrapper[4772]: I0127 15:29:02.094328 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2869a695-9773-4816-90d1-34f45555b442","Type":"ContainerDied","Data":"dd17ba31af2f6ba28c2a4b9d132f086ba22d76034e3e3228e522e489759ecad1"} Jan 27 15:29:02 crc kubenswrapper[4772]: I0127 15:29:02.094523 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2869a695-9773-4816-90d1-34f45555b442","Type":"ContainerDied","Data":"ab34d41bdb3da5c2ca1372f2b064ab426f9d0b9c15e4bb1bb989886eacb17cd9"} Jan 27 15:29:02 crc kubenswrapper[4772]: I0127 15:29:02.094537 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2869a695-9773-4816-90d1-34f45555b442","Type":"ContainerDied","Data":"98f0561ee38d8d108638c05ac609999f7f05dffc671417770e7d54c659d7c770"} Jan 27 15:29:03 crc kubenswrapper[4772]: I0127 15:29:03.716686 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:29:03 crc kubenswrapper[4772]: I0127 15:29:03.856001 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2869a695-9773-4816-90d1-34f45555b442-sg-core-conf-yaml\") pod \"2869a695-9773-4816-90d1-34f45555b442\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " Jan 27 15:29:03 crc kubenswrapper[4772]: I0127 15:29:03.856213 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2869a695-9773-4816-90d1-34f45555b442-run-httpd\") pod \"2869a695-9773-4816-90d1-34f45555b442\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " Jan 27 15:29:03 crc kubenswrapper[4772]: I0127 15:29:03.856269 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2869a695-9773-4816-90d1-34f45555b442-scripts\") pod \"2869a695-9773-4816-90d1-34f45555b442\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " Jan 27 15:29:03 crc kubenswrapper[4772]: I0127 15:29:03.856306 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnzcc\" (UniqueName: \"kubernetes.io/projected/2869a695-9773-4816-90d1-34f45555b442-kube-api-access-tnzcc\") pod \"2869a695-9773-4816-90d1-34f45555b442\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " Jan 27 15:29:03 crc kubenswrapper[4772]: I0127 15:29:03.856360 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2869a695-9773-4816-90d1-34f45555b442-log-httpd\") pod \"2869a695-9773-4816-90d1-34f45555b442\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " Jan 27 15:29:03 crc kubenswrapper[4772]: I0127 15:29:03.856462 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2869a695-9773-4816-90d1-34f45555b442-combined-ca-bundle\") pod \"2869a695-9773-4816-90d1-34f45555b442\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " Jan 27 15:29:03 crc kubenswrapper[4772]: I0127 15:29:03.856554 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2869a695-9773-4816-90d1-34f45555b442-config-data\") pod \"2869a695-9773-4816-90d1-34f45555b442\" (UID: \"2869a695-9773-4816-90d1-34f45555b442\") " Jan 27 15:29:03 crc kubenswrapper[4772]: I0127 15:29:03.856800 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2869a695-9773-4816-90d1-34f45555b442-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2869a695-9773-4816-90d1-34f45555b442" (UID: "2869a695-9773-4816-90d1-34f45555b442"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:29:03 crc kubenswrapper[4772]: I0127 15:29:03.856919 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2869a695-9773-4816-90d1-34f45555b442-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2869a695-9773-4816-90d1-34f45555b442" (UID: "2869a695-9773-4816-90d1-34f45555b442"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:29:03 crc kubenswrapper[4772]: I0127 15:29:03.857582 4772 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2869a695-9773-4816-90d1-34f45555b442-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:03 crc kubenswrapper[4772]: I0127 15:29:03.857614 4772 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2869a695-9773-4816-90d1-34f45555b442-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:03 crc kubenswrapper[4772]: I0127 15:29:03.865537 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2869a695-9773-4816-90d1-34f45555b442-scripts" (OuterVolumeSpecName: "scripts") pod "2869a695-9773-4816-90d1-34f45555b442" (UID: "2869a695-9773-4816-90d1-34f45555b442"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:29:03 crc kubenswrapper[4772]: I0127 15:29:03.865590 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2869a695-9773-4816-90d1-34f45555b442-kube-api-access-tnzcc" (OuterVolumeSpecName: "kube-api-access-tnzcc") pod "2869a695-9773-4816-90d1-34f45555b442" (UID: "2869a695-9773-4816-90d1-34f45555b442"). InnerVolumeSpecName "kube-api-access-tnzcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:29:03 crc kubenswrapper[4772]: I0127 15:29:03.905492 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2869a695-9773-4816-90d1-34f45555b442-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2869a695-9773-4816-90d1-34f45555b442" (UID: "2869a695-9773-4816-90d1-34f45555b442"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:29:03 crc kubenswrapper[4772]: I0127 15:29:03.941361 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2869a695-9773-4816-90d1-34f45555b442-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2869a695-9773-4816-90d1-34f45555b442" (UID: "2869a695-9773-4816-90d1-34f45555b442"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:29:03 crc kubenswrapper[4772]: I0127 15:29:03.958711 4772 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2869a695-9773-4816-90d1-34f45555b442-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:03 crc kubenswrapper[4772]: I0127 15:29:03.958742 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2869a695-9773-4816-90d1-34f45555b442-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:03 crc kubenswrapper[4772]: I0127 15:29:03.958753 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnzcc\" (UniqueName: \"kubernetes.io/projected/2869a695-9773-4816-90d1-34f45555b442-kube-api-access-tnzcc\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:03 crc kubenswrapper[4772]: I0127 15:29:03.958762 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2869a695-9773-4816-90d1-34f45555b442-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:03 crc kubenswrapper[4772]: I0127 15:29:03.977302 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2869a695-9773-4816-90d1-34f45555b442-config-data" (OuterVolumeSpecName: "config-data") pod "2869a695-9773-4816-90d1-34f45555b442" (UID: "2869a695-9773-4816-90d1-34f45555b442"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.060510 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2869a695-9773-4816-90d1-34f45555b442-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.117923 4772 generic.go:334] "Generic (PLEG): container finished" podID="2869a695-9773-4816-90d1-34f45555b442" containerID="ede1e3bccf115909c4db09d1e983ec3d2f0212adfce0f8676afa3715415fb310" exitCode=0 Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.117966 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2869a695-9773-4816-90d1-34f45555b442","Type":"ContainerDied","Data":"ede1e3bccf115909c4db09d1e983ec3d2f0212adfce0f8676afa3715415fb310"} Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.117994 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2869a695-9773-4816-90d1-34f45555b442","Type":"ContainerDied","Data":"6e415f02d11a2c8f15d05a7114c5f8606abeaaf34e280ab6666002c1dea01ba4"} Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.117999 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.118013 4772 scope.go:117] "RemoveContainer" containerID="dd17ba31af2f6ba28c2a4b9d132f086ba22d76034e3e3228e522e489759ecad1" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.156218 4772 scope.go:117] "RemoveContainer" containerID="ab34d41bdb3da5c2ca1372f2b064ab426f9d0b9c15e4bb1bb989886eacb17cd9" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.162574 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.176295 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.190283 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:29:04 crc kubenswrapper[4772]: E0127 15:29:04.190794 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2869a695-9773-4816-90d1-34f45555b442" containerName="proxy-httpd" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.190809 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="2869a695-9773-4816-90d1-34f45555b442" containerName="proxy-httpd" Jan 27 15:29:04 crc kubenswrapper[4772]: E0127 15:29:04.190825 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2869a695-9773-4816-90d1-34f45555b442" containerName="sg-core" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.190832 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="2869a695-9773-4816-90d1-34f45555b442" containerName="sg-core" Jan 27 15:29:04 crc kubenswrapper[4772]: E0127 15:29:04.190859 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2869a695-9773-4816-90d1-34f45555b442" containerName="ceilometer-central-agent" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.190867 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="2869a695-9773-4816-90d1-34f45555b442" containerName="ceilometer-central-agent" Jan 27 15:29:04 crc kubenswrapper[4772]: E0127 15:29:04.190887 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2869a695-9773-4816-90d1-34f45555b442" containerName="ceilometer-notification-agent" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.190895 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="2869a695-9773-4816-90d1-34f45555b442" containerName="ceilometer-notification-agent" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.191122 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="2869a695-9773-4816-90d1-34f45555b442" containerName="sg-core" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.191158 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="2869a695-9773-4816-90d1-34f45555b442" containerName="proxy-httpd" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.191677 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="2869a695-9773-4816-90d1-34f45555b442" containerName="ceilometer-central-agent" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.191700 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="2869a695-9773-4816-90d1-34f45555b442" containerName="ceilometer-notification-agent" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.196235 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.198039 4772 scope.go:117] "RemoveContainer" containerID="98f0561ee38d8d108638c05ac609999f7f05dffc671417770e7d54c659d7c770" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.200742 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.201241 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.203398 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.230583 4772 scope.go:117] "RemoveContainer" containerID="ede1e3bccf115909c4db09d1e983ec3d2f0212adfce0f8676afa3715415fb310" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.247767 4772 scope.go:117] "RemoveContainer" containerID="dd17ba31af2f6ba28c2a4b9d132f086ba22d76034e3e3228e522e489759ecad1" Jan 27 15:29:04 crc kubenswrapper[4772]: E0127 15:29:04.248747 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd17ba31af2f6ba28c2a4b9d132f086ba22d76034e3e3228e522e489759ecad1\": container with ID starting with dd17ba31af2f6ba28c2a4b9d132f086ba22d76034e3e3228e522e489759ecad1 not found: ID does not exist" containerID="dd17ba31af2f6ba28c2a4b9d132f086ba22d76034e3e3228e522e489759ecad1" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.248806 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd17ba31af2f6ba28c2a4b9d132f086ba22d76034e3e3228e522e489759ecad1"} err="failed to get container status \"dd17ba31af2f6ba28c2a4b9d132f086ba22d76034e3e3228e522e489759ecad1\": rpc error: code = NotFound desc = could not find container \"dd17ba31af2f6ba28c2a4b9d132f086ba22d76034e3e3228e522e489759ecad1\": container with ID starting with dd17ba31af2f6ba28c2a4b9d132f086ba22d76034e3e3228e522e489759ecad1 not found: ID does not exist" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.248827 4772 scope.go:117] "RemoveContainer" containerID="ab34d41bdb3da5c2ca1372f2b064ab426f9d0b9c15e4bb1bb989886eacb17cd9" Jan 27 15:29:04 crc kubenswrapper[4772]: E0127 15:29:04.249123 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab34d41bdb3da5c2ca1372f2b064ab426f9d0b9c15e4bb1bb989886eacb17cd9\": container with ID starting with ab34d41bdb3da5c2ca1372f2b064ab426f9d0b9c15e4bb1bb989886eacb17cd9 not found: ID does not exist" containerID="ab34d41bdb3da5c2ca1372f2b064ab426f9d0b9c15e4bb1bb989886eacb17cd9" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.249273 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab34d41bdb3da5c2ca1372f2b064ab426f9d0b9c15e4bb1bb989886eacb17cd9"} err="failed to get container status \"ab34d41bdb3da5c2ca1372f2b064ab426f9d0b9c15e4bb1bb989886eacb17cd9\": rpc error: code = NotFound desc = could not find container \"ab34d41bdb3da5c2ca1372f2b064ab426f9d0b9c15e4bb1bb989886eacb17cd9\": container with ID starting with ab34d41bdb3da5c2ca1372f2b064ab426f9d0b9c15e4bb1bb989886eacb17cd9 not found: ID does not exist" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.249371 4772 scope.go:117] "RemoveContainer" containerID="98f0561ee38d8d108638c05ac609999f7f05dffc671417770e7d54c659d7c770" Jan 27 15:29:04 crc kubenswrapper[4772]: E0127 15:29:04.249860 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98f0561ee38d8d108638c05ac609999f7f05dffc671417770e7d54c659d7c770\": container with ID starting with 98f0561ee38d8d108638c05ac609999f7f05dffc671417770e7d54c659d7c770 not found: ID does not exist" containerID="98f0561ee38d8d108638c05ac609999f7f05dffc671417770e7d54c659d7c770" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.249991 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98f0561ee38d8d108638c05ac609999f7f05dffc671417770e7d54c659d7c770"} err="failed to get container status \"98f0561ee38d8d108638c05ac609999f7f05dffc671417770e7d54c659d7c770\": rpc error: code = NotFound desc = could not find container \"98f0561ee38d8d108638c05ac609999f7f05dffc671417770e7d54c659d7c770\": container with ID starting with 98f0561ee38d8d108638c05ac609999f7f05dffc671417770e7d54c659d7c770 not found: ID does not exist" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.250100 4772 scope.go:117] "RemoveContainer" containerID="ede1e3bccf115909c4db09d1e983ec3d2f0212adfce0f8676afa3715415fb310" Jan 27 15:29:04 crc kubenswrapper[4772]: E0127 15:29:04.250613 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ede1e3bccf115909c4db09d1e983ec3d2f0212adfce0f8676afa3715415fb310\": container with ID starting with ede1e3bccf115909c4db09d1e983ec3d2f0212adfce0f8676afa3715415fb310 not found: ID does not exist" containerID="ede1e3bccf115909c4db09d1e983ec3d2f0212adfce0f8676afa3715415fb310" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.250653 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ede1e3bccf115909c4db09d1e983ec3d2f0212adfce0f8676afa3715415fb310"} err="failed to get container status \"ede1e3bccf115909c4db09d1e983ec3d2f0212adfce0f8676afa3715415fb310\": rpc error: code = NotFound desc = could not find container \"ede1e3bccf115909c4db09d1e983ec3d2f0212adfce0f8676afa3715415fb310\": container with ID starting with ede1e3bccf115909c4db09d1e983ec3d2f0212adfce0f8676afa3715415fb310 not found: ID does not exist" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.365613 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ebbc7c3-09c3-4524-854b-e0d64400ab93-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " pod="openstack/ceilometer-0" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.365715 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ebbc7c3-09c3-4524-854b-e0d64400ab93-run-httpd\") pod \"ceilometer-0\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " pod="openstack/ceilometer-0" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.365792 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27d6x\" (UniqueName: \"kubernetes.io/projected/3ebbc7c3-09c3-4524-854b-e0d64400ab93-kube-api-access-27d6x\") pod \"ceilometer-0\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " pod="openstack/ceilometer-0" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.365920 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ebbc7c3-09c3-4524-854b-e0d64400ab93-scripts\") pod \"ceilometer-0\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " pod="openstack/ceilometer-0" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.366008 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ebbc7c3-09c3-4524-854b-e0d64400ab93-log-httpd\") pod \"ceilometer-0\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " pod="openstack/ceilometer-0" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.366046 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ebbc7c3-09c3-4524-854b-e0d64400ab93-config-data\") pod \"ceilometer-0\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " pod="openstack/ceilometer-0" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.366102 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ebbc7c3-09c3-4524-854b-e0d64400ab93-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " pod="openstack/ceilometer-0" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.467655 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ebbc7c3-09c3-4524-854b-e0d64400ab93-run-httpd\") pod \"ceilometer-0\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " pod="openstack/ceilometer-0" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.468009 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27d6x\" (UniqueName: \"kubernetes.io/projected/3ebbc7c3-09c3-4524-854b-e0d64400ab93-kube-api-access-27d6x\") pod \"ceilometer-0\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " pod="openstack/ceilometer-0" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.468133 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ebbc7c3-09c3-4524-854b-e0d64400ab93-run-httpd\") pod \"ceilometer-0\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " pod="openstack/ceilometer-0" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.468352 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ebbc7c3-09c3-4524-854b-e0d64400ab93-scripts\") pod \"ceilometer-0\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " pod="openstack/ceilometer-0" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.468547 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ebbc7c3-09c3-4524-854b-e0d64400ab93-log-httpd\") pod \"ceilometer-0\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " pod="openstack/ceilometer-0" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.468726 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ebbc7c3-09c3-4524-854b-e0d64400ab93-config-data\") pod \"ceilometer-0\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " pod="openstack/ceilometer-0" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.468832 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ebbc7c3-09c3-4524-854b-e0d64400ab93-log-httpd\") pod \"ceilometer-0\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " pod="openstack/ceilometer-0" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.468913 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ebbc7c3-09c3-4524-854b-e0d64400ab93-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " pod="openstack/ceilometer-0" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.469248 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ebbc7c3-09c3-4524-854b-e0d64400ab93-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " pod="openstack/ceilometer-0" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.473260 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ebbc7c3-09c3-4524-854b-e0d64400ab93-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " pod="openstack/ceilometer-0" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.474293 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ebbc7c3-09c3-4524-854b-e0d64400ab93-scripts\") pod \"ceilometer-0\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " pod="openstack/ceilometer-0" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.476130 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ebbc7c3-09c3-4524-854b-e0d64400ab93-config-data\") pod \"ceilometer-0\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " pod="openstack/ceilometer-0" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.482885 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ebbc7c3-09c3-4524-854b-e0d64400ab93-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " pod="openstack/ceilometer-0" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.486474 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27d6x\" (UniqueName: \"kubernetes.io/projected/3ebbc7c3-09c3-4524-854b-e0d64400ab93-kube-api-access-27d6x\") pod \"ceilometer-0\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " pod="openstack/ceilometer-0" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.519927 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:29:04 crc kubenswrapper[4772]: I0127 15:29:04.686385 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2869a695-9773-4816-90d1-34f45555b442" path="/var/lib/kubelet/pods/2869a695-9773-4816-90d1-34f45555b442/volumes" Jan 27 15:29:05 crc kubenswrapper[4772]: I0127 15:29:05.020611 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:29:05 crc kubenswrapper[4772]: I0127 15:29:05.128499 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ebbc7c3-09c3-4524-854b-e0d64400ab93","Type":"ContainerStarted","Data":"78cb40c125f03659c88535f65c13efc1a1776dd09e65791b32311c60251b3ac1"} Jan 27 15:29:06 crc kubenswrapper[4772]: I0127 15:29:06.140651 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ebbc7c3-09c3-4524-854b-e0d64400ab93","Type":"ContainerStarted","Data":"c83d73369af2ff787dcb6d403df3f89cc18f8c344ca9a168a243d6708efe0e93"} Jan 27 15:29:07 crc kubenswrapper[4772]: I0127 15:29:07.154358 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ebbc7c3-09c3-4524-854b-e0d64400ab93","Type":"ContainerStarted","Data":"83aac8ec33dfd5bc2c51a863d771b4120a23b62bf31ab0c79c0cd15ea11a7398"} Jan 27 15:29:07 crc kubenswrapper[4772]: I0127 15:29:07.154861 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ebbc7c3-09c3-4524-854b-e0d64400ab93","Type":"ContainerStarted","Data":"3b39e9533e1851e42935b988f031239d201f08489a2d644eb389253a63ab74a9"} Jan 27 15:29:08 crc kubenswrapper[4772]: E0127 15:29:08.467774 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d0ebea91d7cab43bb0777a4dc747e6d73a7eab735ebf1eea102733e045c246c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Jan 27 15:29:08 crc kubenswrapper[4772]: E0127 15:29:08.469378 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d0ebea91d7cab43bb0777a4dc747e6d73a7eab735ebf1eea102733e045c246c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Jan 27 15:29:08 crc kubenswrapper[4772]: E0127 15:29:08.471056 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d0ebea91d7cab43bb0777a4dc747e6d73a7eab735ebf1eea102733e045c246c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Jan 27 15:29:08 crc kubenswrapper[4772]: E0127 15:29:08.471103 4772 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="bd9ac534-7732-417d-81a3-573fe821b26d" containerName="nova-cell0-conductor-conductor" Jan 27 15:29:09 crc kubenswrapper[4772]: I0127 15:29:09.172014 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ebbc7c3-09c3-4524-854b-e0d64400ab93","Type":"ContainerStarted","Data":"c37e3598e7c9e4e7088ad7d7ea7e00efe5c231d10378509f82fb7d2b9df0b502"} Jan 27 15:29:09 crc kubenswrapper[4772]: I0127 15:29:09.172527 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 27 15:29:09 crc kubenswrapper[4772]: I0127 15:29:09.208981 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.653556471 podStartE2EDuration="5.208950824s" podCreationTimestamp="2026-01-27 15:29:04 +0000 UTC" firstStartedPulling="2026-01-27 15:29:05.019574665 +0000 UTC m=+1331.000183763" lastFinishedPulling="2026-01-27 15:29:08.574969008 +0000 UTC m=+1334.555578116" observedRunningTime="2026-01-27 15:29:09.194892705 +0000 UTC m=+1335.175501863" watchObservedRunningTime="2026-01-27 15:29:09.208950824 +0000 UTC m=+1335.189559962" Jan 27 15:29:12 crc kubenswrapper[4772]: I0127 15:29:12.058697 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:29:12 crc kubenswrapper[4772]: I0127 15:29:12.059470 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:29:13 crc kubenswrapper[4772]: E0127 15:29:13.467120 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d0ebea91d7cab43bb0777a4dc747e6d73a7eab735ebf1eea102733e045c246c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Jan 27 15:29:13 crc kubenswrapper[4772]: E0127 15:29:13.468486 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d0ebea91d7cab43bb0777a4dc747e6d73a7eab735ebf1eea102733e045c246c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Jan 27 15:29:13 crc kubenswrapper[4772]: E0127 15:29:13.469730 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d0ebea91d7cab43bb0777a4dc747e6d73a7eab735ebf1eea102733e045c246c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Jan 27 15:29:13 crc kubenswrapper[4772]: E0127 15:29:13.469784 4772 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="bd9ac534-7732-417d-81a3-573fe821b26d" containerName="nova-cell0-conductor-conductor" Jan 27 15:29:18 crc kubenswrapper[4772]: E0127 15:29:18.466939 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d0ebea91d7cab43bb0777a4dc747e6d73a7eab735ebf1eea102733e045c246c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Jan 27 15:29:18 crc kubenswrapper[4772]: E0127 15:29:18.469292 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d0ebea91d7cab43bb0777a4dc747e6d73a7eab735ebf1eea102733e045c246c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Jan 27 15:29:18 crc kubenswrapper[4772]: E0127 15:29:18.475896 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d0ebea91d7cab43bb0777a4dc747e6d73a7eab735ebf1eea102733e045c246c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Jan 27 15:29:18 crc kubenswrapper[4772]: E0127 15:29:18.475985 4772 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="bd9ac534-7732-417d-81a3-573fe821b26d" containerName="nova-cell0-conductor-conductor" Jan 27 15:29:23 crc kubenswrapper[4772]: E0127 15:29:23.467379 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d0ebea91d7cab43bb0777a4dc747e6d73a7eab735ebf1eea102733e045c246c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Jan 27 15:29:23 crc kubenswrapper[4772]: E0127 15:29:23.471993 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d0ebea91d7cab43bb0777a4dc747e6d73a7eab735ebf1eea102733e045c246c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Jan 27 15:29:23 crc kubenswrapper[4772]: E0127 15:29:23.474686 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d0ebea91d7cab43bb0777a4dc747e6d73a7eab735ebf1eea102733e045c246c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Jan 27 15:29:23 crc kubenswrapper[4772]: E0127 15:29:23.474797 4772 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="bd9ac534-7732-417d-81a3-573fe821b26d" containerName="nova-cell0-conductor-conductor" Jan 27 15:29:28 crc kubenswrapper[4772]: E0127 15:29:28.467670 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d0ebea91d7cab43bb0777a4dc747e6d73a7eab735ebf1eea102733e045c246c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Jan 27 15:29:28 crc kubenswrapper[4772]: E0127 15:29:28.469868 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d0ebea91d7cab43bb0777a4dc747e6d73a7eab735ebf1eea102733e045c246c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Jan 27 15:29:28 crc kubenswrapper[4772]: E0127 15:29:28.471927 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d0ebea91d7cab43bb0777a4dc747e6d73a7eab735ebf1eea102733e045c246c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Jan 27 15:29:28 crc kubenswrapper[4772]: E0127 15:29:28.472011 4772 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="bd9ac534-7732-417d-81a3-573fe821b26d" containerName="nova-cell0-conductor-conductor" Jan 27 15:29:30 crc kubenswrapper[4772]: I0127 15:29:30.374344 4772 generic.go:334] "Generic (PLEG): container finished" podID="bd9ac534-7732-417d-81a3-573fe821b26d" containerID="d0ebea91d7cab43bb0777a4dc747e6d73a7eab735ebf1eea102733e045c246c7" exitCode=137 Jan 27 15:29:30 crc kubenswrapper[4772]: I0127 15:29:30.374432 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"bd9ac534-7732-417d-81a3-573fe821b26d","Type":"ContainerDied","Data":"d0ebea91d7cab43bb0777a4dc747e6d73a7eab735ebf1eea102733e045c246c7"} Jan 27 15:29:30 crc kubenswrapper[4772]: I0127 15:29:30.485379 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 27 15:29:30 crc kubenswrapper[4772]: I0127 15:29:30.586412 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd9ac534-7732-417d-81a3-573fe821b26d-config-data\") pod \"bd9ac534-7732-417d-81a3-573fe821b26d\" (UID: \"bd9ac534-7732-417d-81a3-573fe821b26d\") " Jan 27 15:29:30 crc kubenswrapper[4772]: I0127 15:29:30.586614 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txnx2\" (UniqueName: \"kubernetes.io/projected/bd9ac534-7732-417d-81a3-573fe821b26d-kube-api-access-txnx2\") pod \"bd9ac534-7732-417d-81a3-573fe821b26d\" (UID: \"bd9ac534-7732-417d-81a3-573fe821b26d\") " Jan 27 15:29:30 crc kubenswrapper[4772]: I0127 15:29:30.586659 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd9ac534-7732-417d-81a3-573fe821b26d-combined-ca-bundle\") pod \"bd9ac534-7732-417d-81a3-573fe821b26d\" (UID: \"bd9ac534-7732-417d-81a3-573fe821b26d\") " Jan 27 15:29:30 crc kubenswrapper[4772]: I0127 15:29:30.608679 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd9ac534-7732-417d-81a3-573fe821b26d-kube-api-access-txnx2" (OuterVolumeSpecName: "kube-api-access-txnx2") pod "bd9ac534-7732-417d-81a3-573fe821b26d" (UID: "bd9ac534-7732-417d-81a3-573fe821b26d"). InnerVolumeSpecName "kube-api-access-txnx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:29:30 crc kubenswrapper[4772]: E0127 15:29:30.659079 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bd9ac534-7732-417d-81a3-573fe821b26d-combined-ca-bundle podName:bd9ac534-7732-417d-81a3-573fe821b26d nodeName:}" failed. No retries permitted until 2026-01-27 15:29:31.159049248 +0000 UTC m=+1357.139658356 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/bd9ac534-7732-417d-81a3-573fe821b26d-combined-ca-bundle") pod "bd9ac534-7732-417d-81a3-573fe821b26d" (UID: "bd9ac534-7732-417d-81a3-573fe821b26d") : error deleting /var/lib/kubelet/pods/bd9ac534-7732-417d-81a3-573fe821b26d/volume-subpaths: remove /var/lib/kubelet/pods/bd9ac534-7732-417d-81a3-573fe821b26d/volume-subpaths: no such file or directory Jan 27 15:29:30 crc kubenswrapper[4772]: I0127 15:29:30.663013 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd9ac534-7732-417d-81a3-573fe821b26d-config-data" (OuterVolumeSpecName: "config-data") pod "bd9ac534-7732-417d-81a3-573fe821b26d" (UID: "bd9ac534-7732-417d-81a3-573fe821b26d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:29:30 crc kubenswrapper[4772]: I0127 15:29:30.688971 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd9ac534-7732-417d-81a3-573fe821b26d-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:30 crc kubenswrapper[4772]: I0127 15:29:30.689019 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txnx2\" (UniqueName: \"kubernetes.io/projected/bd9ac534-7732-417d-81a3-573fe821b26d-kube-api-access-txnx2\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.198935 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd9ac534-7732-417d-81a3-573fe821b26d-combined-ca-bundle\") pod \"bd9ac534-7732-417d-81a3-573fe821b26d\" (UID: \"bd9ac534-7732-417d-81a3-573fe821b26d\") " Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.202436 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd9ac534-7732-417d-81a3-573fe821b26d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd9ac534-7732-417d-81a3-573fe821b26d" (UID: "bd9ac534-7732-417d-81a3-573fe821b26d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.301062 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd9ac534-7732-417d-81a3-573fe821b26d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.385201 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"bd9ac534-7732-417d-81a3-573fe821b26d","Type":"ContainerDied","Data":"1fa0fda721794873849780c939c36b0ae92fda6799dd0ff519efa3b4fb4008b6"} Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.385525 4772 scope.go:117] "RemoveContainer" containerID="d0ebea91d7cab43bb0777a4dc747e6d73a7eab735ebf1eea102733e045c246c7" Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.385274 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.420615 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.430906 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.447714 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 27 15:29:31 crc kubenswrapper[4772]: E0127 15:29:31.448291 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd9ac534-7732-417d-81a3-573fe821b26d" containerName="nova-cell0-conductor-conductor" Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.448311 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd9ac534-7732-417d-81a3-573fe821b26d" containerName="nova-cell0-conductor-conductor" Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.448523 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd9ac534-7732-417d-81a3-573fe821b26d" containerName="nova-cell0-conductor-conductor" Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.449235 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.451554 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bq7vb" Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.451606 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.456691 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.607357 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kbc8\" (UniqueName: \"kubernetes.io/projected/b20b9215-5398-4100-bac4-763daa5ed222-kube-api-access-8kbc8\") pod \"nova-cell0-conductor-0\" (UID: \"b20b9215-5398-4100-bac4-763daa5ed222\") " pod="openstack/nova-cell0-conductor-0" Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.607449 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b20b9215-5398-4100-bac4-763daa5ed222-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b20b9215-5398-4100-bac4-763daa5ed222\") " pod="openstack/nova-cell0-conductor-0" Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.607577 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b20b9215-5398-4100-bac4-763daa5ed222-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b20b9215-5398-4100-bac4-763daa5ed222\") " pod="openstack/nova-cell0-conductor-0" Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.709137 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kbc8\" (UniqueName: \"kubernetes.io/projected/b20b9215-5398-4100-bac4-763daa5ed222-kube-api-access-8kbc8\") pod \"nova-cell0-conductor-0\" (UID: \"b20b9215-5398-4100-bac4-763daa5ed222\") " pod="openstack/nova-cell0-conductor-0" Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.709508 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b20b9215-5398-4100-bac4-763daa5ed222-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b20b9215-5398-4100-bac4-763daa5ed222\") " pod="openstack/nova-cell0-conductor-0" Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.709653 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b20b9215-5398-4100-bac4-763daa5ed222-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b20b9215-5398-4100-bac4-763daa5ed222\") " pod="openstack/nova-cell0-conductor-0" Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.716848 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b20b9215-5398-4100-bac4-763daa5ed222-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b20b9215-5398-4100-bac4-763daa5ed222\") " pod="openstack/nova-cell0-conductor-0" Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.717101 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b20b9215-5398-4100-bac4-763daa5ed222-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b20b9215-5398-4100-bac4-763daa5ed222\") " pod="openstack/nova-cell0-conductor-0" Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.734702 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kbc8\" (UniqueName: \"kubernetes.io/projected/b20b9215-5398-4100-bac4-763daa5ed222-kube-api-access-8kbc8\") pod \"nova-cell0-conductor-0\" (UID: \"b20b9215-5398-4100-bac4-763daa5ed222\") " pod="openstack/nova-cell0-conductor-0" Jan 27 15:29:31 crc kubenswrapper[4772]: I0127 15:29:31.770595 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 27 15:29:32 crc kubenswrapper[4772]: I0127 15:29:32.285029 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 27 15:29:32 crc kubenswrapper[4772]: I0127 15:29:32.402238 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"b20b9215-5398-4100-bac4-763daa5ed222","Type":"ContainerStarted","Data":"09c4f5b1f70267b595eb90b2b27556cd2ace28d9594da5210f17e313d0d8a29a"} Jan 27 15:29:32 crc kubenswrapper[4772]: I0127 15:29:32.681520 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd9ac534-7732-417d-81a3-573fe821b26d" path="/var/lib/kubelet/pods/bd9ac534-7732-417d-81a3-573fe821b26d/volumes" Jan 27 15:29:33 crc kubenswrapper[4772]: I0127 15:29:33.411573 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"b20b9215-5398-4100-bac4-763daa5ed222","Type":"ContainerStarted","Data":"2e74d40bce110215c2607c9a9b716bf9d9db61e446fe99bb897518adbdc86d83"} Jan 27 15:29:33 crc kubenswrapper[4772]: I0127 15:29:33.411692 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Jan 27 15:29:33 crc kubenswrapper[4772]: I0127 15:29:33.435156 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.435130801 podStartE2EDuration="2.435130801s" podCreationTimestamp="2026-01-27 15:29:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:29:33.425531332 +0000 UTC m=+1359.406140440" watchObservedRunningTime="2026-01-27 15:29:33.435130801 +0000 UTC m=+1359.415739899" Jan 27 15:29:34 crc kubenswrapper[4772]: I0127 15:29:34.536822 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Jan 27 15:29:38 crc kubenswrapper[4772]: I0127 15:29:38.007106 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 27 15:29:38 crc kubenswrapper[4772]: I0127 15:29:38.007969 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="1ef66151-0ea7-4696-9db0-7b6665731670" containerName="kube-state-metrics" containerID="cri-o://e93f9f446173d4fd985d40db28827a7f313c9dbe0522a2d3003fa93c8ac7de5e" gracePeriod=30 Jan 27 15:29:38 crc kubenswrapper[4772]: I0127 15:29:38.471197 4772 generic.go:334] "Generic (PLEG): container finished" podID="1ef66151-0ea7-4696-9db0-7b6665731670" containerID="e93f9f446173d4fd985d40db28827a7f313c9dbe0522a2d3003fa93c8ac7de5e" exitCode=2 Jan 27 15:29:38 crc kubenswrapper[4772]: I0127 15:29:38.471302 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1ef66151-0ea7-4696-9db0-7b6665731670","Type":"ContainerDied","Data":"e93f9f446173d4fd985d40db28827a7f313c9dbe0522a2d3003fa93c8ac7de5e"} Jan 27 15:29:38 crc kubenswrapper[4772]: I0127 15:29:38.471611 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1ef66151-0ea7-4696-9db0-7b6665731670","Type":"ContainerDied","Data":"fabcd309d9b92ca01d4a1240a11210e76a8e365a872f6471e0b9d641c3e1ff39"} Jan 27 15:29:38 crc kubenswrapper[4772]: I0127 15:29:38.471633 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fabcd309d9b92ca01d4a1240a11210e76a8e365a872f6471e0b9d641c3e1ff39" Jan 27 15:29:38 crc kubenswrapper[4772]: I0127 15:29:38.520699 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 27 15:29:38 crc kubenswrapper[4772]: I0127 15:29:38.680770 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znb2g\" (UniqueName: \"kubernetes.io/projected/1ef66151-0ea7-4696-9db0-7b6665731670-kube-api-access-znb2g\") pod \"1ef66151-0ea7-4696-9db0-7b6665731670\" (UID: \"1ef66151-0ea7-4696-9db0-7b6665731670\") " Jan 27 15:29:38 crc kubenswrapper[4772]: I0127 15:29:38.688683 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ef66151-0ea7-4696-9db0-7b6665731670-kube-api-access-znb2g" (OuterVolumeSpecName: "kube-api-access-znb2g") pod "1ef66151-0ea7-4696-9db0-7b6665731670" (UID: "1ef66151-0ea7-4696-9db0-7b6665731670"). InnerVolumeSpecName "kube-api-access-znb2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:29:38 crc kubenswrapper[4772]: I0127 15:29:38.782628 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znb2g\" (UniqueName: \"kubernetes.io/projected/1ef66151-0ea7-4696-9db0-7b6665731670-kube-api-access-znb2g\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.479204 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.516225 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.526585 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.545831 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Jan 27 15:29:39 crc kubenswrapper[4772]: E0127 15:29:39.546611 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ef66151-0ea7-4696-9db0-7b6665731670" containerName="kube-state-metrics" Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.546935 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ef66151-0ea7-4696-9db0-7b6665731670" containerName="kube-state-metrics" Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.547263 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ef66151-0ea7-4696-9db0-7b6665731670" containerName="kube-state-metrics" Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.548085 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.550428 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.550593 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.564738 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.698980 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/21f54218-5889-4ae9-a7a1-7ed4895ad63c-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"21f54218-5889-4ae9-a7a1-7ed4895ad63c\") " pod="openstack/kube-state-metrics-0" Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.699043 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21f54218-5889-4ae9-a7a1-7ed4895ad63c-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"21f54218-5889-4ae9-a7a1-7ed4895ad63c\") " pod="openstack/kube-state-metrics-0" Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.699081 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/21f54218-5889-4ae9-a7a1-7ed4895ad63c-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"21f54218-5889-4ae9-a7a1-7ed4895ad63c\") " pod="openstack/kube-state-metrics-0" Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.699191 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nblgh\" (UniqueName: \"kubernetes.io/projected/21f54218-5889-4ae9-a7a1-7ed4895ad63c-kube-api-access-nblgh\") pod \"kube-state-metrics-0\" (UID: \"21f54218-5889-4ae9-a7a1-7ed4895ad63c\") " pod="openstack/kube-state-metrics-0" Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.745272 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.745597 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3ebbc7c3-09c3-4524-854b-e0d64400ab93" containerName="ceilometer-central-agent" containerID="cri-o://c83d73369af2ff787dcb6d403df3f89cc18f8c344ca9a168a243d6708efe0e93" gracePeriod=30 Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.745735 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3ebbc7c3-09c3-4524-854b-e0d64400ab93" containerName="proxy-httpd" containerID="cri-o://c37e3598e7c9e4e7088ad7d7ea7e00efe5c231d10378509f82fb7d2b9df0b502" gracePeriod=30 Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.745795 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3ebbc7c3-09c3-4524-854b-e0d64400ab93" containerName="sg-core" containerID="cri-o://83aac8ec33dfd5bc2c51a863d771b4120a23b62bf31ab0c79c0cd15ea11a7398" gracePeriod=30 Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.745841 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3ebbc7c3-09c3-4524-854b-e0d64400ab93" containerName="ceilometer-notification-agent" containerID="cri-o://3b39e9533e1851e42935b988f031239d201f08489a2d644eb389253a63ab74a9" gracePeriod=30 Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.800826 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nblgh\" (UniqueName: \"kubernetes.io/projected/21f54218-5889-4ae9-a7a1-7ed4895ad63c-kube-api-access-nblgh\") pod \"kube-state-metrics-0\" (UID: \"21f54218-5889-4ae9-a7a1-7ed4895ad63c\") " pod="openstack/kube-state-metrics-0" Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.801288 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/21f54218-5889-4ae9-a7a1-7ed4895ad63c-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"21f54218-5889-4ae9-a7a1-7ed4895ad63c\") " pod="openstack/kube-state-metrics-0" Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.801400 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21f54218-5889-4ae9-a7a1-7ed4895ad63c-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"21f54218-5889-4ae9-a7a1-7ed4895ad63c\") " pod="openstack/kube-state-metrics-0" Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.801507 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/21f54218-5889-4ae9-a7a1-7ed4895ad63c-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"21f54218-5889-4ae9-a7a1-7ed4895ad63c\") " pod="openstack/kube-state-metrics-0" Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.807322 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/21f54218-5889-4ae9-a7a1-7ed4895ad63c-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"21f54218-5889-4ae9-a7a1-7ed4895ad63c\") " pod="openstack/kube-state-metrics-0" Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.807418 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21f54218-5889-4ae9-a7a1-7ed4895ad63c-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"21f54218-5889-4ae9-a7a1-7ed4895ad63c\") " pod="openstack/kube-state-metrics-0" Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.813418 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/21f54218-5889-4ae9-a7a1-7ed4895ad63c-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"21f54218-5889-4ae9-a7a1-7ed4895ad63c\") " pod="openstack/kube-state-metrics-0" Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.822623 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nblgh\" (UniqueName: \"kubernetes.io/projected/21f54218-5889-4ae9-a7a1-7ed4895ad63c-kube-api-access-nblgh\") pod \"kube-state-metrics-0\" (UID: \"21f54218-5889-4ae9-a7a1-7ed4895ad63c\") " pod="openstack/kube-state-metrics-0" Jan 27 15:29:39 crc kubenswrapper[4772]: I0127 15:29:39.876239 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 27 15:29:40 crc kubenswrapper[4772]: W0127 15:29:40.329570 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21f54218_5889_4ae9_a7a1_7ed4895ad63c.slice/crio-cf3bc864ff0528c25cfa09a147802c26a644517c099a85fc5bafd7c4da9534c3 WatchSource:0}: Error finding container cf3bc864ff0528c25cfa09a147802c26a644517c099a85fc5bafd7c4da9534c3: Status 404 returned error can't find the container with id cf3bc864ff0528c25cfa09a147802c26a644517c099a85fc5bafd7c4da9534c3 Jan 27 15:29:40 crc kubenswrapper[4772]: I0127 15:29:40.331733 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 27 15:29:40 crc kubenswrapper[4772]: I0127 15:29:40.332115 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 15:29:40 crc kubenswrapper[4772]: I0127 15:29:40.492594 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"21f54218-5889-4ae9-a7a1-7ed4895ad63c","Type":"ContainerStarted","Data":"cf3bc864ff0528c25cfa09a147802c26a644517c099a85fc5bafd7c4da9534c3"} Jan 27 15:29:40 crc kubenswrapper[4772]: I0127 15:29:40.496966 4772 generic.go:334] "Generic (PLEG): container finished" podID="3ebbc7c3-09c3-4524-854b-e0d64400ab93" containerID="c37e3598e7c9e4e7088ad7d7ea7e00efe5c231d10378509f82fb7d2b9df0b502" exitCode=0 Jan 27 15:29:40 crc kubenswrapper[4772]: I0127 15:29:40.497002 4772 generic.go:334] "Generic (PLEG): container finished" podID="3ebbc7c3-09c3-4524-854b-e0d64400ab93" containerID="83aac8ec33dfd5bc2c51a863d771b4120a23b62bf31ab0c79c0cd15ea11a7398" exitCode=2 Jan 27 15:29:40 crc kubenswrapper[4772]: I0127 15:29:40.497014 4772 generic.go:334] "Generic (PLEG): container finished" podID="3ebbc7c3-09c3-4524-854b-e0d64400ab93" containerID="c83d73369af2ff787dcb6d403df3f89cc18f8c344ca9a168a243d6708efe0e93" exitCode=0 Jan 27 15:29:40 crc kubenswrapper[4772]: I0127 15:29:40.497036 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ebbc7c3-09c3-4524-854b-e0d64400ab93","Type":"ContainerDied","Data":"c37e3598e7c9e4e7088ad7d7ea7e00efe5c231d10378509f82fb7d2b9df0b502"} Jan 27 15:29:40 crc kubenswrapper[4772]: I0127 15:29:40.497064 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ebbc7c3-09c3-4524-854b-e0d64400ab93","Type":"ContainerDied","Data":"83aac8ec33dfd5bc2c51a863d771b4120a23b62bf31ab0c79c0cd15ea11a7398"} Jan 27 15:29:40 crc kubenswrapper[4772]: I0127 15:29:40.497076 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ebbc7c3-09c3-4524-854b-e0d64400ab93","Type":"ContainerDied","Data":"c83d73369af2ff787dcb6d403df3f89cc18f8c344ca9a168a243d6708efe0e93"} Jan 27 15:29:40 crc kubenswrapper[4772]: I0127 15:29:40.674751 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ef66151-0ea7-4696-9db0-7b6665731670" path="/var/lib/kubelet/pods/1ef66151-0ea7-4696-9db0-7b6665731670/volumes" Jan 27 15:29:41 crc kubenswrapper[4772]: I0127 15:29:41.506826 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"21f54218-5889-4ae9-a7a1-7ed4895ad63c","Type":"ContainerStarted","Data":"670d5287e2a9882bc2137122191964eb76c57b36df9c904f50db621c1141ab98"} Jan 27 15:29:41 crc kubenswrapper[4772]: I0127 15:29:41.507411 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Jan 27 15:29:41 crc kubenswrapper[4772]: I0127 15:29:41.530936 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.128304677 podStartE2EDuration="2.530913812s" podCreationTimestamp="2026-01-27 15:29:39 +0000 UTC" firstStartedPulling="2026-01-27 15:29:40.331783248 +0000 UTC m=+1366.312392346" lastFinishedPulling="2026-01-27 15:29:40.734392373 +0000 UTC m=+1366.715001481" observedRunningTime="2026-01-27 15:29:41.523129397 +0000 UTC m=+1367.503738505" watchObservedRunningTime="2026-01-27 15:29:41.530913812 +0000 UTC m=+1367.511522910" Jan 27 15:29:41 crc kubenswrapper[4772]: I0127 15:29:41.799622 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.059071 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.059147 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.059214 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.059730 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1d1c45659af37dbb5fcad6152d119ca4f804c58006a54555795ff000f3b7aea9"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.059788 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://1d1c45659af37dbb5fcad6152d119ca4f804c58006a54555795ff000f3b7aea9" gracePeriod=600 Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.439038 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-h5ch7"] Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.440100 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-h5ch7" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.443316 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.448664 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.453381 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-h5ch7"] Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.519664 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="1d1c45659af37dbb5fcad6152d119ca4f804c58006a54555795ff000f3b7aea9" exitCode=0 Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.519737 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"1d1c45659af37dbb5fcad6152d119ca4f804c58006a54555795ff000f3b7aea9"} Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.519794 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2"} Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.519814 4772 scope.go:117] "RemoveContainer" containerID="ed9bc8d4920540552bc96f7af996996e69c893224418d74c897e7298ed107163" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.549214 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea2e7e0f-aef9-4687-932c-d21f24fd4bff-scripts\") pod \"nova-cell0-cell-mapping-h5ch7\" (UID: \"ea2e7e0f-aef9-4687-932c-d21f24fd4bff\") " pod="openstack/nova-cell0-cell-mapping-h5ch7" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.549303 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea2e7e0f-aef9-4687-932c-d21f24fd4bff-config-data\") pod \"nova-cell0-cell-mapping-h5ch7\" (UID: \"ea2e7e0f-aef9-4687-932c-d21f24fd4bff\") " pod="openstack/nova-cell0-cell-mapping-h5ch7" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.549345 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea2e7e0f-aef9-4687-932c-d21f24fd4bff-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-h5ch7\" (UID: \"ea2e7e0f-aef9-4687-932c-d21f24fd4bff\") " pod="openstack/nova-cell0-cell-mapping-h5ch7" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.549376 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftl7j\" (UniqueName: \"kubernetes.io/projected/ea2e7e0f-aef9-4687-932c-d21f24fd4bff-kube-api-access-ftl7j\") pod \"nova-cell0-cell-mapping-h5ch7\" (UID: \"ea2e7e0f-aef9-4687-932c-d21f24fd4bff\") " pod="openstack/nova-cell0-cell-mapping-h5ch7" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.638936 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.640337 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.644951 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.651544 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea2e7e0f-aef9-4687-932c-d21f24fd4bff-scripts\") pod \"nova-cell0-cell-mapping-h5ch7\" (UID: \"ea2e7e0f-aef9-4687-932c-d21f24fd4bff\") " pod="openstack/nova-cell0-cell-mapping-h5ch7" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.651651 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea2e7e0f-aef9-4687-932c-d21f24fd4bff-config-data\") pod \"nova-cell0-cell-mapping-h5ch7\" (UID: \"ea2e7e0f-aef9-4687-932c-d21f24fd4bff\") " pod="openstack/nova-cell0-cell-mapping-h5ch7" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.651737 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea2e7e0f-aef9-4687-932c-d21f24fd4bff-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-h5ch7\" (UID: \"ea2e7e0f-aef9-4687-932c-d21f24fd4bff\") " pod="openstack/nova-cell0-cell-mapping-h5ch7" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.651804 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftl7j\" (UniqueName: \"kubernetes.io/projected/ea2e7e0f-aef9-4687-932c-d21f24fd4bff-kube-api-access-ftl7j\") pod \"nova-cell0-cell-mapping-h5ch7\" (UID: \"ea2e7e0f-aef9-4687-932c-d21f24fd4bff\") " pod="openstack/nova-cell0-cell-mapping-h5ch7" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.659402 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.660855 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.667797 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea2e7e0f-aef9-4687-932c-d21f24fd4bff-scripts\") pod \"nova-cell0-cell-mapping-h5ch7\" (UID: \"ea2e7e0f-aef9-4687-932c-d21f24fd4bff\") " pod="openstack/nova-cell0-cell-mapping-h5ch7" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.668051 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.668339 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea2e7e0f-aef9-4687-932c-d21f24fd4bff-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-h5ch7\" (UID: \"ea2e7e0f-aef9-4687-932c-d21f24fd4bff\") " pod="openstack/nova-cell0-cell-mapping-h5ch7" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.690861 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea2e7e0f-aef9-4687-932c-d21f24fd4bff-config-data\") pod \"nova-cell0-cell-mapping-h5ch7\" (UID: \"ea2e7e0f-aef9-4687-932c-d21f24fd4bff\") " pod="openstack/nova-cell0-cell-mapping-h5ch7" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.701400 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.722187 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftl7j\" (UniqueName: \"kubernetes.io/projected/ea2e7e0f-aef9-4687-932c-d21f24fd4bff-kube-api-access-ftl7j\") pod \"nova-cell0-cell-mapping-h5ch7\" (UID: \"ea2e7e0f-aef9-4687-932c-d21f24fd4bff\") " pod="openstack/nova-cell0-cell-mapping-h5ch7" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.726241 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.756465 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgjrt\" (UniqueName: \"kubernetes.io/projected/8a218ca3-a163-4f9c-8e73-f630b2228bb2-kube-api-access-hgjrt\") pod \"nova-cell1-novncproxy-0\" (UID: \"8a218ca3-a163-4f9c-8e73-f630b2228bb2\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.756591 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a79db711-df56-46f6-93c2-7e1e5c914ba6-config-data\") pod \"nova-api-0\" (UID: \"a79db711-df56-46f6-93c2-7e1e5c914ba6\") " pod="openstack/nova-api-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.756619 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2vn6\" (UniqueName: \"kubernetes.io/projected/a79db711-df56-46f6-93c2-7e1e5c914ba6-kube-api-access-t2vn6\") pod \"nova-api-0\" (UID: \"a79db711-df56-46f6-93c2-7e1e5c914ba6\") " pod="openstack/nova-api-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.756675 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a79db711-df56-46f6-93c2-7e1e5c914ba6-logs\") pod \"nova-api-0\" (UID: \"a79db711-df56-46f6-93c2-7e1e5c914ba6\") " pod="openstack/nova-api-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.756762 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a218ca3-a163-4f9c-8e73-f630b2228bb2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8a218ca3-a163-4f9c-8e73-f630b2228bb2\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.756813 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a79db711-df56-46f6-93c2-7e1e5c914ba6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a79db711-df56-46f6-93c2-7e1e5c914ba6\") " pod="openstack/nova-api-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.756927 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a218ca3-a163-4f9c-8e73-f630b2228bb2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8a218ca3-a163-4f9c-8e73-f630b2228bb2\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.767789 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-h5ch7" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.781053 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.783495 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.787682 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.834452 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.858430 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a218ca3-a163-4f9c-8e73-f630b2228bb2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8a218ca3-a163-4f9c-8e73-f630b2228bb2\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.858885 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgjrt\" (UniqueName: \"kubernetes.io/projected/8a218ca3-a163-4f9c-8e73-f630b2228bb2-kube-api-access-hgjrt\") pod \"nova-cell1-novncproxy-0\" (UID: \"8a218ca3-a163-4f9c-8e73-f630b2228bb2\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.858946 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a79db711-df56-46f6-93c2-7e1e5c914ba6-config-data\") pod \"nova-api-0\" (UID: \"a79db711-df56-46f6-93c2-7e1e5c914ba6\") " pod="openstack/nova-api-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.858972 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2vn6\" (UniqueName: \"kubernetes.io/projected/a79db711-df56-46f6-93c2-7e1e5c914ba6-kube-api-access-t2vn6\") pod \"nova-api-0\" (UID: \"a79db711-df56-46f6-93c2-7e1e5c914ba6\") " pod="openstack/nova-api-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.859004 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a79db711-df56-46f6-93c2-7e1e5c914ba6-logs\") pod \"nova-api-0\" (UID: \"a79db711-df56-46f6-93c2-7e1e5c914ba6\") " pod="openstack/nova-api-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.859061 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a218ca3-a163-4f9c-8e73-f630b2228bb2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8a218ca3-a163-4f9c-8e73-f630b2228bb2\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.859103 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a79db711-df56-46f6-93c2-7e1e5c914ba6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a79db711-df56-46f6-93c2-7e1e5c914ba6\") " pod="openstack/nova-api-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.861292 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a79db711-df56-46f6-93c2-7e1e5c914ba6-logs\") pod \"nova-api-0\" (UID: \"a79db711-df56-46f6-93c2-7e1e5c914ba6\") " pod="openstack/nova-api-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.877128 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a218ca3-a163-4f9c-8e73-f630b2228bb2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8a218ca3-a163-4f9c-8e73-f630b2228bb2\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.887597 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a79db711-df56-46f6-93c2-7e1e5c914ba6-config-data\") pod \"nova-api-0\" (UID: \"a79db711-df56-46f6-93c2-7e1e5c914ba6\") " pod="openstack/nova-api-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.889890 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a79db711-df56-46f6-93c2-7e1e5c914ba6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a79db711-df56-46f6-93c2-7e1e5c914ba6\") " pod="openstack/nova-api-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.890321 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a218ca3-a163-4f9c-8e73-f630b2228bb2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8a218ca3-a163-4f9c-8e73-f630b2228bb2\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.907707 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgjrt\" (UniqueName: \"kubernetes.io/projected/8a218ca3-a163-4f9c-8e73-f630b2228bb2-kube-api-access-hgjrt\") pod \"nova-cell1-novncproxy-0\" (UID: \"8a218ca3-a163-4f9c-8e73-f630b2228bb2\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.908295 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2vn6\" (UniqueName: \"kubernetes.io/projected/a79db711-df56-46f6-93c2-7e1e5c914ba6-kube-api-access-t2vn6\") pod \"nova-api-0\" (UID: \"a79db711-df56-46f6-93c2-7e1e5c914ba6\") " pod="openstack/nova-api-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.912928 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.914386 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.917898 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.941436 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.964400 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/325e0266-afde-43a7-b77c-4b29a2d55c3a-config-data\") pod \"nova-metadata-0\" (UID: \"325e0266-afde-43a7-b77c-4b29a2d55c3a\") " pod="openstack/nova-metadata-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.964441 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/325e0266-afde-43a7-b77c-4b29a2d55c3a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"325e0266-afde-43a7-b77c-4b29a2d55c3a\") " pod="openstack/nova-metadata-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.964481 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xml5\" (UniqueName: \"kubernetes.io/projected/325e0266-afde-43a7-b77c-4b29a2d55c3a-kube-api-access-6xml5\") pod \"nova-metadata-0\" (UID: \"325e0266-afde-43a7-b77c-4b29a2d55c3a\") " pod="openstack/nova-metadata-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.964604 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/325e0266-afde-43a7-b77c-4b29a2d55c3a-logs\") pod \"nova-metadata-0\" (UID: \"325e0266-afde-43a7-b77c-4b29a2d55c3a\") " pod="openstack/nova-metadata-0" Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.983982 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-88t2p"] Jan 27 15:29:42 crc kubenswrapper[4772]: I0127 15:29:42.985553 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.002020 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-88t2p"] Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.065952 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/325e0266-afde-43a7-b77c-4b29a2d55c3a-config-data\") pod \"nova-metadata-0\" (UID: \"325e0266-afde-43a7-b77c-4b29a2d55c3a\") " pod="openstack/nova-metadata-0" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.065999 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/325e0266-afde-43a7-b77c-4b29a2d55c3a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"325e0266-afde-43a7-b77c-4b29a2d55c3a\") " pod="openstack/nova-metadata-0" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.066038 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bfae3d5-7017-4ced-9691-4255769c51f6-config-data\") pod \"nova-scheduler-0\" (UID: \"2bfae3d5-7017-4ced-9691-4255769c51f6\") " pod="openstack/nova-scheduler-0" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.066063 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xml5\" (UniqueName: \"kubernetes.io/projected/325e0266-afde-43a7-b77c-4b29a2d55c3a-kube-api-access-6xml5\") pod \"nova-metadata-0\" (UID: \"325e0266-afde-43a7-b77c-4b29a2d55c3a\") " pod="openstack/nova-metadata-0" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.066084 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/325e0266-afde-43a7-b77c-4b29a2d55c3a-logs\") pod \"nova-metadata-0\" (UID: \"325e0266-afde-43a7-b77c-4b29a2d55c3a\") " pod="openstack/nova-metadata-0" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.066191 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx5gw\" (UniqueName: \"kubernetes.io/projected/2bfae3d5-7017-4ced-9691-4255769c51f6-kube-api-access-cx5gw\") pod \"nova-scheduler-0\" (UID: \"2bfae3d5-7017-4ced-9691-4255769c51f6\") " pod="openstack/nova-scheduler-0" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.066213 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bfae3d5-7017-4ced-9691-4255769c51f6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2bfae3d5-7017-4ced-9691-4255769c51f6\") " pod="openstack/nova-scheduler-0" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.068595 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/325e0266-afde-43a7-b77c-4b29a2d55c3a-logs\") pod \"nova-metadata-0\" (UID: \"325e0266-afde-43a7-b77c-4b29a2d55c3a\") " pod="openstack/nova-metadata-0" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.070023 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/325e0266-afde-43a7-b77c-4b29a2d55c3a-config-data\") pod \"nova-metadata-0\" (UID: \"325e0266-afde-43a7-b77c-4b29a2d55c3a\") " pod="openstack/nova-metadata-0" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.070096 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/325e0266-afde-43a7-b77c-4b29a2d55c3a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"325e0266-afde-43a7-b77c-4b29a2d55c3a\") " pod="openstack/nova-metadata-0" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.098553 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.098831 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xml5\" (UniqueName: \"kubernetes.io/projected/325e0266-afde-43a7-b77c-4b29a2d55c3a-kube-api-access-6xml5\") pod \"nova-metadata-0\" (UID: \"325e0266-afde-43a7-b77c-4b29a2d55c3a\") " pod="openstack/nova-metadata-0" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.136635 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.161272 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.172682 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-88t2p\" (UID: \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\") " pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.172743 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bfae3d5-7017-4ced-9691-4255769c51f6-config-data\") pod \"nova-scheduler-0\" (UID: \"2bfae3d5-7017-4ced-9691-4255769c51f6\") " pod="openstack/nova-scheduler-0" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.172780 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-config\") pod \"dnsmasq-dns-845d6d6f59-88t2p\" (UID: \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\") " pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.172882 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-88t2p\" (UID: \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\") " pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.173188 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-88t2p\" (UID: \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\") " pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.173296 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx5gw\" (UniqueName: \"kubernetes.io/projected/2bfae3d5-7017-4ced-9691-4255769c51f6-kube-api-access-cx5gw\") pod \"nova-scheduler-0\" (UID: \"2bfae3d5-7017-4ced-9691-4255769c51f6\") " pod="openstack/nova-scheduler-0" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.173341 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bfae3d5-7017-4ced-9691-4255769c51f6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2bfae3d5-7017-4ced-9691-4255769c51f6\") " pod="openstack/nova-scheduler-0" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.173435 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-88t2p\" (UID: \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\") " pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.173466 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf2fc\" (UniqueName: \"kubernetes.io/projected/73ee81ee-57fa-466a-8ada-2fa4da5987a0-kube-api-access-vf2fc\") pod \"dnsmasq-dns-845d6d6f59-88t2p\" (UID: \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\") " pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.182414 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bfae3d5-7017-4ced-9691-4255769c51f6-config-data\") pod \"nova-scheduler-0\" (UID: \"2bfae3d5-7017-4ced-9691-4255769c51f6\") " pod="openstack/nova-scheduler-0" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.182718 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bfae3d5-7017-4ced-9691-4255769c51f6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2bfae3d5-7017-4ced-9691-4255769c51f6\") " pod="openstack/nova-scheduler-0" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.210989 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx5gw\" (UniqueName: \"kubernetes.io/projected/2bfae3d5-7017-4ced-9691-4255769c51f6-kube-api-access-cx5gw\") pod \"nova-scheduler-0\" (UID: \"2bfae3d5-7017-4ced-9691-4255769c51f6\") " pod="openstack/nova-scheduler-0" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.275408 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-88t2p\" (UID: \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\") " pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.275531 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-88t2p\" (UID: \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\") " pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.275585 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-88t2p\" (UID: \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\") " pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.275605 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf2fc\" (UniqueName: \"kubernetes.io/projected/73ee81ee-57fa-466a-8ada-2fa4da5987a0-kube-api-access-vf2fc\") pod \"dnsmasq-dns-845d6d6f59-88t2p\" (UID: \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\") " pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.275657 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-88t2p\" (UID: \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\") " pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.275699 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-config\") pod \"dnsmasq-dns-845d6d6f59-88t2p\" (UID: \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\") " pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.277339 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-config\") pod \"dnsmasq-dns-845d6d6f59-88t2p\" (UID: \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\") " pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.277938 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-88t2p\" (UID: \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\") " pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.278586 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-88t2p\" (UID: \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\") " pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.280476 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-88t2p\" (UID: \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\") " pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.283557 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-88t2p\" (UID: \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\") " pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.283932 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.329811 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf2fc\" (UniqueName: \"kubernetes.io/projected/73ee81ee-57fa-466a-8ada-2fa4da5987a0-kube-api-access-vf2fc\") pod \"dnsmasq-dns-845d6d6f59-88t2p\" (UID: \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\") " pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.478515 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.553021 4772 generic.go:334] "Generic (PLEG): container finished" podID="3ebbc7c3-09c3-4524-854b-e0d64400ab93" containerID="3b39e9533e1851e42935b988f031239d201f08489a2d644eb389253a63ab74a9" exitCode=0 Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.553320 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.553348 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ebbc7c3-09c3-4524-854b-e0d64400ab93","Type":"ContainerDied","Data":"3b39e9533e1851e42935b988f031239d201f08489a2d644eb389253a63ab74a9"} Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.553684 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ebbc7c3-09c3-4524-854b-e0d64400ab93","Type":"ContainerDied","Data":"78cb40c125f03659c88535f65c13efc1a1776dd09e65791b32311c60251b3ac1"} Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.553713 4772 scope.go:117] "RemoveContainer" containerID="c37e3598e7c9e4e7088ad7d7ea7e00efe5c231d10378509f82fb7d2b9df0b502" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.583487 4772 scope.go:117] "RemoveContainer" containerID="83aac8ec33dfd5bc2c51a863d771b4120a23b62bf31ab0c79c0cd15ea11a7398" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.584213 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ebbc7c3-09c3-4524-854b-e0d64400ab93-run-httpd\") pod \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.584307 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ebbc7c3-09c3-4524-854b-e0d64400ab93-config-data\") pod \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.584370 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ebbc7c3-09c3-4524-854b-e0d64400ab93-log-httpd\") pod \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.584479 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ebbc7c3-09c3-4524-854b-e0d64400ab93-sg-core-conf-yaml\") pod \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.584549 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ebbc7c3-09c3-4524-854b-e0d64400ab93-scripts\") pod \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.584663 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ebbc7c3-09c3-4524-854b-e0d64400ab93-combined-ca-bundle\") pod \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.584697 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27d6x\" (UniqueName: \"kubernetes.io/projected/3ebbc7c3-09c3-4524-854b-e0d64400ab93-kube-api-access-27d6x\") pod \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\" (UID: \"3ebbc7c3-09c3-4524-854b-e0d64400ab93\") " Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.585428 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ebbc7c3-09c3-4524-854b-e0d64400ab93-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3ebbc7c3-09c3-4524-854b-e0d64400ab93" (UID: "3ebbc7c3-09c3-4524-854b-e0d64400ab93"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.585791 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ebbc7c3-09c3-4524-854b-e0d64400ab93-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3ebbc7c3-09c3-4524-854b-e0d64400ab93" (UID: "3ebbc7c3-09c3-4524-854b-e0d64400ab93"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.602456 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ebbc7c3-09c3-4524-854b-e0d64400ab93-scripts" (OuterVolumeSpecName: "scripts") pod "3ebbc7c3-09c3-4524-854b-e0d64400ab93" (UID: "3ebbc7c3-09c3-4524-854b-e0d64400ab93"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.625865 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.640705 4772 scope.go:117] "RemoveContainer" containerID="3b39e9533e1851e42935b988f031239d201f08489a2d644eb389253a63ab74a9" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.640932 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ebbc7c3-09c3-4524-854b-e0d64400ab93-kube-api-access-27d6x" (OuterVolumeSpecName: "kube-api-access-27d6x") pod "3ebbc7c3-09c3-4524-854b-e0d64400ab93" (UID: "3ebbc7c3-09c3-4524-854b-e0d64400ab93"). InnerVolumeSpecName "kube-api-access-27d6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.652557 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ebbc7c3-09c3-4524-854b-e0d64400ab93-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3ebbc7c3-09c3-4524-854b-e0d64400ab93" (UID: "3ebbc7c3-09c3-4524-854b-e0d64400ab93"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.694671 4772 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ebbc7c3-09c3-4524-854b-e0d64400ab93-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.694703 4772 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ebbc7c3-09c3-4524-854b-e0d64400ab93-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.694718 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ebbc7c3-09c3-4524-854b-e0d64400ab93-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.694732 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27d6x\" (UniqueName: \"kubernetes.io/projected/3ebbc7c3-09c3-4524-854b-e0d64400ab93-kube-api-access-27d6x\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.694744 4772 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ebbc7c3-09c3-4524-854b-e0d64400ab93-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.697122 4772 scope.go:117] "RemoveContainer" containerID="c83d73369af2ff787dcb6d403df3f89cc18f8c344ca9a168a243d6708efe0e93" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.761439 4772 scope.go:117] "RemoveContainer" containerID="c37e3598e7c9e4e7088ad7d7ea7e00efe5c231d10378509f82fb7d2b9df0b502" Jan 27 15:29:43 crc kubenswrapper[4772]: E0127 15:29:43.763914 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c37e3598e7c9e4e7088ad7d7ea7e00efe5c231d10378509f82fb7d2b9df0b502\": container with ID starting with c37e3598e7c9e4e7088ad7d7ea7e00efe5c231d10378509f82fb7d2b9df0b502 not found: ID does not exist" containerID="c37e3598e7c9e4e7088ad7d7ea7e00efe5c231d10378509f82fb7d2b9df0b502" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.763957 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c37e3598e7c9e4e7088ad7d7ea7e00efe5c231d10378509f82fb7d2b9df0b502"} err="failed to get container status \"c37e3598e7c9e4e7088ad7d7ea7e00efe5c231d10378509f82fb7d2b9df0b502\": rpc error: code = NotFound desc = could not find container \"c37e3598e7c9e4e7088ad7d7ea7e00efe5c231d10378509f82fb7d2b9df0b502\": container with ID starting with c37e3598e7c9e4e7088ad7d7ea7e00efe5c231d10378509f82fb7d2b9df0b502 not found: ID does not exist" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.763986 4772 scope.go:117] "RemoveContainer" containerID="83aac8ec33dfd5bc2c51a863d771b4120a23b62bf31ab0c79c0cd15ea11a7398" Jan 27 15:29:43 crc kubenswrapper[4772]: E0127 15:29:43.764555 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83aac8ec33dfd5bc2c51a863d771b4120a23b62bf31ab0c79c0cd15ea11a7398\": container with ID starting with 83aac8ec33dfd5bc2c51a863d771b4120a23b62bf31ab0c79c0cd15ea11a7398 not found: ID does not exist" containerID="83aac8ec33dfd5bc2c51a863d771b4120a23b62bf31ab0c79c0cd15ea11a7398" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.764607 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83aac8ec33dfd5bc2c51a863d771b4120a23b62bf31ab0c79c0cd15ea11a7398"} err="failed to get container status \"83aac8ec33dfd5bc2c51a863d771b4120a23b62bf31ab0c79c0cd15ea11a7398\": rpc error: code = NotFound desc = could not find container \"83aac8ec33dfd5bc2c51a863d771b4120a23b62bf31ab0c79c0cd15ea11a7398\": container with ID starting with 83aac8ec33dfd5bc2c51a863d771b4120a23b62bf31ab0c79c0cd15ea11a7398 not found: ID does not exist" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.764621 4772 scope.go:117] "RemoveContainer" containerID="3b39e9533e1851e42935b988f031239d201f08489a2d644eb389253a63ab74a9" Jan 27 15:29:43 crc kubenswrapper[4772]: E0127 15:29:43.764965 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b39e9533e1851e42935b988f031239d201f08489a2d644eb389253a63ab74a9\": container with ID starting with 3b39e9533e1851e42935b988f031239d201f08489a2d644eb389253a63ab74a9 not found: ID does not exist" containerID="3b39e9533e1851e42935b988f031239d201f08489a2d644eb389253a63ab74a9" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.764994 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b39e9533e1851e42935b988f031239d201f08489a2d644eb389253a63ab74a9"} err="failed to get container status \"3b39e9533e1851e42935b988f031239d201f08489a2d644eb389253a63ab74a9\": rpc error: code = NotFound desc = could not find container \"3b39e9533e1851e42935b988f031239d201f08489a2d644eb389253a63ab74a9\": container with ID starting with 3b39e9533e1851e42935b988f031239d201f08489a2d644eb389253a63ab74a9 not found: ID does not exist" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.765007 4772 scope.go:117] "RemoveContainer" containerID="c83d73369af2ff787dcb6d403df3f89cc18f8c344ca9a168a243d6708efe0e93" Jan 27 15:29:43 crc kubenswrapper[4772]: E0127 15:29:43.765813 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c83d73369af2ff787dcb6d403df3f89cc18f8c344ca9a168a243d6708efe0e93\": container with ID starting with c83d73369af2ff787dcb6d403df3f89cc18f8c344ca9a168a243d6708efe0e93 not found: ID does not exist" containerID="c83d73369af2ff787dcb6d403df3f89cc18f8c344ca9a168a243d6708efe0e93" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.765860 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c83d73369af2ff787dcb6d403df3f89cc18f8c344ca9a168a243d6708efe0e93"} err="failed to get container status \"c83d73369af2ff787dcb6d403df3f89cc18f8c344ca9a168a243d6708efe0e93\": rpc error: code = NotFound desc = could not find container \"c83d73369af2ff787dcb6d403df3f89cc18f8c344ca9a168a243d6708efe0e93\": container with ID starting with c83d73369af2ff787dcb6d403df3f89cc18f8c344ca9a168a243d6708efe0e93 not found: ID does not exist" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.783495 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ebbc7c3-09c3-4524-854b-e0d64400ab93-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ebbc7c3-09c3-4524-854b-e0d64400ab93" (UID: "3ebbc7c3-09c3-4524-854b-e0d64400ab93"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.790553 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ebbc7c3-09c3-4524-854b-e0d64400ab93-config-data" (OuterVolumeSpecName: "config-data") pod "3ebbc7c3-09c3-4524-854b-e0d64400ab93" (UID: "3ebbc7c3-09c3-4524-854b-e0d64400ab93"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.822369 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ebbc7c3-09c3-4524-854b-e0d64400ab93-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.822435 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ebbc7c3-09c3-4524-854b-e0d64400ab93-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:43 crc kubenswrapper[4772]: W0127 15:29:43.822582 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea2e7e0f_aef9_4687_932c_d21f24fd4bff.slice/crio-68f7be70d41a760b9715eb4d90633fa6d0b418abb8ac102df2c4b1b0cee8fa86 WatchSource:0}: Error finding container 68f7be70d41a760b9715eb4d90633fa6d0b418abb8ac102df2c4b1b0cee8fa86: Status 404 returned error can't find the container with id 68f7be70d41a760b9715eb4d90633fa6d0b418abb8ac102df2c4b1b0cee8fa86 Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.845595 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-h5ch7"] Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.873986 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 27 15:29:43 crc kubenswrapper[4772]: W0127 15:29:43.877809 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda79db711_df56_46f6_93c2_7e1e5c914ba6.slice/crio-ea5b86feefaf42849d7aeffa2e275c403579d3f00ecb406df1ac2ae06ceefa44 WatchSource:0}: Error finding container ea5b86feefaf42849d7aeffa2e275c403579d3f00ecb406df1ac2ae06ceefa44: Status 404 returned error can't find the container with id ea5b86feefaf42849d7aeffa2e275c403579d3f00ecb406df1ac2ae06ceefa44 Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.885886 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 27 15:29:43 crc kubenswrapper[4772]: W0127 15:29:43.910719 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a218ca3_a163_4f9c_8e73_f630b2228bb2.slice/crio-d661ce2ba7ef10af0b86479b7eca76159de7a8a82826e702dc8a584607cd118c WatchSource:0}: Error finding container d661ce2ba7ef10af0b86479b7eca76159de7a8a82826e702dc8a584607cd118c: Status 404 returned error can't find the container with id d661ce2ba7ef10af0b86479b7eca76159de7a8a82826e702dc8a584607cd118c Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.935661 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gjwh2"] Jan 27 15:29:43 crc kubenswrapper[4772]: E0127 15:29:43.936086 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ebbc7c3-09c3-4524-854b-e0d64400ab93" containerName="sg-core" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.936098 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ebbc7c3-09c3-4524-854b-e0d64400ab93" containerName="sg-core" Jan 27 15:29:43 crc kubenswrapper[4772]: E0127 15:29:43.936112 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ebbc7c3-09c3-4524-854b-e0d64400ab93" containerName="ceilometer-notification-agent" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.936118 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ebbc7c3-09c3-4524-854b-e0d64400ab93" containerName="ceilometer-notification-agent" Jan 27 15:29:43 crc kubenswrapper[4772]: E0127 15:29:43.936127 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ebbc7c3-09c3-4524-854b-e0d64400ab93" containerName="proxy-httpd" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.936133 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ebbc7c3-09c3-4524-854b-e0d64400ab93" containerName="proxy-httpd" Jan 27 15:29:43 crc kubenswrapper[4772]: E0127 15:29:43.936161 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ebbc7c3-09c3-4524-854b-e0d64400ab93" containerName="ceilometer-central-agent" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.936179 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ebbc7c3-09c3-4524-854b-e0d64400ab93" containerName="ceilometer-central-agent" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.936431 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ebbc7c3-09c3-4524-854b-e0d64400ab93" containerName="sg-core" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.936453 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ebbc7c3-09c3-4524-854b-e0d64400ab93" containerName="proxy-httpd" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.936463 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ebbc7c3-09c3-4524-854b-e0d64400ab93" containerName="ceilometer-central-agent" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.936475 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ebbc7c3-09c3-4524-854b-e0d64400ab93" containerName="ceilometer-notification-agent" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.937132 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gjwh2" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.942046 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.942278 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Jan 27 15:29:43 crc kubenswrapper[4772]: I0127 15:29:43.971476 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gjwh2"] Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.001308 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.021258 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.028350 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f91bfd1b-6386-444f-95da-045fbe957f5c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-gjwh2\" (UID: \"f91bfd1b-6386-444f-95da-045fbe957f5c\") " pod="openstack/nova-cell1-conductor-db-sync-gjwh2" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.028434 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nh52\" (UniqueName: \"kubernetes.io/projected/f91bfd1b-6386-444f-95da-045fbe957f5c-kube-api-access-7nh52\") pod \"nova-cell1-conductor-db-sync-gjwh2\" (UID: \"f91bfd1b-6386-444f-95da-045fbe957f5c\") " pod="openstack/nova-cell1-conductor-db-sync-gjwh2" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.028550 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f91bfd1b-6386-444f-95da-045fbe957f5c-config-data\") pod \"nova-cell1-conductor-db-sync-gjwh2\" (UID: \"f91bfd1b-6386-444f-95da-045fbe957f5c\") " pod="openstack/nova-cell1-conductor-db-sync-gjwh2" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.028629 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f91bfd1b-6386-444f-95da-045fbe957f5c-scripts\") pod \"nova-cell1-conductor-db-sync-gjwh2\" (UID: \"f91bfd1b-6386-444f-95da-045fbe957f5c\") " pod="openstack/nova-cell1-conductor-db-sync-gjwh2" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.040258 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.042921 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.045675 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.045938 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.046140 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.056534 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.084653 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.108978 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.130324 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f91bfd1b-6386-444f-95da-045fbe957f5c-config-data\") pod \"nova-cell1-conductor-db-sync-gjwh2\" (UID: \"f91bfd1b-6386-444f-95da-045fbe957f5c\") " pod="openstack/nova-cell1-conductor-db-sync-gjwh2" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.130361 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-scripts\") pod \"ceilometer-0\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.130399 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.130421 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/445d3e38-8f68-4dad-9e97-d927d60ee1e4-log-httpd\") pod \"ceilometer-0\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.130453 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f91bfd1b-6386-444f-95da-045fbe957f5c-scripts\") pod \"nova-cell1-conductor-db-sync-gjwh2\" (UID: \"f91bfd1b-6386-444f-95da-045fbe957f5c\") " pod="openstack/nova-cell1-conductor-db-sync-gjwh2" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.130520 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/445d3e38-8f68-4dad-9e97-d927d60ee1e4-run-httpd\") pod \"ceilometer-0\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.130541 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f91bfd1b-6386-444f-95da-045fbe957f5c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-gjwh2\" (UID: \"f91bfd1b-6386-444f-95da-045fbe957f5c\") " pod="openstack/nova-cell1-conductor-db-sync-gjwh2" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.130563 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-config-data\") pod \"ceilometer-0\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.130584 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nh52\" (UniqueName: \"kubernetes.io/projected/f91bfd1b-6386-444f-95da-045fbe957f5c-kube-api-access-7nh52\") pod \"nova-cell1-conductor-db-sync-gjwh2\" (UID: \"f91bfd1b-6386-444f-95da-045fbe957f5c\") " pod="openstack/nova-cell1-conductor-db-sync-gjwh2" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.130601 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.130641 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br29b\" (UniqueName: \"kubernetes.io/projected/445d3e38-8f68-4dad-9e97-d927d60ee1e4-kube-api-access-br29b\") pod \"ceilometer-0\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.130663 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.134613 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f91bfd1b-6386-444f-95da-045fbe957f5c-config-data\") pod \"nova-cell1-conductor-db-sync-gjwh2\" (UID: \"f91bfd1b-6386-444f-95da-045fbe957f5c\") " pod="openstack/nova-cell1-conductor-db-sync-gjwh2" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.135811 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f91bfd1b-6386-444f-95da-045fbe957f5c-scripts\") pod \"nova-cell1-conductor-db-sync-gjwh2\" (UID: \"f91bfd1b-6386-444f-95da-045fbe957f5c\") " pod="openstack/nova-cell1-conductor-db-sync-gjwh2" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.136030 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f91bfd1b-6386-444f-95da-045fbe957f5c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-gjwh2\" (UID: \"f91bfd1b-6386-444f-95da-045fbe957f5c\") " pod="openstack/nova-cell1-conductor-db-sync-gjwh2" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.149865 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nh52\" (UniqueName: \"kubernetes.io/projected/f91bfd1b-6386-444f-95da-045fbe957f5c-kube-api-access-7nh52\") pod \"nova-cell1-conductor-db-sync-gjwh2\" (UID: \"f91bfd1b-6386-444f-95da-045fbe957f5c\") " pod="openstack/nova-cell1-conductor-db-sync-gjwh2" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.232761 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/445d3e38-8f68-4dad-9e97-d927d60ee1e4-run-httpd\") pod \"ceilometer-0\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.232830 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-config-data\") pod \"ceilometer-0\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.232865 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.232929 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br29b\" (UniqueName: \"kubernetes.io/projected/445d3e38-8f68-4dad-9e97-d927d60ee1e4-kube-api-access-br29b\") pod \"ceilometer-0\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.232969 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.233012 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-scripts\") pod \"ceilometer-0\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.233058 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.233089 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/445d3e38-8f68-4dad-9e97-d927d60ee1e4-log-httpd\") pod \"ceilometer-0\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.233778 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/445d3e38-8f68-4dad-9e97-d927d60ee1e4-log-httpd\") pod \"ceilometer-0\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.235139 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/445d3e38-8f68-4dad-9e97-d927d60ee1e4-run-httpd\") pod \"ceilometer-0\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.242521 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.245387 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.248271 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-config-data\") pod \"ceilometer-0\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.251857 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.252804 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-scripts\") pod \"ceilometer-0\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.269073 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br29b\" (UniqueName: \"kubernetes.io/projected/445d3e38-8f68-4dad-9e97-d927d60ee1e4-kube-api-access-br29b\") pod \"ceilometer-0\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.276811 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gjwh2" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.304726 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-88t2p"] Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.399076 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.597833 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8a218ca3-a163-4f9c-8e73-f630b2228bb2","Type":"ContainerStarted","Data":"d661ce2ba7ef10af0b86479b7eca76159de7a8a82826e702dc8a584607cd118c"} Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.610256 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2bfae3d5-7017-4ced-9691-4255769c51f6","Type":"ContainerStarted","Data":"1941b03ee46ef8231c28748227a7ca5084953c3a384610a809f7eb266c2d086a"} Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.626142 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-h5ch7" event={"ID":"ea2e7e0f-aef9-4687-932c-d21f24fd4bff","Type":"ContainerStarted","Data":"bea9ecc5c8bd7f22996f379a16987a5468d25478afcbfdd986751cd73382ded7"} Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.626221 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-h5ch7" event={"ID":"ea2e7e0f-aef9-4687-932c-d21f24fd4bff","Type":"ContainerStarted","Data":"68f7be70d41a760b9715eb4d90633fa6d0b418abb8ac102df2c4b1b0cee8fa86"} Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.659882 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a79db711-df56-46f6-93c2-7e1e5c914ba6","Type":"ContainerStarted","Data":"ea5b86feefaf42849d7aeffa2e275c403579d3f00ecb406df1ac2ae06ceefa44"} Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.660309 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-h5ch7" podStartSLOduration=2.6602895159999997 podStartE2EDuration="2.660289516s" podCreationTimestamp="2026-01-27 15:29:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:29:44.649274608 +0000 UTC m=+1370.629883706" watchObservedRunningTime="2026-01-27 15:29:44.660289516 +0000 UTC m=+1370.640898614" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.720538 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ebbc7c3-09c3-4524-854b-e0d64400ab93" path="/var/lib/kubelet/pods/3ebbc7c3-09c3-4524-854b-e0d64400ab93/volumes" Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.721389 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" event={"ID":"73ee81ee-57fa-466a-8ada-2fa4da5987a0","Type":"ContainerStarted","Data":"79cc249e145e1f853047b2402d998a0aa111ca080edd461a766047221e313d63"} Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.721414 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"325e0266-afde-43a7-b77c-4b29a2d55c3a","Type":"ContainerStarted","Data":"0fbb6b12d9611d132f4209f04274b5892a26d8ef07f59ee5c52d51242f78a833"} Jan 27 15:29:44 crc kubenswrapper[4772]: I0127 15:29:44.857019 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gjwh2"] Jan 27 15:29:45 crc kubenswrapper[4772]: I0127 15:29:45.005610 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:29:45 crc kubenswrapper[4772]: W0127 15:29:45.007460 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod445d3e38_8f68_4dad_9e97_d927d60ee1e4.slice/crio-5f55b8adcf12a6b66983b6beb58f8b720085a1c824b9e5e763f7ea5a2b511df2 WatchSource:0}: Error finding container 5f55b8adcf12a6b66983b6beb58f8b720085a1c824b9e5e763f7ea5a2b511df2: Status 404 returned error can't find the container with id 5f55b8adcf12a6b66983b6beb58f8b720085a1c824b9e5e763f7ea5a2b511df2 Jan 27 15:29:45 crc kubenswrapper[4772]: I0127 15:29:45.699081 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gjwh2" event={"ID":"f91bfd1b-6386-444f-95da-045fbe957f5c","Type":"ContainerStarted","Data":"e1df482be0829e766abad6c9eb6842ba0e9d9f6fb517127a47f819ce8b296c7d"} Jan 27 15:29:45 crc kubenswrapper[4772]: I0127 15:29:45.699589 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gjwh2" event={"ID":"f91bfd1b-6386-444f-95da-045fbe957f5c","Type":"ContainerStarted","Data":"a1515e80dd84c3d7eeadb4bf57668100c5ba9a6a2123be07618feb4932b00619"} Jan 27 15:29:45 crc kubenswrapper[4772]: I0127 15:29:45.703348 4772 generic.go:334] "Generic (PLEG): container finished" podID="73ee81ee-57fa-466a-8ada-2fa4da5987a0" containerID="4148bc33036bbf1369f8777e53357b14d6ea084f15c60f06da4d4195151a1ddd" exitCode=0 Jan 27 15:29:45 crc kubenswrapper[4772]: I0127 15:29:45.703531 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" event={"ID":"73ee81ee-57fa-466a-8ada-2fa4da5987a0","Type":"ContainerDied","Data":"4148bc33036bbf1369f8777e53357b14d6ea084f15c60f06da4d4195151a1ddd"} Jan 27 15:29:45 crc kubenswrapper[4772]: I0127 15:29:45.709586 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"445d3e38-8f68-4dad-9e97-d927d60ee1e4","Type":"ContainerStarted","Data":"5f55b8adcf12a6b66983b6beb58f8b720085a1c824b9e5e763f7ea5a2b511df2"} Jan 27 15:29:45 crc kubenswrapper[4772]: I0127 15:29:45.726681 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-gjwh2" podStartSLOduration=2.726665913 podStartE2EDuration="2.726665913s" podCreationTimestamp="2026-01-27 15:29:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:29:45.724714007 +0000 UTC m=+1371.705323115" watchObservedRunningTime="2026-01-27 15:29:45.726665913 +0000 UTC m=+1371.707275011" Jan 27 15:29:46 crc kubenswrapper[4772]: I0127 15:29:46.734892 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" event={"ID":"73ee81ee-57fa-466a-8ada-2fa4da5987a0","Type":"ContainerStarted","Data":"e6ff45d04539aabc690d9ba73108fbdca7b2759433b3c528ba073567c047da4f"} Jan 27 15:29:46 crc kubenswrapper[4772]: I0127 15:29:46.766613 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" podStartSLOduration=4.766594985 podStartE2EDuration="4.766594985s" podCreationTimestamp="2026-01-27 15:29:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:29:46.760643863 +0000 UTC m=+1372.741252981" watchObservedRunningTime="2026-01-27 15:29:46.766594985 +0000 UTC m=+1372.747204083" Jan 27 15:29:46 crc kubenswrapper[4772]: I0127 15:29:46.983639 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 15:29:46 crc kubenswrapper[4772]: I0127 15:29:46.994158 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 27 15:29:47 crc kubenswrapper[4772]: I0127 15:29:47.750873 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" Jan 27 15:29:48 crc kubenswrapper[4772]: I0127 15:29:48.762531 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"445d3e38-8f68-4dad-9e97-d927d60ee1e4","Type":"ContainerStarted","Data":"7d804eabf1929c1b2bbdc21b8685f4ea2301f718b175dd856db56adb377cd8ed"} Jan 27 15:29:48 crc kubenswrapper[4772]: I0127 15:29:48.764359 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"325e0266-afde-43a7-b77c-4b29a2d55c3a","Type":"ContainerStarted","Data":"f65fe58604f2e802f9b737bcbd1b97003be50f6c492456675251d85de107de5f"} Jan 27 15:29:48 crc kubenswrapper[4772]: I0127 15:29:48.764406 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"325e0266-afde-43a7-b77c-4b29a2d55c3a","Type":"ContainerStarted","Data":"dea9ce36986750defb2b297fb876194a33a2f0aaee4b8e21995a37e01e57d98a"} Jan 27 15:29:48 crc kubenswrapper[4772]: I0127 15:29:48.764440 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="325e0266-afde-43a7-b77c-4b29a2d55c3a" containerName="nova-metadata-log" containerID="cri-o://dea9ce36986750defb2b297fb876194a33a2f0aaee4b8e21995a37e01e57d98a" gracePeriod=30 Jan 27 15:29:48 crc kubenswrapper[4772]: I0127 15:29:48.764503 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="325e0266-afde-43a7-b77c-4b29a2d55c3a" containerName="nova-metadata-metadata" containerID="cri-o://f65fe58604f2e802f9b737bcbd1b97003be50f6c492456675251d85de107de5f" gracePeriod=30 Jan 27 15:29:48 crc kubenswrapper[4772]: I0127 15:29:48.767331 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8a218ca3-a163-4f9c-8e73-f630b2228bb2","Type":"ContainerStarted","Data":"dfac9725f6b1a542e430af600747ad2b7e4c5c445357868cef0cc0fe2f4dae49"} Jan 27 15:29:48 crc kubenswrapper[4772]: I0127 15:29:48.767452 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="8a218ca3-a163-4f9c-8e73-f630b2228bb2" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://dfac9725f6b1a542e430af600747ad2b7e4c5c445357868cef0cc0fe2f4dae49" gracePeriod=30 Jan 27 15:29:48 crc kubenswrapper[4772]: I0127 15:29:48.769966 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2bfae3d5-7017-4ced-9691-4255769c51f6","Type":"ContainerStarted","Data":"63a9ec332e4c6035116beb1bf14c43d1297eb6d5bd2c76280996d80acf99516a"} Jan 27 15:29:48 crc kubenswrapper[4772]: I0127 15:29:48.777682 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a79db711-df56-46f6-93c2-7e1e5c914ba6","Type":"ContainerStarted","Data":"441883730f3253404e2e897259d3ab7a0d38be7920d10151b1f2a7d3a8e23ae9"} Jan 27 15:29:48 crc kubenswrapper[4772]: I0127 15:29:48.796632 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.741714889 podStartE2EDuration="6.79660952s" podCreationTimestamp="2026-01-27 15:29:42 +0000 UTC" firstStartedPulling="2026-01-27 15:29:44.047129026 +0000 UTC m=+1370.027738134" lastFinishedPulling="2026-01-27 15:29:48.102023667 +0000 UTC m=+1374.082632765" observedRunningTime="2026-01-27 15:29:48.789619088 +0000 UTC m=+1374.770228186" watchObservedRunningTime="2026-01-27 15:29:48.79660952 +0000 UTC m=+1374.777218618" Jan 27 15:29:48 crc kubenswrapper[4772]: I0127 15:29:48.811566 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.651844293 podStartE2EDuration="6.811543182s" podCreationTimestamp="2026-01-27 15:29:42 +0000 UTC" firstStartedPulling="2026-01-27 15:29:43.945438058 +0000 UTC m=+1369.926047166" lastFinishedPulling="2026-01-27 15:29:48.105136947 +0000 UTC m=+1374.085746055" observedRunningTime="2026-01-27 15:29:48.809161663 +0000 UTC m=+1374.789770761" watchObservedRunningTime="2026-01-27 15:29:48.811543182 +0000 UTC m=+1374.792152280" Jan 27 15:29:48 crc kubenswrapper[4772]: I0127 15:29:48.837371 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.832456401 podStartE2EDuration="6.837350387s" podCreationTimestamp="2026-01-27 15:29:42 +0000 UTC" firstStartedPulling="2026-01-27 15:29:44.120439365 +0000 UTC m=+1370.101048463" lastFinishedPulling="2026-01-27 15:29:48.125333351 +0000 UTC m=+1374.105942449" observedRunningTime="2026-01-27 15:29:48.830329584 +0000 UTC m=+1374.810938692" watchObservedRunningTime="2026-01-27 15:29:48.837350387 +0000 UTC m=+1374.817959485" Jan 27 15:29:49 crc kubenswrapper[4772]: I0127 15:29:49.788492 4772 generic.go:334] "Generic (PLEG): container finished" podID="325e0266-afde-43a7-b77c-4b29a2d55c3a" containerID="dea9ce36986750defb2b297fb876194a33a2f0aaee4b8e21995a37e01e57d98a" exitCode=143 Jan 27 15:29:49 crc kubenswrapper[4772]: I0127 15:29:49.788605 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"325e0266-afde-43a7-b77c-4b29a2d55c3a","Type":"ContainerDied","Data":"dea9ce36986750defb2b297fb876194a33a2f0aaee4b8e21995a37e01e57d98a"} Jan 27 15:29:49 crc kubenswrapper[4772]: I0127 15:29:49.790960 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a79db711-df56-46f6-93c2-7e1e5c914ba6","Type":"ContainerStarted","Data":"b80017bd45d285abaa0954756e4b6dd746142e879b1c063421cc148826375061"} Jan 27 15:29:49 crc kubenswrapper[4772]: I0127 15:29:49.795177 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"445d3e38-8f68-4dad-9e97-d927d60ee1e4","Type":"ContainerStarted","Data":"20892681e030eeb363de84f7efb9def934d634922b8bfc86ec12ed76e131d504"} Jan 27 15:29:49 crc kubenswrapper[4772]: I0127 15:29:49.810424 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.600531868 podStartE2EDuration="7.810404997s" podCreationTimestamp="2026-01-27 15:29:42 +0000 UTC" firstStartedPulling="2026-01-27 15:29:43.893499067 +0000 UTC m=+1369.874108165" lastFinishedPulling="2026-01-27 15:29:48.103372196 +0000 UTC m=+1374.083981294" observedRunningTime="2026-01-27 15:29:49.809279324 +0000 UTC m=+1375.789888442" watchObservedRunningTime="2026-01-27 15:29:49.810404997 +0000 UTC m=+1375.791014095" Jan 27 15:29:49 crc kubenswrapper[4772]: I0127 15:29:49.901662 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Jan 27 15:29:50 crc kubenswrapper[4772]: I0127 15:29:50.808156 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"445d3e38-8f68-4dad-9e97-d927d60ee1e4","Type":"ContainerStarted","Data":"a36cda53da6b26fa4dc152f0abf33d17f95ff93ab0b84474109c123bbd0176a8"} Jan 27 15:29:52 crc kubenswrapper[4772]: I0127 15:29:52.828503 4772 generic.go:334] "Generic (PLEG): container finished" podID="ea2e7e0f-aef9-4687-932c-d21f24fd4bff" containerID="bea9ecc5c8bd7f22996f379a16987a5468d25478afcbfdd986751cd73382ded7" exitCode=0 Jan 27 15:29:52 crc kubenswrapper[4772]: I0127 15:29:52.828586 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-h5ch7" event={"ID":"ea2e7e0f-aef9-4687-932c-d21f24fd4bff","Type":"ContainerDied","Data":"bea9ecc5c8bd7f22996f379a16987a5468d25478afcbfdd986751cd73382ded7"} Jan 27 15:29:52 crc kubenswrapper[4772]: I0127 15:29:52.832403 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"445d3e38-8f68-4dad-9e97-d927d60ee1e4","Type":"ContainerStarted","Data":"a2f054d9dabda4ad71c0c80285f0ce4245d4a9aa17475b812ef0cd6382dca481"} Jan 27 15:29:52 crc kubenswrapper[4772]: I0127 15:29:52.833619 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 27 15:29:52 crc kubenswrapper[4772]: I0127 15:29:52.869816 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.143749896 podStartE2EDuration="9.869795749s" podCreationTimestamp="2026-01-27 15:29:43 +0000 UTC" firstStartedPulling="2026-01-27 15:29:45.012108703 +0000 UTC m=+1370.992717801" lastFinishedPulling="2026-01-27 15:29:51.738154556 +0000 UTC m=+1377.718763654" observedRunningTime="2026-01-27 15:29:52.86911618 +0000 UTC m=+1378.849725288" watchObservedRunningTime="2026-01-27 15:29:52.869795749 +0000 UTC m=+1378.850404847" Jan 27 15:29:53 crc kubenswrapper[4772]: I0127 15:29:53.100056 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 27 15:29:53 crc kubenswrapper[4772]: I0127 15:29:53.100560 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 27 15:29:53 crc kubenswrapper[4772]: I0127 15:29:53.138722 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:29:53 crc kubenswrapper[4772]: I0127 15:29:53.164182 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 27 15:29:53 crc kubenswrapper[4772]: I0127 15:29:53.164259 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 27 15:29:53 crc kubenswrapper[4772]: I0127 15:29:53.285329 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Jan 27 15:29:53 crc kubenswrapper[4772]: I0127 15:29:53.285374 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 27 15:29:53 crc kubenswrapper[4772]: I0127 15:29:53.327086 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Jan 27 15:29:53 crc kubenswrapper[4772]: I0127 15:29:53.628375 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" Jan 27 15:29:53 crc kubenswrapper[4772]: I0127 15:29:53.701758 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-2849v"] Jan 27 15:29:53 crc kubenswrapper[4772]: I0127 15:29:53.702042 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-2849v" podUID="4ab060da-8587-413a-a410-ee0e9cec40c6" containerName="dnsmasq-dns" containerID="cri-o://b3b0580b2d9a989010c2055ae938a024c281531976b862414fc303ffddcf01e5" gracePeriod=10 Jan 27 15:29:53 crc kubenswrapper[4772]: I0127 15:29:53.851586 4772 generic.go:334] "Generic (PLEG): container finished" podID="4ab060da-8587-413a-a410-ee0e9cec40c6" containerID="b3b0580b2d9a989010c2055ae938a024c281531976b862414fc303ffddcf01e5" exitCode=0 Jan 27 15:29:53 crc kubenswrapper[4772]: I0127 15:29:53.851676 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-2849v" event={"ID":"4ab060da-8587-413a-a410-ee0e9cec40c6","Type":"ContainerDied","Data":"b3b0580b2d9a989010c2055ae938a024c281531976b862414fc303ffddcf01e5"} Jan 27 15:29:53 crc kubenswrapper[4772]: I0127 15:29:53.906055 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.184376 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a79db711-df56-46f6-93c2-7e1e5c914ba6" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.188:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.185285 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a79db711-df56-46f6-93c2-7e1e5c914ba6" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.188:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.413378 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-h5ch7" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.423647 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-2849v" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.534523 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftl7j\" (UniqueName: \"kubernetes.io/projected/ea2e7e0f-aef9-4687-932c-d21f24fd4bff-kube-api-access-ftl7j\") pod \"ea2e7e0f-aef9-4687-932c-d21f24fd4bff\" (UID: \"ea2e7e0f-aef9-4687-932c-d21f24fd4bff\") " Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.534609 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-dns-svc\") pod \"4ab060da-8587-413a-a410-ee0e9cec40c6\" (UID: \"4ab060da-8587-413a-a410-ee0e9cec40c6\") " Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.534689 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea2e7e0f-aef9-4687-932c-d21f24fd4bff-config-data\") pod \"ea2e7e0f-aef9-4687-932c-d21f24fd4bff\" (UID: \"ea2e7e0f-aef9-4687-932c-d21f24fd4bff\") " Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.534757 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-ovsdbserver-nb\") pod \"4ab060da-8587-413a-a410-ee0e9cec40c6\" (UID: \"4ab060da-8587-413a-a410-ee0e9cec40c6\") " Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.534788 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-config\") pod \"4ab060da-8587-413a-a410-ee0e9cec40c6\" (UID: \"4ab060da-8587-413a-a410-ee0e9cec40c6\") " Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.534803 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-dns-swift-storage-0\") pod \"4ab060da-8587-413a-a410-ee0e9cec40c6\" (UID: \"4ab060da-8587-413a-a410-ee0e9cec40c6\") " Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.534827 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea2e7e0f-aef9-4687-932c-d21f24fd4bff-scripts\") pod \"ea2e7e0f-aef9-4687-932c-d21f24fd4bff\" (UID: \"ea2e7e0f-aef9-4687-932c-d21f24fd4bff\") " Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.534856 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s774r\" (UniqueName: \"kubernetes.io/projected/4ab060da-8587-413a-a410-ee0e9cec40c6-kube-api-access-s774r\") pod \"4ab060da-8587-413a-a410-ee0e9cec40c6\" (UID: \"4ab060da-8587-413a-a410-ee0e9cec40c6\") " Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.534887 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-ovsdbserver-sb\") pod \"4ab060da-8587-413a-a410-ee0e9cec40c6\" (UID: \"4ab060da-8587-413a-a410-ee0e9cec40c6\") " Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.534907 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea2e7e0f-aef9-4687-932c-d21f24fd4bff-combined-ca-bundle\") pod \"ea2e7e0f-aef9-4687-932c-d21f24fd4bff\" (UID: \"ea2e7e0f-aef9-4687-932c-d21f24fd4bff\") " Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.541124 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea2e7e0f-aef9-4687-932c-d21f24fd4bff-kube-api-access-ftl7j" (OuterVolumeSpecName: "kube-api-access-ftl7j") pod "ea2e7e0f-aef9-4687-932c-d21f24fd4bff" (UID: "ea2e7e0f-aef9-4687-932c-d21f24fd4bff"). InnerVolumeSpecName "kube-api-access-ftl7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.544707 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea2e7e0f-aef9-4687-932c-d21f24fd4bff-scripts" (OuterVolumeSpecName: "scripts") pod "ea2e7e0f-aef9-4687-932c-d21f24fd4bff" (UID: "ea2e7e0f-aef9-4687-932c-d21f24fd4bff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.545054 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ab060da-8587-413a-a410-ee0e9cec40c6-kube-api-access-s774r" (OuterVolumeSpecName: "kube-api-access-s774r") pod "4ab060da-8587-413a-a410-ee0e9cec40c6" (UID: "4ab060da-8587-413a-a410-ee0e9cec40c6"). InnerVolumeSpecName "kube-api-access-s774r". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.586281 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea2e7e0f-aef9-4687-932c-d21f24fd4bff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ea2e7e0f-aef9-4687-932c-d21f24fd4bff" (UID: "ea2e7e0f-aef9-4687-932c-d21f24fd4bff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.604612 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea2e7e0f-aef9-4687-932c-d21f24fd4bff-config-data" (OuterVolumeSpecName: "config-data") pod "ea2e7e0f-aef9-4687-932c-d21f24fd4bff" (UID: "ea2e7e0f-aef9-4687-932c-d21f24fd4bff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.623973 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4ab060da-8587-413a-a410-ee0e9cec40c6" (UID: "4ab060da-8587-413a-a410-ee0e9cec40c6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.627809 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4ab060da-8587-413a-a410-ee0e9cec40c6" (UID: "4ab060da-8587-413a-a410-ee0e9cec40c6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.628690 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4ab060da-8587-413a-a410-ee0e9cec40c6" (UID: "4ab060da-8587-413a-a410-ee0e9cec40c6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.638521 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.638558 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea2e7e0f-aef9-4687-932c-d21f24fd4bff-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.638572 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s774r\" (UniqueName: \"kubernetes.io/projected/4ab060da-8587-413a-a410-ee0e9cec40c6-kube-api-access-s774r\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.638586 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.638601 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea2e7e0f-aef9-4687-932c-d21f24fd4bff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.638613 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftl7j\" (UniqueName: \"kubernetes.io/projected/ea2e7e0f-aef9-4687-932c-d21f24fd4bff-kube-api-access-ftl7j\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.638625 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.638635 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea2e7e0f-aef9-4687-932c-d21f24fd4bff-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.646514 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4ab060da-8587-413a-a410-ee0e9cec40c6" (UID: "4ab060da-8587-413a-a410-ee0e9cec40c6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.651629 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-config" (OuterVolumeSpecName: "config") pod "4ab060da-8587-413a-a410-ee0e9cec40c6" (UID: "4ab060da-8587-413a-a410-ee0e9cec40c6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.740058 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.740097 4772 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4ab060da-8587-413a-a410-ee0e9cec40c6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.864467 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-h5ch7" event={"ID":"ea2e7e0f-aef9-4687-932c-d21f24fd4bff","Type":"ContainerDied","Data":"68f7be70d41a760b9715eb4d90633fa6d0b418abb8ac102df2c4b1b0cee8fa86"} Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.864544 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68f7be70d41a760b9715eb4d90633fa6d0b418abb8ac102df2c4b1b0cee8fa86" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.864618 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-h5ch7" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.869637 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-2849v" event={"ID":"4ab060da-8587-413a-a410-ee0e9cec40c6","Type":"ContainerDied","Data":"894acb53cf18e87f7e2c3c3b36e872c3494f3f5477487f3a865082faee222b95"} Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.869688 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-2849v" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.869730 4772 scope.go:117] "RemoveContainer" containerID="b3b0580b2d9a989010c2055ae938a024c281531976b862414fc303ffddcf01e5" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.935467 4772 scope.go:117] "RemoveContainer" containerID="d9accb6fdf89d9e80604718e7ef1a89b03857412edcadace6f8fed5be8e5dfab" Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.946229 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-2849v"] Jan 27 15:29:54 crc kubenswrapper[4772]: I0127 15:29:54.953920 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-2849v"] Jan 27 15:29:55 crc kubenswrapper[4772]: I0127 15:29:55.078720 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 27 15:29:55 crc kubenswrapper[4772]: I0127 15:29:55.079471 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a79db711-df56-46f6-93c2-7e1e5c914ba6" containerName="nova-api-log" containerID="cri-o://441883730f3253404e2e897259d3ab7a0d38be7920d10151b1f2a7d3a8e23ae9" gracePeriod=30 Jan 27 15:29:55 crc kubenswrapper[4772]: I0127 15:29:55.080013 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a79db711-df56-46f6-93c2-7e1e5c914ba6" containerName="nova-api-api" containerID="cri-o://b80017bd45d285abaa0954756e4b6dd746142e879b1c063421cc148826375061" gracePeriod=30 Jan 27 15:29:55 crc kubenswrapper[4772]: I0127 15:29:55.096811 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 15:29:55 crc kubenswrapper[4772]: I0127 15:29:55.882296 4772 generic.go:334] "Generic (PLEG): container finished" podID="a79db711-df56-46f6-93c2-7e1e5c914ba6" containerID="441883730f3253404e2e897259d3ab7a0d38be7920d10151b1f2a7d3a8e23ae9" exitCode=143 Jan 27 15:29:55 crc kubenswrapper[4772]: I0127 15:29:55.882377 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a79db711-df56-46f6-93c2-7e1e5c914ba6","Type":"ContainerDied","Data":"441883730f3253404e2e897259d3ab7a0d38be7920d10151b1f2a7d3a8e23ae9"} Jan 27 15:29:55 crc kubenswrapper[4772]: I0127 15:29:55.883531 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="2bfae3d5-7017-4ced-9691-4255769c51f6" containerName="nova-scheduler-scheduler" containerID="cri-o://63a9ec332e4c6035116beb1bf14c43d1297eb6d5bd2c76280996d80acf99516a" gracePeriod=30 Jan 27 15:29:56 crc kubenswrapper[4772]: I0127 15:29:56.674918 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ab060da-8587-413a-a410-ee0e9cec40c6" path="/var/lib/kubelet/pods/4ab060da-8587-413a-a410-ee0e9cec40c6/volumes" Jan 27 15:29:56 crc kubenswrapper[4772]: I0127 15:29:56.896835 4772 generic.go:334] "Generic (PLEG): container finished" podID="f91bfd1b-6386-444f-95da-045fbe957f5c" containerID="e1df482be0829e766abad6c9eb6842ba0e9d9f6fb517127a47f819ce8b296c7d" exitCode=0 Jan 27 15:29:56 crc kubenswrapper[4772]: I0127 15:29:56.896934 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gjwh2" event={"ID":"f91bfd1b-6386-444f-95da-045fbe957f5c","Type":"ContainerDied","Data":"e1df482be0829e766abad6c9eb6842ba0e9d9f6fb517127a47f819ce8b296c7d"} Jan 27 15:29:56 crc kubenswrapper[4772]: I0127 15:29:56.899316 4772 generic.go:334] "Generic (PLEG): container finished" podID="2bfae3d5-7017-4ced-9691-4255769c51f6" containerID="63a9ec332e4c6035116beb1bf14c43d1297eb6d5bd2c76280996d80acf99516a" exitCode=0 Jan 27 15:29:56 crc kubenswrapper[4772]: I0127 15:29:56.899352 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2bfae3d5-7017-4ced-9691-4255769c51f6","Type":"ContainerDied","Data":"63a9ec332e4c6035116beb1bf14c43d1297eb6d5bd2c76280996d80acf99516a"} Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.183615 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.291507 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cx5gw\" (UniqueName: \"kubernetes.io/projected/2bfae3d5-7017-4ced-9691-4255769c51f6-kube-api-access-cx5gw\") pod \"2bfae3d5-7017-4ced-9691-4255769c51f6\" (UID: \"2bfae3d5-7017-4ced-9691-4255769c51f6\") " Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.291610 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bfae3d5-7017-4ced-9691-4255769c51f6-combined-ca-bundle\") pod \"2bfae3d5-7017-4ced-9691-4255769c51f6\" (UID: \"2bfae3d5-7017-4ced-9691-4255769c51f6\") " Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.291632 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bfae3d5-7017-4ced-9691-4255769c51f6-config-data\") pod \"2bfae3d5-7017-4ced-9691-4255769c51f6\" (UID: \"2bfae3d5-7017-4ced-9691-4255769c51f6\") " Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.349666 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bfae3d5-7017-4ced-9691-4255769c51f6-kube-api-access-cx5gw" (OuterVolumeSpecName: "kube-api-access-cx5gw") pod "2bfae3d5-7017-4ced-9691-4255769c51f6" (UID: "2bfae3d5-7017-4ced-9691-4255769c51f6"). InnerVolumeSpecName "kube-api-access-cx5gw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.353245 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bfae3d5-7017-4ced-9691-4255769c51f6-config-data" (OuterVolumeSpecName: "config-data") pod "2bfae3d5-7017-4ced-9691-4255769c51f6" (UID: "2bfae3d5-7017-4ced-9691-4255769c51f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.356928 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bfae3d5-7017-4ced-9691-4255769c51f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2bfae3d5-7017-4ced-9691-4255769c51f6" (UID: "2bfae3d5-7017-4ced-9691-4255769c51f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.398452 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cx5gw\" (UniqueName: \"kubernetes.io/projected/2bfae3d5-7017-4ced-9691-4255769c51f6-kube-api-access-cx5gw\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.398658 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bfae3d5-7017-4ced-9691-4255769c51f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.398795 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bfae3d5-7017-4ced-9691-4255769c51f6-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.910298 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2bfae3d5-7017-4ced-9691-4255769c51f6","Type":"ContainerDied","Data":"1941b03ee46ef8231c28748227a7ca5084953c3a384610a809f7eb266c2d086a"} Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.910327 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.910362 4772 scope.go:117] "RemoveContainer" containerID="63a9ec332e4c6035116beb1bf14c43d1297eb6d5bd2c76280996d80acf99516a" Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.954522 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.972397 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.986257 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 15:29:57 crc kubenswrapper[4772]: E0127 15:29:57.986803 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea2e7e0f-aef9-4687-932c-d21f24fd4bff" containerName="nova-manage" Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.986829 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea2e7e0f-aef9-4687-932c-d21f24fd4bff" containerName="nova-manage" Jan 27 15:29:57 crc kubenswrapper[4772]: E0127 15:29:57.986849 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab060da-8587-413a-a410-ee0e9cec40c6" containerName="dnsmasq-dns" Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.986857 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab060da-8587-413a-a410-ee0e9cec40c6" containerName="dnsmasq-dns" Jan 27 15:29:57 crc kubenswrapper[4772]: E0127 15:29:57.986868 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bfae3d5-7017-4ced-9691-4255769c51f6" containerName="nova-scheduler-scheduler" Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.986875 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bfae3d5-7017-4ced-9691-4255769c51f6" containerName="nova-scheduler-scheduler" Jan 27 15:29:57 crc kubenswrapper[4772]: E0127 15:29:57.986888 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab060da-8587-413a-a410-ee0e9cec40c6" containerName="init" Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.986896 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab060da-8587-413a-a410-ee0e9cec40c6" containerName="init" Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.987126 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea2e7e0f-aef9-4687-932c-d21f24fd4bff" containerName="nova-manage" Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.987150 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ab060da-8587-413a-a410-ee0e9cec40c6" containerName="dnsmasq-dns" Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.987188 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bfae3d5-7017-4ced-9691-4255769c51f6" containerName="nova-scheduler-scheduler" Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.987942 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.997091 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 15:29:57 crc kubenswrapper[4772]: I0127 15:29:57.998762 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.111289 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0a8fd83-de04-4458-8ef8-ebe7ae60194f-config-data\") pod \"nova-scheduler-0\" (UID: \"c0a8fd83-de04-4458-8ef8-ebe7ae60194f\") " pod="openstack/nova-scheduler-0" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.111455 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0a8fd83-de04-4458-8ef8-ebe7ae60194f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c0a8fd83-de04-4458-8ef8-ebe7ae60194f\") " pod="openstack/nova-scheduler-0" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.111499 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt2mz\" (UniqueName: \"kubernetes.io/projected/c0a8fd83-de04-4458-8ef8-ebe7ae60194f-kube-api-access-rt2mz\") pod \"nova-scheduler-0\" (UID: \"c0a8fd83-de04-4458-8ef8-ebe7ae60194f\") " pod="openstack/nova-scheduler-0" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.213513 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0a8fd83-de04-4458-8ef8-ebe7ae60194f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c0a8fd83-de04-4458-8ef8-ebe7ae60194f\") " pod="openstack/nova-scheduler-0" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.213919 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt2mz\" (UniqueName: \"kubernetes.io/projected/c0a8fd83-de04-4458-8ef8-ebe7ae60194f-kube-api-access-rt2mz\") pod \"nova-scheduler-0\" (UID: \"c0a8fd83-de04-4458-8ef8-ebe7ae60194f\") " pod="openstack/nova-scheduler-0" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.214002 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0a8fd83-de04-4458-8ef8-ebe7ae60194f-config-data\") pod \"nova-scheduler-0\" (UID: \"c0a8fd83-de04-4458-8ef8-ebe7ae60194f\") " pod="openstack/nova-scheduler-0" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.220028 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0a8fd83-de04-4458-8ef8-ebe7ae60194f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c0a8fd83-de04-4458-8ef8-ebe7ae60194f\") " pod="openstack/nova-scheduler-0" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.220206 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0a8fd83-de04-4458-8ef8-ebe7ae60194f-config-data\") pod \"nova-scheduler-0\" (UID: \"c0a8fd83-de04-4458-8ef8-ebe7ae60194f\") " pod="openstack/nova-scheduler-0" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.232953 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt2mz\" (UniqueName: \"kubernetes.io/projected/c0a8fd83-de04-4458-8ef8-ebe7ae60194f-kube-api-access-rt2mz\") pod \"nova-scheduler-0\" (UID: \"c0a8fd83-de04-4458-8ef8-ebe7ae60194f\") " pod="openstack/nova-scheduler-0" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.308290 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gjwh2" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.311053 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.427859 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f91bfd1b-6386-444f-95da-045fbe957f5c-combined-ca-bundle\") pod \"f91bfd1b-6386-444f-95da-045fbe957f5c\" (UID: \"f91bfd1b-6386-444f-95da-045fbe957f5c\") " Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.427970 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nh52\" (UniqueName: \"kubernetes.io/projected/f91bfd1b-6386-444f-95da-045fbe957f5c-kube-api-access-7nh52\") pod \"f91bfd1b-6386-444f-95da-045fbe957f5c\" (UID: \"f91bfd1b-6386-444f-95da-045fbe957f5c\") " Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.428145 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f91bfd1b-6386-444f-95da-045fbe957f5c-config-data\") pod \"f91bfd1b-6386-444f-95da-045fbe957f5c\" (UID: \"f91bfd1b-6386-444f-95da-045fbe957f5c\") " Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.428283 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f91bfd1b-6386-444f-95da-045fbe957f5c-scripts\") pod \"f91bfd1b-6386-444f-95da-045fbe957f5c\" (UID: \"f91bfd1b-6386-444f-95da-045fbe957f5c\") " Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.435656 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f91bfd1b-6386-444f-95da-045fbe957f5c-scripts" (OuterVolumeSpecName: "scripts") pod "f91bfd1b-6386-444f-95da-045fbe957f5c" (UID: "f91bfd1b-6386-444f-95da-045fbe957f5c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.435720 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f91bfd1b-6386-444f-95da-045fbe957f5c-kube-api-access-7nh52" (OuterVolumeSpecName: "kube-api-access-7nh52") pod "f91bfd1b-6386-444f-95da-045fbe957f5c" (UID: "f91bfd1b-6386-444f-95da-045fbe957f5c"). InnerVolumeSpecName "kube-api-access-7nh52". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.459548 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f91bfd1b-6386-444f-95da-045fbe957f5c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f91bfd1b-6386-444f-95da-045fbe957f5c" (UID: "f91bfd1b-6386-444f-95da-045fbe957f5c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.459818 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f91bfd1b-6386-444f-95da-045fbe957f5c-config-data" (OuterVolumeSpecName: "config-data") pod "f91bfd1b-6386-444f-95da-045fbe957f5c" (UID: "f91bfd1b-6386-444f-95da-045fbe957f5c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.530917 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f91bfd1b-6386-444f-95da-045fbe957f5c-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.531077 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f91bfd1b-6386-444f-95da-045fbe957f5c-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.531087 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f91bfd1b-6386-444f-95da-045fbe957f5c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.531097 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nh52\" (UniqueName: \"kubernetes.io/projected/f91bfd1b-6386-444f-95da-045fbe957f5c-kube-api-access-7nh52\") on node \"crc\" DevicePath \"\"" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.693578 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bfae3d5-7017-4ced-9691-4255769c51f6" path="/var/lib/kubelet/pods/2bfae3d5-7017-4ced-9691-4255769c51f6/volumes" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.768975 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.933624 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gjwh2" event={"ID":"f91bfd1b-6386-444f-95da-045fbe957f5c","Type":"ContainerDied","Data":"a1515e80dd84c3d7eeadb4bf57668100c5ba9a6a2123be07618feb4932b00619"} Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.933662 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1515e80dd84c3d7eeadb4bf57668100c5ba9a6a2123be07618feb4932b00619" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.934894 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gjwh2" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.935395 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c0a8fd83-de04-4458-8ef8-ebe7ae60194f","Type":"ContainerStarted","Data":"4d588204c22b2f28d3f4e69bf8dac95975b2db88030c7b5662a17b491819b8ab"} Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.996288 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 27 15:29:58 crc kubenswrapper[4772]: E0127 15:29:58.996762 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f91bfd1b-6386-444f-95da-045fbe957f5c" containerName="nova-cell1-conductor-db-sync" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.996779 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f91bfd1b-6386-444f-95da-045fbe957f5c" containerName="nova-cell1-conductor-db-sync" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.996991 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="f91bfd1b-6386-444f-95da-045fbe957f5c" containerName="nova-cell1-conductor-db-sync" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.997684 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Jan 27 15:29:58 crc kubenswrapper[4772]: I0127 15:29:58.999384 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Jan 27 15:29:59 crc kubenswrapper[4772]: I0127 15:29:59.019916 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 27 15:29:59 crc kubenswrapper[4772]: I0127 15:29:59.041246 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbbd3c83-3fde-4b11-8ef0-add837d393ce-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"dbbd3c83-3fde-4b11-8ef0-add837d393ce\") " pod="openstack/nova-cell1-conductor-0" Jan 27 15:29:59 crc kubenswrapper[4772]: I0127 15:29:59.041376 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbbd3c83-3fde-4b11-8ef0-add837d393ce-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"dbbd3c83-3fde-4b11-8ef0-add837d393ce\") " pod="openstack/nova-cell1-conductor-0" Jan 27 15:29:59 crc kubenswrapper[4772]: I0127 15:29:59.041428 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vh2cp\" (UniqueName: \"kubernetes.io/projected/dbbd3c83-3fde-4b11-8ef0-add837d393ce-kube-api-access-vh2cp\") pod \"nova-cell1-conductor-0\" (UID: \"dbbd3c83-3fde-4b11-8ef0-add837d393ce\") " pod="openstack/nova-cell1-conductor-0" Jan 27 15:29:59 crc kubenswrapper[4772]: I0127 15:29:59.143403 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbbd3c83-3fde-4b11-8ef0-add837d393ce-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"dbbd3c83-3fde-4b11-8ef0-add837d393ce\") " pod="openstack/nova-cell1-conductor-0" Jan 27 15:29:59 crc kubenswrapper[4772]: I0127 15:29:59.143522 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbbd3c83-3fde-4b11-8ef0-add837d393ce-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"dbbd3c83-3fde-4b11-8ef0-add837d393ce\") " pod="openstack/nova-cell1-conductor-0" Jan 27 15:29:59 crc kubenswrapper[4772]: I0127 15:29:59.143556 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vh2cp\" (UniqueName: \"kubernetes.io/projected/dbbd3c83-3fde-4b11-8ef0-add837d393ce-kube-api-access-vh2cp\") pod \"nova-cell1-conductor-0\" (UID: \"dbbd3c83-3fde-4b11-8ef0-add837d393ce\") " pod="openstack/nova-cell1-conductor-0" Jan 27 15:29:59 crc kubenswrapper[4772]: I0127 15:29:59.148986 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbbd3c83-3fde-4b11-8ef0-add837d393ce-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"dbbd3c83-3fde-4b11-8ef0-add837d393ce\") " pod="openstack/nova-cell1-conductor-0" Jan 27 15:29:59 crc kubenswrapper[4772]: I0127 15:29:59.150041 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbbd3c83-3fde-4b11-8ef0-add837d393ce-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"dbbd3c83-3fde-4b11-8ef0-add837d393ce\") " pod="openstack/nova-cell1-conductor-0" Jan 27 15:29:59 crc kubenswrapper[4772]: I0127 15:29:59.164846 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vh2cp\" (UniqueName: \"kubernetes.io/projected/dbbd3c83-3fde-4b11-8ef0-add837d393ce-kube-api-access-vh2cp\") pod \"nova-cell1-conductor-0\" (UID: \"dbbd3c83-3fde-4b11-8ef0-add837d393ce\") " pod="openstack/nova-cell1-conductor-0" Jan 27 15:29:59 crc kubenswrapper[4772]: I0127 15:29:59.366569 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Jan 27 15:29:59 crc kubenswrapper[4772]: I0127 15:29:59.813352 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 27 15:29:59 crc kubenswrapper[4772]: I0127 15:29:59.957472 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c0a8fd83-de04-4458-8ef8-ebe7ae60194f","Type":"ContainerStarted","Data":"caa4ecd4bd8e1eda4d938d7b43a4c7d54c06fe65669c022620e5448e98a4584e"} Jan 27 15:29:59 crc kubenswrapper[4772]: I0127 15:29:59.958409 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"dbbd3c83-3fde-4b11-8ef0-add837d393ce","Type":"ContainerStarted","Data":"4ac8efb7b8696b151d1bdc121a58850ad086edea13390d08276f3048f0eea493"} Jan 27 15:29:59 crc kubenswrapper[4772]: I0127 15:29:59.977498 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.977473201 podStartE2EDuration="2.977473201s" podCreationTimestamp="2026-01-27 15:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:29:59.9736025 +0000 UTC m=+1385.954211598" watchObservedRunningTime="2026-01-27 15:29:59.977473201 +0000 UTC m=+1385.958082299" Jan 27 15:30:00 crc kubenswrapper[4772]: I0127 15:30:00.137411 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492130-7lgz6"] Jan 27 15:30:00 crc kubenswrapper[4772]: I0127 15:30:00.138853 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492130-7lgz6" Jan 27 15:30:00 crc kubenswrapper[4772]: I0127 15:30:00.182854 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 27 15:30:00 crc kubenswrapper[4772]: I0127 15:30:00.183086 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 27 15:30:00 crc kubenswrapper[4772]: I0127 15:30:00.213199 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492130-7lgz6"] Jan 27 15:30:00 crc kubenswrapper[4772]: I0127 15:30:00.265864 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d103a19-1490-433a-abdb-3ebd279265f5-secret-volume\") pod \"collect-profiles-29492130-7lgz6\" (UID: \"6d103a19-1490-433a-abdb-3ebd279265f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492130-7lgz6" Jan 27 15:30:00 crc kubenswrapper[4772]: I0127 15:30:00.266242 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d103a19-1490-433a-abdb-3ebd279265f5-config-volume\") pod \"collect-profiles-29492130-7lgz6\" (UID: \"6d103a19-1490-433a-abdb-3ebd279265f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492130-7lgz6" Jan 27 15:30:00 crc kubenswrapper[4772]: I0127 15:30:00.266333 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk6pv\" (UniqueName: \"kubernetes.io/projected/6d103a19-1490-433a-abdb-3ebd279265f5-kube-api-access-qk6pv\") pod \"collect-profiles-29492130-7lgz6\" (UID: \"6d103a19-1490-433a-abdb-3ebd279265f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492130-7lgz6" Jan 27 15:30:00 crc kubenswrapper[4772]: I0127 15:30:00.368212 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d103a19-1490-433a-abdb-3ebd279265f5-secret-volume\") pod \"collect-profiles-29492130-7lgz6\" (UID: \"6d103a19-1490-433a-abdb-3ebd279265f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492130-7lgz6" Jan 27 15:30:00 crc kubenswrapper[4772]: I0127 15:30:00.368276 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d103a19-1490-433a-abdb-3ebd279265f5-config-volume\") pod \"collect-profiles-29492130-7lgz6\" (UID: \"6d103a19-1490-433a-abdb-3ebd279265f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492130-7lgz6" Jan 27 15:30:00 crc kubenswrapper[4772]: I0127 15:30:00.368364 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk6pv\" (UniqueName: \"kubernetes.io/projected/6d103a19-1490-433a-abdb-3ebd279265f5-kube-api-access-qk6pv\") pod \"collect-profiles-29492130-7lgz6\" (UID: \"6d103a19-1490-433a-abdb-3ebd279265f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492130-7lgz6" Jan 27 15:30:00 crc kubenswrapper[4772]: I0127 15:30:00.369397 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d103a19-1490-433a-abdb-3ebd279265f5-config-volume\") pod \"collect-profiles-29492130-7lgz6\" (UID: \"6d103a19-1490-433a-abdb-3ebd279265f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492130-7lgz6" Jan 27 15:30:00 crc kubenswrapper[4772]: I0127 15:30:00.376965 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d103a19-1490-433a-abdb-3ebd279265f5-secret-volume\") pod \"collect-profiles-29492130-7lgz6\" (UID: \"6d103a19-1490-433a-abdb-3ebd279265f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492130-7lgz6" Jan 27 15:30:00 crc kubenswrapper[4772]: I0127 15:30:00.401941 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk6pv\" (UniqueName: \"kubernetes.io/projected/6d103a19-1490-433a-abdb-3ebd279265f5-kube-api-access-qk6pv\") pod \"collect-profiles-29492130-7lgz6\" (UID: \"6d103a19-1490-433a-abdb-3ebd279265f5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492130-7lgz6" Jan 27 15:30:00 crc kubenswrapper[4772]: I0127 15:30:00.513671 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492130-7lgz6" Jan 27 15:30:00 crc kubenswrapper[4772]: I0127 15:30:00.982197 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492130-7lgz6"] Jan 27 15:30:00 crc kubenswrapper[4772]: I0127 15:30:00.997016 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"dbbd3c83-3fde-4b11-8ef0-add837d393ce","Type":"ContainerStarted","Data":"788384a3ae6b89b3eeabbb3fe7578f4cb514172f7c7e0c341ec2b75ed4d75a29"} Jan 27 15:30:00 crc kubenswrapper[4772]: I0127 15:30:00.997078 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Jan 27 15:30:01 crc kubenswrapper[4772]: I0127 15:30:01.001132 4772 generic.go:334] "Generic (PLEG): container finished" podID="a79db711-df56-46f6-93c2-7e1e5c914ba6" containerID="b80017bd45d285abaa0954756e4b6dd746142e879b1c063421cc148826375061" exitCode=0 Jan 27 15:30:01 crc kubenswrapper[4772]: I0127 15:30:01.003492 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a79db711-df56-46f6-93c2-7e1e5c914ba6","Type":"ContainerDied","Data":"b80017bd45d285abaa0954756e4b6dd746142e879b1c063421cc148826375061"} Jan 27 15:30:01 crc kubenswrapper[4772]: I0127 15:30:01.028319 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.028298589 podStartE2EDuration="3.028298589s" podCreationTimestamp="2026-01-27 15:29:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:30:01.027699231 +0000 UTC m=+1387.008308339" watchObservedRunningTime="2026-01-27 15:30:01.028298589 +0000 UTC m=+1387.008907697" Jan 27 15:30:01 crc kubenswrapper[4772]: I0127 15:30:01.227681 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 15:30:01 crc kubenswrapper[4772]: I0127 15:30:01.292245 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2vn6\" (UniqueName: \"kubernetes.io/projected/a79db711-df56-46f6-93c2-7e1e5c914ba6-kube-api-access-t2vn6\") pod \"a79db711-df56-46f6-93c2-7e1e5c914ba6\" (UID: \"a79db711-df56-46f6-93c2-7e1e5c914ba6\") " Jan 27 15:30:01 crc kubenswrapper[4772]: I0127 15:30:01.292482 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a79db711-df56-46f6-93c2-7e1e5c914ba6-logs\") pod \"a79db711-df56-46f6-93c2-7e1e5c914ba6\" (UID: \"a79db711-df56-46f6-93c2-7e1e5c914ba6\") " Jan 27 15:30:01 crc kubenswrapper[4772]: I0127 15:30:01.292511 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a79db711-df56-46f6-93c2-7e1e5c914ba6-config-data\") pod \"a79db711-df56-46f6-93c2-7e1e5c914ba6\" (UID: \"a79db711-df56-46f6-93c2-7e1e5c914ba6\") " Jan 27 15:30:01 crc kubenswrapper[4772]: I0127 15:30:01.292656 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a79db711-df56-46f6-93c2-7e1e5c914ba6-combined-ca-bundle\") pod \"a79db711-df56-46f6-93c2-7e1e5c914ba6\" (UID: \"a79db711-df56-46f6-93c2-7e1e5c914ba6\") " Jan 27 15:30:01 crc kubenswrapper[4772]: I0127 15:30:01.293098 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a79db711-df56-46f6-93c2-7e1e5c914ba6-logs" (OuterVolumeSpecName: "logs") pod "a79db711-df56-46f6-93c2-7e1e5c914ba6" (UID: "a79db711-df56-46f6-93c2-7e1e5c914ba6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:30:01 crc kubenswrapper[4772]: I0127 15:30:01.293344 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a79db711-df56-46f6-93c2-7e1e5c914ba6-logs\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:01 crc kubenswrapper[4772]: I0127 15:30:01.329309 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a79db711-df56-46f6-93c2-7e1e5c914ba6-kube-api-access-t2vn6" (OuterVolumeSpecName: "kube-api-access-t2vn6") pod "a79db711-df56-46f6-93c2-7e1e5c914ba6" (UID: "a79db711-df56-46f6-93c2-7e1e5c914ba6"). InnerVolumeSpecName "kube-api-access-t2vn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:30:01 crc kubenswrapper[4772]: I0127 15:30:01.334197 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a79db711-df56-46f6-93c2-7e1e5c914ba6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a79db711-df56-46f6-93c2-7e1e5c914ba6" (UID: "a79db711-df56-46f6-93c2-7e1e5c914ba6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:01 crc kubenswrapper[4772]: I0127 15:30:01.334273 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a79db711-df56-46f6-93c2-7e1e5c914ba6-config-data" (OuterVolumeSpecName: "config-data") pod "a79db711-df56-46f6-93c2-7e1e5c914ba6" (UID: "a79db711-df56-46f6-93c2-7e1e5c914ba6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:01 crc kubenswrapper[4772]: I0127 15:30:01.396460 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a79db711-df56-46f6-93c2-7e1e5c914ba6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:01 crc kubenswrapper[4772]: I0127 15:30:01.396991 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2vn6\" (UniqueName: \"kubernetes.io/projected/a79db711-df56-46f6-93c2-7e1e5c914ba6-kube-api-access-t2vn6\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:01 crc kubenswrapper[4772]: I0127 15:30:01.397008 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a79db711-df56-46f6-93c2-7e1e5c914ba6-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.013344 4772 generic.go:334] "Generic (PLEG): container finished" podID="6d103a19-1490-433a-abdb-3ebd279265f5" containerID="58e0f9aeee1bc53c7d023bfdbaa2444440ab205390cfd9df2a1973966a2ae19f" exitCode=0 Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.013923 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492130-7lgz6" event={"ID":"6d103a19-1490-433a-abdb-3ebd279265f5","Type":"ContainerDied","Data":"58e0f9aeee1bc53c7d023bfdbaa2444440ab205390cfd9df2a1973966a2ae19f"} Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.013955 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492130-7lgz6" event={"ID":"6d103a19-1490-433a-abdb-3ebd279265f5","Type":"ContainerStarted","Data":"5dc4066777884aea20f786297a419a7880d6948c75f7adb1580741a73f3871da"} Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.018030 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.019205 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a79db711-df56-46f6-93c2-7e1e5c914ba6","Type":"ContainerDied","Data":"ea5b86feefaf42849d7aeffa2e275c403579d3f00ecb406df1ac2ae06ceefa44"} Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.019280 4772 scope.go:117] "RemoveContainer" containerID="b80017bd45d285abaa0954756e4b6dd746142e879b1c063421cc148826375061" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.059508 4772 scope.go:117] "RemoveContainer" containerID="441883730f3253404e2e897259d3ab7a0d38be7920d10151b1f2a7d3a8e23ae9" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.109862 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.153008 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.170820 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 27 15:30:02 crc kubenswrapper[4772]: E0127 15:30:02.171617 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a79db711-df56-46f6-93c2-7e1e5c914ba6" containerName="nova-api-api" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.171708 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a79db711-df56-46f6-93c2-7e1e5c914ba6" containerName="nova-api-api" Jan 27 15:30:02 crc kubenswrapper[4772]: E0127 15:30:02.171897 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a79db711-df56-46f6-93c2-7e1e5c914ba6" containerName="nova-api-log" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.171959 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a79db711-df56-46f6-93c2-7e1e5c914ba6" containerName="nova-api-log" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.172259 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="a79db711-df56-46f6-93c2-7e1e5c914ba6" containerName="nova-api-api" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.172455 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="a79db711-df56-46f6-93c2-7e1e5c914ba6" containerName="nova-api-log" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.173832 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.176542 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.182604 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.314523 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc238129-30ce-43ab-be89-045e2a9ae8e4-logs\") pod \"nova-api-0\" (UID: \"dc238129-30ce-43ab-be89-045e2a9ae8e4\") " pod="openstack/nova-api-0" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.314628 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc238129-30ce-43ab-be89-045e2a9ae8e4-config-data\") pod \"nova-api-0\" (UID: \"dc238129-30ce-43ab-be89-045e2a9ae8e4\") " pod="openstack/nova-api-0" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.314766 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhrvf\" (UniqueName: \"kubernetes.io/projected/dc238129-30ce-43ab-be89-045e2a9ae8e4-kube-api-access-lhrvf\") pod \"nova-api-0\" (UID: \"dc238129-30ce-43ab-be89-045e2a9ae8e4\") " pod="openstack/nova-api-0" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.314831 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc238129-30ce-43ab-be89-045e2a9ae8e4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dc238129-30ce-43ab-be89-045e2a9ae8e4\") " pod="openstack/nova-api-0" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.418057 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhrvf\" (UniqueName: \"kubernetes.io/projected/dc238129-30ce-43ab-be89-045e2a9ae8e4-kube-api-access-lhrvf\") pod \"nova-api-0\" (UID: \"dc238129-30ce-43ab-be89-045e2a9ae8e4\") " pod="openstack/nova-api-0" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.418689 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc238129-30ce-43ab-be89-045e2a9ae8e4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dc238129-30ce-43ab-be89-045e2a9ae8e4\") " pod="openstack/nova-api-0" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.418773 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc238129-30ce-43ab-be89-045e2a9ae8e4-logs\") pod \"nova-api-0\" (UID: \"dc238129-30ce-43ab-be89-045e2a9ae8e4\") " pod="openstack/nova-api-0" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.418826 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc238129-30ce-43ab-be89-045e2a9ae8e4-config-data\") pod \"nova-api-0\" (UID: \"dc238129-30ce-43ab-be89-045e2a9ae8e4\") " pod="openstack/nova-api-0" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.419902 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc238129-30ce-43ab-be89-045e2a9ae8e4-logs\") pod \"nova-api-0\" (UID: \"dc238129-30ce-43ab-be89-045e2a9ae8e4\") " pod="openstack/nova-api-0" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.425204 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc238129-30ce-43ab-be89-045e2a9ae8e4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dc238129-30ce-43ab-be89-045e2a9ae8e4\") " pod="openstack/nova-api-0" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.425416 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc238129-30ce-43ab-be89-045e2a9ae8e4-config-data\") pod \"nova-api-0\" (UID: \"dc238129-30ce-43ab-be89-045e2a9ae8e4\") " pod="openstack/nova-api-0" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.448728 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhrvf\" (UniqueName: \"kubernetes.io/projected/dc238129-30ce-43ab-be89-045e2a9ae8e4-kube-api-access-lhrvf\") pod \"nova-api-0\" (UID: \"dc238129-30ce-43ab-be89-045e2a9ae8e4\") " pod="openstack/nova-api-0" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.510506 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.681766 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a79db711-df56-46f6-93c2-7e1e5c914ba6" path="/var/lib/kubelet/pods/a79db711-df56-46f6-93c2-7e1e5c914ba6/volumes" Jan 27 15:30:02 crc kubenswrapper[4772]: I0127 15:30:02.967014 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 27 15:30:03 crc kubenswrapper[4772]: I0127 15:30:03.031337 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dc238129-30ce-43ab-be89-045e2a9ae8e4","Type":"ContainerStarted","Data":"73f31dc4802f0c84c1162163b7afaf3ae7604a699c756292b15d9d1058c35cf9"} Jan 27 15:30:03 crc kubenswrapper[4772]: I0127 15:30:03.311816 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 27 15:30:03 crc kubenswrapper[4772]: I0127 15:30:03.429073 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492130-7lgz6" Jan 27 15:30:03 crc kubenswrapper[4772]: I0127 15:30:03.548121 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d103a19-1490-433a-abdb-3ebd279265f5-config-volume\") pod \"6d103a19-1490-433a-abdb-3ebd279265f5\" (UID: \"6d103a19-1490-433a-abdb-3ebd279265f5\") " Jan 27 15:30:03 crc kubenswrapper[4772]: I0127 15:30:03.548160 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qk6pv\" (UniqueName: \"kubernetes.io/projected/6d103a19-1490-433a-abdb-3ebd279265f5-kube-api-access-qk6pv\") pod \"6d103a19-1490-433a-abdb-3ebd279265f5\" (UID: \"6d103a19-1490-433a-abdb-3ebd279265f5\") " Jan 27 15:30:03 crc kubenswrapper[4772]: I0127 15:30:03.548361 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d103a19-1490-433a-abdb-3ebd279265f5-secret-volume\") pod \"6d103a19-1490-433a-abdb-3ebd279265f5\" (UID: \"6d103a19-1490-433a-abdb-3ebd279265f5\") " Jan 27 15:30:03 crc kubenswrapper[4772]: I0127 15:30:03.549188 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d103a19-1490-433a-abdb-3ebd279265f5-config-volume" (OuterVolumeSpecName: "config-volume") pod "6d103a19-1490-433a-abdb-3ebd279265f5" (UID: "6d103a19-1490-433a-abdb-3ebd279265f5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:30:03 crc kubenswrapper[4772]: I0127 15:30:03.551849 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d103a19-1490-433a-abdb-3ebd279265f5-kube-api-access-qk6pv" (OuterVolumeSpecName: "kube-api-access-qk6pv") pod "6d103a19-1490-433a-abdb-3ebd279265f5" (UID: "6d103a19-1490-433a-abdb-3ebd279265f5"). InnerVolumeSpecName "kube-api-access-qk6pv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:30:03 crc kubenswrapper[4772]: I0127 15:30:03.552871 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d103a19-1490-433a-abdb-3ebd279265f5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6d103a19-1490-433a-abdb-3ebd279265f5" (UID: "6d103a19-1490-433a-abdb-3ebd279265f5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:03 crc kubenswrapper[4772]: I0127 15:30:03.650763 4772 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d103a19-1490-433a-abdb-3ebd279265f5-config-volume\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:03 crc kubenswrapper[4772]: I0127 15:30:03.651108 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qk6pv\" (UniqueName: \"kubernetes.io/projected/6d103a19-1490-433a-abdb-3ebd279265f5-kube-api-access-qk6pv\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:03 crc kubenswrapper[4772]: I0127 15:30:03.651123 4772 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d103a19-1490-433a-abdb-3ebd279265f5-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:04 crc kubenswrapper[4772]: I0127 15:30:04.043772 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dc238129-30ce-43ab-be89-045e2a9ae8e4","Type":"ContainerStarted","Data":"ce0174bb725abb785d0b27fe956a5f1ba8818f4c6e84ec3b0dd1b5155087cf7f"} Jan 27 15:30:04 crc kubenswrapper[4772]: I0127 15:30:04.044815 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dc238129-30ce-43ab-be89-045e2a9ae8e4","Type":"ContainerStarted","Data":"b83e870244cf87f2c8284ca6688ebf8db7da16dcd1b693d6c904bb5747128773"} Jan 27 15:30:04 crc kubenswrapper[4772]: I0127 15:30:04.047071 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492130-7lgz6" event={"ID":"6d103a19-1490-433a-abdb-3ebd279265f5","Type":"ContainerDied","Data":"5dc4066777884aea20f786297a419a7880d6948c75f7adb1580741a73f3871da"} Jan 27 15:30:04 crc kubenswrapper[4772]: I0127 15:30:04.047199 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5dc4066777884aea20f786297a419a7880d6948c75f7adb1580741a73f3871da" Jan 27 15:30:04 crc kubenswrapper[4772]: I0127 15:30:04.047127 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492130-7lgz6" Jan 27 15:30:04 crc kubenswrapper[4772]: I0127 15:30:04.481143 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.48109183 podStartE2EDuration="2.48109183s" podCreationTimestamp="2026-01-27 15:30:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:30:04.06414747 +0000 UTC m=+1390.044756588" watchObservedRunningTime="2026-01-27 15:30:04.48109183 +0000 UTC m=+1390.461700928" Jan 27 15:30:08 crc kubenswrapper[4772]: I0127 15:30:08.312070 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Jan 27 15:30:08 crc kubenswrapper[4772]: I0127 15:30:08.338300 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Jan 27 15:30:09 crc kubenswrapper[4772]: I0127 15:30:09.123509 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Jan 27 15:30:09 crc kubenswrapper[4772]: I0127 15:30:09.469741 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Jan 27 15:30:12 crc kubenswrapper[4772]: I0127 15:30:12.511112 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 27 15:30:12 crc kubenswrapper[4772]: I0127 15:30:12.511546 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 27 15:30:13 crc kubenswrapper[4772]: I0127 15:30:13.593428 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="dc238129-30ce-43ab-be89-045e2a9ae8e4" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.198:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 27 15:30:13 crc kubenswrapper[4772]: I0127 15:30:13.593440 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="dc238129-30ce-43ab-be89-045e2a9ae8e4" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.198:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 27 15:30:14 crc kubenswrapper[4772]: I0127 15:30:14.430690 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Jan 27 15:30:18 crc kubenswrapper[4772]: W0127 15:30:18.819389 4772 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d103a19_1490_433a_abdb_3ebd279265f5.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d103a19_1490_433a_abdb_3ebd279265f5.slice: no such file or directory Jan 27 15:30:19 crc kubenswrapper[4772]: E0127 15:30:19.111006 4772 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a218ca3_a163_4f9c_8e73_f630b2228bb2.slice/crio-conmon-dfac9725f6b1a542e430af600747ad2b7e4c5c445357868cef0cc0fe2f4dae49.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a218ca3_a163_4f9c_8e73_f630b2228bb2.slice/crio-dfac9725f6b1a542e430af600747ad2b7e4c5c445357868cef0cc0fe2f4dae49.scope\": RecentStats: unable to find data in memory cache]" Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.194645 4772 generic.go:334] "Generic (PLEG): container finished" podID="325e0266-afde-43a7-b77c-4b29a2d55c3a" containerID="f65fe58604f2e802f9b737bcbd1b97003be50f6c492456675251d85de107de5f" exitCode=137 Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.194713 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"325e0266-afde-43a7-b77c-4b29a2d55c3a","Type":"ContainerDied","Data":"f65fe58604f2e802f9b737bcbd1b97003be50f6c492456675251d85de107de5f"} Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.202089 4772 generic.go:334] "Generic (PLEG): container finished" podID="8a218ca3-a163-4f9c-8e73-f630b2228bb2" containerID="dfac9725f6b1a542e430af600747ad2b7e4c5c445357868cef0cc0fe2f4dae49" exitCode=137 Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.202140 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8a218ca3-a163-4f9c-8e73-f630b2228bb2","Type":"ContainerDied","Data":"dfac9725f6b1a542e430af600747ad2b7e4c5c445357868cef0cc0fe2f4dae49"} Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.202185 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8a218ca3-a163-4f9c-8e73-f630b2228bb2","Type":"ContainerDied","Data":"d661ce2ba7ef10af0b86479b7eca76159de7a8a82826e702dc8a584607cd118c"} Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.202198 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d661ce2ba7ef10af0b86479b7eca76159de7a8a82826e702dc8a584607cd118c" Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.302238 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.312324 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.383861 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a218ca3-a163-4f9c-8e73-f630b2228bb2-config-data\") pod \"8a218ca3-a163-4f9c-8e73-f630b2228bb2\" (UID: \"8a218ca3-a163-4f9c-8e73-f630b2228bb2\") " Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.383912 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/325e0266-afde-43a7-b77c-4b29a2d55c3a-logs\") pod \"325e0266-afde-43a7-b77c-4b29a2d55c3a\" (UID: \"325e0266-afde-43a7-b77c-4b29a2d55c3a\") " Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.383961 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a218ca3-a163-4f9c-8e73-f630b2228bb2-combined-ca-bundle\") pod \"8a218ca3-a163-4f9c-8e73-f630b2228bb2\" (UID: \"8a218ca3-a163-4f9c-8e73-f630b2228bb2\") " Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.384029 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xml5\" (UniqueName: \"kubernetes.io/projected/325e0266-afde-43a7-b77c-4b29a2d55c3a-kube-api-access-6xml5\") pod \"325e0266-afde-43a7-b77c-4b29a2d55c3a\" (UID: \"325e0266-afde-43a7-b77c-4b29a2d55c3a\") " Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.384060 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/325e0266-afde-43a7-b77c-4b29a2d55c3a-config-data\") pod \"325e0266-afde-43a7-b77c-4b29a2d55c3a\" (UID: \"325e0266-afde-43a7-b77c-4b29a2d55c3a\") " Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.384155 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/325e0266-afde-43a7-b77c-4b29a2d55c3a-combined-ca-bundle\") pod \"325e0266-afde-43a7-b77c-4b29a2d55c3a\" (UID: \"325e0266-afde-43a7-b77c-4b29a2d55c3a\") " Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.384205 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgjrt\" (UniqueName: \"kubernetes.io/projected/8a218ca3-a163-4f9c-8e73-f630b2228bb2-kube-api-access-hgjrt\") pod \"8a218ca3-a163-4f9c-8e73-f630b2228bb2\" (UID: \"8a218ca3-a163-4f9c-8e73-f630b2228bb2\") " Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.385916 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/325e0266-afde-43a7-b77c-4b29a2d55c3a-logs" (OuterVolumeSpecName: "logs") pod "325e0266-afde-43a7-b77c-4b29a2d55c3a" (UID: "325e0266-afde-43a7-b77c-4b29a2d55c3a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.390107 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/325e0266-afde-43a7-b77c-4b29a2d55c3a-kube-api-access-6xml5" (OuterVolumeSpecName: "kube-api-access-6xml5") pod "325e0266-afde-43a7-b77c-4b29a2d55c3a" (UID: "325e0266-afde-43a7-b77c-4b29a2d55c3a"). InnerVolumeSpecName "kube-api-access-6xml5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.391638 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a218ca3-a163-4f9c-8e73-f630b2228bb2-kube-api-access-hgjrt" (OuterVolumeSpecName: "kube-api-access-hgjrt") pod "8a218ca3-a163-4f9c-8e73-f630b2228bb2" (UID: "8a218ca3-a163-4f9c-8e73-f630b2228bb2"). InnerVolumeSpecName "kube-api-access-hgjrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.414452 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a218ca3-a163-4f9c-8e73-f630b2228bb2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a218ca3-a163-4f9c-8e73-f630b2228bb2" (UID: "8a218ca3-a163-4f9c-8e73-f630b2228bb2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.415838 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a218ca3-a163-4f9c-8e73-f630b2228bb2-config-data" (OuterVolumeSpecName: "config-data") pod "8a218ca3-a163-4f9c-8e73-f630b2228bb2" (UID: "8a218ca3-a163-4f9c-8e73-f630b2228bb2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.418572 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/325e0266-afde-43a7-b77c-4b29a2d55c3a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "325e0266-afde-43a7-b77c-4b29a2d55c3a" (UID: "325e0266-afde-43a7-b77c-4b29a2d55c3a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.430559 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/325e0266-afde-43a7-b77c-4b29a2d55c3a-config-data" (OuterVolumeSpecName: "config-data") pod "325e0266-afde-43a7-b77c-4b29a2d55c3a" (UID: "325e0266-afde-43a7-b77c-4b29a2d55c3a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.486207 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a218ca3-a163-4f9c-8e73-f630b2228bb2-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.486261 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/325e0266-afde-43a7-b77c-4b29a2d55c3a-logs\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.486278 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a218ca3-a163-4f9c-8e73-f630b2228bb2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.486297 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xml5\" (UniqueName: \"kubernetes.io/projected/325e0266-afde-43a7-b77c-4b29a2d55c3a-kube-api-access-6xml5\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.486309 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/325e0266-afde-43a7-b77c-4b29a2d55c3a-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.486329 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/325e0266-afde-43a7-b77c-4b29a2d55c3a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:19 crc kubenswrapper[4772]: I0127 15:30:19.486340 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgjrt\" (UniqueName: \"kubernetes.io/projected/8a218ca3-a163-4f9c-8e73-f630b2228bb2-kube-api-access-hgjrt\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.212081 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"325e0266-afde-43a7-b77c-4b29a2d55c3a","Type":"ContainerDied","Data":"0fbb6b12d9611d132f4209f04274b5892a26d8ef07f59ee5c52d51242f78a833"} Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.212525 4772 scope.go:117] "RemoveContainer" containerID="f65fe58604f2e802f9b737bcbd1b97003be50f6c492456675251d85de107de5f" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.212109 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.212100 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.239897 4772 scope.go:117] "RemoveContainer" containerID="dea9ce36986750defb2b297fb876194a33a2f0aaee4b8e21995a37e01e57d98a" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.256701 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.283765 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.298598 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.312157 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 27 15:30:20 crc kubenswrapper[4772]: E0127 15:30:20.312695 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a218ca3-a163-4f9c-8e73-f630b2228bb2" containerName="nova-cell1-novncproxy-novncproxy" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.312717 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a218ca3-a163-4f9c-8e73-f630b2228bb2" containerName="nova-cell1-novncproxy-novncproxy" Jan 27 15:30:20 crc kubenswrapper[4772]: E0127 15:30:20.312739 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d103a19-1490-433a-abdb-3ebd279265f5" containerName="collect-profiles" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.312748 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d103a19-1490-433a-abdb-3ebd279265f5" containerName="collect-profiles" Jan 27 15:30:20 crc kubenswrapper[4772]: E0127 15:30:20.312773 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="325e0266-afde-43a7-b77c-4b29a2d55c3a" containerName="nova-metadata-metadata" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.312783 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="325e0266-afde-43a7-b77c-4b29a2d55c3a" containerName="nova-metadata-metadata" Jan 27 15:30:20 crc kubenswrapper[4772]: E0127 15:30:20.312812 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="325e0266-afde-43a7-b77c-4b29a2d55c3a" containerName="nova-metadata-log" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.312819 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="325e0266-afde-43a7-b77c-4b29a2d55c3a" containerName="nova-metadata-log" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.313025 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d103a19-1490-433a-abdb-3ebd279265f5" containerName="collect-profiles" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.313053 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="325e0266-afde-43a7-b77c-4b29a2d55c3a" containerName="nova-metadata-log" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.313080 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="325e0266-afde-43a7-b77c-4b29a2d55c3a" containerName="nova-metadata-metadata" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.313093 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a218ca3-a163-4f9c-8e73-f630b2228bb2" containerName="nova-cell1-novncproxy-novncproxy" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.315497 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.317465 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.323405 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.324367 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.334772 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.346283 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.347823 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.350585 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.352436 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.352645 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.365292 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.404092 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3293f51d-380b-4247-b1ca-5d1f4b831e52-config-data\") pod \"nova-metadata-0\" (UID: \"3293f51d-380b-4247-b1ca-5d1f4b831e52\") " pod="openstack/nova-metadata-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.404128 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e69643a-e8c2-4057-a993-d5506ceeec1b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e69643a-e8c2-4057-a993-d5506ceeec1b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.404339 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3293f51d-380b-4247-b1ca-5d1f4b831e52-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3293f51d-380b-4247-b1ca-5d1f4b831e52\") " pod="openstack/nova-metadata-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.404609 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85ttx\" (UniqueName: \"kubernetes.io/projected/5e69643a-e8c2-4057-a993-d5506ceeec1b-kube-api-access-85ttx\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e69643a-e8c2-4057-a993-d5506ceeec1b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.404709 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3293f51d-380b-4247-b1ca-5d1f4b831e52-logs\") pod \"nova-metadata-0\" (UID: \"3293f51d-380b-4247-b1ca-5d1f4b831e52\") " pod="openstack/nova-metadata-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.404812 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e69643a-e8c2-4057-a993-d5506ceeec1b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e69643a-e8c2-4057-a993-d5506ceeec1b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.404857 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6whx\" (UniqueName: \"kubernetes.io/projected/3293f51d-380b-4247-b1ca-5d1f4b831e52-kube-api-access-v6whx\") pod \"nova-metadata-0\" (UID: \"3293f51d-380b-4247-b1ca-5d1f4b831e52\") " pod="openstack/nova-metadata-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.404889 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3293f51d-380b-4247-b1ca-5d1f4b831e52-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3293f51d-380b-4247-b1ca-5d1f4b831e52\") " pod="openstack/nova-metadata-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.404925 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e69643a-e8c2-4057-a993-d5506ceeec1b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e69643a-e8c2-4057-a993-d5506ceeec1b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.405001 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e69643a-e8c2-4057-a993-d5506ceeec1b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e69643a-e8c2-4057-a993-d5506ceeec1b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.506746 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85ttx\" (UniqueName: \"kubernetes.io/projected/5e69643a-e8c2-4057-a993-d5506ceeec1b-kube-api-access-85ttx\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e69643a-e8c2-4057-a993-d5506ceeec1b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.506830 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3293f51d-380b-4247-b1ca-5d1f4b831e52-logs\") pod \"nova-metadata-0\" (UID: \"3293f51d-380b-4247-b1ca-5d1f4b831e52\") " pod="openstack/nova-metadata-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.506880 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e69643a-e8c2-4057-a993-d5506ceeec1b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e69643a-e8c2-4057-a993-d5506ceeec1b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.506912 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6whx\" (UniqueName: \"kubernetes.io/projected/3293f51d-380b-4247-b1ca-5d1f4b831e52-kube-api-access-v6whx\") pod \"nova-metadata-0\" (UID: \"3293f51d-380b-4247-b1ca-5d1f4b831e52\") " pod="openstack/nova-metadata-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.507436 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3293f51d-380b-4247-b1ca-5d1f4b831e52-logs\") pod \"nova-metadata-0\" (UID: \"3293f51d-380b-4247-b1ca-5d1f4b831e52\") " pod="openstack/nova-metadata-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.507792 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3293f51d-380b-4247-b1ca-5d1f4b831e52-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3293f51d-380b-4247-b1ca-5d1f4b831e52\") " pod="openstack/nova-metadata-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.507824 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e69643a-e8c2-4057-a993-d5506ceeec1b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e69643a-e8c2-4057-a993-d5506ceeec1b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.507860 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e69643a-e8c2-4057-a993-d5506ceeec1b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e69643a-e8c2-4057-a993-d5506ceeec1b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.507901 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3293f51d-380b-4247-b1ca-5d1f4b831e52-config-data\") pod \"nova-metadata-0\" (UID: \"3293f51d-380b-4247-b1ca-5d1f4b831e52\") " pod="openstack/nova-metadata-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.507915 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e69643a-e8c2-4057-a993-d5506ceeec1b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e69643a-e8c2-4057-a993-d5506ceeec1b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.507936 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3293f51d-380b-4247-b1ca-5d1f4b831e52-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3293f51d-380b-4247-b1ca-5d1f4b831e52\") " pod="openstack/nova-metadata-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.512514 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e69643a-e8c2-4057-a993-d5506ceeec1b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e69643a-e8c2-4057-a993-d5506ceeec1b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.514860 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3293f51d-380b-4247-b1ca-5d1f4b831e52-config-data\") pod \"nova-metadata-0\" (UID: \"3293f51d-380b-4247-b1ca-5d1f4b831e52\") " pod="openstack/nova-metadata-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.515125 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3293f51d-380b-4247-b1ca-5d1f4b831e52-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3293f51d-380b-4247-b1ca-5d1f4b831e52\") " pod="openstack/nova-metadata-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.515305 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e69643a-e8c2-4057-a993-d5506ceeec1b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e69643a-e8c2-4057-a993-d5506ceeec1b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.515704 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3293f51d-380b-4247-b1ca-5d1f4b831e52-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3293f51d-380b-4247-b1ca-5d1f4b831e52\") " pod="openstack/nova-metadata-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.521545 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e69643a-e8c2-4057-a993-d5506ceeec1b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e69643a-e8c2-4057-a993-d5506ceeec1b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.527759 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e69643a-e8c2-4057-a993-d5506ceeec1b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e69643a-e8c2-4057-a993-d5506ceeec1b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.528515 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85ttx\" (UniqueName: \"kubernetes.io/projected/5e69643a-e8c2-4057-a993-d5506ceeec1b-kube-api-access-85ttx\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e69643a-e8c2-4057-a993-d5506ceeec1b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.529721 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6whx\" (UniqueName: \"kubernetes.io/projected/3293f51d-380b-4247-b1ca-5d1f4b831e52-kube-api-access-v6whx\") pod \"nova-metadata-0\" (UID: \"3293f51d-380b-4247-b1ca-5d1f4b831e52\") " pod="openstack/nova-metadata-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.639068 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.674451 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.676708 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="325e0266-afde-43a7-b77c-4b29a2d55c3a" path="/var/lib/kubelet/pods/325e0266-afde-43a7-b77c-4b29a2d55c3a/volumes" Jan 27 15:30:20 crc kubenswrapper[4772]: I0127 15:30:20.677852 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a218ca3-a163-4f9c-8e73-f630b2228bb2" path="/var/lib/kubelet/pods/8a218ca3-a163-4f9c-8e73-f630b2228bb2/volumes" Jan 27 15:30:21 crc kubenswrapper[4772]: I0127 15:30:21.151877 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 15:30:21 crc kubenswrapper[4772]: I0127 15:30:21.248473 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3293f51d-380b-4247-b1ca-5d1f4b831e52","Type":"ContainerStarted","Data":"bc2f5b8265c22782479449303b876d41140ba5bbf29e25aebf07297950204c86"} Jan 27 15:30:21 crc kubenswrapper[4772]: I0127 15:30:21.272458 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 27 15:30:21 crc kubenswrapper[4772]: W0127 15:30:21.273480 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e69643a_e8c2_4057_a993_d5506ceeec1b.slice/crio-6444b6c25763e568fb1ce306052e7e5dc898559abb4fc82fb393de7a9f4a2b66 WatchSource:0}: Error finding container 6444b6c25763e568fb1ce306052e7e5dc898559abb4fc82fb393de7a9f4a2b66: Status 404 returned error can't find the container with id 6444b6c25763e568fb1ce306052e7e5dc898559abb4fc82fb393de7a9f4a2b66 Jan 27 15:30:22 crc kubenswrapper[4772]: I0127 15:30:22.264905 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3293f51d-380b-4247-b1ca-5d1f4b831e52","Type":"ContainerStarted","Data":"5263e57d66912aeb6763dff6b7e8221a46b29b91e8c0bceeae9c0497237e7ec2"} Jan 27 15:30:22 crc kubenswrapper[4772]: I0127 15:30:22.265531 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3293f51d-380b-4247-b1ca-5d1f4b831e52","Type":"ContainerStarted","Data":"2fca126b35386587e45cc8336f3cbcb29951790c38ccf77f0aca8a45525bbaba"} Jan 27 15:30:22 crc kubenswrapper[4772]: I0127 15:30:22.267546 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5e69643a-e8c2-4057-a993-d5506ceeec1b","Type":"ContainerStarted","Data":"7a1429ee13edd2169e8a683ea45dcb648c58812d36d48307ba37a8f39d0a67ce"} Jan 27 15:30:22 crc kubenswrapper[4772]: I0127 15:30:22.267586 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5e69643a-e8c2-4057-a993-d5506ceeec1b","Type":"ContainerStarted","Data":"6444b6c25763e568fb1ce306052e7e5dc898559abb4fc82fb393de7a9f4a2b66"} Jan 27 15:30:22 crc kubenswrapper[4772]: I0127 15:30:22.314033 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.314010716 podStartE2EDuration="2.314010716s" podCreationTimestamp="2026-01-27 15:30:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:30:22.2854577 +0000 UTC m=+1408.266066818" watchObservedRunningTime="2026-01-27 15:30:22.314010716 +0000 UTC m=+1408.294619814" Jan 27 15:30:22 crc kubenswrapper[4772]: I0127 15:30:22.314379 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.314371656 podStartE2EDuration="2.314371656s" podCreationTimestamp="2026-01-27 15:30:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:30:22.308452085 +0000 UTC m=+1408.289061183" watchObservedRunningTime="2026-01-27 15:30:22.314371656 +0000 UTC m=+1408.294980764" Jan 27 15:30:22 crc kubenswrapper[4772]: I0127 15:30:22.515491 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 27 15:30:22 crc kubenswrapper[4772]: I0127 15:30:22.516109 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 27 15:30:22 crc kubenswrapper[4772]: I0127 15:30:22.518094 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 27 15:30:22 crc kubenswrapper[4772]: I0127 15:30:22.518692 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 27 15:30:23 crc kubenswrapper[4772]: I0127 15:30:23.277449 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 27 15:30:23 crc kubenswrapper[4772]: I0127 15:30:23.282454 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 27 15:30:23 crc kubenswrapper[4772]: I0127 15:30:23.475138 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-2hd4f"] Jan 27 15:30:23 crc kubenswrapper[4772]: I0127 15:30:23.476753 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" Jan 27 15:30:23 crc kubenswrapper[4772]: I0127 15:30:23.502241 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-2hd4f"] Jan 27 15:30:23 crc kubenswrapper[4772]: I0127 15:30:23.579500 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-2hd4f\" (UID: \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\") " pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" Jan 27 15:30:23 crc kubenswrapper[4772]: I0127 15:30:23.579571 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-2hd4f\" (UID: \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\") " pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" Jan 27 15:30:23 crc kubenswrapper[4772]: I0127 15:30:23.579609 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fqp8\" (UniqueName: \"kubernetes.io/projected/fddd5e59-3124-4a05-aafd-92d6aea05f7e-kube-api-access-6fqp8\") pod \"dnsmasq-dns-59cf4bdb65-2hd4f\" (UID: \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\") " pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" Jan 27 15:30:23 crc kubenswrapper[4772]: I0127 15:30:23.579650 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-2hd4f\" (UID: \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\") " pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" Jan 27 15:30:23 crc kubenswrapper[4772]: I0127 15:30:23.579872 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-2hd4f\" (UID: \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\") " pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" Jan 27 15:30:23 crc kubenswrapper[4772]: I0127 15:30:23.580152 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-config\") pod \"dnsmasq-dns-59cf4bdb65-2hd4f\" (UID: \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\") " pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" Jan 27 15:30:23 crc kubenswrapper[4772]: I0127 15:30:23.682609 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-2hd4f\" (UID: \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\") " pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" Jan 27 15:30:23 crc kubenswrapper[4772]: I0127 15:30:23.682730 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-config\") pod \"dnsmasq-dns-59cf4bdb65-2hd4f\" (UID: \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\") " pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" Jan 27 15:30:23 crc kubenswrapper[4772]: I0127 15:30:23.682793 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-2hd4f\" (UID: \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\") " pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" Jan 27 15:30:23 crc kubenswrapper[4772]: I0127 15:30:23.682835 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-2hd4f\" (UID: \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\") " pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" Jan 27 15:30:23 crc kubenswrapper[4772]: I0127 15:30:23.682881 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fqp8\" (UniqueName: \"kubernetes.io/projected/fddd5e59-3124-4a05-aafd-92d6aea05f7e-kube-api-access-6fqp8\") pod \"dnsmasq-dns-59cf4bdb65-2hd4f\" (UID: \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\") " pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" Jan 27 15:30:23 crc kubenswrapper[4772]: I0127 15:30:23.682932 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-2hd4f\" (UID: \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\") " pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" Jan 27 15:30:23 crc kubenswrapper[4772]: I0127 15:30:23.684049 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-2hd4f\" (UID: \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\") " pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" Jan 27 15:30:23 crc kubenswrapper[4772]: I0127 15:30:23.684706 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-2hd4f\" (UID: \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\") " pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" Jan 27 15:30:23 crc kubenswrapper[4772]: I0127 15:30:23.685370 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-config\") pod \"dnsmasq-dns-59cf4bdb65-2hd4f\" (UID: \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\") " pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" Jan 27 15:30:23 crc kubenswrapper[4772]: I0127 15:30:23.686007 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-2hd4f\" (UID: \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\") " pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" Jan 27 15:30:23 crc kubenswrapper[4772]: I0127 15:30:23.686956 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-2hd4f\" (UID: \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\") " pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" Jan 27 15:30:23 crc kubenswrapper[4772]: I0127 15:30:23.716209 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fqp8\" (UniqueName: \"kubernetes.io/projected/fddd5e59-3124-4a05-aafd-92d6aea05f7e-kube-api-access-6fqp8\") pod \"dnsmasq-dns-59cf4bdb65-2hd4f\" (UID: \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\") " pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" Jan 27 15:30:23 crc kubenswrapper[4772]: I0127 15:30:23.804429 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" Jan 27 15:30:24 crc kubenswrapper[4772]: W0127 15:30:24.277762 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfddd5e59_3124_4a05_aafd_92d6aea05f7e.slice/crio-c5c94af58b0cd6c043cac9ed46da0616cb74fd66aa5279858fb42cf515ba3aa1 WatchSource:0}: Error finding container c5c94af58b0cd6c043cac9ed46da0616cb74fd66aa5279858fb42cf515ba3aa1: Status 404 returned error can't find the container with id c5c94af58b0cd6c043cac9ed46da0616cb74fd66aa5279858fb42cf515ba3aa1 Jan 27 15:30:24 crc kubenswrapper[4772]: I0127 15:30:24.279835 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-2hd4f"] Jan 27 15:30:25 crc kubenswrapper[4772]: I0127 15:30:25.304134 4772 generic.go:334] "Generic (PLEG): container finished" podID="fddd5e59-3124-4a05-aafd-92d6aea05f7e" containerID="71013e440a971fe3ef401d90a82408249df3d1180b65da0eb4683442d6023500" exitCode=0 Jan 27 15:30:25 crc kubenswrapper[4772]: I0127 15:30:25.304217 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" event={"ID":"fddd5e59-3124-4a05-aafd-92d6aea05f7e","Type":"ContainerDied","Data":"71013e440a971fe3ef401d90a82408249df3d1180b65da0eb4683442d6023500"} Jan 27 15:30:25 crc kubenswrapper[4772]: I0127 15:30:25.304711 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" event={"ID":"fddd5e59-3124-4a05-aafd-92d6aea05f7e","Type":"ContainerStarted","Data":"c5c94af58b0cd6c043cac9ed46da0616cb74fd66aa5279858fb42cf515ba3aa1"} Jan 27 15:30:25 crc kubenswrapper[4772]: I0127 15:30:25.589340 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:30:25 crc kubenswrapper[4772]: I0127 15:30:25.590276 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="445d3e38-8f68-4dad-9e97-d927d60ee1e4" containerName="ceilometer-central-agent" containerID="cri-o://7d804eabf1929c1b2bbdc21b8685f4ea2301f718b175dd856db56adb377cd8ed" gracePeriod=30 Jan 27 15:30:25 crc kubenswrapper[4772]: I0127 15:30:25.590642 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="445d3e38-8f68-4dad-9e97-d927d60ee1e4" containerName="sg-core" containerID="cri-o://a36cda53da6b26fa4dc152f0abf33d17f95ff93ab0b84474109c123bbd0176a8" gracePeriod=30 Jan 27 15:30:25 crc kubenswrapper[4772]: I0127 15:30:25.590675 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="445d3e38-8f68-4dad-9e97-d927d60ee1e4" containerName="ceilometer-notification-agent" containerID="cri-o://20892681e030eeb363de84f7efb9def934d634922b8bfc86ec12ed76e131d504" gracePeriod=30 Jan 27 15:30:25 crc kubenswrapper[4772]: I0127 15:30:25.590682 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="445d3e38-8f68-4dad-9e97-d927d60ee1e4" containerName="proxy-httpd" containerID="cri-o://a2f054d9dabda4ad71c0c80285f0ce4245d4a9aa17475b812ef0cd6382dca481" gracePeriod=30 Jan 27 15:30:25 crc kubenswrapper[4772]: I0127 15:30:25.639666 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 27 15:30:25 crc kubenswrapper[4772]: I0127 15:30:25.641097 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 27 15:30:25 crc kubenswrapper[4772]: I0127 15:30:25.675157 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:30:26 crc kubenswrapper[4772]: I0127 15:30:26.314264 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" event={"ID":"fddd5e59-3124-4a05-aafd-92d6aea05f7e","Type":"ContainerStarted","Data":"0ee5661567fa3ca13869262f2ac472811c7a59976cae6fbfe300747e324b4e86"} Jan 27 15:30:26 crc kubenswrapper[4772]: I0127 15:30:26.315606 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" Jan 27 15:30:26 crc kubenswrapper[4772]: I0127 15:30:26.319693 4772 generic.go:334] "Generic (PLEG): container finished" podID="445d3e38-8f68-4dad-9e97-d927d60ee1e4" containerID="a2f054d9dabda4ad71c0c80285f0ce4245d4a9aa17475b812ef0cd6382dca481" exitCode=0 Jan 27 15:30:26 crc kubenswrapper[4772]: I0127 15:30:26.319722 4772 generic.go:334] "Generic (PLEG): container finished" podID="445d3e38-8f68-4dad-9e97-d927d60ee1e4" containerID="a36cda53da6b26fa4dc152f0abf33d17f95ff93ab0b84474109c123bbd0176a8" exitCode=2 Jan 27 15:30:26 crc kubenswrapper[4772]: I0127 15:30:26.319732 4772 generic.go:334] "Generic (PLEG): container finished" podID="445d3e38-8f68-4dad-9e97-d927d60ee1e4" containerID="7d804eabf1929c1b2bbdc21b8685f4ea2301f718b175dd856db56adb377cd8ed" exitCode=0 Jan 27 15:30:26 crc kubenswrapper[4772]: I0127 15:30:26.319869 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"445d3e38-8f68-4dad-9e97-d927d60ee1e4","Type":"ContainerDied","Data":"a2f054d9dabda4ad71c0c80285f0ce4245d4a9aa17475b812ef0cd6382dca481"} Jan 27 15:30:26 crc kubenswrapper[4772]: I0127 15:30:26.319899 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"445d3e38-8f68-4dad-9e97-d927d60ee1e4","Type":"ContainerDied","Data":"a36cda53da6b26fa4dc152f0abf33d17f95ff93ab0b84474109c123bbd0176a8"} Jan 27 15:30:26 crc kubenswrapper[4772]: I0127 15:30:26.319914 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"445d3e38-8f68-4dad-9e97-d927d60ee1e4","Type":"ContainerDied","Data":"7d804eabf1929c1b2bbdc21b8685f4ea2301f718b175dd856db56adb377cd8ed"} Jan 27 15:30:26 crc kubenswrapper[4772]: I0127 15:30:26.339814 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" podStartSLOduration=3.339794555 podStartE2EDuration="3.339794555s" podCreationTimestamp="2026-01-27 15:30:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:30:26.337914971 +0000 UTC m=+1412.318524089" watchObservedRunningTime="2026-01-27 15:30:26.339794555 +0000 UTC m=+1412.320403653" Jan 27 15:30:27 crc kubenswrapper[4772]: I0127 15:30:27.405596 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 27 15:30:27 crc kubenswrapper[4772]: I0127 15:30:27.406210 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="dc238129-30ce-43ab-be89-045e2a9ae8e4" containerName="nova-api-log" containerID="cri-o://b83e870244cf87f2c8284ca6688ebf8db7da16dcd1b693d6c904bb5747128773" gracePeriod=30 Jan 27 15:30:27 crc kubenswrapper[4772]: I0127 15:30:27.406312 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="dc238129-30ce-43ab-be89-045e2a9ae8e4" containerName="nova-api-api" containerID="cri-o://ce0174bb725abb785d0b27fe956a5f1ba8818f4c6e84ec3b0dd1b5155087cf7f" gracePeriod=30 Jan 27 15:30:28 crc kubenswrapper[4772]: I0127 15:30:28.343598 4772 generic.go:334] "Generic (PLEG): container finished" podID="dc238129-30ce-43ab-be89-045e2a9ae8e4" containerID="b83e870244cf87f2c8284ca6688ebf8db7da16dcd1b693d6c904bb5747128773" exitCode=143 Jan 27 15:30:28 crc kubenswrapper[4772]: I0127 15:30:28.344282 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dc238129-30ce-43ab-be89-045e2a9ae8e4","Type":"ContainerDied","Data":"b83e870244cf87f2c8284ca6688ebf8db7da16dcd1b693d6c904bb5747128773"} Jan 27 15:30:30 crc kubenswrapper[4772]: I0127 15:30:30.639634 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 27 15:30:30 crc kubenswrapper[4772]: I0127 15:30:30.640028 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 27 15:30:30 crc kubenswrapper[4772]: I0127 15:30:30.675022 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:30:30 crc kubenswrapper[4772]: I0127 15:30:30.694136 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.090490 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.245009 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc238129-30ce-43ab-be89-045e2a9ae8e4-config-data\") pod \"dc238129-30ce-43ab-be89-045e2a9ae8e4\" (UID: \"dc238129-30ce-43ab-be89-045e2a9ae8e4\") " Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.245072 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhrvf\" (UniqueName: \"kubernetes.io/projected/dc238129-30ce-43ab-be89-045e2a9ae8e4-kube-api-access-lhrvf\") pod \"dc238129-30ce-43ab-be89-045e2a9ae8e4\" (UID: \"dc238129-30ce-43ab-be89-045e2a9ae8e4\") " Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.245183 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc238129-30ce-43ab-be89-045e2a9ae8e4-logs\") pod \"dc238129-30ce-43ab-be89-045e2a9ae8e4\" (UID: \"dc238129-30ce-43ab-be89-045e2a9ae8e4\") " Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.245241 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc238129-30ce-43ab-be89-045e2a9ae8e4-combined-ca-bundle\") pod \"dc238129-30ce-43ab-be89-045e2a9ae8e4\" (UID: \"dc238129-30ce-43ab-be89-045e2a9ae8e4\") " Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.246916 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc238129-30ce-43ab-be89-045e2a9ae8e4-logs" (OuterVolumeSpecName: "logs") pod "dc238129-30ce-43ab-be89-045e2a9ae8e4" (UID: "dc238129-30ce-43ab-be89-045e2a9ae8e4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.270319 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc238129-30ce-43ab-be89-045e2a9ae8e4-kube-api-access-lhrvf" (OuterVolumeSpecName: "kube-api-access-lhrvf") pod "dc238129-30ce-43ab-be89-045e2a9ae8e4" (UID: "dc238129-30ce-43ab-be89-045e2a9ae8e4"). InnerVolumeSpecName "kube-api-access-lhrvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.283882 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc238129-30ce-43ab-be89-045e2a9ae8e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dc238129-30ce-43ab-be89-045e2a9ae8e4" (UID: "dc238129-30ce-43ab-be89-045e2a9ae8e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.306502 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc238129-30ce-43ab-be89-045e2a9ae8e4-config-data" (OuterVolumeSpecName: "config-data") pod "dc238129-30ce-43ab-be89-045e2a9ae8e4" (UID: "dc238129-30ce-43ab-be89-045e2a9ae8e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.347729 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc238129-30ce-43ab-be89-045e2a9ae8e4-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.348165 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhrvf\" (UniqueName: \"kubernetes.io/projected/dc238129-30ce-43ab-be89-045e2a9ae8e4-kube-api-access-lhrvf\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.348203 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc238129-30ce-43ab-be89-045e2a9ae8e4-logs\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.348216 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc238129-30ce-43ab-be89-045e2a9ae8e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.374915 4772 generic.go:334] "Generic (PLEG): container finished" podID="dc238129-30ce-43ab-be89-045e2a9ae8e4" containerID="ce0174bb725abb785d0b27fe956a5f1ba8818f4c6e84ec3b0dd1b5155087cf7f" exitCode=0 Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.375000 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.375040 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dc238129-30ce-43ab-be89-045e2a9ae8e4","Type":"ContainerDied","Data":"ce0174bb725abb785d0b27fe956a5f1ba8818f4c6e84ec3b0dd1b5155087cf7f"} Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.375070 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dc238129-30ce-43ab-be89-045e2a9ae8e4","Type":"ContainerDied","Data":"73f31dc4802f0c84c1162163b7afaf3ae7604a699c756292b15d9d1058c35cf9"} Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.375087 4772 scope.go:117] "RemoveContainer" containerID="ce0174bb725abb785d0b27fe956a5f1ba8818f4c6e84ec3b0dd1b5155087cf7f" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.396149 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.403324 4772 scope.go:117] "RemoveContainer" containerID="b83e870244cf87f2c8284ca6688ebf8db7da16dcd1b693d6c904bb5747128773" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.435965 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.457268 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.487310 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 27 15:30:31 crc kubenswrapper[4772]: E0127 15:30:31.489895 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc238129-30ce-43ab-be89-045e2a9ae8e4" containerName="nova-api-api" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.489928 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc238129-30ce-43ab-be89-045e2a9ae8e4" containerName="nova-api-api" Jan 27 15:30:31 crc kubenswrapper[4772]: E0127 15:30:31.489949 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc238129-30ce-43ab-be89-045e2a9ae8e4" containerName="nova-api-log" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.489955 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc238129-30ce-43ab-be89-045e2a9ae8e4" containerName="nova-api-log" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.490498 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc238129-30ce-43ab-be89-045e2a9ae8e4" containerName="nova-api-api" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.490513 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc238129-30ce-43ab-be89-045e2a9ae8e4" containerName="nova-api-log" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.493737 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.497280 4772 scope.go:117] "RemoveContainer" containerID="ce0174bb725abb785d0b27fe956a5f1ba8818f4c6e84ec3b0dd1b5155087cf7f" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.497550 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.497577 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.497603 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Jan 27 15:30:31 crc kubenswrapper[4772]: E0127 15:30:31.499681 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce0174bb725abb785d0b27fe956a5f1ba8818f4c6e84ec3b0dd1b5155087cf7f\": container with ID starting with ce0174bb725abb785d0b27fe956a5f1ba8818f4c6e84ec3b0dd1b5155087cf7f not found: ID does not exist" containerID="ce0174bb725abb785d0b27fe956a5f1ba8818f4c6e84ec3b0dd1b5155087cf7f" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.499755 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce0174bb725abb785d0b27fe956a5f1ba8818f4c6e84ec3b0dd1b5155087cf7f"} err="failed to get container status \"ce0174bb725abb785d0b27fe956a5f1ba8818f4c6e84ec3b0dd1b5155087cf7f\": rpc error: code = NotFound desc = could not find container \"ce0174bb725abb785d0b27fe956a5f1ba8818f4c6e84ec3b0dd1b5155087cf7f\": container with ID starting with ce0174bb725abb785d0b27fe956a5f1ba8818f4c6e84ec3b0dd1b5155087cf7f not found: ID does not exist" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.499785 4772 scope.go:117] "RemoveContainer" containerID="b83e870244cf87f2c8284ca6688ebf8db7da16dcd1b693d6c904bb5747128773" Jan 27 15:30:31 crc kubenswrapper[4772]: E0127 15:30:31.501516 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b83e870244cf87f2c8284ca6688ebf8db7da16dcd1b693d6c904bb5747128773\": container with ID starting with b83e870244cf87f2c8284ca6688ebf8db7da16dcd1b693d6c904bb5747128773 not found: ID does not exist" containerID="b83e870244cf87f2c8284ca6688ebf8db7da16dcd1b693d6c904bb5747128773" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.511929 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b83e870244cf87f2c8284ca6688ebf8db7da16dcd1b693d6c904bb5747128773"} err="failed to get container status \"b83e870244cf87f2c8284ca6688ebf8db7da16dcd1b693d6c904bb5747128773\": rpc error: code = NotFound desc = could not find container \"b83e870244cf87f2c8284ca6688ebf8db7da16dcd1b693d6c904bb5747128773\": container with ID starting with b83e870244cf87f2c8284ca6688ebf8db7da16dcd1b693d6c904bb5747128773 not found: ID does not exist" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.528683 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.587244 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-dr2w8"] Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.588947 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dr2w8" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.592887 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.594797 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.599966 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-dr2w8"] Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.652406 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3293f51d-380b-4247-b1ca-5d1f4b831e52" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.652431 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3293f51d-380b-4247-b1ca-5d1f4b831e52" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.658950 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xjcn\" (UniqueName: \"kubernetes.io/projected/2d10501c-aefc-4b6b-934a-cd53db7aa029-kube-api-access-4xjcn\") pod \"nova-cell1-cell-mapping-dr2w8\" (UID: \"2d10501c-aefc-4b6b-934a-cd53db7aa029\") " pod="openstack/nova-cell1-cell-mapping-dr2w8" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.659106 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/838e8a63-fd3f-4f03-9030-a2c3a4db7393-logs\") pod \"nova-api-0\" (UID: \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\") " pod="openstack/nova-api-0" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.659136 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d10501c-aefc-4b6b-934a-cd53db7aa029-scripts\") pod \"nova-cell1-cell-mapping-dr2w8\" (UID: \"2d10501c-aefc-4b6b-934a-cd53db7aa029\") " pod="openstack/nova-cell1-cell-mapping-dr2w8" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.659191 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwgqb\" (UniqueName: \"kubernetes.io/projected/838e8a63-fd3f-4f03-9030-a2c3a4db7393-kube-api-access-kwgqb\") pod \"nova-api-0\" (UID: \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\") " pod="openstack/nova-api-0" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.659251 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/838e8a63-fd3f-4f03-9030-a2c3a4db7393-internal-tls-certs\") pod \"nova-api-0\" (UID: \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\") " pod="openstack/nova-api-0" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.659278 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d10501c-aefc-4b6b-934a-cd53db7aa029-config-data\") pod \"nova-cell1-cell-mapping-dr2w8\" (UID: \"2d10501c-aefc-4b6b-934a-cd53db7aa029\") " pod="openstack/nova-cell1-cell-mapping-dr2w8" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.659321 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/838e8a63-fd3f-4f03-9030-a2c3a4db7393-config-data\") pod \"nova-api-0\" (UID: \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\") " pod="openstack/nova-api-0" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.659367 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/838e8a63-fd3f-4f03-9030-a2c3a4db7393-public-tls-certs\") pod \"nova-api-0\" (UID: \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\") " pod="openstack/nova-api-0" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.659397 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/838e8a63-fd3f-4f03-9030-a2c3a4db7393-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\") " pod="openstack/nova-api-0" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.659440 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d10501c-aefc-4b6b-934a-cd53db7aa029-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dr2w8\" (UID: \"2d10501c-aefc-4b6b-934a-cd53db7aa029\") " pod="openstack/nova-cell1-cell-mapping-dr2w8" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.760937 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/838e8a63-fd3f-4f03-9030-a2c3a4db7393-logs\") pod \"nova-api-0\" (UID: \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\") " pod="openstack/nova-api-0" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.760992 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d10501c-aefc-4b6b-934a-cd53db7aa029-scripts\") pod \"nova-cell1-cell-mapping-dr2w8\" (UID: \"2d10501c-aefc-4b6b-934a-cd53db7aa029\") " pod="openstack/nova-cell1-cell-mapping-dr2w8" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.761019 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwgqb\" (UniqueName: \"kubernetes.io/projected/838e8a63-fd3f-4f03-9030-a2c3a4db7393-kube-api-access-kwgqb\") pod \"nova-api-0\" (UID: \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\") " pod="openstack/nova-api-0" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.761475 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/838e8a63-fd3f-4f03-9030-a2c3a4db7393-logs\") pod \"nova-api-0\" (UID: \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\") " pod="openstack/nova-api-0" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.762309 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/838e8a63-fd3f-4f03-9030-a2c3a4db7393-internal-tls-certs\") pod \"nova-api-0\" (UID: \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\") " pod="openstack/nova-api-0" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.762346 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d10501c-aefc-4b6b-934a-cd53db7aa029-config-data\") pod \"nova-cell1-cell-mapping-dr2w8\" (UID: \"2d10501c-aefc-4b6b-934a-cd53db7aa029\") " pod="openstack/nova-cell1-cell-mapping-dr2w8" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.762398 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/838e8a63-fd3f-4f03-9030-a2c3a4db7393-config-data\") pod \"nova-api-0\" (UID: \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\") " pod="openstack/nova-api-0" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.762459 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/838e8a63-fd3f-4f03-9030-a2c3a4db7393-public-tls-certs\") pod \"nova-api-0\" (UID: \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\") " pod="openstack/nova-api-0" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.762483 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/838e8a63-fd3f-4f03-9030-a2c3a4db7393-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\") " pod="openstack/nova-api-0" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.762539 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d10501c-aefc-4b6b-934a-cd53db7aa029-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dr2w8\" (UID: \"2d10501c-aefc-4b6b-934a-cd53db7aa029\") " pod="openstack/nova-cell1-cell-mapping-dr2w8" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.762570 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xjcn\" (UniqueName: \"kubernetes.io/projected/2d10501c-aefc-4b6b-934a-cd53db7aa029-kube-api-access-4xjcn\") pod \"nova-cell1-cell-mapping-dr2w8\" (UID: \"2d10501c-aefc-4b6b-934a-cd53db7aa029\") " pod="openstack/nova-cell1-cell-mapping-dr2w8" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.767110 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/838e8a63-fd3f-4f03-9030-a2c3a4db7393-public-tls-certs\") pod \"nova-api-0\" (UID: \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\") " pod="openstack/nova-api-0" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.767323 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d10501c-aefc-4b6b-934a-cd53db7aa029-scripts\") pod \"nova-cell1-cell-mapping-dr2w8\" (UID: \"2d10501c-aefc-4b6b-934a-cd53db7aa029\") " pod="openstack/nova-cell1-cell-mapping-dr2w8" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.767820 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/838e8a63-fd3f-4f03-9030-a2c3a4db7393-internal-tls-certs\") pod \"nova-api-0\" (UID: \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\") " pod="openstack/nova-api-0" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.769008 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d10501c-aefc-4b6b-934a-cd53db7aa029-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dr2w8\" (UID: \"2d10501c-aefc-4b6b-934a-cd53db7aa029\") " pod="openstack/nova-cell1-cell-mapping-dr2w8" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.769381 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/838e8a63-fd3f-4f03-9030-a2c3a4db7393-config-data\") pod \"nova-api-0\" (UID: \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\") " pod="openstack/nova-api-0" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.772853 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/838e8a63-fd3f-4f03-9030-a2c3a4db7393-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\") " pod="openstack/nova-api-0" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.776967 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d10501c-aefc-4b6b-934a-cd53db7aa029-config-data\") pod \"nova-cell1-cell-mapping-dr2w8\" (UID: \"2d10501c-aefc-4b6b-934a-cd53db7aa029\") " pod="openstack/nova-cell1-cell-mapping-dr2w8" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.782115 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xjcn\" (UniqueName: \"kubernetes.io/projected/2d10501c-aefc-4b6b-934a-cd53db7aa029-kube-api-access-4xjcn\") pod \"nova-cell1-cell-mapping-dr2w8\" (UID: \"2d10501c-aefc-4b6b-934a-cd53db7aa029\") " pod="openstack/nova-cell1-cell-mapping-dr2w8" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.794700 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwgqb\" (UniqueName: \"kubernetes.io/projected/838e8a63-fd3f-4f03-9030-a2c3a4db7393-kube-api-access-kwgqb\") pod \"nova-api-0\" (UID: \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\") " pod="openstack/nova-api-0" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.829515 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 15:30:31 crc kubenswrapper[4772]: I0127 15:30:31.909778 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dr2w8" Jan 27 15:30:32 crc kubenswrapper[4772]: I0127 15:30:32.369864 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 27 15:30:32 crc kubenswrapper[4772]: I0127 15:30:32.390377 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"838e8a63-fd3f-4f03-9030-a2c3a4db7393","Type":"ContainerStarted","Data":"f7d96b6a64da6759e6a7a8bc3df6aebfecfa7e5ec0b668d726902cda154a1888"} Jan 27 15:30:32 crc kubenswrapper[4772]: W0127 15:30:32.457470 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d10501c_aefc_4b6b_934a_cd53db7aa029.slice/crio-33208e54b5b97ccefb163c65dd3c0ecf279de8f0faa126d781f9e471398f64fd WatchSource:0}: Error finding container 33208e54b5b97ccefb163c65dd3c0ecf279de8f0faa126d781f9e471398f64fd: Status 404 returned error can't find the container with id 33208e54b5b97ccefb163c65dd3c0ecf279de8f0faa126d781f9e471398f64fd Jan 27 15:30:32 crc kubenswrapper[4772]: I0127 15:30:32.461899 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-dr2w8"] Jan 27 15:30:32 crc kubenswrapper[4772]: I0127 15:30:32.674004 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc238129-30ce-43ab-be89-045e2a9ae8e4" path="/var/lib/kubelet/pods/dc238129-30ce-43ab-be89-045e2a9ae8e4/volumes" Jan 27 15:30:33 crc kubenswrapper[4772]: I0127 15:30:33.401628 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dr2w8" event={"ID":"2d10501c-aefc-4b6b-934a-cd53db7aa029","Type":"ContainerStarted","Data":"7c4a200cbf0e299c55e6a175696503b17a34325960db8f2addd09db07bdebe00"} Jan 27 15:30:33 crc kubenswrapper[4772]: I0127 15:30:33.401989 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dr2w8" event={"ID":"2d10501c-aefc-4b6b-934a-cd53db7aa029","Type":"ContainerStarted","Data":"33208e54b5b97ccefb163c65dd3c0ecf279de8f0faa126d781f9e471398f64fd"} Jan 27 15:30:33 crc kubenswrapper[4772]: I0127 15:30:33.406267 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"838e8a63-fd3f-4f03-9030-a2c3a4db7393","Type":"ContainerStarted","Data":"9356a4887a357fc1448800688f7032704f830e8c9f7df6f0b3fe3c97ddfa3bb7"} Jan 27 15:30:33 crc kubenswrapper[4772]: I0127 15:30:33.424618 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-dr2w8" podStartSLOduration=2.424594896 podStartE2EDuration="2.424594896s" podCreationTimestamp="2026-01-27 15:30:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:30:33.419064116 +0000 UTC m=+1419.399673214" watchObservedRunningTime="2026-01-27 15:30:33.424594896 +0000 UTC m=+1419.405203994" Jan 27 15:30:33 crc kubenswrapper[4772]: I0127 15:30:33.806000 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" Jan 27 15:30:33 crc kubenswrapper[4772]: I0127 15:30:33.882847 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-88t2p"] Jan 27 15:30:33 crc kubenswrapper[4772]: I0127 15:30:33.884259 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" podUID="73ee81ee-57fa-466a-8ada-2fa4da5987a0" containerName="dnsmasq-dns" containerID="cri-o://e6ff45d04539aabc690d9ba73108fbdca7b2759433b3c528ba073567c047da4f" gracePeriod=10 Jan 27 15:30:34 crc kubenswrapper[4772]: I0127 15:30:34.418151 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"838e8a63-fd3f-4f03-9030-a2c3a4db7393","Type":"ContainerStarted","Data":"596e46cd823d75132da321dc3e1e49f4c351f7ee177713e0655112d945b2390c"} Jan 27 15:30:34 crc kubenswrapper[4772]: I0127 15:30:34.420934 4772 generic.go:334] "Generic (PLEG): container finished" podID="73ee81ee-57fa-466a-8ada-2fa4da5987a0" containerID="e6ff45d04539aabc690d9ba73108fbdca7b2759433b3c528ba073567c047da4f" exitCode=0 Jan 27 15:30:34 crc kubenswrapper[4772]: I0127 15:30:34.421146 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" event={"ID":"73ee81ee-57fa-466a-8ada-2fa4da5987a0","Type":"ContainerDied","Data":"e6ff45d04539aabc690d9ba73108fbdca7b2759433b3c528ba073567c047da4f"} Jan 27 15:30:34 crc kubenswrapper[4772]: I0127 15:30:34.440662 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.4406461 podStartE2EDuration="3.4406461s" podCreationTimestamp="2026-01-27 15:30:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:30:34.435905663 +0000 UTC m=+1420.416514761" watchObservedRunningTime="2026-01-27 15:30:34.4406461 +0000 UTC m=+1420.421255198" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.005984 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.043792 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-ovsdbserver-sb\") pod \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\" (UID: \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\") " Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.043971 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vf2fc\" (UniqueName: \"kubernetes.io/projected/73ee81ee-57fa-466a-8ada-2fa4da5987a0-kube-api-access-vf2fc\") pod \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\" (UID: \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\") " Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.044038 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-dns-swift-storage-0\") pod \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\" (UID: \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\") " Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.044105 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-dns-svc\") pod \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\" (UID: \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\") " Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.044133 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-config\") pod \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\" (UID: \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\") " Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.044182 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-ovsdbserver-nb\") pod \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\" (UID: \"73ee81ee-57fa-466a-8ada-2fa4da5987a0\") " Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.050827 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73ee81ee-57fa-466a-8ada-2fa4da5987a0-kube-api-access-vf2fc" (OuterVolumeSpecName: "kube-api-access-vf2fc") pod "73ee81ee-57fa-466a-8ada-2fa4da5987a0" (UID: "73ee81ee-57fa-466a-8ada-2fa4da5987a0"). InnerVolumeSpecName "kube-api-access-vf2fc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.104618 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "73ee81ee-57fa-466a-8ada-2fa4da5987a0" (UID: "73ee81ee-57fa-466a-8ada-2fa4da5987a0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.112789 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "73ee81ee-57fa-466a-8ada-2fa4da5987a0" (UID: "73ee81ee-57fa-466a-8ada-2fa4da5987a0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.114299 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-config" (OuterVolumeSpecName: "config") pod "73ee81ee-57fa-466a-8ada-2fa4da5987a0" (UID: "73ee81ee-57fa-466a-8ada-2fa4da5987a0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.118953 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "73ee81ee-57fa-466a-8ada-2fa4da5987a0" (UID: "73ee81ee-57fa-466a-8ada-2fa4da5987a0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.147511 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.147547 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.147556 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.147567 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.147575 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vf2fc\" (UniqueName: \"kubernetes.io/projected/73ee81ee-57fa-466a-8ada-2fa4da5987a0-kube-api-access-vf2fc\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.151413 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "73ee81ee-57fa-466a-8ada-2fa4da5987a0" (UID: "73ee81ee-57fa-466a-8ada-2fa4da5987a0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.235299 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.248193 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/445d3e38-8f68-4dad-9e97-d927d60ee1e4-run-httpd\") pod \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.248253 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/445d3e38-8f68-4dad-9e97-d927d60ee1e4-log-httpd\") pod \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.248304 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-br29b\" (UniqueName: \"kubernetes.io/projected/445d3e38-8f68-4dad-9e97-d927d60ee1e4-kube-api-access-br29b\") pod \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.248346 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-scripts\") pod \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.248369 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-combined-ca-bundle\") pod \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.248399 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-ceilometer-tls-certs\") pod \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.248483 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-sg-core-conf-yaml\") pod \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.248516 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-config-data\") pod \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\" (UID: \"445d3e38-8f68-4dad-9e97-d927d60ee1e4\") " Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.248869 4772 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73ee81ee-57fa-466a-8ada-2fa4da5987a0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.249880 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/445d3e38-8f68-4dad-9e97-d927d60ee1e4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "445d3e38-8f68-4dad-9e97-d927d60ee1e4" (UID: "445d3e38-8f68-4dad-9e97-d927d60ee1e4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.249990 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/445d3e38-8f68-4dad-9e97-d927d60ee1e4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "445d3e38-8f68-4dad-9e97-d927d60ee1e4" (UID: "445d3e38-8f68-4dad-9e97-d927d60ee1e4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.256352 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/445d3e38-8f68-4dad-9e97-d927d60ee1e4-kube-api-access-br29b" (OuterVolumeSpecName: "kube-api-access-br29b") pod "445d3e38-8f68-4dad-9e97-d927d60ee1e4" (UID: "445d3e38-8f68-4dad-9e97-d927d60ee1e4"). InnerVolumeSpecName "kube-api-access-br29b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.257333 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-scripts" (OuterVolumeSpecName: "scripts") pod "445d3e38-8f68-4dad-9e97-d927d60ee1e4" (UID: "445d3e38-8f68-4dad-9e97-d927d60ee1e4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.297571 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "445d3e38-8f68-4dad-9e97-d927d60ee1e4" (UID: "445d3e38-8f68-4dad-9e97-d927d60ee1e4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.343341 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "445d3e38-8f68-4dad-9e97-d927d60ee1e4" (UID: "445d3e38-8f68-4dad-9e97-d927d60ee1e4"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.354116 4772 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.354152 4772 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/445d3e38-8f68-4dad-9e97-d927d60ee1e4-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.354233 4772 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/445d3e38-8f68-4dad-9e97-d927d60ee1e4-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.354245 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-br29b\" (UniqueName: \"kubernetes.io/projected/445d3e38-8f68-4dad-9e97-d927d60ee1e4-kube-api-access-br29b\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.354255 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.354264 4772 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.366567 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "445d3e38-8f68-4dad-9e97-d927d60ee1e4" (UID: "445d3e38-8f68-4dad-9e97-d927d60ee1e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.397811 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-config-data" (OuterVolumeSpecName: "config-data") pod "445d3e38-8f68-4dad-9e97-d927d60ee1e4" (UID: "445d3e38-8f68-4dad-9e97-d927d60ee1e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.431361 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" event={"ID":"73ee81ee-57fa-466a-8ada-2fa4da5987a0","Type":"ContainerDied","Data":"79cc249e145e1f853047b2402d998a0aa111ca080edd461a766047221e313d63"} Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.431378 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-88t2p" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.431425 4772 scope.go:117] "RemoveContainer" containerID="e6ff45d04539aabc690d9ba73108fbdca7b2759433b3c528ba073567c047da4f" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.444849 4772 generic.go:334] "Generic (PLEG): container finished" podID="445d3e38-8f68-4dad-9e97-d927d60ee1e4" containerID="20892681e030eeb363de84f7efb9def934d634922b8bfc86ec12ed76e131d504" exitCode=0 Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.444909 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.444956 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"445d3e38-8f68-4dad-9e97-d927d60ee1e4","Type":"ContainerDied","Data":"20892681e030eeb363de84f7efb9def934d634922b8bfc86ec12ed76e131d504"} Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.444989 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"445d3e38-8f68-4dad-9e97-d927d60ee1e4","Type":"ContainerDied","Data":"5f55b8adcf12a6b66983b6beb58f8b720085a1c824b9e5e763f7ea5a2b511df2"} Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.463555 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.463588 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/445d3e38-8f68-4dad-9e97-d927d60ee1e4-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.479549 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-88t2p"] Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.488142 4772 scope.go:117] "RemoveContainer" containerID="4148bc33036bbf1369f8777e53357b14d6ea084f15c60f06da4d4195151a1ddd" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.492971 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-88t2p"] Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.516139 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.538767 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.558575 4772 scope.go:117] "RemoveContainer" containerID="a2f054d9dabda4ad71c0c80285f0ce4245d4a9aa17475b812ef0cd6382dca481" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.564637 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:30:35 crc kubenswrapper[4772]: E0127 15:30:35.565122 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73ee81ee-57fa-466a-8ada-2fa4da5987a0" containerName="init" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.565145 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="73ee81ee-57fa-466a-8ada-2fa4da5987a0" containerName="init" Jan 27 15:30:35 crc kubenswrapper[4772]: E0127 15:30:35.565159 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="445d3e38-8f68-4dad-9e97-d927d60ee1e4" containerName="sg-core" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.565184 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="445d3e38-8f68-4dad-9e97-d927d60ee1e4" containerName="sg-core" Jan 27 15:30:35 crc kubenswrapper[4772]: E0127 15:30:35.565209 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="445d3e38-8f68-4dad-9e97-d927d60ee1e4" containerName="ceilometer-central-agent" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.565217 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="445d3e38-8f68-4dad-9e97-d927d60ee1e4" containerName="ceilometer-central-agent" Jan 27 15:30:35 crc kubenswrapper[4772]: E0127 15:30:35.565237 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73ee81ee-57fa-466a-8ada-2fa4da5987a0" containerName="dnsmasq-dns" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.565246 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="73ee81ee-57fa-466a-8ada-2fa4da5987a0" containerName="dnsmasq-dns" Jan 27 15:30:35 crc kubenswrapper[4772]: E0127 15:30:35.565259 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="445d3e38-8f68-4dad-9e97-d927d60ee1e4" containerName="proxy-httpd" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.565266 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="445d3e38-8f68-4dad-9e97-d927d60ee1e4" containerName="proxy-httpd" Jan 27 15:30:35 crc kubenswrapper[4772]: E0127 15:30:35.565281 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="445d3e38-8f68-4dad-9e97-d927d60ee1e4" containerName="ceilometer-notification-agent" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.565289 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="445d3e38-8f68-4dad-9e97-d927d60ee1e4" containerName="ceilometer-notification-agent" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.565539 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="73ee81ee-57fa-466a-8ada-2fa4da5987a0" containerName="dnsmasq-dns" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.565562 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="445d3e38-8f68-4dad-9e97-d927d60ee1e4" containerName="ceilometer-central-agent" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.565573 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="445d3e38-8f68-4dad-9e97-d927d60ee1e4" containerName="ceilometer-notification-agent" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.565582 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="445d3e38-8f68-4dad-9e97-d927d60ee1e4" containerName="sg-core" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.565602 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="445d3e38-8f68-4dad-9e97-d927d60ee1e4" containerName="proxy-httpd" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.568819 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.570642 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.574070 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.574337 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.579370 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.592100 4772 scope.go:117] "RemoveContainer" containerID="a36cda53da6b26fa4dc152f0abf33d17f95ff93ab0b84474109c123bbd0176a8" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.675515 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-config-data\") pod \"ceilometer-0\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.675923 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72jk7\" (UniqueName: \"kubernetes.io/projected/aea5ee43-36e3-437d-8aca-b2faedd87c5b-kube-api-access-72jk7\") pod \"ceilometer-0\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.676015 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.676052 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aea5ee43-36e3-437d-8aca-b2faedd87c5b-log-httpd\") pod \"ceilometer-0\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.676250 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aea5ee43-36e3-437d-8aca-b2faedd87c5b-run-httpd\") pod \"ceilometer-0\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.676280 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.676344 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-scripts\") pod \"ceilometer-0\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.676534 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.689347 4772 scope.go:117] "RemoveContainer" containerID="20892681e030eeb363de84f7efb9def934d634922b8bfc86ec12ed76e131d504" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.754241 4772 scope.go:117] "RemoveContainer" containerID="7d804eabf1929c1b2bbdc21b8685f4ea2301f718b175dd856db56adb377cd8ed" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.772787 4772 scope.go:117] "RemoveContainer" containerID="a2f054d9dabda4ad71c0c80285f0ce4245d4a9aa17475b812ef0cd6382dca481" Jan 27 15:30:35 crc kubenswrapper[4772]: E0127 15:30:35.773205 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2f054d9dabda4ad71c0c80285f0ce4245d4a9aa17475b812ef0cd6382dca481\": container with ID starting with a2f054d9dabda4ad71c0c80285f0ce4245d4a9aa17475b812ef0cd6382dca481 not found: ID does not exist" containerID="a2f054d9dabda4ad71c0c80285f0ce4245d4a9aa17475b812ef0cd6382dca481" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.773248 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2f054d9dabda4ad71c0c80285f0ce4245d4a9aa17475b812ef0cd6382dca481"} err="failed to get container status \"a2f054d9dabda4ad71c0c80285f0ce4245d4a9aa17475b812ef0cd6382dca481\": rpc error: code = NotFound desc = could not find container \"a2f054d9dabda4ad71c0c80285f0ce4245d4a9aa17475b812ef0cd6382dca481\": container with ID starting with a2f054d9dabda4ad71c0c80285f0ce4245d4a9aa17475b812ef0cd6382dca481 not found: ID does not exist" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.773268 4772 scope.go:117] "RemoveContainer" containerID="a36cda53da6b26fa4dc152f0abf33d17f95ff93ab0b84474109c123bbd0176a8" Jan 27 15:30:35 crc kubenswrapper[4772]: E0127 15:30:35.773613 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a36cda53da6b26fa4dc152f0abf33d17f95ff93ab0b84474109c123bbd0176a8\": container with ID starting with a36cda53da6b26fa4dc152f0abf33d17f95ff93ab0b84474109c123bbd0176a8 not found: ID does not exist" containerID="a36cda53da6b26fa4dc152f0abf33d17f95ff93ab0b84474109c123bbd0176a8" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.773636 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a36cda53da6b26fa4dc152f0abf33d17f95ff93ab0b84474109c123bbd0176a8"} err="failed to get container status \"a36cda53da6b26fa4dc152f0abf33d17f95ff93ab0b84474109c123bbd0176a8\": rpc error: code = NotFound desc = could not find container \"a36cda53da6b26fa4dc152f0abf33d17f95ff93ab0b84474109c123bbd0176a8\": container with ID starting with a36cda53da6b26fa4dc152f0abf33d17f95ff93ab0b84474109c123bbd0176a8 not found: ID does not exist" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.773652 4772 scope.go:117] "RemoveContainer" containerID="20892681e030eeb363de84f7efb9def934d634922b8bfc86ec12ed76e131d504" Jan 27 15:30:35 crc kubenswrapper[4772]: E0127 15:30:35.773868 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20892681e030eeb363de84f7efb9def934d634922b8bfc86ec12ed76e131d504\": container with ID starting with 20892681e030eeb363de84f7efb9def934d634922b8bfc86ec12ed76e131d504 not found: ID does not exist" containerID="20892681e030eeb363de84f7efb9def934d634922b8bfc86ec12ed76e131d504" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.773892 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20892681e030eeb363de84f7efb9def934d634922b8bfc86ec12ed76e131d504"} err="failed to get container status \"20892681e030eeb363de84f7efb9def934d634922b8bfc86ec12ed76e131d504\": rpc error: code = NotFound desc = could not find container \"20892681e030eeb363de84f7efb9def934d634922b8bfc86ec12ed76e131d504\": container with ID starting with 20892681e030eeb363de84f7efb9def934d634922b8bfc86ec12ed76e131d504 not found: ID does not exist" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.773907 4772 scope.go:117] "RemoveContainer" containerID="7d804eabf1929c1b2bbdc21b8685f4ea2301f718b175dd856db56adb377cd8ed" Jan 27 15:30:35 crc kubenswrapper[4772]: E0127 15:30:35.774249 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d804eabf1929c1b2bbdc21b8685f4ea2301f718b175dd856db56adb377cd8ed\": container with ID starting with 7d804eabf1929c1b2bbdc21b8685f4ea2301f718b175dd856db56adb377cd8ed not found: ID does not exist" containerID="7d804eabf1929c1b2bbdc21b8685f4ea2301f718b175dd856db56adb377cd8ed" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.774388 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d804eabf1929c1b2bbdc21b8685f4ea2301f718b175dd856db56adb377cd8ed"} err="failed to get container status \"7d804eabf1929c1b2bbdc21b8685f4ea2301f718b175dd856db56adb377cd8ed\": rpc error: code = NotFound desc = could not find container \"7d804eabf1929c1b2bbdc21b8685f4ea2301f718b175dd856db56adb377cd8ed\": container with ID starting with 7d804eabf1929c1b2bbdc21b8685f4ea2301f718b175dd856db56adb377cd8ed not found: ID does not exist" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.778437 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.778496 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-config-data\") pod \"ceilometer-0\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.778532 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72jk7\" (UniqueName: \"kubernetes.io/projected/aea5ee43-36e3-437d-8aca-b2faedd87c5b-kube-api-access-72jk7\") pod \"ceilometer-0\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.778555 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.778590 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aea5ee43-36e3-437d-8aca-b2faedd87c5b-log-httpd\") pod \"ceilometer-0\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.778642 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aea5ee43-36e3-437d-8aca-b2faedd87c5b-run-httpd\") pod \"ceilometer-0\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.778659 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.778692 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-scripts\") pod \"ceilometer-0\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.780505 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aea5ee43-36e3-437d-8aca-b2faedd87c5b-log-httpd\") pod \"ceilometer-0\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.780720 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aea5ee43-36e3-437d-8aca-b2faedd87c5b-run-httpd\") pod \"ceilometer-0\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.782981 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.783719 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-scripts\") pod \"ceilometer-0\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.783951 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.785343 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-config-data\") pod \"ceilometer-0\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.785600 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.796639 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72jk7\" (UniqueName: \"kubernetes.io/projected/aea5ee43-36e3-437d-8aca-b2faedd87c5b-kube-api-access-72jk7\") pod \"ceilometer-0\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " pod="openstack/ceilometer-0" Jan 27 15:30:35 crc kubenswrapper[4772]: I0127 15:30:35.888529 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:30:36 crc kubenswrapper[4772]: I0127 15:30:36.369095 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:30:36 crc kubenswrapper[4772]: W0127 15:30:36.375537 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaea5ee43_36e3_437d_8aca_b2faedd87c5b.slice/crio-0d9e1d64ee2212bcbce9b483a76517d64478f416567bc79c87cd9fc874d3b4e1 WatchSource:0}: Error finding container 0d9e1d64ee2212bcbce9b483a76517d64478f416567bc79c87cd9fc874d3b4e1: Status 404 returned error can't find the container with id 0d9e1d64ee2212bcbce9b483a76517d64478f416567bc79c87cd9fc874d3b4e1 Jan 27 15:30:36 crc kubenswrapper[4772]: I0127 15:30:36.458913 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aea5ee43-36e3-437d-8aca-b2faedd87c5b","Type":"ContainerStarted","Data":"0d9e1d64ee2212bcbce9b483a76517d64478f416567bc79c87cd9fc874d3b4e1"} Jan 27 15:30:36 crc kubenswrapper[4772]: I0127 15:30:36.677028 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="445d3e38-8f68-4dad-9e97-d927d60ee1e4" path="/var/lib/kubelet/pods/445d3e38-8f68-4dad-9e97-d927d60ee1e4/volumes" Jan 27 15:30:36 crc kubenswrapper[4772]: I0127 15:30:36.678183 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73ee81ee-57fa-466a-8ada-2fa4da5987a0" path="/var/lib/kubelet/pods/73ee81ee-57fa-466a-8ada-2fa4da5987a0/volumes" Jan 27 15:30:38 crc kubenswrapper[4772]: I0127 15:30:38.481407 4772 generic.go:334] "Generic (PLEG): container finished" podID="2d10501c-aefc-4b6b-934a-cd53db7aa029" containerID="7c4a200cbf0e299c55e6a175696503b17a34325960db8f2addd09db07bdebe00" exitCode=0 Jan 27 15:30:38 crc kubenswrapper[4772]: I0127 15:30:38.481511 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dr2w8" event={"ID":"2d10501c-aefc-4b6b-934a-cd53db7aa029","Type":"ContainerDied","Data":"7c4a200cbf0e299c55e6a175696503b17a34325960db8f2addd09db07bdebe00"} Jan 27 15:30:38 crc kubenswrapper[4772]: I0127 15:30:38.485103 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aea5ee43-36e3-437d-8aca-b2faedd87c5b","Type":"ContainerStarted","Data":"a4293d3cbd138216987430f5dab62fa26e55c56743eee0b42dd4fc7797a52afd"} Jan 27 15:30:39 crc kubenswrapper[4772]: I0127 15:30:39.506514 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aea5ee43-36e3-437d-8aca-b2faedd87c5b","Type":"ContainerStarted","Data":"81bb10c06283521cef14702be02bc4e89a7f82e4ae6c7d56b76d0d05f92797d0"} Jan 27 15:30:39 crc kubenswrapper[4772]: I0127 15:30:39.906622 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dr2w8" Jan 27 15:30:39 crc kubenswrapper[4772]: I0127 15:30:39.953877 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d10501c-aefc-4b6b-934a-cd53db7aa029-combined-ca-bundle\") pod \"2d10501c-aefc-4b6b-934a-cd53db7aa029\" (UID: \"2d10501c-aefc-4b6b-934a-cd53db7aa029\") " Jan 27 15:30:39 crc kubenswrapper[4772]: I0127 15:30:39.954514 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d10501c-aefc-4b6b-934a-cd53db7aa029-config-data\") pod \"2d10501c-aefc-4b6b-934a-cd53db7aa029\" (UID: \"2d10501c-aefc-4b6b-934a-cd53db7aa029\") " Jan 27 15:30:39 crc kubenswrapper[4772]: I0127 15:30:39.954572 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d10501c-aefc-4b6b-934a-cd53db7aa029-scripts\") pod \"2d10501c-aefc-4b6b-934a-cd53db7aa029\" (UID: \"2d10501c-aefc-4b6b-934a-cd53db7aa029\") " Jan 27 15:30:39 crc kubenswrapper[4772]: I0127 15:30:39.954714 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xjcn\" (UniqueName: \"kubernetes.io/projected/2d10501c-aefc-4b6b-934a-cd53db7aa029-kube-api-access-4xjcn\") pod \"2d10501c-aefc-4b6b-934a-cd53db7aa029\" (UID: \"2d10501c-aefc-4b6b-934a-cd53db7aa029\") " Jan 27 15:30:39 crc kubenswrapper[4772]: I0127 15:30:39.970531 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d10501c-aefc-4b6b-934a-cd53db7aa029-scripts" (OuterVolumeSpecName: "scripts") pod "2d10501c-aefc-4b6b-934a-cd53db7aa029" (UID: "2d10501c-aefc-4b6b-934a-cd53db7aa029"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:39 crc kubenswrapper[4772]: I0127 15:30:39.987205 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d10501c-aefc-4b6b-934a-cd53db7aa029-kube-api-access-4xjcn" (OuterVolumeSpecName: "kube-api-access-4xjcn") pod "2d10501c-aefc-4b6b-934a-cd53db7aa029" (UID: "2d10501c-aefc-4b6b-934a-cd53db7aa029"). InnerVolumeSpecName "kube-api-access-4xjcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:30:39 crc kubenswrapper[4772]: I0127 15:30:39.988870 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d10501c-aefc-4b6b-934a-cd53db7aa029-config-data" (OuterVolumeSpecName: "config-data") pod "2d10501c-aefc-4b6b-934a-cd53db7aa029" (UID: "2d10501c-aefc-4b6b-934a-cd53db7aa029"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:39 crc kubenswrapper[4772]: I0127 15:30:39.990861 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d10501c-aefc-4b6b-934a-cd53db7aa029-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d10501c-aefc-4b6b-934a-cd53db7aa029" (UID: "2d10501c-aefc-4b6b-934a-cd53db7aa029"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:40 crc kubenswrapper[4772]: I0127 15:30:40.056687 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d10501c-aefc-4b6b-934a-cd53db7aa029-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:40 crc kubenswrapper[4772]: I0127 15:30:40.056753 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d10501c-aefc-4b6b-934a-cd53db7aa029-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:40 crc kubenswrapper[4772]: I0127 15:30:40.056773 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d10501c-aefc-4b6b-934a-cd53db7aa029-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:40 crc kubenswrapper[4772]: I0127 15:30:40.056791 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xjcn\" (UniqueName: \"kubernetes.io/projected/2d10501c-aefc-4b6b-934a-cd53db7aa029-kube-api-access-4xjcn\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:40 crc kubenswrapper[4772]: I0127 15:30:40.520423 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dr2w8" event={"ID":"2d10501c-aefc-4b6b-934a-cd53db7aa029","Type":"ContainerDied","Data":"33208e54b5b97ccefb163c65dd3c0ecf279de8f0faa126d781f9e471398f64fd"} Jan 27 15:30:40 crc kubenswrapper[4772]: I0127 15:30:40.520469 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dr2w8" Jan 27 15:30:40 crc kubenswrapper[4772]: I0127 15:30:40.520475 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33208e54b5b97ccefb163c65dd3c0ecf279de8f0faa126d781f9e471398f64fd" Jan 27 15:30:40 crc kubenswrapper[4772]: I0127 15:30:40.651949 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 27 15:30:40 crc kubenswrapper[4772]: I0127 15:30:40.661774 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 27 15:30:40 crc kubenswrapper[4772]: I0127 15:30:40.673468 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 27 15:30:40 crc kubenswrapper[4772]: I0127 15:30:40.692469 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 27 15:30:40 crc kubenswrapper[4772]: I0127 15:30:40.692754 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="838e8a63-fd3f-4f03-9030-a2c3a4db7393" containerName="nova-api-log" containerID="cri-o://9356a4887a357fc1448800688f7032704f830e8c9f7df6f0b3fe3c97ddfa3bb7" gracePeriod=30 Jan 27 15:30:40 crc kubenswrapper[4772]: I0127 15:30:40.692821 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="838e8a63-fd3f-4f03-9030-a2c3a4db7393" containerName="nova-api-api" containerID="cri-o://596e46cd823d75132da321dc3e1e49f4c351f7ee177713e0655112d945b2390c" gracePeriod=30 Jan 27 15:30:40 crc kubenswrapper[4772]: I0127 15:30:40.717591 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 15:30:40 crc kubenswrapper[4772]: I0127 15:30:40.717885 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c0a8fd83-de04-4458-8ef8-ebe7ae60194f" containerName="nova-scheduler-scheduler" containerID="cri-o://caa4ecd4bd8e1eda4d938d7b43a4c7d54c06fe65669c022620e5448e98a4584e" gracePeriod=30 Jan 27 15:30:40 crc kubenswrapper[4772]: I0127 15:30:40.741080 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 15:30:41 crc kubenswrapper[4772]: I0127 15:30:41.530703 4772 generic.go:334] "Generic (PLEG): container finished" podID="838e8a63-fd3f-4f03-9030-a2c3a4db7393" containerID="596e46cd823d75132da321dc3e1e49f4c351f7ee177713e0655112d945b2390c" exitCode=0 Jan 27 15:30:41 crc kubenswrapper[4772]: I0127 15:30:41.531011 4772 generic.go:334] "Generic (PLEG): container finished" podID="838e8a63-fd3f-4f03-9030-a2c3a4db7393" containerID="9356a4887a357fc1448800688f7032704f830e8c9f7df6f0b3fe3c97ddfa3bb7" exitCode=143 Jan 27 15:30:41 crc kubenswrapper[4772]: I0127 15:30:41.530795 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"838e8a63-fd3f-4f03-9030-a2c3a4db7393","Type":"ContainerDied","Data":"596e46cd823d75132da321dc3e1e49f4c351f7ee177713e0655112d945b2390c"} Jan 27 15:30:41 crc kubenswrapper[4772]: I0127 15:30:41.531062 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"838e8a63-fd3f-4f03-9030-a2c3a4db7393","Type":"ContainerDied","Data":"9356a4887a357fc1448800688f7032704f830e8c9f7df6f0b3fe3c97ddfa3bb7"} Jan 27 15:30:41 crc kubenswrapper[4772]: I0127 15:30:41.588683 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.013226 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.103489 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/838e8a63-fd3f-4f03-9030-a2c3a4db7393-logs\") pod \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\" (UID: \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\") " Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.103610 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/838e8a63-fd3f-4f03-9030-a2c3a4db7393-config-data\") pod \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\" (UID: \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\") " Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.103634 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwgqb\" (UniqueName: \"kubernetes.io/projected/838e8a63-fd3f-4f03-9030-a2c3a4db7393-kube-api-access-kwgqb\") pod \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\" (UID: \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\") " Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.103691 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/838e8a63-fd3f-4f03-9030-a2c3a4db7393-internal-tls-certs\") pod \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\" (UID: \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\") " Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.103752 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/838e8a63-fd3f-4f03-9030-a2c3a4db7393-combined-ca-bundle\") pod \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\" (UID: \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\") " Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.103792 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/838e8a63-fd3f-4f03-9030-a2c3a4db7393-public-tls-certs\") pod \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\" (UID: \"838e8a63-fd3f-4f03-9030-a2c3a4db7393\") " Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.104736 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/838e8a63-fd3f-4f03-9030-a2c3a4db7393-logs" (OuterVolumeSpecName: "logs") pod "838e8a63-fd3f-4f03-9030-a2c3a4db7393" (UID: "838e8a63-fd3f-4f03-9030-a2c3a4db7393"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.117418 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/838e8a63-fd3f-4f03-9030-a2c3a4db7393-kube-api-access-kwgqb" (OuterVolumeSpecName: "kube-api-access-kwgqb") pod "838e8a63-fd3f-4f03-9030-a2c3a4db7393" (UID: "838e8a63-fd3f-4f03-9030-a2c3a4db7393"). InnerVolumeSpecName "kube-api-access-kwgqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.147813 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/838e8a63-fd3f-4f03-9030-a2c3a4db7393-config-data" (OuterVolumeSpecName: "config-data") pod "838e8a63-fd3f-4f03-9030-a2c3a4db7393" (UID: "838e8a63-fd3f-4f03-9030-a2c3a4db7393"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.147932 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/838e8a63-fd3f-4f03-9030-a2c3a4db7393-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "838e8a63-fd3f-4f03-9030-a2c3a4db7393" (UID: "838e8a63-fd3f-4f03-9030-a2c3a4db7393"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.181966 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/838e8a63-fd3f-4f03-9030-a2c3a4db7393-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "838e8a63-fd3f-4f03-9030-a2c3a4db7393" (UID: "838e8a63-fd3f-4f03-9030-a2c3a4db7393"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.205712 4772 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/838e8a63-fd3f-4f03-9030-a2c3a4db7393-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.205740 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/838e8a63-fd3f-4f03-9030-a2c3a4db7393-logs\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.205750 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/838e8a63-fd3f-4f03-9030-a2c3a4db7393-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.205759 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwgqb\" (UniqueName: \"kubernetes.io/projected/838e8a63-fd3f-4f03-9030-a2c3a4db7393-kube-api-access-kwgqb\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.205771 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/838e8a63-fd3f-4f03-9030-a2c3a4db7393-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.239918 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/838e8a63-fd3f-4f03-9030-a2c3a4db7393-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "838e8a63-fd3f-4f03-9030-a2c3a4db7393" (UID: "838e8a63-fd3f-4f03-9030-a2c3a4db7393"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.307482 4772 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/838e8a63-fd3f-4f03-9030-a2c3a4db7393-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.544773 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aea5ee43-36e3-437d-8aca-b2faedd87c5b","Type":"ContainerStarted","Data":"7b0085db2ce3021657d7773e88196b66b6759beeca3bff2b51fc3fdf5d6b4bd2"} Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.546133 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"838e8a63-fd3f-4f03-9030-a2c3a4db7393","Type":"ContainerDied","Data":"f7d96b6a64da6759e6a7a8bc3df6aebfecfa7e5ec0b668d726902cda154a1888"} Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.546253 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.546283 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3293f51d-380b-4247-b1ca-5d1f4b831e52" containerName="nova-metadata-log" containerID="cri-o://2fca126b35386587e45cc8336f3cbcb29951790c38ccf77f0aca8a45525bbaba" gracePeriod=30 Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.546339 4772 scope.go:117] "RemoveContainer" containerID="596e46cd823d75132da321dc3e1e49f4c351f7ee177713e0655112d945b2390c" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.546390 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3293f51d-380b-4247-b1ca-5d1f4b831e52" containerName="nova-metadata-metadata" containerID="cri-o://5263e57d66912aeb6763dff6b7e8221a46b29b91e8c0bceeae9c0497237e7ec2" gracePeriod=30 Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.587437 4772 scope.go:117] "RemoveContainer" containerID="9356a4887a357fc1448800688f7032704f830e8c9f7df6f0b3fe3c97ddfa3bb7" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.638880 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.675919 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.679292 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 27 15:30:42 crc kubenswrapper[4772]: E0127 15:30:42.679793 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="838e8a63-fd3f-4f03-9030-a2c3a4db7393" containerName="nova-api-log" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.679816 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="838e8a63-fd3f-4f03-9030-a2c3a4db7393" containerName="nova-api-log" Jan 27 15:30:42 crc kubenswrapper[4772]: E0127 15:30:42.679831 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d10501c-aefc-4b6b-934a-cd53db7aa029" containerName="nova-manage" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.679838 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d10501c-aefc-4b6b-934a-cd53db7aa029" containerName="nova-manage" Jan 27 15:30:42 crc kubenswrapper[4772]: E0127 15:30:42.679891 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="838e8a63-fd3f-4f03-9030-a2c3a4db7393" containerName="nova-api-api" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.679905 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="838e8a63-fd3f-4f03-9030-a2c3a4db7393" containerName="nova-api-api" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.680132 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d10501c-aefc-4b6b-934a-cd53db7aa029" containerName="nova-manage" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.680154 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="838e8a63-fd3f-4f03-9030-a2c3a4db7393" containerName="nova-api-api" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.680183 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="838e8a63-fd3f-4f03-9030-a2c3a4db7393" containerName="nova-api-log" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.681229 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.683442 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.683788 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.684300 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.696046 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.716260 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-config-data\") pod \"nova-api-0\" (UID: \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\") " pod="openstack/nova-api-0" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.716344 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\") " pod="openstack/nova-api-0" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.716375 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-logs\") pod \"nova-api-0\" (UID: \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\") " pod="openstack/nova-api-0" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.716478 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\") " pod="openstack/nova-api-0" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.716511 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmnhr\" (UniqueName: \"kubernetes.io/projected/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-kube-api-access-pmnhr\") pod \"nova-api-0\" (UID: \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\") " pod="openstack/nova-api-0" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.716536 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-public-tls-certs\") pod \"nova-api-0\" (UID: \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\") " pod="openstack/nova-api-0" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.817707 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\") " pod="openstack/nova-api-0" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.817756 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmnhr\" (UniqueName: \"kubernetes.io/projected/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-kube-api-access-pmnhr\") pod \"nova-api-0\" (UID: \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\") " pod="openstack/nova-api-0" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.817781 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-public-tls-certs\") pod \"nova-api-0\" (UID: \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\") " pod="openstack/nova-api-0" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.817879 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-config-data\") pod \"nova-api-0\" (UID: \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\") " pod="openstack/nova-api-0" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.817927 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\") " pod="openstack/nova-api-0" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.817941 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-logs\") pod \"nova-api-0\" (UID: \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\") " pod="openstack/nova-api-0" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.818427 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-logs\") pod \"nova-api-0\" (UID: \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\") " pod="openstack/nova-api-0" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.821729 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\") " pod="openstack/nova-api-0" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.822917 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\") " pod="openstack/nova-api-0" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.822970 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-config-data\") pod \"nova-api-0\" (UID: \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\") " pod="openstack/nova-api-0" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.824257 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-public-tls-certs\") pod \"nova-api-0\" (UID: \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\") " pod="openstack/nova-api-0" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.834383 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmnhr\" (UniqueName: \"kubernetes.io/projected/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-kube-api-access-pmnhr\") pod \"nova-api-0\" (UID: \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\") " pod="openstack/nova-api-0" Jan 27 15:30:42 crc kubenswrapper[4772]: I0127 15:30:42.999157 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 15:30:43 crc kubenswrapper[4772]: E0127 15:30:43.315219 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="caa4ecd4bd8e1eda4d938d7b43a4c7d54c06fe65669c022620e5448e98a4584e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 27 15:30:43 crc kubenswrapper[4772]: E0127 15:30:43.318087 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="caa4ecd4bd8e1eda4d938d7b43a4c7d54c06fe65669c022620e5448e98a4584e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 27 15:30:43 crc kubenswrapper[4772]: E0127 15:30:43.319589 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="caa4ecd4bd8e1eda4d938d7b43a4c7d54c06fe65669c022620e5448e98a4584e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 27 15:30:43 crc kubenswrapper[4772]: E0127 15:30:43.319679 4772 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="c0a8fd83-de04-4458-8ef8-ebe7ae60194f" containerName="nova-scheduler-scheduler" Jan 27 15:30:43 crc kubenswrapper[4772]: I0127 15:30:43.475981 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 27 15:30:43 crc kubenswrapper[4772]: I0127 15:30:43.559336 4772 generic.go:334] "Generic (PLEG): container finished" podID="3293f51d-380b-4247-b1ca-5d1f4b831e52" containerID="2fca126b35386587e45cc8336f3cbcb29951790c38ccf77f0aca8a45525bbaba" exitCode=143 Jan 27 15:30:43 crc kubenswrapper[4772]: I0127 15:30:43.559425 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3293f51d-380b-4247-b1ca-5d1f4b831e52","Type":"ContainerDied","Data":"2fca126b35386587e45cc8336f3cbcb29951790c38ccf77f0aca8a45525bbaba"} Jan 27 15:30:43 crc kubenswrapper[4772]: I0127 15:30:43.561432 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"93c8f9a4-c6ef-42b8-8543-ff8b5347977e","Type":"ContainerStarted","Data":"dfddffa6f559c177ea99d7f7fef5a8fb81a5dc7c7f2005faaf77278166e23279"} Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.382395 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.466996 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rt2mz\" (UniqueName: \"kubernetes.io/projected/c0a8fd83-de04-4458-8ef8-ebe7ae60194f-kube-api-access-rt2mz\") pod \"c0a8fd83-de04-4458-8ef8-ebe7ae60194f\" (UID: \"c0a8fd83-de04-4458-8ef8-ebe7ae60194f\") " Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.467099 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0a8fd83-de04-4458-8ef8-ebe7ae60194f-config-data\") pod \"c0a8fd83-de04-4458-8ef8-ebe7ae60194f\" (UID: \"c0a8fd83-de04-4458-8ef8-ebe7ae60194f\") " Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.467205 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0a8fd83-de04-4458-8ef8-ebe7ae60194f-combined-ca-bundle\") pod \"c0a8fd83-de04-4458-8ef8-ebe7ae60194f\" (UID: \"c0a8fd83-de04-4458-8ef8-ebe7ae60194f\") " Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.475020 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0a8fd83-de04-4458-8ef8-ebe7ae60194f-kube-api-access-rt2mz" (OuterVolumeSpecName: "kube-api-access-rt2mz") pod "c0a8fd83-de04-4458-8ef8-ebe7ae60194f" (UID: "c0a8fd83-de04-4458-8ef8-ebe7ae60194f"). InnerVolumeSpecName "kube-api-access-rt2mz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.501340 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0a8fd83-de04-4458-8ef8-ebe7ae60194f-config-data" (OuterVolumeSpecName: "config-data") pod "c0a8fd83-de04-4458-8ef8-ebe7ae60194f" (UID: "c0a8fd83-de04-4458-8ef8-ebe7ae60194f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.507359 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0a8fd83-de04-4458-8ef8-ebe7ae60194f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c0a8fd83-de04-4458-8ef8-ebe7ae60194f" (UID: "c0a8fd83-de04-4458-8ef8-ebe7ae60194f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.569777 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0a8fd83-de04-4458-8ef8-ebe7ae60194f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.569809 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rt2mz\" (UniqueName: \"kubernetes.io/projected/c0a8fd83-de04-4458-8ef8-ebe7ae60194f-kube-api-access-rt2mz\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.569822 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0a8fd83-de04-4458-8ef8-ebe7ae60194f-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.580385 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"93c8f9a4-c6ef-42b8-8543-ff8b5347977e","Type":"ContainerStarted","Data":"abfb528c89657cd0985ff90de17dace11a1be4c50ae49dc95a4a7ec03d093246"} Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.580438 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"93c8f9a4-c6ef-42b8-8543-ff8b5347977e","Type":"ContainerStarted","Data":"db0ed28f713318389578164d5ba7364ff5e0ca569d4bd32de0483eb615fe7597"} Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.584653 4772 generic.go:334] "Generic (PLEG): container finished" podID="c0a8fd83-de04-4458-8ef8-ebe7ae60194f" containerID="caa4ecd4bd8e1eda4d938d7b43a4c7d54c06fe65669c022620e5448e98a4584e" exitCode=0 Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.584729 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c0a8fd83-de04-4458-8ef8-ebe7ae60194f","Type":"ContainerDied","Data":"caa4ecd4bd8e1eda4d938d7b43a4c7d54c06fe65669c022620e5448e98a4584e"} Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.584829 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c0a8fd83-de04-4458-8ef8-ebe7ae60194f","Type":"ContainerDied","Data":"4d588204c22b2f28d3f4e69bf8dac95975b2db88030c7b5662a17b491819b8ab"} Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.584866 4772 scope.go:117] "RemoveContainer" containerID="caa4ecd4bd8e1eda4d938d7b43a4c7d54c06fe65669c022620e5448e98a4584e" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.584775 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.592624 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aea5ee43-36e3-437d-8aca-b2faedd87c5b","Type":"ContainerStarted","Data":"0447c2ea1d147e4cee27fce146e4edc38d746774dc492452f5da3c48df7973bb"} Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.593661 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.618484 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.618460994 podStartE2EDuration="2.618460994s" podCreationTimestamp="2026-01-27 15:30:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:30:44.60480723 +0000 UTC m=+1430.585416328" watchObservedRunningTime="2026-01-27 15:30:44.618460994 +0000 UTC m=+1430.599070102" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.622936 4772 scope.go:117] "RemoveContainer" containerID="caa4ecd4bd8e1eda4d938d7b43a4c7d54c06fe65669c022620e5448e98a4584e" Jan 27 15:30:44 crc kubenswrapper[4772]: E0127 15:30:44.625756 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"caa4ecd4bd8e1eda4d938d7b43a4c7d54c06fe65669c022620e5448e98a4584e\": container with ID starting with caa4ecd4bd8e1eda4d938d7b43a4c7d54c06fe65669c022620e5448e98a4584e not found: ID does not exist" containerID="caa4ecd4bd8e1eda4d938d7b43a4c7d54c06fe65669c022620e5448e98a4584e" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.625805 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"caa4ecd4bd8e1eda4d938d7b43a4c7d54c06fe65669c022620e5448e98a4584e"} err="failed to get container status \"caa4ecd4bd8e1eda4d938d7b43a4c7d54c06fe65669c022620e5448e98a4584e\": rpc error: code = NotFound desc = could not find container \"caa4ecd4bd8e1eda4d938d7b43a4c7d54c06fe65669c022620e5448e98a4584e\": container with ID starting with caa4ecd4bd8e1eda4d938d7b43a4c7d54c06fe65669c022620e5448e98a4584e not found: ID does not exist" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.643498 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.226216639 podStartE2EDuration="9.643468717s" podCreationTimestamp="2026-01-27 15:30:35 +0000 UTC" firstStartedPulling="2026-01-27 15:30:36.378154601 +0000 UTC m=+1422.358763709" lastFinishedPulling="2026-01-27 15:30:43.795406689 +0000 UTC m=+1429.776015787" observedRunningTime="2026-01-27 15:30:44.628873715 +0000 UTC m=+1430.609482813" watchObservedRunningTime="2026-01-27 15:30:44.643468717 +0000 UTC m=+1430.624077915" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.679595 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="838e8a63-fd3f-4f03-9030-a2c3a4db7393" path="/var/lib/kubelet/pods/838e8a63-fd3f-4f03-9030-a2c3a4db7393/volumes" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.680208 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.687626 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.710833 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 15:30:44 crc kubenswrapper[4772]: E0127 15:30:44.711294 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0a8fd83-de04-4458-8ef8-ebe7ae60194f" containerName="nova-scheduler-scheduler" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.711313 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0a8fd83-de04-4458-8ef8-ebe7ae60194f" containerName="nova-scheduler-scheduler" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.711505 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0a8fd83-de04-4458-8ef8-ebe7ae60194f" containerName="nova-scheduler-scheduler" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.712105 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.713949 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.719738 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.773136 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b83f7578-8113-46c8-be24-5968aa0ca563-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b83f7578-8113-46c8-be24-5968aa0ca563\") " pod="openstack/nova-scheduler-0" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.773349 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b83f7578-8113-46c8-be24-5968aa0ca563-config-data\") pod \"nova-scheduler-0\" (UID: \"b83f7578-8113-46c8-be24-5968aa0ca563\") " pod="openstack/nova-scheduler-0" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.773449 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq4lp\" (UniqueName: \"kubernetes.io/projected/b83f7578-8113-46c8-be24-5968aa0ca563-kube-api-access-hq4lp\") pod \"nova-scheduler-0\" (UID: \"b83f7578-8113-46c8-be24-5968aa0ca563\") " pod="openstack/nova-scheduler-0" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.875224 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b83f7578-8113-46c8-be24-5968aa0ca563-config-data\") pod \"nova-scheduler-0\" (UID: \"b83f7578-8113-46c8-be24-5968aa0ca563\") " pod="openstack/nova-scheduler-0" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.875296 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq4lp\" (UniqueName: \"kubernetes.io/projected/b83f7578-8113-46c8-be24-5968aa0ca563-kube-api-access-hq4lp\") pod \"nova-scheduler-0\" (UID: \"b83f7578-8113-46c8-be24-5968aa0ca563\") " pod="openstack/nova-scheduler-0" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.875416 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b83f7578-8113-46c8-be24-5968aa0ca563-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b83f7578-8113-46c8-be24-5968aa0ca563\") " pod="openstack/nova-scheduler-0" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.879426 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b83f7578-8113-46c8-be24-5968aa0ca563-config-data\") pod \"nova-scheduler-0\" (UID: \"b83f7578-8113-46c8-be24-5968aa0ca563\") " pod="openstack/nova-scheduler-0" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.879658 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b83f7578-8113-46c8-be24-5968aa0ca563-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b83f7578-8113-46c8-be24-5968aa0ca563\") " pod="openstack/nova-scheduler-0" Jan 27 15:30:44 crc kubenswrapper[4772]: I0127 15:30:44.906776 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq4lp\" (UniqueName: \"kubernetes.io/projected/b83f7578-8113-46c8-be24-5968aa0ca563-kube-api-access-hq4lp\") pod \"nova-scheduler-0\" (UID: \"b83f7578-8113-46c8-be24-5968aa0ca563\") " pod="openstack/nova-scheduler-0" Jan 27 15:30:45 crc kubenswrapper[4772]: I0127 15:30:45.038811 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 15:30:45 crc kubenswrapper[4772]: I0127 15:30:45.516199 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 15:30:45 crc kubenswrapper[4772]: W0127 15:30:45.520338 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb83f7578_8113_46c8_be24_5968aa0ca563.slice/crio-ee6d65efb4f3df3d96335b7b6b58d4ee20a12c71c0ca644b8c8c4208300d2710 WatchSource:0}: Error finding container ee6d65efb4f3df3d96335b7b6b58d4ee20a12c71c0ca644b8c8c4208300d2710: Status 404 returned error can't find the container with id ee6d65efb4f3df3d96335b7b6b58d4ee20a12c71c0ca644b8c8c4208300d2710 Jan 27 15:30:45 crc kubenswrapper[4772]: I0127 15:30:45.617581 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b83f7578-8113-46c8-be24-5968aa0ca563","Type":"ContainerStarted","Data":"ee6d65efb4f3df3d96335b7b6b58d4ee20a12c71c0ca644b8c8c4208300d2710"} Jan 27 15:30:45 crc kubenswrapper[4772]: I0127 15:30:45.702973 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="3293f51d-380b-4247-b1ca-5d1f4b831e52" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": read tcp 10.217.0.2:37576->10.217.0.199:8775: read: connection reset by peer" Jan 27 15:30:45 crc kubenswrapper[4772]: I0127 15:30:45.703507 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="3293f51d-380b-4247-b1ca-5d1f4b831e52" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": read tcp 10.217.0.2:37582->10.217.0.199:8775: read: connection reset by peer" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.175109 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.306449 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3293f51d-380b-4247-b1ca-5d1f4b831e52-nova-metadata-tls-certs\") pod \"3293f51d-380b-4247-b1ca-5d1f4b831e52\" (UID: \"3293f51d-380b-4247-b1ca-5d1f4b831e52\") " Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.306603 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3293f51d-380b-4247-b1ca-5d1f4b831e52-logs\") pod \"3293f51d-380b-4247-b1ca-5d1f4b831e52\" (UID: \"3293f51d-380b-4247-b1ca-5d1f4b831e52\") " Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.307052 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3293f51d-380b-4247-b1ca-5d1f4b831e52-logs" (OuterVolumeSpecName: "logs") pod "3293f51d-380b-4247-b1ca-5d1f4b831e52" (UID: "3293f51d-380b-4247-b1ca-5d1f4b831e52"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.307094 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6whx\" (UniqueName: \"kubernetes.io/projected/3293f51d-380b-4247-b1ca-5d1f4b831e52-kube-api-access-v6whx\") pod \"3293f51d-380b-4247-b1ca-5d1f4b831e52\" (UID: \"3293f51d-380b-4247-b1ca-5d1f4b831e52\") " Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.307194 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3293f51d-380b-4247-b1ca-5d1f4b831e52-combined-ca-bundle\") pod \"3293f51d-380b-4247-b1ca-5d1f4b831e52\" (UID: \"3293f51d-380b-4247-b1ca-5d1f4b831e52\") " Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.307550 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3293f51d-380b-4247-b1ca-5d1f4b831e52-config-data\") pod \"3293f51d-380b-4247-b1ca-5d1f4b831e52\" (UID: \"3293f51d-380b-4247-b1ca-5d1f4b831e52\") " Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.307954 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3293f51d-380b-4247-b1ca-5d1f4b831e52-logs\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.332468 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3293f51d-380b-4247-b1ca-5d1f4b831e52-kube-api-access-v6whx" (OuterVolumeSpecName: "kube-api-access-v6whx") pod "3293f51d-380b-4247-b1ca-5d1f4b831e52" (UID: "3293f51d-380b-4247-b1ca-5d1f4b831e52"). InnerVolumeSpecName "kube-api-access-v6whx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.335509 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3293f51d-380b-4247-b1ca-5d1f4b831e52-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3293f51d-380b-4247-b1ca-5d1f4b831e52" (UID: "3293f51d-380b-4247-b1ca-5d1f4b831e52"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.342231 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3293f51d-380b-4247-b1ca-5d1f4b831e52-config-data" (OuterVolumeSpecName: "config-data") pod "3293f51d-380b-4247-b1ca-5d1f4b831e52" (UID: "3293f51d-380b-4247-b1ca-5d1f4b831e52"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.360561 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3293f51d-380b-4247-b1ca-5d1f4b831e52-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "3293f51d-380b-4247-b1ca-5d1f4b831e52" (UID: "3293f51d-380b-4247-b1ca-5d1f4b831e52"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.409489 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3293f51d-380b-4247-b1ca-5d1f4b831e52-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.409519 4772 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3293f51d-380b-4247-b1ca-5d1f4b831e52-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.409529 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6whx\" (UniqueName: \"kubernetes.io/projected/3293f51d-380b-4247-b1ca-5d1f4b831e52-kube-api-access-v6whx\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.409537 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3293f51d-380b-4247-b1ca-5d1f4b831e52-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.638570 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b83f7578-8113-46c8-be24-5968aa0ca563","Type":"ContainerStarted","Data":"e92037085b98ccc46bbd64416f98018d2426a17d0883dd17b830d5574a8a0f4e"} Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.644970 4772 generic.go:334] "Generic (PLEG): container finished" podID="3293f51d-380b-4247-b1ca-5d1f4b831e52" containerID="5263e57d66912aeb6763dff6b7e8221a46b29b91e8c0bceeae9c0497237e7ec2" exitCode=0 Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.645862 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.653102 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3293f51d-380b-4247-b1ca-5d1f4b831e52","Type":"ContainerDied","Data":"5263e57d66912aeb6763dff6b7e8221a46b29b91e8c0bceeae9c0497237e7ec2"} Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.653204 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3293f51d-380b-4247-b1ca-5d1f4b831e52","Type":"ContainerDied","Data":"bc2f5b8265c22782479449303b876d41140ba5bbf29e25aebf07297950204c86"} Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.653231 4772 scope.go:117] "RemoveContainer" containerID="5263e57d66912aeb6763dff6b7e8221a46b29b91e8c0bceeae9c0497237e7ec2" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.663764 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.6637495810000003 podStartE2EDuration="2.663749581s" podCreationTimestamp="2026-01-27 15:30:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:30:46.658765107 +0000 UTC m=+1432.639374255" watchObservedRunningTime="2026-01-27 15:30:46.663749581 +0000 UTC m=+1432.644358679" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.684365 4772 scope.go:117] "RemoveContainer" containerID="2fca126b35386587e45cc8336f3cbcb29951790c38ccf77f0aca8a45525bbaba" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.687570 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0a8fd83-de04-4458-8ef8-ebe7ae60194f" path="/var/lib/kubelet/pods/c0a8fd83-de04-4458-8ef8-ebe7ae60194f/volumes" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.712428 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.713668 4772 scope.go:117] "RemoveContainer" containerID="5263e57d66912aeb6763dff6b7e8221a46b29b91e8c0bceeae9c0497237e7ec2" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.725468 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 15:30:46 crc kubenswrapper[4772]: E0127 15:30:46.725630 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5263e57d66912aeb6763dff6b7e8221a46b29b91e8c0bceeae9c0497237e7ec2\": container with ID starting with 5263e57d66912aeb6763dff6b7e8221a46b29b91e8c0bceeae9c0497237e7ec2 not found: ID does not exist" containerID="5263e57d66912aeb6763dff6b7e8221a46b29b91e8c0bceeae9c0497237e7ec2" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.725803 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5263e57d66912aeb6763dff6b7e8221a46b29b91e8c0bceeae9c0497237e7ec2"} err="failed to get container status \"5263e57d66912aeb6763dff6b7e8221a46b29b91e8c0bceeae9c0497237e7ec2\": rpc error: code = NotFound desc = could not find container \"5263e57d66912aeb6763dff6b7e8221a46b29b91e8c0bceeae9c0497237e7ec2\": container with ID starting with 5263e57d66912aeb6763dff6b7e8221a46b29b91e8c0bceeae9c0497237e7ec2 not found: ID does not exist" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.725847 4772 scope.go:117] "RemoveContainer" containerID="2fca126b35386587e45cc8336f3cbcb29951790c38ccf77f0aca8a45525bbaba" Jan 27 15:30:46 crc kubenswrapper[4772]: E0127 15:30:46.726327 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fca126b35386587e45cc8336f3cbcb29951790c38ccf77f0aca8a45525bbaba\": container with ID starting with 2fca126b35386587e45cc8336f3cbcb29951790c38ccf77f0aca8a45525bbaba not found: ID does not exist" containerID="2fca126b35386587e45cc8336f3cbcb29951790c38ccf77f0aca8a45525bbaba" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.726370 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fca126b35386587e45cc8336f3cbcb29951790c38ccf77f0aca8a45525bbaba"} err="failed to get container status \"2fca126b35386587e45cc8336f3cbcb29951790c38ccf77f0aca8a45525bbaba\": rpc error: code = NotFound desc = could not find container \"2fca126b35386587e45cc8336f3cbcb29951790c38ccf77f0aca8a45525bbaba\": container with ID starting with 2fca126b35386587e45cc8336f3cbcb29951790c38ccf77f0aca8a45525bbaba not found: ID does not exist" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.734356 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 27 15:30:46 crc kubenswrapper[4772]: E0127 15:30:46.734725 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3293f51d-380b-4247-b1ca-5d1f4b831e52" containerName="nova-metadata-log" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.734740 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3293f51d-380b-4247-b1ca-5d1f4b831e52" containerName="nova-metadata-log" Jan 27 15:30:46 crc kubenswrapper[4772]: E0127 15:30:46.734761 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3293f51d-380b-4247-b1ca-5d1f4b831e52" containerName="nova-metadata-metadata" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.734769 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3293f51d-380b-4247-b1ca-5d1f4b831e52" containerName="nova-metadata-metadata" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.734952 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3293f51d-380b-4247-b1ca-5d1f4b831e52" containerName="nova-metadata-log" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.734973 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3293f51d-380b-4247-b1ca-5d1f4b831e52" containerName="nova-metadata-metadata" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.736056 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.739398 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.739675 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.740517 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.820286 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\") " pod="openstack/nova-metadata-0" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.820400 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-config-data\") pod \"nova-metadata-0\" (UID: \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\") " pod="openstack/nova-metadata-0" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.820430 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\") " pod="openstack/nova-metadata-0" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.820451 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b44p\" (UniqueName: \"kubernetes.io/projected/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-kube-api-access-6b44p\") pod \"nova-metadata-0\" (UID: \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\") " pod="openstack/nova-metadata-0" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.820484 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-logs\") pod \"nova-metadata-0\" (UID: \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\") " pod="openstack/nova-metadata-0" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.921641 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\") " pod="openstack/nova-metadata-0" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.921745 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-config-data\") pod \"nova-metadata-0\" (UID: \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\") " pod="openstack/nova-metadata-0" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.921764 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\") " pod="openstack/nova-metadata-0" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.921790 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b44p\" (UniqueName: \"kubernetes.io/projected/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-kube-api-access-6b44p\") pod \"nova-metadata-0\" (UID: \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\") " pod="openstack/nova-metadata-0" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.921820 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-logs\") pod \"nova-metadata-0\" (UID: \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\") " pod="openstack/nova-metadata-0" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.922293 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-logs\") pod \"nova-metadata-0\" (UID: \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\") " pod="openstack/nova-metadata-0" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.926449 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\") " pod="openstack/nova-metadata-0" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.926888 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-config-data\") pod \"nova-metadata-0\" (UID: \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\") " pod="openstack/nova-metadata-0" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.928911 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\") " pod="openstack/nova-metadata-0" Jan 27 15:30:46 crc kubenswrapper[4772]: I0127 15:30:46.943821 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b44p\" (UniqueName: \"kubernetes.io/projected/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-kube-api-access-6b44p\") pod \"nova-metadata-0\" (UID: \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\") " pod="openstack/nova-metadata-0" Jan 27 15:30:47 crc kubenswrapper[4772]: I0127 15:30:47.051652 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 15:30:47 crc kubenswrapper[4772]: I0127 15:30:47.517567 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 15:30:47 crc kubenswrapper[4772]: I0127 15:30:47.666572 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7","Type":"ContainerStarted","Data":"a940184dde4998665ff3925c8d268f050f05912d1265137578965cd151d251c3"} Jan 27 15:30:48 crc kubenswrapper[4772]: I0127 15:30:48.737682 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3293f51d-380b-4247-b1ca-5d1f4b831e52" path="/var/lib/kubelet/pods/3293f51d-380b-4247-b1ca-5d1f4b831e52/volumes" Jan 27 15:30:48 crc kubenswrapper[4772]: I0127 15:30:48.740662 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7","Type":"ContainerStarted","Data":"7343cd6a2a5cf705b558b4cc862d749d392235682218489d0106143cb8a5d4bc"} Jan 27 15:30:49 crc kubenswrapper[4772]: I0127 15:30:49.737283 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7","Type":"ContainerStarted","Data":"db38347574e8ea3471da74617b5c2b8fd8e23430f530dbd434f5aba2a153f9bb"} Jan 27 15:30:49 crc kubenswrapper[4772]: I0127 15:30:49.769234 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.769205734 podStartE2EDuration="3.769205734s" podCreationTimestamp="2026-01-27 15:30:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:30:49.76144535 +0000 UTC m=+1435.742054468" watchObservedRunningTime="2026-01-27 15:30:49.769205734 +0000 UTC m=+1435.749814832" Jan 27 15:30:50 crc kubenswrapper[4772]: I0127 15:30:50.039222 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 27 15:30:52 crc kubenswrapper[4772]: I0127 15:30:52.052568 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 27 15:30:52 crc kubenswrapper[4772]: I0127 15:30:52.052828 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 27 15:30:53 crc kubenswrapper[4772]: I0127 15:30:53.000903 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 27 15:30:53 crc kubenswrapper[4772]: I0127 15:30:53.000985 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 27 15:30:54 crc kubenswrapper[4772]: I0127 15:30:54.013355 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="93c8f9a4-c6ef-42b8-8543-ff8b5347977e" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 27 15:30:54 crc kubenswrapper[4772]: I0127 15:30:54.013373 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="93c8f9a4-c6ef-42b8-8543-ff8b5347977e" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 27 15:30:55 crc kubenswrapper[4772]: I0127 15:30:55.039700 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Jan 27 15:30:55 crc kubenswrapper[4772]: I0127 15:30:55.066710 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Jan 27 15:30:55 crc kubenswrapper[4772]: I0127 15:30:55.836380 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Jan 27 15:30:57 crc kubenswrapper[4772]: I0127 15:30:57.052425 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 27 15:30:57 crc kubenswrapper[4772]: I0127 15:30:57.053467 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 27 15:30:58 crc kubenswrapper[4772]: I0127 15:30:58.066460 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f63bf600-ff03-43a3-92b4-fe8ac68a9bb7" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.207:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 27 15:30:58 crc kubenswrapper[4772]: I0127 15:30:58.067113 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f63bf600-ff03-43a3-92b4-fe8ac68a9bb7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.207:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 27 15:31:03 crc kubenswrapper[4772]: I0127 15:31:03.007151 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 27 15:31:03 crc kubenswrapper[4772]: I0127 15:31:03.008099 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 27 15:31:03 crc kubenswrapper[4772]: I0127 15:31:03.013510 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 27 15:31:03 crc kubenswrapper[4772]: I0127 15:31:03.016132 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 27 15:31:03 crc kubenswrapper[4772]: I0127 15:31:03.886277 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 27 15:31:03 crc kubenswrapper[4772]: I0127 15:31:03.892936 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 27 15:31:05 crc kubenswrapper[4772]: I0127 15:31:05.897268 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Jan 27 15:31:07 crc kubenswrapper[4772]: I0127 15:31:07.056758 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 27 15:31:07 crc kubenswrapper[4772]: I0127 15:31:07.058023 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 27 15:31:07 crc kubenswrapper[4772]: I0127 15:31:07.070576 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 27 15:31:07 crc kubenswrapper[4772]: I0127 15:31:07.926279 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.060962 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.061841 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="0edf6707-14dd-4986-8d64-0e48a31d6a39" containerName="openstackclient" containerID="cri-o://0c37dc673e475cc4ca1e8b831b0543b26650ceedc799dace964e07fb4c7c7ffa" gracePeriod=2 Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.080543 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.108614 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-547fc"] Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.123130 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-547fc"] Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.251629 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-qmppl"] Jan 27 15:31:26 crc kubenswrapper[4772]: E0127 15:31:26.252129 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0edf6707-14dd-4986-8d64-0e48a31d6a39" containerName="openstackclient" Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.252152 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0edf6707-14dd-4986-8d64-0e48a31d6a39" containerName="openstackclient" Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.252432 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="0edf6707-14dd-4986-8d64-0e48a31d6a39" containerName="openstackclient" Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.253124 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-qmppl" Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.259159 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cbf7469-816d-4e54-a7ad-b5b76d0d59d6-operator-scripts\") pod \"root-account-create-update-qmppl\" (UID: \"4cbf7469-816d-4e54-a7ad-b5b76d0d59d6\") " pod="openstack/root-account-create-update-qmppl" Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.259278 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thp4z\" (UniqueName: \"kubernetes.io/projected/4cbf7469-816d-4e54-a7ad-b5b76d0d59d6-kube-api-access-thp4z\") pod \"root-account-create-update-qmppl\" (UID: \"4cbf7469-816d-4e54-a7ad-b5b76d0d59d6\") " pod="openstack/root-account-create-update-qmppl" Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.264732 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.290545 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-qmppl"] Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.311809 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-97c3-account-create-update-xlghl"] Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.413334 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cbf7469-816d-4e54-a7ad-b5b76d0d59d6-operator-scripts\") pod \"root-account-create-update-qmppl\" (UID: \"4cbf7469-816d-4e54-a7ad-b5b76d0d59d6\") " pod="openstack/root-account-create-update-qmppl" Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.417727 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cbf7469-816d-4e54-a7ad-b5b76d0d59d6-operator-scripts\") pod \"root-account-create-update-qmppl\" (UID: \"4cbf7469-816d-4e54-a7ad-b5b76d0d59d6\") " pod="openstack/root-account-create-update-qmppl" Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.418196 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thp4z\" (UniqueName: \"kubernetes.io/projected/4cbf7469-816d-4e54-a7ad-b5b76d0d59d6-kube-api-access-thp4z\") pod \"root-account-create-update-qmppl\" (UID: \"4cbf7469-816d-4e54-a7ad-b5b76d0d59d6\") " pod="openstack/root-account-create-update-qmppl" Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.427244 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-97c3-account-create-update-xlghl"] Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.499261 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thp4z\" (UniqueName: \"kubernetes.io/projected/4cbf7469-816d-4e54-a7ad-b5b76d0d59d6-kube-api-access-thp4z\") pod \"root-account-create-update-qmppl\" (UID: \"4cbf7469-816d-4e54-a7ad-b5b76d0d59d6\") " pod="openstack/root-account-create-update-qmppl" Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.549858 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-97c3-account-create-update-bvlvs"] Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.551534 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-97c3-account-create-update-bvlvs" Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.567841 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.583185 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-qmppl" Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.632214 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j94f\" (UniqueName: \"kubernetes.io/projected/ef060591-3809-4f0b-974f-0785261db9b9-kube-api-access-2j94f\") pod \"barbican-97c3-account-create-update-bvlvs\" (UID: \"ef060591-3809-4f0b-974f-0785261db9b9\") " pod="openstack/barbican-97c3-account-create-update-bvlvs" Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.632359 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef060591-3809-4f0b-974f-0785261db9b9-operator-scripts\") pod \"barbican-97c3-account-create-update-bvlvs\" (UID: \"ef060591-3809-4f0b-974f-0785261db9b9\") " pod="openstack/barbican-97c3-account-create-update-bvlvs" Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.654182 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-97c3-account-create-update-bvlvs"] Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.694951 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af61fb8e-e749-4872-8dc6-c590e4b9787a" path="/var/lib/kubelet/pods/af61fb8e-e749-4872-8dc6-c590e4b9787a/volumes" Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.695636 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd1eaad6-cd29-4189-8ecd-62b7658e69ef" path="/var/lib/kubelet/pods/dd1eaad6-cd29-4189-8ecd-62b7658e69ef/volumes" Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.730280 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-a648-account-create-update-qhx8z"] Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.738407 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef060591-3809-4f0b-974f-0785261db9b9-operator-scripts\") pod \"barbican-97c3-account-create-update-bvlvs\" (UID: \"ef060591-3809-4f0b-974f-0785261db9b9\") " pod="openstack/barbican-97c3-account-create-update-bvlvs" Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.738637 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j94f\" (UniqueName: \"kubernetes.io/projected/ef060591-3809-4f0b-974f-0785261db9b9-kube-api-access-2j94f\") pod \"barbican-97c3-account-create-update-bvlvs\" (UID: \"ef060591-3809-4f0b-974f-0785261db9b9\") " pod="openstack/barbican-97c3-account-create-update-bvlvs" Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.739783 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef060591-3809-4f0b-974f-0785261db9b9-operator-scripts\") pod \"barbican-97c3-account-create-update-bvlvs\" (UID: \"ef060591-3809-4f0b-974f-0785261db9b9\") " pod="openstack/barbican-97c3-account-create-update-bvlvs" Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.800576 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-a648-account-create-update-qhx8z"] Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.834372 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j94f\" (UniqueName: \"kubernetes.io/projected/ef060591-3809-4f0b-974f-0785261db9b9-kube-api-access-2j94f\") pod \"barbican-97c3-account-create-update-bvlvs\" (UID: \"ef060591-3809-4f0b-974f-0785261db9b9\") " pod="openstack/barbican-97c3-account-create-update-bvlvs" Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.864274 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.908912 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-97c3-account-create-update-bvlvs" Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.930271 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.930619 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="8b8101bc-2ddf-48ed-9b92-e8f9e5e71938" containerName="ovn-northd" containerID="cri-o://f351431c9793a13f48f307e65178046dd4ccdc52ebd7ba269a580599ff0da01a" gracePeriod=30 Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.931119 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="8b8101bc-2ddf-48ed-9b92-e8f9e5e71938" containerName="openstack-network-exporter" containerID="cri-o://b1542ba131aec1cffd5520f2969b843d3aa12fe7b4cd60022addce3e73977b99" gracePeriod=30 Jan 27 15:31:26 crc kubenswrapper[4772]: I0127 15:31:26.989238 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-d2a3-account-create-update-hfkkb"] Jan 27 15:31:27 crc kubenswrapper[4772]: E0127 15:31:27.023355 4772 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Jan 27 15:31:27 crc kubenswrapper[4772]: E0127 15:31:27.023743 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-config-data podName:76fdbdb1-d48a-4cd1-8372-78887671dce8 nodeName:}" failed. No retries permitted until 2026-01-27 15:31:27.523722509 +0000 UTC m=+1473.504331597 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-config-data") pod "rabbitmq-cell1-server-0" (UID: "76fdbdb1-d48a-4cd1-8372-78887671dce8") : configmap "rabbitmq-cell1-config-data" not found Jan 27 15:31:27 crc kubenswrapper[4772]: I0127 15:31:27.029968 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-e8b1-account-create-update-8rlww"] Jan 27 15:31:27 crc kubenswrapper[4772]: I0127 15:31:27.065816 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-cqx7r"] Jan 27 15:31:27 crc kubenswrapper[4772]: I0127 15:31:27.107235 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-d2a3-account-create-update-hfkkb"] Jan 27 15:31:27 crc kubenswrapper[4772]: I0127 15:31:27.114032 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-e8b1-account-create-update-8rlww"] Jan 27 15:31:27 crc kubenswrapper[4772]: I0127 15:31:27.147616 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-gxjzh"] Jan 27 15:31:27 crc kubenswrapper[4772]: E0127 15:31:27.156710 4772 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.134:35824->38.129.56.134:35895: write tcp 38.129.56.134:35824->38.129.56.134:35895: write: broken pipe Jan 27 15:31:27 crc kubenswrapper[4772]: I0127 15:31:27.200357 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-vqpfg"] Jan 27 15:31:27 crc kubenswrapper[4772]: I0127 15:31:27.200538 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-vqpfg" podUID="a490a71b-c33d-4c94-9592-f97d1d315e81" containerName="openstack-network-exporter" containerID="cri-o://b93ad84c922746d427d3e2a2deb04a875a239fcafbecb5146ae05b1b11e36a09" gracePeriod=30 Jan 27 15:31:27 crc kubenswrapper[4772]: I0127 15:31:27.254319 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-vdmv7"] Jan 27 15:31:27 crc kubenswrapper[4772]: I0127 15:31:27.286377 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-vdmv7"] Jan 27 15:31:27 crc kubenswrapper[4772]: I0127 15:31:27.326226 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-8l85z"] Jan 27 15:31:27 crc kubenswrapper[4772]: I0127 15:31:27.348835 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-pmk27"] Jan 27 15:31:27 crc kubenswrapper[4772]: I0127 15:31:27.386217 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-8l85z"] Jan 27 15:31:27 crc kubenswrapper[4772]: I0127 15:31:27.416732 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-pmk27"] Jan 27 15:31:27 crc kubenswrapper[4772]: I0127 15:31:27.465278 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-zf2tx"] Jan 27 15:31:27 crc kubenswrapper[4772]: I0127 15:31:27.498315 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-zf2tx"] Jan 27 15:31:27 crc kubenswrapper[4772]: I0127 15:31:27.542407 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-d4llz"] Jan 27 15:31:27 crc kubenswrapper[4772]: E0127 15:31:27.543559 4772 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Jan 27 15:31:27 crc kubenswrapper[4772]: E0127 15:31:27.543608 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-config-data podName:76fdbdb1-d48a-4cd1-8372-78887671dce8 nodeName:}" failed. No retries permitted until 2026-01-27 15:31:28.543593843 +0000 UTC m=+1474.524202941 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-config-data") pod "rabbitmq-cell1-server-0" (UID: "76fdbdb1-d48a-4cd1-8372-78887671dce8") : configmap "rabbitmq-cell1-config-data" not found Jan 27 15:31:27 crc kubenswrapper[4772]: I0127 15:31:27.566877 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-d4llz"] Jan 27 15:31:27 crc kubenswrapper[4772]: I0127 15:31:27.598259 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-5104-account-create-update-vp7x7"] Jan 27 15:31:27 crc kubenswrapper[4772]: I0127 15:31:27.632600 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-5104-account-create-update-vp7x7"] Jan 27 15:31:27 crc kubenswrapper[4772]: I0127 15:31:27.657934 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-v689b"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:27.700513 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:27.755594 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-v689b"] Jan 27 15:31:28 crc kubenswrapper[4772]: E0127 15:31:27.805665 4772 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err="command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: " execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-gxjzh" message="Exiting ovn-controller (1) " Jan 27 15:31:28 crc kubenswrapper[4772]: E0127 15:31:27.805738 4772 kuberuntime_container.go:691] "PreStop hook failed" err="command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: " pod="openstack/ovn-controller-gxjzh" podUID="220011f2-8778-4a14-82d4-33a07bd33379" containerName="ovn-controller" containerID="cri-o://afc8ab10fea0840566de64c53bc97d22454ee25e120ead660e5999b0da009daf" Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:27.813860 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-gxjzh" podUID="220011f2-8778-4a14-82d4-33a07bd33379" containerName="ovn-controller" containerID="cri-o://afc8ab10fea0840566de64c53bc97d22454ee25e120ead660e5999b0da009daf" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:27.821619 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:27.823025 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2" containerName="openstack-network-exporter" containerID="cri-o://9ab2ac6bce7a8071ec2b4cecbc76933f6c63344bca73557900280dd89a9b1ef4" gracePeriod=300 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:27.891459 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-dr2w8"] Jan 27 15:31:28 crc kubenswrapper[4772]: E0127 15:31:27.895644 4772 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Jan 27 15:31:28 crc kubenswrapper[4772]: E0127 15:31:27.895698 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-config-data podName:508c3d5b-212a-46da-9a55-de3f35d7019b nodeName:}" failed. No retries permitted until 2026-01-27 15:31:28.395682848 +0000 UTC m=+1474.376291946 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-config-data") pod "rabbitmq-server-0" (UID: "508c3d5b-212a-46da-9a55-de3f35d7019b") : configmap "rabbitmq-config-data" not found Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:27.925668 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-dr2w8"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:27.935685 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-6af8-account-create-update-ltwnh"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:27.952356 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-6af8-account-create-update-ltwnh"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:27.971293 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-h5ch7"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:27.982509 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-h5ch7"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.002237 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-2hd4f"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.002508 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" podUID="fddd5e59-3124-4a05-aafd-92d6aea05f7e" containerName="dnsmasq-dns" containerID="cri-o://0ee5661567fa3ca13869262f2ac472811c7a59976cae6fbfe300747e324b4e86" gracePeriod=10 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.003983 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.004345 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="dc34a3a4-ad0b-4154-82c9-728227b19732" containerName="openstack-network-exporter" containerID="cri-o://fc9316436d0d6826797760ceeb255662e4eca1649200864d00e921be6f2e6eed" gracePeriod=300 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.040613 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2" containerName="ovsdbserver-nb" containerID="cri-o://abb84f069b7ba6556a04c96fbef42abc5bac570c75f402b32a5f9f20ac96046c" gracePeriod=300 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.077498 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.078016 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="account-server" containerID="cri-o://d35aa807e61d39133b8319305719556fcfa6889495c80253864eaf2dc48a450b" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.082606 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="container-updater" containerID="cri-o://99c9f47c0720632dfecbfc5e9152885ab96d751677b561767c79f0a032ca5cf5" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.082763 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="swift-recon-cron" containerID="cri-o://0b50101071feccad5793667a8f4849d22482c6d522fac228c249d69d6d557cdf" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.082808 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="rsync" containerID="cri-o://8d889567d10b3e8868d76680ff442da2a14216919aae766c356918ec9960b9a4" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.082846 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="object-expirer" containerID="cri-o://c1cf3012e8501ba3a809e028a1ab49c960d95fb090a04b4dbca6cd01d2de9524" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.082900 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="object-updater" containerID="cri-o://b0a7c137687a720a7d8c3f84cc586f4b9d3bde7c9bc9e2e0c83a325c2ae23322" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.082935 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="object-auditor" containerID="cri-o://8bbb31c1be222187b0e9b27f07c1ac0fe66d8ad583df4ff6b26fec62ab98cf87" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.082966 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="object-replicator" containerID="cri-o://71b4242b9081be055bfb8bd2db6959d32259cd0c3ee2b95ddde1c1d2154be74b" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.082996 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="object-server" containerID="cri-o://bc57f117c387fb10832190ea21f63cdb319308d9390292395fb515e28966d217" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.083044 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="account-reaper" containerID="cri-o://ac32767b3784713a66fbfe32a337398a7461aa8ffad58bbfea7ccf6e3c4ee19d" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.083076 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="container-auditor" containerID="cri-o://0c6f6ecf89a4947c23560538762ca73dfe5e13c4acb04e206d91772a3cfc9c49" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.083113 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="container-replicator" containerID="cri-o://94e4c588a745acb16ce919a52f7150cf54119c1c41e94c9e658206e6b58958ed" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.083188 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="account-auditor" containerID="cri-o://c3f602f5b8fe5f978c40989adc1d0130c6aaae0dce0fc13d5e34bbe819e8eccb" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.083226 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="account-replicator" containerID="cri-o://5f271cd2dcb6b658cde722402c5b2945c28f4d7486cab8c56e064081779416a1" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.083188 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="container-server" containerID="cri-o://494d3ebaeddb756bf375d2bc394a4b4086ee3e25d9a76747552d41c1f40a9737" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.121389 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.121661 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9a02b617-28a7-4262-a110-f1c71763ad19" containerName="glance-log" containerID="cri-o://d767e789b4befb7b8caac693075691222c00bb6ae1189417345706dad41621f9" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.122039 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9a02b617-28a7-4262-a110-f1c71763ad19" containerName="glance-httpd" containerID="cri-o://3114715e24bc63a93ce31ec7ec2cc2fdeaad0a6c7647de22f23d06ac45e3d864" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.163287 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.163647 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="683f458e-44e9-49ea-a66b-4ac91a3f2bc1" containerName="cinder-scheduler" containerID="cri-o://3e806373a2604b5465de7a3913d6865c82f0689bac61f26c430950d7d4efb948" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.164156 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="683f458e-44e9-49ea-a66b-4ac91a3f2bc1" containerName="probe" containerID="cri-o://112ddc6068b3694383f83c1ffece42788a7623920d1c02ff9f46202f7c8c0d7e" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.209200 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-dpr42"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.251759 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-dpr42"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.267595 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-597699949b-q6msx"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.267819 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-597699949b-q6msx" podUID="4205dfea-7dc7-496a-9745-fc5e3d0a418a" containerName="placement-log" containerID="cri-o://f10ed54f4ea68e56be83b8d8387a9768612b5c035b1fc42928132066af5bd689" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.268149 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-597699949b-q6msx" podUID="4205dfea-7dc7-496a-9745-fc5e3d0a418a" containerName="placement-api" containerID="cri-o://ad26ca4835a223df0b0aa3065e02d9e54b67030d2b6d0436f1f1a0dd7bf06415" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.276039 4772 generic.go:334] "Generic (PLEG): container finished" podID="220011f2-8778-4a14-82d4-33a07bd33379" containerID="afc8ab10fea0840566de64c53bc97d22454ee25e120ead660e5999b0da009daf" exitCode=0 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.276118 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gxjzh" event={"ID":"220011f2-8778-4a14-82d4-33a07bd33379","Type":"ContainerDied","Data":"afc8ab10fea0840566de64c53bc97d22454ee25e120ead660e5999b0da009daf"} Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.313325 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2/ovsdbserver-nb/0.log" Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.313381 4772 generic.go:334] "Generic (PLEG): container finished" podID="4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2" containerID="9ab2ac6bce7a8071ec2b4cecbc76933f6c63344bca73557900280dd89a9b1ef4" exitCode=2 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.313472 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2","Type":"ContainerDied","Data":"9ab2ac6bce7a8071ec2b4cecbc76933f6c63344bca73557900280dd89a9b1ef4"} Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.333837 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.334094 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="be772158-a71c-448d-8972-014f0d3a9ab8" containerName="cinder-api-log" containerID="cri-o://26cc6d1f580535edc969fb0f7d0d2e7d716fa8450f944ca1657554f90801529b" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.334384 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="be772158-a71c-448d-8972-014f0d3a9ab8" containerName="cinder-api" containerID="cri-o://c47159ab0aee5087f5a44073988d2ad8d6aaaa0e47ba7702dc2a03eab229b375" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.339290 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-vqpfg_a490a71b-c33d-4c94-9592-f97d1d315e81/openstack-network-exporter/0.log" Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.339350 4772 generic.go:334] "Generic (PLEG): container finished" podID="a490a71b-c33d-4c94-9592-f97d1d315e81" containerID="b93ad84c922746d427d3e2a2deb04a875a239fcafbecb5146ae05b1b11e36a09" exitCode=2 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.339468 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-vqpfg" event={"ID":"a490a71b-c33d-4c94-9592-f97d1d315e81","Type":"ContainerDied","Data":"b93ad84c922746d427d3e2a2deb04a875a239fcafbecb5146ae05b1b11e36a09"} Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.363917 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.364228 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e7e78641-77e6-4c89-b5c9-0d6f3c9a9343" containerName="glance-log" containerID="cri-o://3454f9899adaff309b52934e71697924735c1f269fb473444cba03b5baf4e1e5" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.364767 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e7e78641-77e6-4c89-b5c9-0d6f3c9a9343" containerName="glance-httpd" containerID="cri-o://6481b50eed7f8997cc197c4b50a1b5d1b9aa395b3745aa30ff2d6ee451d23215" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.366471 4772 generic.go:334] "Generic (PLEG): container finished" podID="8b8101bc-2ddf-48ed-9b92-e8f9e5e71938" containerID="b1542ba131aec1cffd5520f2969b843d3aa12fe7b4cd60022addce3e73977b99" exitCode=2 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.366515 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938","Type":"ContainerDied","Data":"b1542ba131aec1cffd5520f2969b843d3aa12fe7b4cd60022addce3e73977b99"} Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.373745 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-647c88bb6f-wzf82"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.374036 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-647c88bb6f-wzf82" podUID="6cf131c4-a5bd-452b-8598-42312c3a0270" containerName="neutron-api" containerID="cri-o://72824ad39b806a2254b462f1a46f766a404dd5dd1e5172059745c7930bd54b37" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.374724 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-647c88bb6f-wzf82" podUID="6cf131c4-a5bd-452b-8598-42312c3a0270" containerName="neutron-httpd" containerID="cri-o://b96f34157cbed4eef2143feeb0fd51ea8ff8193f4fa6d28ad6a9487061aba8b2" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: E0127 15:31:28.417422 4772 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Jan 27 15:31:28 crc kubenswrapper[4772]: E0127 15:31:28.417510 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-config-data podName:508c3d5b-212a-46da-9a55-de3f35d7019b nodeName:}" failed. No retries permitted until 2026-01-27 15:31:29.417476347 +0000 UTC m=+1475.398085445 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-config-data") pod "rabbitmq-server-0" (UID: "508c3d5b-212a-46da-9a55-de3f35d7019b") : configmap "rabbitmq-config-data" not found Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.429039 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="dc34a3a4-ad0b-4154-82c9-728227b19732" containerName="ovsdbserver-sb" containerID="cri-o://abdb27873ea97363386820f9e29ffa55e1d51fea483628db2a98ad8d2f8fc593" gracePeriod=300 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.448263 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.448639 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="93c8f9a4-c6ef-42b8-8543-ff8b5347977e" containerName="nova-api-log" containerID="cri-o://db0ed28f713318389578164d5ba7364ff5e0ca569d4bd32de0483eb615fe7597" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.480256 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="93c8f9a4-c6ef-42b8-8543-ff8b5347977e" containerName="nova-api-api" containerID="cri-o://abfb528c89657cd0985ff90de17dace11a1be4c50ae49dc95a4a7ec03d093246" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.501566 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-ed9a-account-create-update-b7pnl"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.515710 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-ed9a-account-create-update-b7pnl"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.530084 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-z224f"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.543287 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-z224f"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.559215 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-556764fb84-r628x"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.559479 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-556764fb84-r628x" podUID="4ce27714-673f-47de-acc3-b6902b534bdd" containerName="barbican-keystone-listener-log" containerID="cri-o://f068099f2f85afe1f1db1c1b4191de3b3198e413724471d516ae5586de30eb8f" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.561582 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-556764fb84-r628x" podUID="4ce27714-673f-47de-acc3-b6902b534bdd" containerName="barbican-keystone-listener" containerID="cri-o://aa76ea75f91196a6ccffd5d7e7d149b5efe900bfae2e86e19fa1ec88171321cc" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.584260 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-v7ncm"] Jan 27 15:31:28 crc kubenswrapper[4772]: E0127 15:31:28.622490 4772 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Jan 27 15:31:28 crc kubenswrapper[4772]: E0127 15:31:28.622548 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-config-data podName:76fdbdb1-d48a-4cd1-8372-78887671dce8 nodeName:}" failed. No retries permitted until 2026-01-27 15:31:30.622534893 +0000 UTC m=+1476.603143991 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-config-data") pod "rabbitmq-cell1-server-0" (UID: "76fdbdb1-d48a-4cd1-8372-78887671dce8") : configmap "rabbitmq-cell1-config-data" not found Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.648789 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-v7ncm"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.656904 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.658043 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f63bf600-ff03-43a3-92b4-fe8ac68a9bb7" containerName="nova-metadata-log" containerID="cri-o://7343cd6a2a5cf705b558b4cc862d749d392235682218489d0106143cb8a5d4bc" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.658194 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f63bf600-ff03-43a3-92b4-fe8ac68a9bb7" containerName="nova-metadata-metadata" containerID="cri-o://db38347574e8ea3471da74617b5c2b8fd8e23430f530dbd434f5aba2a153f9bb" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.683606 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2313c291-4eb5-4b79-ad9b-b04cd06a1ee9" path="/var/lib/kubelet/pods/2313c291-4eb5-4b79-ad9b-b04cd06a1ee9/volumes" Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.703733 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d10501c-aefc-4b6b-934a-cd53db7aa029" path="/var/lib/kubelet/pods/2d10501c-aefc-4b6b-934a-cd53db7aa029/volumes" Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.704634 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="564de425-5170-45df-9080-5b02579483ee" path="/var/lib/kubelet/pods/564de425-5170-45df-9080-5b02579483ee/volumes" Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.705197 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a423229-06be-4934-9715-58105e1af686" path="/var/lib/kubelet/pods/5a423229-06be-4934-9715-58105e1af686/volumes" Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.705711 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69f24c00-a64a-4e82-a125-c0ee3fe8fa8f" path="/var/lib/kubelet/pods/69f24c00-a64a-4e82-a125-c0ee3fe8fa8f/volumes" Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.706707 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="752279e5-88ff-469d-a4db-2942659c7e24" path="/var/lib/kubelet/pods/752279e5-88ff-469d-a4db-2942659c7e24/volumes" Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.707221 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86d0241f-ae16-400f-837c-3b43c904c91e" path="/var/lib/kubelet/pods/86d0241f-ae16-400f-837c-3b43c904c91e/volumes" Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.707912 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ae05919-68bf-43d1-abd9-9908ec287bd0" path="/var/lib/kubelet/pods/9ae05919-68bf-43d1-abd9-9908ec287bd0/volumes" Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.715639 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cbda9cc-3ec5-4193-a7fb-ff06bdd20846" path="/var/lib/kubelet/pods/9cbda9cc-3ec5-4193-a7fb-ff06bdd20846/volumes" Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.726510 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af586fb2-38ff-4e17-86bc-a7793cb3ac45" path="/var/lib/kubelet/pods/af586fb2-38ff-4e17-86bc-a7793cb3ac45/volumes" Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.738442 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0625578-3b48-44c7-9082-174fce3a7e74" path="/var/lib/kubelet/pods/b0625578-3b48-44c7-9082-174fce3a7e74/volumes" Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.739593 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b163780a-6dd7-4232-b0da-a22f18d36fcc" path="/var/lib/kubelet/pods/b163780a-6dd7-4232-b0da-a22f18d36fcc/volumes" Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.740187 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b412abae-93af-4ae0-8cd8-7c0a827da4b3" path="/var/lib/kubelet/pods/b412abae-93af-4ae0-8cd8-7c0a827da4b3/volumes" Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.741081 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c54b2036-d943-4f0d-b1c4-8a47dfab5099" path="/var/lib/kubelet/pods/c54b2036-d943-4f0d-b1c4-8a47dfab5099/volumes" Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.742282 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea2e7e0f-aef9-4687-932c-d21f24fd4bff" path="/var/lib/kubelet/pods/ea2e7e0f-aef9-4687-932c-d21f24fd4bff/volumes" Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.752254 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1746148-2e3f-476f-9a1f-f3656d44fb0b" path="/var/lib/kubelet/pods/f1746148-2e3f-476f-9a1f-f3656d44fb0b/volumes" Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.752859 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6" path="/var/lib/kubelet/pods/f2cbbc00-4796-4a0a-943d-a8d0c5cd11c6/volumes" Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.753412 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.753444 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.753456 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-659485ddbb-5bnzg"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.753686 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-659485ddbb-5bnzg" podUID="766c2a26-46ea-41b2-ba0c-2101ec9477d5" containerName="barbican-api-log" containerID="cri-o://23481794981b6875427087c492230b72f248918903ae28fa47bb73190cfa8573" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.754379 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-659485ddbb-5bnzg" podUID="766c2a26-46ea-41b2-ba0c-2101ec9477d5" containerName="barbican-api" containerID="cri-o://ffbe05081a83d720881627c45e3d405aaf574d1db8fa63481da9c229023c0874" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.774636 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-6748df9c8c-zk7zp"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.780546 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-6748df9c8c-zk7zp" podUID="710edaa6-ba83-4b1f-a49a-769ca1911c9b" containerName="barbican-worker-log" containerID="cri-o://b1ca77abcb5dfa41040a6625bbdc220ae80143a0714b3ff9a856057794a6d02a" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.780560 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-6748df9c8c-zk7zp" podUID="710edaa6-ba83-4b1f-a49a-769ca1911c9b" containerName="barbican-worker" containerID="cri-o://a063d80b4cd5f0199157f5e139c54f744514f0203001f47e4bce93805443a546" gracePeriod=30 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.806125 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-97c3-account-create-update-bvlvs"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.806387 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" podUID="fddd5e59-3124-4a05-aafd-92d6aea05f7e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.201:5353: connect: connection refused" Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.831394 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-cg94r"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.849743 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-12a3-account-create-update-mdv84"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.852772 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-cg94r"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.864206 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-12a3-account-create-update-mdv84"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.883800 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-xpbb6"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.893502 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-6w7p7"] Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.907245 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="76fdbdb1-d48a-4cd1-8372-78887671dce8" containerName="rabbitmq" containerID="cri-o://d8699d4e2fb6bcbb97c43048a20aeda8d17be226258e1acddb0364ab41c23e4a" gracePeriod=604800 Jan 27 15:31:28 crc kubenswrapper[4772]: I0127 15:31:28.960242 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-xpbb6"] Jan 27 15:31:29 crc kubenswrapper[4772]: E0127 15:31:29.010359 4772 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Jan 27 15:31:29 crc kubenswrapper[4772]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Jan 27 15:31:29 crc kubenswrapper[4772]: + source /usr/local/bin/container-scripts/functions Jan 27 15:31:29 crc kubenswrapper[4772]: ++ OVNBridge=br-int Jan 27 15:31:29 crc kubenswrapper[4772]: ++ OVNRemote=tcp:localhost:6642 Jan 27 15:31:29 crc kubenswrapper[4772]: ++ OVNEncapType=geneve Jan 27 15:31:29 crc kubenswrapper[4772]: ++ OVNAvailabilityZones= Jan 27 15:31:29 crc kubenswrapper[4772]: ++ EnableChassisAsGateway=true Jan 27 15:31:29 crc kubenswrapper[4772]: ++ PhysicalNetworks= Jan 27 15:31:29 crc kubenswrapper[4772]: ++ OVNHostName= Jan 27 15:31:29 crc kubenswrapper[4772]: ++ DB_FILE=/etc/openvswitch/conf.db Jan 27 15:31:29 crc kubenswrapper[4772]: ++ ovs_dir=/var/lib/openvswitch Jan 27 15:31:29 crc kubenswrapper[4772]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Jan 27 15:31:29 crc kubenswrapper[4772]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Jan 27 15:31:29 crc kubenswrapper[4772]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Jan 27 15:31:29 crc kubenswrapper[4772]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Jan 27 15:31:29 crc kubenswrapper[4772]: + sleep 0.5 Jan 27 15:31:29 crc kubenswrapper[4772]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Jan 27 15:31:29 crc kubenswrapper[4772]: + sleep 0.5 Jan 27 15:31:29 crc kubenswrapper[4772]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Jan 27 15:31:29 crc kubenswrapper[4772]: + sleep 0.5 Jan 27 15:31:29 crc kubenswrapper[4772]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Jan 27 15:31:29 crc kubenswrapper[4772]: + cleanup_ovsdb_server_semaphore Jan 27 15:31:29 crc kubenswrapper[4772]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Jan 27 15:31:29 crc kubenswrapper[4772]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Jan 27 15:31:29 crc kubenswrapper[4772]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-cqx7r" message=< Jan 27 15:31:29 crc kubenswrapper[4772]: Exiting ovsdb-server (5) [ OK ] Jan 27 15:31:29 crc kubenswrapper[4772]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Jan 27 15:31:29 crc kubenswrapper[4772]: + source /usr/local/bin/container-scripts/functions Jan 27 15:31:29 crc kubenswrapper[4772]: ++ OVNBridge=br-int Jan 27 15:31:29 crc kubenswrapper[4772]: ++ OVNRemote=tcp:localhost:6642 Jan 27 15:31:29 crc kubenswrapper[4772]: ++ OVNEncapType=geneve Jan 27 15:31:29 crc kubenswrapper[4772]: ++ OVNAvailabilityZones= Jan 27 15:31:29 crc kubenswrapper[4772]: ++ EnableChassisAsGateway=true Jan 27 15:31:29 crc kubenswrapper[4772]: ++ PhysicalNetworks= Jan 27 15:31:29 crc kubenswrapper[4772]: ++ OVNHostName= Jan 27 15:31:29 crc kubenswrapper[4772]: ++ DB_FILE=/etc/openvswitch/conf.db Jan 27 15:31:29 crc kubenswrapper[4772]: ++ ovs_dir=/var/lib/openvswitch Jan 27 15:31:29 crc kubenswrapper[4772]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Jan 27 15:31:29 crc kubenswrapper[4772]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Jan 27 15:31:29 crc kubenswrapper[4772]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Jan 27 15:31:29 crc kubenswrapper[4772]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Jan 27 15:31:29 crc kubenswrapper[4772]: + sleep 0.5 Jan 27 15:31:29 crc kubenswrapper[4772]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Jan 27 15:31:29 crc kubenswrapper[4772]: + sleep 0.5 Jan 27 15:31:29 crc kubenswrapper[4772]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Jan 27 15:31:29 crc kubenswrapper[4772]: + sleep 0.5 Jan 27 15:31:29 crc kubenswrapper[4772]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Jan 27 15:31:29 crc kubenswrapper[4772]: + cleanup_ovsdb_server_semaphore Jan 27 15:31:29 crc kubenswrapper[4772]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Jan 27 15:31:29 crc kubenswrapper[4772]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Jan 27 15:31:29 crc kubenswrapper[4772]: > Jan 27 15:31:29 crc kubenswrapper[4772]: E0127 15:31:29.010418 4772 kuberuntime_container.go:691] "PreStop hook failed" err=< Jan 27 15:31:29 crc kubenswrapper[4772]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Jan 27 15:31:29 crc kubenswrapper[4772]: + source /usr/local/bin/container-scripts/functions Jan 27 15:31:29 crc kubenswrapper[4772]: ++ OVNBridge=br-int Jan 27 15:31:29 crc kubenswrapper[4772]: ++ OVNRemote=tcp:localhost:6642 Jan 27 15:31:29 crc kubenswrapper[4772]: ++ OVNEncapType=geneve Jan 27 15:31:29 crc kubenswrapper[4772]: ++ OVNAvailabilityZones= Jan 27 15:31:29 crc kubenswrapper[4772]: ++ EnableChassisAsGateway=true Jan 27 15:31:29 crc kubenswrapper[4772]: ++ PhysicalNetworks= Jan 27 15:31:29 crc kubenswrapper[4772]: ++ OVNHostName= Jan 27 15:31:29 crc kubenswrapper[4772]: ++ DB_FILE=/etc/openvswitch/conf.db Jan 27 15:31:29 crc kubenswrapper[4772]: ++ ovs_dir=/var/lib/openvswitch Jan 27 15:31:29 crc kubenswrapper[4772]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Jan 27 15:31:29 crc kubenswrapper[4772]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Jan 27 15:31:29 crc kubenswrapper[4772]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Jan 27 15:31:29 crc kubenswrapper[4772]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Jan 27 15:31:29 crc kubenswrapper[4772]: + sleep 0.5 Jan 27 15:31:29 crc kubenswrapper[4772]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Jan 27 15:31:29 crc kubenswrapper[4772]: + sleep 0.5 Jan 27 15:31:29 crc kubenswrapper[4772]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Jan 27 15:31:29 crc kubenswrapper[4772]: + sleep 0.5 Jan 27 15:31:29 crc kubenswrapper[4772]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Jan 27 15:31:29 crc kubenswrapper[4772]: + cleanup_ovsdb_server_semaphore Jan 27 15:31:29 crc kubenswrapper[4772]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Jan 27 15:31:29 crc kubenswrapper[4772]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Jan 27 15:31:29 crc kubenswrapper[4772]: > pod="openstack/ovn-controller-ovs-cqx7r" podUID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerName="ovsdb-server" containerID="cri-o://4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.010477 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-cqx7r" podUID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerName="ovsdb-server" containerID="cri-o://4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b" gracePeriod=29 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.016516 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-6w7p7"] Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.038858 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-gbrww"] Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.050241 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-pszgr"] Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.070417 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-gbrww"] Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.087879 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-pszgr"] Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.133364 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="cf619242-7348-4de4-a37e-8ebdc4ca54d7" containerName="galera" containerID="cri-o://2e743dfaa62b788cb68a4d553d64cf9affaf8ef6e4da1308fddf4dc259167b69" gracePeriod=30 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.135039 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.135241 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="5e69643a-e8c2-4057-a993-d5506ceeec1b" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://7a1429ee13edd2169e8a683ea45dcb648c58812d36d48307ba37a8f39d0a67ce" gracePeriod=30 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.146652 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.155675 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gjwh2"] Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.182990 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.183239 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="dbbd3c83-3fde-4b11-8ef0-add837d393ce" containerName="nova-cell1-conductor-conductor" containerID="cri-o://788384a3ae6b89b3eeabbb3fe7578f4cb514172f7c7e0c341ec2b75ed4d75a29" gracePeriod=30 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.207726 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gjwh2"] Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.233394 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.233635 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="b20b9215-5398-4100-bac4-763daa5ed222" containerName="nova-cell0-conductor-conductor" containerID="cri-o://2e74d40bce110215c2607c9a9b716bf9d9db61e446fe99bb897518adbdc86d83" gracePeriod=30 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.240606 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-v9mqp"] Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.258229 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-v9mqp"] Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.259703 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-cqx7r" podUID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerName="ovs-vswitchd" containerID="cri-o://d6579efc0c6a14eb40a1349e6b5e9e288881435286dc3fff811374b436b48c91" gracePeriod=28 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.268263 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-qmppl"] Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.279830 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.280702 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.280852 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="b83f7578-8113-46c8-be24-5968aa0ca563" containerName="nova-scheduler-scheduler" containerID="cri-o://e92037085b98ccc46bbd64416f98018d2426a17d0883dd17b830d5574a8a0f4e" gracePeriod=30 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.293599 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="508c3d5b-212a-46da-9a55-de3f35d7019b" containerName="rabbitmq" containerID="cri-o://f002759dea4443f7600e0f76f24481c1604449a5ee31bd8aa53171a2121ec4b2" gracePeriod=604800 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.321282 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gxjzh" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.343544 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-dns-svc\") pod \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\" (UID: \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.343822 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fqp8\" (UniqueName: \"kubernetes.io/projected/fddd5e59-3124-4a05-aafd-92d6aea05f7e-kube-api-access-6fqp8\") pod \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\" (UID: \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.343960 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-dns-swift-storage-0\") pod \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\" (UID: \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.344093 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-ovsdbserver-sb\") pod \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\" (UID: \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.344293 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-ovsdbserver-nb\") pod \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\" (UID: \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.344694 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-config\") pod \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\" (UID: \"fddd5e59-3124-4a05-aafd-92d6aea05f7e\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.352988 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fddd5e59-3124-4a05-aafd-92d6aea05f7e-kube-api-access-6fqp8" (OuterVolumeSpecName: "kube-api-access-6fqp8") pod "fddd5e59-3124-4a05-aafd-92d6aea05f7e" (UID: "fddd5e59-3124-4a05-aafd-92d6aea05f7e"). InnerVolumeSpecName "kube-api-access-6fqp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.353483 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2/ovsdbserver-nb/0.log" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.353738 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.365413 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.366800 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-vqpfg_a490a71b-c33d-4c94-9592-f97d1d315e81/openstack-network-exporter/0.log" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.366846 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-vqpfg" Jan 27 15:31:29 crc kubenswrapper[4772]: E0127 15:31:29.381457 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="788384a3ae6b89b3eeabbb3fe7578f4cb514172f7c7e0c341ec2b75ed4d75a29" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Jan 27 15:31:29 crc kubenswrapper[4772]: E0127 15:31:29.395513 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="788384a3ae6b89b3eeabbb3fe7578f4cb514172f7c7e0c341ec2b75ed4d75a29" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Jan 27 15:31:29 crc kubenswrapper[4772]: E0127 15:31:29.419265 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="788384a3ae6b89b3eeabbb3fe7578f4cb514172f7c7e0c341ec2b75ed4d75a29" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Jan 27 15:31:29 crc kubenswrapper[4772]: E0127 15:31:29.419329 4772 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="dbbd3c83-3fde-4b11-8ef0-add837d393ce" containerName="nova-cell1-conductor-conductor" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.432330 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_dc34a3a4-ad0b-4154-82c9-728227b19732/ovsdbserver-sb/0.log" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.432421 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.440534 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fddd5e59-3124-4a05-aafd-92d6aea05f7e" (UID: "fddd5e59-3124-4a05-aafd-92d6aea05f7e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.444693 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fddd5e59-3124-4a05-aafd-92d6aea05f7e" (UID: "fddd5e59-3124-4a05-aafd-92d6aea05f7e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.448821 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-ovsdb-rundir\") pod \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.448877 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a490a71b-c33d-4c94-9592-f97d1d315e81-ovs-rundir\") pod \"a490a71b-c33d-4c94-9592-f97d1d315e81\" (UID: \"a490a71b-c33d-4c94-9592-f97d1d315e81\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.448909 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0edf6707-14dd-4986-8d64-0e48a31d6a39-combined-ca-bundle\") pod \"0edf6707-14dd-4986-8d64-0e48a31d6a39\" (UID: \"0edf6707-14dd-4986-8d64-0e48a31d6a39\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.448943 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdkbw\" (UniqueName: \"kubernetes.io/projected/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-kube-api-access-vdkbw\") pod \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.449008 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-metrics-certs-tls-certs\") pod \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.449042 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-scripts\") pod \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.449069 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0edf6707-14dd-4986-8d64-0e48a31d6a39-openstack-config\") pod \"0edf6707-14dd-4986-8d64-0e48a31d6a39\" (UID: \"0edf6707-14dd-4986-8d64-0e48a31d6a39\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.449099 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/220011f2-8778-4a14-82d4-33a07bd33379-var-run-ovn\") pod \"220011f2-8778-4a14-82d4-33a07bd33379\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.449204 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/220011f2-8778-4a14-82d4-33a07bd33379-scripts\") pod \"220011f2-8778-4a14-82d4-33a07bd33379\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.449231 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0edf6707-14dd-4986-8d64-0e48a31d6a39-openstack-config-secret\") pod \"0edf6707-14dd-4986-8d64-0e48a31d6a39\" (UID: \"0edf6707-14dd-4986-8d64-0e48a31d6a39\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.449257 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmhcm\" (UniqueName: \"kubernetes.io/projected/a490a71b-c33d-4c94-9592-f97d1d315e81-kube-api-access-dmhcm\") pod \"a490a71b-c33d-4c94-9592-f97d1d315e81\" (UID: \"a490a71b-c33d-4c94-9592-f97d1d315e81\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.449290 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkhnj\" (UniqueName: \"kubernetes.io/projected/0edf6707-14dd-4986-8d64-0e48a31d6a39-kube-api-access-xkhnj\") pod \"0edf6707-14dd-4986-8d64-0e48a31d6a39\" (UID: \"0edf6707-14dd-4986-8d64-0e48a31d6a39\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.449325 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/220011f2-8778-4a14-82d4-33a07bd33379-ovn-controller-tls-certs\") pod \"220011f2-8778-4a14-82d4-33a07bd33379\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.449353 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a490a71b-c33d-4c94-9592-f97d1d315e81-metrics-certs-tls-certs\") pod \"a490a71b-c33d-4c94-9592-f97d1d315e81\" (UID: \"a490a71b-c33d-4c94-9592-f97d1d315e81\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.449416 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sclc\" (UniqueName: \"kubernetes.io/projected/220011f2-8778-4a14-82d4-33a07bd33379-kube-api-access-5sclc\") pod \"220011f2-8778-4a14-82d4-33a07bd33379\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.449444 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/220011f2-8778-4a14-82d4-33a07bd33379-var-log-ovn\") pod \"220011f2-8778-4a14-82d4-33a07bd33379\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.449483 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-ovsdbserver-nb-tls-certs\") pod \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.449513 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a490a71b-c33d-4c94-9592-f97d1d315e81-combined-ca-bundle\") pod \"a490a71b-c33d-4c94-9592-f97d1d315e81\" (UID: \"a490a71b-c33d-4c94-9592-f97d1d315e81\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.449536 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a490a71b-c33d-4c94-9592-f97d1d315e81-ovn-rundir\") pod \"a490a71b-c33d-4c94-9592-f97d1d315e81\" (UID: \"a490a71b-c33d-4c94-9592-f97d1d315e81\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.449576 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/220011f2-8778-4a14-82d4-33a07bd33379-var-run\") pod \"220011f2-8778-4a14-82d4-33a07bd33379\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.449632 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/220011f2-8778-4a14-82d4-33a07bd33379-combined-ca-bundle\") pod \"220011f2-8778-4a14-82d4-33a07bd33379\" (UID: \"220011f2-8778-4a14-82d4-33a07bd33379\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.449658 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-combined-ca-bundle\") pod \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.449681 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a490a71b-c33d-4c94-9592-f97d1d315e81-config\") pod \"a490a71b-c33d-4c94-9592-f97d1d315e81\" (UID: \"a490a71b-c33d-4c94-9592-f97d1d315e81\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.449731 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.449779 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-config\") pod \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\" (UID: \"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.450975 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fqp8\" (UniqueName: \"kubernetes.io/projected/fddd5e59-3124-4a05-aafd-92d6aea05f7e-kube-api-access-6fqp8\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.451022 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.451032 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: E0127 15:31:29.451098 4772 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Jan 27 15:31:29 crc kubenswrapper[4772]: E0127 15:31:29.451148 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-config-data podName:508c3d5b-212a-46da-9a55-de3f35d7019b nodeName:}" failed. No retries permitted until 2026-01-27 15:31:31.451130728 +0000 UTC m=+1477.431739826 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-config-data") pod "rabbitmq-server-0" (UID: "508c3d5b-212a-46da-9a55-de3f35d7019b") : configmap "rabbitmq-config-data" not found Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.451346 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-config" (OuterVolumeSpecName: "config") pod "4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2" (UID: "4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.452912 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a490a71b-c33d-4c94-9592-f97d1d315e81-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "a490a71b-c33d-4c94-9592-f97d1d315e81" (UID: "a490a71b-c33d-4c94-9592-f97d1d315e81"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.453417 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/220011f2-8778-4a14-82d4-33a07bd33379-var-run" (OuterVolumeSpecName: "var-run") pod "220011f2-8778-4a14-82d4-33a07bd33379" (UID: "220011f2-8778-4a14-82d4-33a07bd33379"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.454374 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fddd5e59-3124-4a05-aafd-92d6aea05f7e" (UID: "fddd5e59-3124-4a05-aafd-92d6aea05f7e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.459631 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a490a71b-c33d-4c94-9592-f97d1d315e81-config" (OuterVolumeSpecName: "config") pod "a490a71b-c33d-4c94-9592-f97d1d315e81" (UID: "a490a71b-c33d-4c94-9592-f97d1d315e81"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.463813 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/220011f2-8778-4a14-82d4-33a07bd33379-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "220011f2-8778-4a14-82d4-33a07bd33379" (UID: "220011f2-8778-4a14-82d4-33a07bd33379"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.464092 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a490a71b-c33d-4c94-9592-f97d1d315e81-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "a490a71b-c33d-4c94-9592-f97d1d315e81" (UID: "a490a71b-c33d-4c94-9592-f97d1d315e81"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.464335 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-scripts" (OuterVolumeSpecName: "scripts") pod "4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2" (UID: "4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.464675 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2" (UID: "4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.466829 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/220011f2-8778-4a14-82d4-33a07bd33379-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "220011f2-8778-4a14-82d4-33a07bd33379" (UID: "220011f2-8778-4a14-82d4-33a07bd33379"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.467116 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/220011f2-8778-4a14-82d4-33a07bd33379-scripts" (OuterVolumeSpecName: "scripts") pod "220011f2-8778-4a14-82d4-33a07bd33379" (UID: "220011f2-8778-4a14-82d4-33a07bd33379"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.472588 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2" (UID: "4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.475632 4772 generic.go:334] "Generic (PLEG): container finished" podID="f63bf600-ff03-43a3-92b4-fe8ac68a9bb7" containerID="7343cd6a2a5cf705b558b4cc862d749d392235682218489d0106143cb8a5d4bc" exitCode=143 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.476253 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/220011f2-8778-4a14-82d4-33a07bd33379-kube-api-access-5sclc" (OuterVolumeSpecName: "kube-api-access-5sclc") pod "220011f2-8778-4a14-82d4-33a07bd33379" (UID: "220011f2-8778-4a14-82d4-33a07bd33379"). InnerVolumeSpecName "kube-api-access-5sclc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.475847 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7","Type":"ContainerDied","Data":"7343cd6a2a5cf705b558b4cc862d749d392235682218489d0106143cb8a5d4bc"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.481416 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a490a71b-c33d-4c94-9592-f97d1d315e81-kube-api-access-dmhcm" (OuterVolumeSpecName: "kube-api-access-dmhcm") pod "a490a71b-c33d-4c94-9592-f97d1d315e81" (UID: "a490a71b-c33d-4c94-9592-f97d1d315e81"). InnerVolumeSpecName "kube-api-access-dmhcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.483539 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-kube-api-access-vdkbw" (OuterVolumeSpecName: "kube-api-access-vdkbw") pod "4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2" (UID: "4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2"). InnerVolumeSpecName "kube-api-access-vdkbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.485398 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_dc34a3a4-ad0b-4154-82c9-728227b19732/ovsdbserver-sb/0.log" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.485542 4772 generic.go:334] "Generic (PLEG): container finished" podID="dc34a3a4-ad0b-4154-82c9-728227b19732" containerID="fc9316436d0d6826797760ceeb255662e4eca1649200864d00e921be6f2e6eed" exitCode=2 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.485662 4772 generic.go:334] "Generic (PLEG): container finished" podID="dc34a3a4-ad0b-4154-82c9-728227b19732" containerID="abdb27873ea97363386820f9e29ffa55e1d51fea483628db2a98ad8d2f8fc593" exitCode=143 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.485801 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"dc34a3a4-ad0b-4154-82c9-728227b19732","Type":"ContainerDied","Data":"fc9316436d0d6826797760ceeb255662e4eca1649200864d00e921be6f2e6eed"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.485914 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"dc34a3a4-ad0b-4154-82c9-728227b19732","Type":"ContainerDied","Data":"abdb27873ea97363386820f9e29ffa55e1d51fea483628db2a98ad8d2f8fc593"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.486012 4772 scope.go:117] "RemoveContainer" containerID="fc9316436d0d6826797760ceeb255662e4eca1649200864d00e921be6f2e6eed" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.486317 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.492985 4772 generic.go:334] "Generic (PLEG): container finished" podID="683f458e-44e9-49ea-a66b-4ac91a3f2bc1" containerID="112ddc6068b3694383f83c1ffece42788a7623920d1c02ff9f46202f7c8c0d7e" exitCode=0 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.493062 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"683f458e-44e9-49ea-a66b-4ac91a3f2bc1","Type":"ContainerDied","Data":"112ddc6068b3694383f83c1ffece42788a7623920d1c02ff9f46202f7c8c0d7e"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.495400 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/220011f2-8778-4a14-82d4-33a07bd33379-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "220011f2-8778-4a14-82d4-33a07bd33379" (UID: "220011f2-8778-4a14-82d4-33a07bd33379"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.499169 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0edf6707-14dd-4986-8d64-0e48a31d6a39-kube-api-access-xkhnj" (OuterVolumeSpecName: "kube-api-access-xkhnj") pod "0edf6707-14dd-4986-8d64-0e48a31d6a39" (UID: "0edf6707-14dd-4986-8d64-0e48a31d6a39"). InnerVolumeSpecName "kube-api-access-xkhnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.501784 4772 generic.go:334] "Generic (PLEG): container finished" podID="710edaa6-ba83-4b1f-a49a-769ca1911c9b" containerID="b1ca77abcb5dfa41040a6625bbdc220ae80143a0714b3ff9a856057794a6d02a" exitCode=143 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.501872 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6748df9c8c-zk7zp" event={"ID":"710edaa6-ba83-4b1f-a49a-769ca1911c9b","Type":"ContainerDied","Data":"b1ca77abcb5dfa41040a6625bbdc220ae80143a0714b3ff9a856057794a6d02a"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.511019 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-qmppl" event={"ID":"4cbf7469-816d-4e54-a7ad-b5b76d0d59d6","Type":"ContainerStarted","Data":"3cdc8204c9c28616053d96ae2843e1dddf8646f9b546be00bd78d90869086025"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.511074 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-qmppl" event={"ID":"4cbf7469-816d-4e54-a7ad-b5b76d0d59d6","Type":"ContainerStarted","Data":"d5deb0f3cfb55cc15d206b2ff6d6a2e4b5ccc8b9efe8608e2073fe3df0f8d559"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.522925 4772 generic.go:334] "Generic (PLEG): container finished" podID="fddd5e59-3124-4a05-aafd-92d6aea05f7e" containerID="0ee5661567fa3ca13869262f2ac472811c7a59976cae6fbfe300747e324b4e86" exitCode=0 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.522985 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" event={"ID":"fddd5e59-3124-4a05-aafd-92d6aea05f7e","Type":"ContainerDied","Data":"0ee5661567fa3ca13869262f2ac472811c7a59976cae6fbfe300747e324b4e86"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.523009 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" event={"ID":"fddd5e59-3124-4a05-aafd-92d6aea05f7e","Type":"ContainerDied","Data":"c5c94af58b0cd6c043cac9ed46da0616cb74fd66aa5279858fb42cf515ba3aa1"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.523066 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-2hd4f" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.541204 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0edf6707-14dd-4986-8d64-0e48a31d6a39-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "0edf6707-14dd-4986-8d64-0e48a31d6a39" (UID: "0edf6707-14dd-4986-8d64-0e48a31d6a39"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.548129 4772 generic.go:334] "Generic (PLEG): container finished" podID="4205dfea-7dc7-496a-9745-fc5e3d0a418a" containerID="f10ed54f4ea68e56be83b8d8387a9768612b5c035b1fc42928132066af5bd689" exitCode=143 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.548229 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-597699949b-q6msx" event={"ID":"4205dfea-7dc7-496a-9745-fc5e3d0a418a","Type":"ContainerDied","Data":"f10ed54f4ea68e56be83b8d8387a9768612b5c035b1fc42928132066af5bd689"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.556294 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc34a3a4-ad0b-4154-82c9-728227b19732-scripts\") pod \"dc34a3a4-ad0b-4154-82c9-728227b19732\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.556367 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"dc34a3a4-ad0b-4154-82c9-728227b19732\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.556451 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc34a3a4-ad0b-4154-82c9-728227b19732-metrics-certs-tls-certs\") pod \"dc34a3a4-ad0b-4154-82c9-728227b19732\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.556542 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc34a3a4-ad0b-4154-82c9-728227b19732-ovsdbserver-sb-tls-certs\") pod \"dc34a3a4-ad0b-4154-82c9-728227b19732\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.556587 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc34a3a4-ad0b-4154-82c9-728227b19732-combined-ca-bundle\") pod \"dc34a3a4-ad0b-4154-82c9-728227b19732\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.556655 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8wt7\" (UniqueName: \"kubernetes.io/projected/dc34a3a4-ad0b-4154-82c9-728227b19732-kube-api-access-g8wt7\") pod \"dc34a3a4-ad0b-4154-82c9-728227b19732\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.556720 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dc34a3a4-ad0b-4154-82c9-728227b19732-ovsdb-rundir\") pod \"dc34a3a4-ad0b-4154-82c9-728227b19732\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.556738 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc34a3a4-ad0b-4154-82c9-728227b19732-config\") pod \"dc34a3a4-ad0b-4154-82c9-728227b19732\" (UID: \"dc34a3a4-ad0b-4154-82c9-728227b19732\") " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.557126 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sclc\" (UniqueName: \"kubernetes.io/projected/220011f2-8778-4a14-82d4-33a07bd33379-kube-api-access-5sclc\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.557136 4772 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/220011f2-8778-4a14-82d4-33a07bd33379-var-log-ovn\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.557146 4772 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a490a71b-c33d-4c94-9592-f97d1d315e81-ovn-rundir\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.557155 4772 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/220011f2-8778-4a14-82d4-33a07bd33379-var-run\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.557166 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/220011f2-8778-4a14-82d4-33a07bd33379-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.557188 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a490a71b-c33d-4c94-9592-f97d1d315e81-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.557206 4772 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.557215 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.557225 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.557234 4772 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a490a71b-c33d-4c94-9592-f97d1d315e81-ovs-rundir\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.557242 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdkbw\" (UniqueName: \"kubernetes.io/projected/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-kube-api-access-vdkbw\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.557250 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.557260 4772 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0edf6707-14dd-4986-8d64-0e48a31d6a39-openstack-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.557270 4772 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/220011f2-8778-4a14-82d4-33a07bd33379-var-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.557278 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.557286 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/220011f2-8778-4a14-82d4-33a07bd33379-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.557294 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmhcm\" (UniqueName: \"kubernetes.io/projected/a490a71b-c33d-4c94-9592-f97d1d315e81-kube-api-access-dmhcm\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.557303 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkhnj\" (UniqueName: \"kubernetes.io/projected/0edf6707-14dd-4986-8d64-0e48a31d6a39-kube-api-access-xkhnj\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.565350 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc34a3a4-ad0b-4154-82c9-728227b19732-config" (OuterVolumeSpecName: "config") pod "dc34a3a4-ad0b-4154-82c9-728227b19732" (UID: "dc34a3a4-ad0b-4154-82c9-728227b19732"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.566359 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc34a3a4-ad0b-4154-82c9-728227b19732-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "dc34a3a4-ad0b-4154-82c9-728227b19732" (UID: "dc34a3a4-ad0b-4154-82c9-728227b19732"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.567042 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc34a3a4-ad0b-4154-82c9-728227b19732-scripts" (OuterVolumeSpecName: "scripts") pod "dc34a3a4-ad0b-4154-82c9-728227b19732" (UID: "dc34a3a4-ad0b-4154-82c9-728227b19732"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.568819 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "dc34a3a4-ad0b-4154-82c9-728227b19732" (UID: "dc34a3a4-ad0b-4154-82c9-728227b19732"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.574387 4772 generic.go:334] "Generic (PLEG): container finished" podID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerID="4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b" exitCode=0 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.574464 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-cqx7r" event={"ID":"38ebd422-35c5-4682-8a4d-ca9073728d7c","Type":"ContainerDied","Data":"4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.576824 4772 generic.go:334] "Generic (PLEG): container finished" podID="4ce27714-673f-47de-acc3-b6902b534bdd" containerID="f068099f2f85afe1f1db1c1b4191de3b3198e413724471d516ae5586de30eb8f" exitCode=143 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.576942 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-556764fb84-r628x" event={"ID":"4ce27714-673f-47de-acc3-b6902b534bdd","Type":"ContainerDied","Data":"f068099f2f85afe1f1db1c1b4191de3b3198e413724471d516ae5586de30eb8f"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.598532 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc34a3a4-ad0b-4154-82c9-728227b19732-kube-api-access-g8wt7" (OuterVolumeSpecName: "kube-api-access-g8wt7") pod "dc34a3a4-ad0b-4154-82c9-728227b19732" (UID: "dc34a3a4-ad0b-4154-82c9-728227b19732"). InnerVolumeSpecName "kube-api-access-g8wt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.604503 4772 generic.go:334] "Generic (PLEG): container finished" podID="766c2a26-46ea-41b2-ba0c-2101ec9477d5" containerID="23481794981b6875427087c492230b72f248918903ae28fa47bb73190cfa8573" exitCode=143 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.604629 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-659485ddbb-5bnzg" event={"ID":"766c2a26-46ea-41b2-ba0c-2101ec9477d5","Type":"ContainerDied","Data":"23481794981b6875427087c492230b72f248918903ae28fa47bb73190cfa8573"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.630599 4772 generic.go:334] "Generic (PLEG): container finished" podID="be772158-a71c-448d-8972-014f0d3a9ab8" containerID="26cc6d1f580535edc969fb0f7d0d2e7d716fa8450f944ca1657554f90801529b" exitCode=143 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.630678 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"be772158-a71c-448d-8972-014f0d3a9ab8","Type":"ContainerDied","Data":"26cc6d1f580535edc969fb0f7d0d2e7d716fa8450f944ca1657554f90801529b"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.631833 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a490a71b-c33d-4c94-9592-f97d1d315e81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a490a71b-c33d-4c94-9592-f97d1d315e81" (UID: "a490a71b-c33d-4c94-9592-f97d1d315e81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.660297 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a490a71b-c33d-4c94-9592-f97d1d315e81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.660662 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8wt7\" (UniqueName: \"kubernetes.io/projected/dc34a3a4-ad0b-4154-82c9-728227b19732-kube-api-access-g8wt7\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.660679 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dc34a3a4-ad0b-4154-82c9-728227b19732-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.660696 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc34a3a4-ad0b-4154-82c9-728227b19732-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.660707 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc34a3a4-ad0b-4154-82c9-728227b19732-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.660733 4772 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.668702 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2/ovsdbserver-nb/0.log" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.668753 4772 generic.go:334] "Generic (PLEG): container finished" podID="4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2" containerID="abb84f069b7ba6556a04c96fbef42abc5bac570c75f402b32a5f9f20ac96046c" exitCode=143 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.668826 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2","Type":"ContainerDied","Data":"abb84f069b7ba6556a04c96fbef42abc5bac570c75f402b32a5f9f20ac96046c"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.668853 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2","Type":"ContainerDied","Data":"2d9f9f123f138892540800ef23f48dae96e200e8a0b42b345d3f87addf089f7e"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.668920 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.691588 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc34a3a4-ad0b-4154-82c9-728227b19732-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dc34a3a4-ad0b-4154-82c9-728227b19732" (UID: "dc34a3a4-ad0b-4154-82c9-728227b19732"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.729347 4772 generic.go:334] "Generic (PLEG): container finished" podID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerID="8d889567d10b3e8868d76680ff442da2a14216919aae766c356918ec9960b9a4" exitCode=0 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.729533 4772 generic.go:334] "Generic (PLEG): container finished" podID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerID="c1cf3012e8501ba3a809e028a1ab49c960d95fb090a04b4dbca6cd01d2de9524" exitCode=0 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.729586 4772 generic.go:334] "Generic (PLEG): container finished" podID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerID="b0a7c137687a720a7d8c3f84cc586f4b9d3bde7c9bc9e2e0c83a325c2ae23322" exitCode=0 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.729633 4772 generic.go:334] "Generic (PLEG): container finished" podID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerID="8bbb31c1be222187b0e9b27f07c1ac0fe66d8ad583df4ff6b26fec62ab98cf87" exitCode=0 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.729678 4772 generic.go:334] "Generic (PLEG): container finished" podID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerID="71b4242b9081be055bfb8bd2db6959d32259cd0c3ee2b95ddde1c1d2154be74b" exitCode=0 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.729757 4772 generic.go:334] "Generic (PLEG): container finished" podID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerID="bc57f117c387fb10832190ea21f63cdb319308d9390292395fb515e28966d217" exitCode=0 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.729807 4772 generic.go:334] "Generic (PLEG): container finished" podID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerID="99c9f47c0720632dfecbfc5e9152885ab96d751677b561767c79f0a032ca5cf5" exitCode=0 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.729859 4772 generic.go:334] "Generic (PLEG): container finished" podID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerID="0c6f6ecf89a4947c23560538762ca73dfe5e13c4acb04e206d91772a3cfc9c49" exitCode=0 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.729940 4772 generic.go:334] "Generic (PLEG): container finished" podID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerID="94e4c588a745acb16ce919a52f7150cf54119c1c41e94c9e658206e6b58958ed" exitCode=0 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.729991 4772 generic.go:334] "Generic (PLEG): container finished" podID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerID="494d3ebaeddb756bf375d2bc394a4b4086ee3e25d9a76747552d41c1f40a9737" exitCode=0 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.730039 4772 generic.go:334] "Generic (PLEG): container finished" podID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerID="ac32767b3784713a66fbfe32a337398a7461aa8ffad58bbfea7ccf6e3c4ee19d" exitCode=0 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.730093 4772 generic.go:334] "Generic (PLEG): container finished" podID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerID="c3f602f5b8fe5f978c40989adc1d0130c6aaae0dce0fc13d5e34bbe819e8eccb" exitCode=0 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.730140 4772 generic.go:334] "Generic (PLEG): container finished" podID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerID="5f271cd2dcb6b658cde722402c5b2945c28f4d7486cab8c56e064081779416a1" exitCode=0 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.730208 4772 generic.go:334] "Generic (PLEG): container finished" podID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerID="d35aa807e61d39133b8319305719556fcfa6889495c80253864eaf2dc48a450b" exitCode=0 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.730300 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerDied","Data":"8d889567d10b3e8868d76680ff442da2a14216919aae766c356918ec9960b9a4"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.730372 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerDied","Data":"c1cf3012e8501ba3a809e028a1ab49c960d95fb090a04b4dbca6cd01d2de9524"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.730436 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerDied","Data":"b0a7c137687a720a7d8c3f84cc586f4b9d3bde7c9bc9e2e0c83a325c2ae23322"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.730490 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerDied","Data":"8bbb31c1be222187b0e9b27f07c1ac0fe66d8ad583df4ff6b26fec62ab98cf87"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.730540 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerDied","Data":"71b4242b9081be055bfb8bd2db6959d32259cd0c3ee2b95ddde1c1d2154be74b"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.730590 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerDied","Data":"bc57f117c387fb10832190ea21f63cdb319308d9390292395fb515e28966d217"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.730641 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerDied","Data":"99c9f47c0720632dfecbfc5e9152885ab96d751677b561767c79f0a032ca5cf5"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.730693 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerDied","Data":"0c6f6ecf89a4947c23560538762ca73dfe5e13c4acb04e206d91772a3cfc9c49"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.730755 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerDied","Data":"94e4c588a745acb16ce919a52f7150cf54119c1c41e94c9e658206e6b58958ed"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.730807 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerDied","Data":"494d3ebaeddb756bf375d2bc394a4b4086ee3e25d9a76747552d41c1f40a9737"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.730857 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerDied","Data":"ac32767b3784713a66fbfe32a337398a7461aa8ffad58bbfea7ccf6e3c4ee19d"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.730907 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerDied","Data":"c3f602f5b8fe5f978c40989adc1d0130c6aaae0dce0fc13d5e34bbe819e8eccb"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.730956 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerDied","Data":"5f271cd2dcb6b658cde722402c5b2945c28f4d7486cab8c56e064081779416a1"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.731010 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerDied","Data":"d35aa807e61d39133b8319305719556fcfa6889495c80253864eaf2dc48a450b"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.745647 4772 generic.go:334] "Generic (PLEG): container finished" podID="93c8f9a4-c6ef-42b8-8543-ff8b5347977e" containerID="db0ed28f713318389578164d5ba7364ff5e0ca569d4bd32de0483eb615fe7597" exitCode=143 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.745742 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"93c8f9a4-c6ef-42b8-8543-ff8b5347977e","Type":"ContainerDied","Data":"db0ed28f713318389578164d5ba7364ff5e0ca569d4bd32de0483eb615fe7597"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.755750 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2" (UID: "4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.760153 4772 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.766631 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gxjzh" event={"ID":"220011f2-8778-4a14-82d4-33a07bd33379","Type":"ContainerDied","Data":"3cb1a1a1b7113cd35f8e36164d72f2c95422afc48122fb52f34747897808a62b"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.766672 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gxjzh" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.779143 4772 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.780339 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc34a3a4-ad0b-4154-82c9-728227b19732-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.780453 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.783325 4772 generic.go:334] "Generic (PLEG): container finished" podID="6cf131c4-a5bd-452b-8598-42312c3a0270" containerID="b96f34157cbed4eef2143feeb0fd51ea8ff8193f4fa6d28ad6a9487061aba8b2" exitCode=0 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.783472 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-647c88bb6f-wzf82" event={"ID":"6cf131c4-a5bd-452b-8598-42312c3a0270","Type":"ContainerDied","Data":"b96f34157cbed4eef2143feeb0fd51ea8ff8193f4fa6d28ad6a9487061aba8b2"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.788939 4772 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.792878 4772 generic.go:334] "Generic (PLEG): container finished" podID="9a02b617-28a7-4262-a110-f1c71763ad19" containerID="d767e789b4befb7b8caac693075691222c00bb6ae1189417345706dad41621f9" exitCode=143 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.792968 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9a02b617-28a7-4262-a110-f1c71763ad19","Type":"ContainerDied","Data":"d767e789b4befb7b8caac693075691222c00bb6ae1189417345706dad41621f9"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.798439 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "fddd5e59-3124-4a05-aafd-92d6aea05f7e" (UID: "fddd5e59-3124-4a05-aafd-92d6aea05f7e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.801663 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-vqpfg_a490a71b-c33d-4c94-9592-f97d1d315e81/openstack-network-exporter/0.log" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.801922 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-vqpfg" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.802230 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-vqpfg" event={"ID":"a490a71b-c33d-4c94-9592-f97d1d315e81","Type":"ContainerDied","Data":"b17e4736d7a1350c0c68f20fb3327f8519a43cb3b16a163a3b8e79d710328aca"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.827108 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0edf6707-14dd-4986-8d64-0e48a31d6a39-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0edf6707-14dd-4986-8d64-0e48a31d6a39" (UID: "0edf6707-14dd-4986-8d64-0e48a31d6a39"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.837237 4772 generic.go:334] "Generic (PLEG): container finished" podID="0edf6707-14dd-4986-8d64-0e48a31d6a39" containerID="0c37dc673e475cc4ca1e8b831b0543b26650ceedc799dace964e07fb4c7c7ffa" exitCode=137 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.837370 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.851928 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-qmppl" podStartSLOduration=3.851904759 podStartE2EDuration="3.851904759s" podCreationTimestamp="2026-01-27 15:31:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:31:29.531512051 +0000 UTC m=+1475.512121149" watchObservedRunningTime="2026-01-27 15:31:29.851904759 +0000 UTC m=+1475.832513867" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.870643 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-97c3-account-create-update-bvlvs"] Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.882379 4772 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.882450 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0edf6707-14dd-4986-8d64-0e48a31d6a39-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.882463 4772 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.887763 4772 generic.go:334] "Generic (PLEG): container finished" podID="e7e78641-77e6-4c89-b5c9-0d6f3c9a9343" containerID="3454f9899adaff309b52934e71697924735c1f269fb473444cba03b5baf4e1e5" exitCode=143 Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.887816 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343","Type":"ContainerDied","Data":"3454f9899adaff309b52934e71697924735c1f269fb473444cba03b5baf4e1e5"} Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.894025 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0edf6707-14dd-4986-8d64-0e48a31d6a39-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "0edf6707-14dd-4986-8d64-0e48a31d6a39" (UID: "0edf6707-14dd-4986-8d64-0e48a31d6a39"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.894694 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/220011f2-8778-4a14-82d4-33a07bd33379-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "220011f2-8778-4a14-82d4-33a07bd33379" (UID: "220011f2-8778-4a14-82d4-33a07bd33379"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.984761 4772 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0edf6707-14dd-4986-8d64-0e48a31d6a39-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:29 crc kubenswrapper[4772]: I0127 15:31:29.984797 4772 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/220011f2-8778-4a14-82d4-33a07bd33379-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.009094 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2" (UID: "4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.025402 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2" (UID: "4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.031284 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-config" (OuterVolumeSpecName: "config") pod "fddd5e59-3124-4a05-aafd-92d6aea05f7e" (UID: "fddd5e59-3124-4a05-aafd-92d6aea05f7e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.035357 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc34a3a4-ad0b-4154-82c9-728227b19732-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "dc34a3a4-ad0b-4154-82c9-728227b19732" (UID: "dc34a3a4-ad0b-4154-82c9-728227b19732"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.043247 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc34a3a4-ad0b-4154-82c9-728227b19732-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "dc34a3a4-ad0b-4154-82c9-728227b19732" (UID: "dc34a3a4-ad0b-4154-82c9-728227b19732"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:30 crc kubenswrapper[4772]: E0127 15:31:30.068388 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e92037085b98ccc46bbd64416f98018d2426a17d0883dd17b830d5574a8a0f4e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 27 15:31:30 crc kubenswrapper[4772]: E0127 15:31:30.078442 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e92037085b98ccc46bbd64416f98018d2426a17d0883dd17b830d5574a8a0f4e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.082364 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a490a71b-c33d-4c94-9592-f97d1d315e81-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "a490a71b-c33d-4c94-9592-f97d1d315e81" (UID: "a490a71b-c33d-4c94-9592-f97d1d315e81"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.095497 4772 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a490a71b-c33d-4c94-9592-f97d1d315e81-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.095551 4772 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc34a3a4-ad0b-4154-82c9-728227b19732-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.095567 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.095580 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc34a3a4-ad0b-4154-82c9-728227b19732-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.095595 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fddd5e59-3124-4a05-aafd-92d6aea05f7e-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.095607 4772 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:30 crc kubenswrapper[4772]: E0127 15:31:30.102769 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e92037085b98ccc46bbd64416f98018d2426a17d0883dd17b830d5574a8a0f4e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 27 15:31:30 crc kubenswrapper[4772]: E0127 15:31:30.102866 4772 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="b83f7578-8113-46c8-be24-5968aa0ca563" containerName="nova-scheduler-scheduler" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.210671 4772 scope.go:117] "RemoveContainer" containerID="abdb27873ea97363386820f9e29ffa55e1d51fea483628db2a98ad8d2f8fc593" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.252545 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-d86f6cfbc-cwfmc"] Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.253278 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-d86f6cfbc-cwfmc" podUID="c16a29a0-7238-4a5e-b892-8f5195a1f486" containerName="proxy-httpd" containerID="cri-o://a476d84a3741734575b073569a645d9d973c5cdbb39812aa454a7257859db22b" gracePeriod=30 Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.253478 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-d86f6cfbc-cwfmc" podUID="c16a29a0-7238-4a5e-b892-8f5195a1f486" containerName="proxy-server" containerID="cri-o://47a1d8c4913044388b407e6a5c05783d2d3731216d7862873425d28265a5fe05" gracePeriod=30 Jan 27 15:31:30 crc kubenswrapper[4772]: W0127 15:31:30.526841 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef060591_3809_4f0b_974f_0785261db9b9.slice/crio-818d39110910f066b35d697f63a51f4883012a3f77256ac3126d09653c3a60e2 WatchSource:0}: Error finding container 818d39110910f066b35d697f63a51f4883012a3f77256ac3126d09653c3a60e2: Status 404 returned error can't find the container with id 818d39110910f066b35d697f63a51f4883012a3f77256ac3126d09653c3a60e2 Jan 27 15:31:30 crc kubenswrapper[4772]: E0127 15:31:30.531467 4772 kuberuntime_manager.go:1274] "Unhandled Error" err=< Jan 27 15:31:30 crc kubenswrapper[4772]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[/bin/sh -c #!/bin/bash Jan 27 15:31:30 crc kubenswrapper[4772]: Jan 27 15:31:30 crc kubenswrapper[4772]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Jan 27 15:31:30 crc kubenswrapper[4772]: Jan 27 15:31:30 crc kubenswrapper[4772]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Jan 27 15:31:30 crc kubenswrapper[4772]: Jan 27 15:31:30 crc kubenswrapper[4772]: MYSQL_CMD="mysql -h -u root -P 3306" Jan 27 15:31:30 crc kubenswrapper[4772]: Jan 27 15:31:30 crc kubenswrapper[4772]: if [ -n "barbican" ]; then Jan 27 15:31:30 crc kubenswrapper[4772]: GRANT_DATABASE="barbican" Jan 27 15:31:30 crc kubenswrapper[4772]: else Jan 27 15:31:30 crc kubenswrapper[4772]: GRANT_DATABASE="*" Jan 27 15:31:30 crc kubenswrapper[4772]: fi Jan 27 15:31:30 crc kubenswrapper[4772]: Jan 27 15:31:30 crc kubenswrapper[4772]: # going for maximum compatibility here: Jan 27 15:31:30 crc kubenswrapper[4772]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Jan 27 15:31:30 crc kubenswrapper[4772]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Jan 27 15:31:30 crc kubenswrapper[4772]: # 3. create user with CREATE but then do all password and TLS with ALTER to Jan 27 15:31:30 crc kubenswrapper[4772]: # support updates Jan 27 15:31:30 crc kubenswrapper[4772]: Jan 27 15:31:30 crc kubenswrapper[4772]: $MYSQL_CMD < logger="UnhandledError" Jan 27 15:31:30 crc kubenswrapper[4772]: E0127 15:31:30.532873 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"barbican-db-secret\\\" not found\"" pod="openstack/barbican-97c3-account-create-update-bvlvs" podUID="ef060591-3809-4f0b-974f-0785261db9b9" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.675002 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="5e69643a-e8c2-4057-a993-d5506ceeec1b" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"https://10.217.0.200:6080/vnc_lite.html\": dial tcp 10.217.0.200:6080: connect: connection refused" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.715344 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08d7e14a-70d3-446e-8250-ca1047b5bc4b" path="/var/lib/kubelet/pods/08d7e14a-70d3-446e-8250-ca1047b5bc4b/volumes" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.716445 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0edf6707-14dd-4986-8d64-0e48a31d6a39" path="/var/lib/kubelet/pods/0edf6707-14dd-4986-8d64-0e48a31d6a39/volumes" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.717344 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b68551f-119d-4d84-9c91-20e013018b7a" path="/var/lib/kubelet/pods/2b68551f-119d-4d84-9c91-20e013018b7a/volumes" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.718852 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54bbbf38-088b-4e4d-8154-569667fcf9a9" path="/var/lib/kubelet/pods/54bbbf38-088b-4e4d-8154-569667fcf9a9/volumes" Jan 27 15:31:30 crc kubenswrapper[4772]: E0127 15:31:30.719279 4772 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Jan 27 15:31:30 crc kubenswrapper[4772]: E0127 15:31:30.719399 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-config-data podName:76fdbdb1-d48a-4cd1-8372-78887671dce8 nodeName:}" failed. No retries permitted until 2026-01-27 15:31:34.719369878 +0000 UTC m=+1480.699978976 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-config-data") pod "rabbitmq-cell1-server-0" (UID: "76fdbdb1-d48a-4cd1-8372-78887671dce8") : configmap "rabbitmq-cell1-config-data" not found Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.719505 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7907cc16-7665-49d3-ad17-f9e6e0fc2f09" path="/var/lib/kubelet/pods/7907cc16-7665-49d3-ad17-f9e6e0fc2f09/volumes" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.720146 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be888039-f158-4d05-9f7d-6d01b2478b08" path="/var/lib/kubelet/pods/be888039-f158-4d05-9f7d-6d01b2478b08/volumes" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.721026 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112" path="/var/lib/kubelet/pods/ee9c9aa3-63e7-49ae-b3f3-f9bc0802f112/volumes" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.722379 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f91bfd1b-6386-444f-95da-045fbe957f5c" path="/var/lib/kubelet/pods/f91bfd1b-6386-444f-95da-045fbe957f5c/volumes" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.723127 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe34fbf1-61c4-46a9-9954-64ed431d2cb7" path="/var/lib/kubelet/pods/fe34fbf1-61c4-46a9-9954-64ed431d2cb7/volumes" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.880978 4772 scope.go:117] "RemoveContainer" containerID="fc9316436d0d6826797760ceeb255662e4eca1649200864d00e921be6f2e6eed" Jan 27 15:31:30 crc kubenswrapper[4772]: E0127 15:31:30.881434 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc9316436d0d6826797760ceeb255662e4eca1649200864d00e921be6f2e6eed\": container with ID starting with fc9316436d0d6826797760ceeb255662e4eca1649200864d00e921be6f2e6eed not found: ID does not exist" containerID="fc9316436d0d6826797760ceeb255662e4eca1649200864d00e921be6f2e6eed" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.881480 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc9316436d0d6826797760ceeb255662e4eca1649200864d00e921be6f2e6eed"} err="failed to get container status \"fc9316436d0d6826797760ceeb255662e4eca1649200864d00e921be6f2e6eed\": rpc error: code = NotFound desc = could not find container \"fc9316436d0d6826797760ceeb255662e4eca1649200864d00e921be6f2e6eed\": container with ID starting with fc9316436d0d6826797760ceeb255662e4eca1649200864d00e921be6f2e6eed not found: ID does not exist" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.881507 4772 scope.go:117] "RemoveContainer" containerID="abdb27873ea97363386820f9e29ffa55e1d51fea483628db2a98ad8d2f8fc593" Jan 27 15:31:30 crc kubenswrapper[4772]: E0127 15:31:30.881860 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abdb27873ea97363386820f9e29ffa55e1d51fea483628db2a98ad8d2f8fc593\": container with ID starting with abdb27873ea97363386820f9e29ffa55e1d51fea483628db2a98ad8d2f8fc593 not found: ID does not exist" containerID="abdb27873ea97363386820f9e29ffa55e1d51fea483628db2a98ad8d2f8fc593" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.881884 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abdb27873ea97363386820f9e29ffa55e1d51fea483628db2a98ad8d2f8fc593"} err="failed to get container status \"abdb27873ea97363386820f9e29ffa55e1d51fea483628db2a98ad8d2f8fc593\": rpc error: code = NotFound desc = could not find container \"abdb27873ea97363386820f9e29ffa55e1d51fea483628db2a98ad8d2f8fc593\": container with ID starting with abdb27873ea97363386820f9e29ffa55e1d51fea483628db2a98ad8d2f8fc593 not found: ID does not exist" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.881897 4772 scope.go:117] "RemoveContainer" containerID="0ee5661567fa3ca13869262f2ac472811c7a59976cae6fbfe300747e324b4e86" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.887050 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-556764fb84-r628x" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.904755 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-gxjzh"] Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.914591 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-gxjzh"] Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.926491 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.930128 4772 generic.go:334] "Generic (PLEG): container finished" podID="4cbf7469-816d-4e54-a7ad-b5b76d0d59d6" containerID="3cdc8204c9c28616053d96ae2843e1dddf8646f9b546be00bd78d90869086025" exitCode=1 Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.930207 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-qmppl" event={"ID":"4cbf7469-816d-4e54-a7ad-b5b76d0d59d6","Type":"ContainerDied","Data":"3cdc8204c9c28616053d96ae2843e1dddf8646f9b546be00bd78d90869086025"} Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.930316 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6748df9c8c-zk7zp" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.931335 4772 scope.go:117] "RemoveContainer" containerID="3cdc8204c9c28616053d96ae2843e1dddf8646f9b546be00bd78d90869086025" Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.934925 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.941448 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-97c3-account-create-update-bvlvs" event={"ID":"ef060591-3809-4f0b-974f-0785261db9b9","Type":"ContainerStarted","Data":"818d39110910f066b35d697f63a51f4883012a3f77256ac3126d09653c3a60e2"} Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.958901 4772 generic.go:334] "Generic (PLEG): container finished" podID="cf619242-7348-4de4-a37e-8ebdc4ca54d7" containerID="2e743dfaa62b788cb68a4d553d64cf9affaf8ef6e4da1308fddf4dc259167b69" exitCode=0 Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.958991 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"cf619242-7348-4de4-a37e-8ebdc4ca54d7","Type":"ContainerDied","Data":"2e743dfaa62b788cb68a4d553d64cf9affaf8ef6e4da1308fddf4dc259167b69"} Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.987105 4772 generic.go:334] "Generic (PLEG): container finished" podID="710edaa6-ba83-4b1f-a49a-769ca1911c9b" containerID="a063d80b4cd5f0199157f5e139c54f744514f0203001f47e4bce93805443a546" exitCode=0 Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.987192 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6748df9c8c-zk7zp" event={"ID":"710edaa6-ba83-4b1f-a49a-769ca1911c9b","Type":"ContainerDied","Data":"a063d80b4cd5f0199157f5e139c54f744514f0203001f47e4bce93805443a546"} Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.987221 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6748df9c8c-zk7zp" event={"ID":"710edaa6-ba83-4b1f-a49a-769ca1911c9b","Type":"ContainerDied","Data":"c83991847bf683630e70d44722d44695d9152a02d09f0a3d6fe39436ebbf262d"} Jan 27 15:31:30 crc kubenswrapper[4772]: I0127 15:31:30.987287 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6748df9c8c-zk7zp" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.004871 4772 generic.go:334] "Generic (PLEG): container finished" podID="c16a29a0-7238-4a5e-b892-8f5195a1f486" containerID="47a1d8c4913044388b407e6a5c05783d2d3731216d7862873425d28265a5fe05" exitCode=0 Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.004916 4772 generic.go:334] "Generic (PLEG): container finished" podID="c16a29a0-7238-4a5e-b892-8f5195a1f486" containerID="a476d84a3741734575b073569a645d9d973c5cdbb39812aa454a7257859db22b" exitCode=0 Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.004915 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-d86f6cfbc-cwfmc" event={"ID":"c16a29a0-7238-4a5e-b892-8f5195a1f486","Type":"ContainerDied","Data":"47a1d8c4913044388b407e6a5c05783d2d3731216d7862873425d28265a5fe05"} Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.005206 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-d86f6cfbc-cwfmc" event={"ID":"c16a29a0-7238-4a5e-b892-8f5195a1f486","Type":"ContainerDied","Data":"a476d84a3741734575b073569a645d9d973c5cdbb39812aa454a7257859db22b"} Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.012641 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.016336 4772 scope.go:117] "RemoveContainer" containerID="71013e440a971fe3ef401d90a82408249df3d1180b65da0eb4683442d6023500" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.029131 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ce27714-673f-47de-acc3-b6902b534bdd-config-data\") pod \"4ce27714-673f-47de-acc3-b6902b534bdd\" (UID: \"4ce27714-673f-47de-acc3-b6902b534bdd\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.029934 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ce27714-673f-47de-acc3-b6902b534bdd-combined-ca-bundle\") pod \"4ce27714-673f-47de-acc3-b6902b534bdd\" (UID: \"4ce27714-673f-47de-acc3-b6902b534bdd\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.030103 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ce27714-673f-47de-acc3-b6902b534bdd-logs\") pod \"4ce27714-673f-47de-acc3-b6902b534bdd\" (UID: \"4ce27714-673f-47de-acc3-b6902b534bdd\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.030241 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmmsl\" (UniqueName: \"kubernetes.io/projected/4ce27714-673f-47de-acc3-b6902b534bdd-kube-api-access-pmmsl\") pod \"4ce27714-673f-47de-acc3-b6902b534bdd\" (UID: \"4ce27714-673f-47de-acc3-b6902b534bdd\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.030525 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvs49\" (UniqueName: \"kubernetes.io/projected/710edaa6-ba83-4b1f-a49a-769ca1911c9b-kube-api-access-nvs49\") pod \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\" (UID: \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.030670 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/710edaa6-ba83-4b1f-a49a-769ca1911c9b-config-data-custom\") pod \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\" (UID: \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.030831 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ce27714-673f-47de-acc3-b6902b534bdd-config-data-custom\") pod \"4ce27714-673f-47de-acc3-b6902b534bdd\" (UID: \"4ce27714-673f-47de-acc3-b6902b534bdd\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.030959 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/710edaa6-ba83-4b1f-a49a-769ca1911c9b-config-data\") pod \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\" (UID: \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.031566 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/710edaa6-ba83-4b1f-a49a-769ca1911c9b-combined-ca-bundle\") pod \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\" (UID: \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.032017 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/710edaa6-ba83-4b1f-a49a-769ca1911c9b-logs\") pod \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\" (UID: \"710edaa6-ba83-4b1f-a49a-769ca1911c9b\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.033591 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ce27714-673f-47de-acc3-b6902b534bdd-logs" (OuterVolumeSpecName: "logs") pod "4ce27714-673f-47de-acc3-b6902b534bdd" (UID: "4ce27714-673f-47de-acc3-b6902b534bdd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.034703 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/710edaa6-ba83-4b1f-a49a-769ca1911c9b-logs" (OuterVolumeSpecName: "logs") pod "710edaa6-ba83-4b1f-a49a-769ca1911c9b" (UID: "710edaa6-ba83-4b1f-a49a-769ca1911c9b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.038076 4772 generic.go:334] "Generic (PLEG): container finished" podID="4ce27714-673f-47de-acc3-b6902b534bdd" containerID="aa76ea75f91196a6ccffd5d7e7d149b5efe900bfae2e86e19fa1ec88171321cc" exitCode=0 Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.038195 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-556764fb84-r628x" event={"ID":"4ce27714-673f-47de-acc3-b6902b534bdd","Type":"ContainerDied","Data":"aa76ea75f91196a6ccffd5d7e7d149b5efe900bfae2e86e19fa1ec88171321cc"} Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.038542 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/710edaa6-ba83-4b1f-a49a-769ca1911c9b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "710edaa6-ba83-4b1f-a49a-769ca1911c9b" (UID: "710edaa6-ba83-4b1f-a49a-769ca1911c9b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.038727 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-556764fb84-r628x" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.040306 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-556764fb84-r628x" event={"ID":"4ce27714-673f-47de-acc3-b6902b534bdd","Type":"ContainerDied","Data":"51e9e5e71be46820f9c3d1564ff14b9e6df8988ed057a1326779c07f7fee3331"} Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.040446 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/710edaa6-ba83-4b1f-a49a-769ca1911c9b-kube-api-access-nvs49" (OuterVolumeSpecName: "kube-api-access-nvs49") pod "710edaa6-ba83-4b1f-a49a-769ca1911c9b" (UID: "710edaa6-ba83-4b1f-a49a-769ca1911c9b"). InnerVolumeSpecName "kube-api-access-nvs49". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.041347 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ce27714-673f-47de-acc3-b6902b534bdd-kube-api-access-pmmsl" (OuterVolumeSpecName: "kube-api-access-pmmsl") pod "4ce27714-673f-47de-acc3-b6902b534bdd" (UID: "4ce27714-673f-47de-acc3-b6902b534bdd"). InnerVolumeSpecName "kube-api-access-pmmsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.045304 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ce27714-673f-47de-acc3-b6902b534bdd-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4ce27714-673f-47de-acc3-b6902b534bdd" (UID: "4ce27714-673f-47de-acc3-b6902b534bdd"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.058976 4772 generic.go:334] "Generic (PLEG): container finished" podID="5e69643a-e8c2-4057-a993-d5506ceeec1b" containerID="7a1429ee13edd2169e8a683ea45dcb648c58812d36d48307ba37a8f39d0a67ce" exitCode=0 Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.059023 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5e69643a-e8c2-4057-a993-d5506ceeec1b","Type":"ContainerDied","Data":"7a1429ee13edd2169e8a683ea45dcb648c58812d36d48307ba37a8f39d0a67ce"} Jan 27 15:31:31 crc kubenswrapper[4772]: E0127 15:31:31.088759 4772 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e69643a_e8c2_4057_a993_d5506ceeec1b.slice/crio-conmon-7a1429ee13edd2169e8a683ea45dcb648c58812d36d48307ba37a8f39d0a67ce.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e69643a_e8c2_4057_a993_d5506ceeec1b.slice/crio-7a1429ee13edd2169e8a683ea45dcb648c58812d36d48307ba37a8f39d0a67ce.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4270ab9b_f4a9_4d48_9cc2_f25152ee5fb2.slice/crio-2d9f9f123f138892540800ef23f48dae96e200e8a0b42b345d3f87addf089f7e\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda490a71b_c33d_4c94_9592_f97d1d315e81.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0edf6707_14dd_4986_8d64_0e48a31d6a39.slice/crio-5c52f5cf3b82427db2a187bbd0708a64e4f14f826b96324500c229ad2e72a4cf\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf619242_7348_4de4_a37e_8ebdc4ca54d7.slice/crio-conmon-2e743dfaa62b788cb68a4d553d64cf9affaf8ef6e4da1308fddf4dc259167b69.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda490a71b_c33d_4c94_9592_f97d1d315e81.slice/crio-b17e4736d7a1350c0c68f20fb3327f8519a43cb3b16a163a3b8e79d710328aca\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0edf6707_14dd_4986_8d64_0e48a31d6a39.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4270ab9b_f4a9_4d48_9cc2_f25152ee5fb2.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfddd5e59_3124_4a05_aafd_92d6aea05f7e.slice/crio-c5c94af58b0cd6c043cac9ed46da0616cb74fd66aa5279858fb42cf515ba3aa1\": RecentStats: unable to find data in memory cache]" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.112775 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.120384 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.129443 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-2hd4f"] Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.134221 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cf619242-7348-4de4-a37e-8ebdc4ca54d7-kolla-config\") pod \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.134405 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.134541 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvhlp\" (UniqueName: \"kubernetes.io/projected/cf619242-7348-4de4-a37e-8ebdc4ca54d7-kube-api-access-kvhlp\") pod \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.134651 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/cf619242-7348-4de4-a37e-8ebdc4ca54d7-config-data-generated\") pod \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.134801 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/cf619242-7348-4de4-a37e-8ebdc4ca54d7-config-data-default\") pod \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.134940 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf619242-7348-4de4-a37e-8ebdc4ca54d7-galera-tls-certs\") pod \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.135079 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf619242-7348-4de4-a37e-8ebdc4ca54d7-operator-scripts\") pod \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.135222 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf619242-7348-4de4-a37e-8ebdc4ca54d7-combined-ca-bundle\") pod \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\" (UID: \"cf619242-7348-4de4-a37e-8ebdc4ca54d7\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.135884 4772 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ce27714-673f-47de-acc3-b6902b534bdd-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.135987 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/710edaa6-ba83-4b1f-a49a-769ca1911c9b-logs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.136067 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ce27714-673f-47de-acc3-b6902b534bdd-logs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.136126 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmmsl\" (UniqueName: \"kubernetes.io/projected/4ce27714-673f-47de-acc3-b6902b534bdd-kube-api-access-pmmsl\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.136201 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvs49\" (UniqueName: \"kubernetes.io/projected/710edaa6-ba83-4b1f-a49a-769ca1911c9b-kube-api-access-nvs49\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.136256 4772 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/710edaa6-ba83-4b1f-a49a-769ca1911c9b-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.136995 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf619242-7348-4de4-a37e-8ebdc4ca54d7-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "cf619242-7348-4de4-a37e-8ebdc4ca54d7" (UID: "cf619242-7348-4de4-a37e-8ebdc4ca54d7"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.137414 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf619242-7348-4de4-a37e-8ebdc4ca54d7-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "cf619242-7348-4de4-a37e-8ebdc4ca54d7" (UID: "cf619242-7348-4de4-a37e-8ebdc4ca54d7"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.137587 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf619242-7348-4de4-a37e-8ebdc4ca54d7-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "cf619242-7348-4de4-a37e-8ebdc4ca54d7" (UID: "cf619242-7348-4de4-a37e-8ebdc4ca54d7"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.140305 4772 scope.go:117] "RemoveContainer" containerID="0ee5661567fa3ca13869262f2ac472811c7a59976cae6fbfe300747e324b4e86" Jan 27 15:31:31 crc kubenswrapper[4772]: E0127 15:31:31.140888 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ee5661567fa3ca13869262f2ac472811c7a59976cae6fbfe300747e324b4e86\": container with ID starting with 0ee5661567fa3ca13869262f2ac472811c7a59976cae6fbfe300747e324b4e86 not found: ID does not exist" containerID="0ee5661567fa3ca13869262f2ac472811c7a59976cae6fbfe300747e324b4e86" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.140929 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ee5661567fa3ca13869262f2ac472811c7a59976cae6fbfe300747e324b4e86"} err="failed to get container status \"0ee5661567fa3ca13869262f2ac472811c7a59976cae6fbfe300747e324b4e86\": rpc error: code = NotFound desc = could not find container \"0ee5661567fa3ca13869262f2ac472811c7a59976cae6fbfe300747e324b4e86\": container with ID starting with 0ee5661567fa3ca13869262f2ac472811c7a59976cae6fbfe300747e324b4e86 not found: ID does not exist" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.140958 4772 scope.go:117] "RemoveContainer" containerID="71013e440a971fe3ef401d90a82408249df3d1180b65da0eb4683442d6023500" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.141034 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf619242-7348-4de4-a37e-8ebdc4ca54d7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cf619242-7348-4de4-a37e-8ebdc4ca54d7" (UID: "cf619242-7348-4de4-a37e-8ebdc4ca54d7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.141193 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:31:31 crc kubenswrapper[4772]: E0127 15:31:31.141536 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71013e440a971fe3ef401d90a82408249df3d1180b65da0eb4683442d6023500\": container with ID starting with 71013e440a971fe3ef401d90a82408249df3d1180b65da0eb4683442d6023500 not found: ID does not exist" containerID="71013e440a971fe3ef401d90a82408249df3d1180b65da0eb4683442d6023500" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.141578 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71013e440a971fe3ef401d90a82408249df3d1180b65da0eb4683442d6023500"} err="failed to get container status \"71013e440a971fe3ef401d90a82408249df3d1180b65da0eb4683442d6023500\": rpc error: code = NotFound desc = could not find container \"71013e440a971fe3ef401d90a82408249df3d1180b65da0eb4683442d6023500\": container with ID starting with 71013e440a971fe3ef401d90a82408249df3d1180b65da0eb4683442d6023500 not found: ID does not exist" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.141607 4772 scope.go:117] "RemoveContainer" containerID="9ab2ac6bce7a8071ec2b4cecbc76933f6c63344bca73557900280dd89a9b1ef4" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.144196 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-2hd4f"] Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.154711 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-vqpfg"] Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.155362 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf619242-7348-4de4-a37e-8ebdc4ca54d7-kube-api-access-kvhlp" (OuterVolumeSpecName: "kube-api-access-kvhlp") pod "cf619242-7348-4de4-a37e-8ebdc4ca54d7" (UID: "cf619242-7348-4de4-a37e-8ebdc4ca54d7"). InnerVolumeSpecName "kube-api-access-kvhlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.168053 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ce27714-673f-47de-acc3-b6902b534bdd-config-data" (OuterVolumeSpecName: "config-data") pod "4ce27714-673f-47de-acc3-b6902b534bdd" (UID: "4ce27714-673f-47de-acc3-b6902b534bdd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.170230 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-vqpfg"] Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.192429 4772 scope.go:117] "RemoveContainer" containerID="abb84f069b7ba6556a04c96fbef42abc5bac570c75f402b32a5f9f20ac96046c" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.194643 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ce27714-673f-47de-acc3-b6902b534bdd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ce27714-673f-47de-acc3-b6902b534bdd" (UID: "4ce27714-673f-47de-acc3-b6902b534bdd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.221592 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "mysql-db") pod "cf619242-7348-4de4-a37e-8ebdc4ca54d7" (UID: "cf619242-7348-4de4-a37e-8ebdc4ca54d7"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.237807 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85ttx\" (UniqueName: \"kubernetes.io/projected/5e69643a-e8c2-4057-a993-d5506ceeec1b-kube-api-access-85ttx\") pod \"5e69643a-e8c2-4057-a993-d5506ceeec1b\" (UID: \"5e69643a-e8c2-4057-a993-d5506ceeec1b\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.237923 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e69643a-e8c2-4057-a993-d5506ceeec1b-nova-novncproxy-tls-certs\") pod \"5e69643a-e8c2-4057-a993-d5506ceeec1b\" (UID: \"5e69643a-e8c2-4057-a993-d5506ceeec1b\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.238039 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e69643a-e8c2-4057-a993-d5506ceeec1b-config-data\") pod \"5e69643a-e8c2-4057-a993-d5506ceeec1b\" (UID: \"5e69643a-e8c2-4057-a993-d5506ceeec1b\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.238094 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e69643a-e8c2-4057-a993-d5506ceeec1b-vencrypt-tls-certs\") pod \"5e69643a-e8c2-4057-a993-d5506ceeec1b\" (UID: \"5e69643a-e8c2-4057-a993-d5506ceeec1b\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.238198 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e69643a-e8c2-4057-a993-d5506ceeec1b-combined-ca-bundle\") pod \"5e69643a-e8c2-4057-a993-d5506ceeec1b\" (UID: \"5e69643a-e8c2-4057-a993-d5506ceeec1b\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.239373 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf619242-7348-4de4-a37e-8ebdc4ca54d7-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.239402 4772 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cf619242-7348-4de4-a37e-8ebdc4ca54d7-kolla-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.239438 4772 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.239459 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvhlp\" (UniqueName: \"kubernetes.io/projected/cf619242-7348-4de4-a37e-8ebdc4ca54d7-kube-api-access-kvhlp\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.239475 4772 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/cf619242-7348-4de4-a37e-8ebdc4ca54d7-config-data-generated\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.239490 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ce27714-673f-47de-acc3-b6902b534bdd-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.239501 4772 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/cf619242-7348-4de4-a37e-8ebdc4ca54d7-config-data-default\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.239513 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ce27714-673f-47de-acc3-b6902b534bdd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.259042 4772 scope.go:117] "RemoveContainer" containerID="9ab2ac6bce7a8071ec2b4cecbc76933f6c63344bca73557900280dd89a9b1ef4" Jan 27 15:31:31 crc kubenswrapper[4772]: E0127 15:31:31.261453 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ab2ac6bce7a8071ec2b4cecbc76933f6c63344bca73557900280dd89a9b1ef4\": container with ID starting with 9ab2ac6bce7a8071ec2b4cecbc76933f6c63344bca73557900280dd89a9b1ef4 not found: ID does not exist" containerID="9ab2ac6bce7a8071ec2b4cecbc76933f6c63344bca73557900280dd89a9b1ef4" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.261504 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ab2ac6bce7a8071ec2b4cecbc76933f6c63344bca73557900280dd89a9b1ef4"} err="failed to get container status \"9ab2ac6bce7a8071ec2b4cecbc76933f6c63344bca73557900280dd89a9b1ef4\": rpc error: code = NotFound desc = could not find container \"9ab2ac6bce7a8071ec2b4cecbc76933f6c63344bca73557900280dd89a9b1ef4\": container with ID starting with 9ab2ac6bce7a8071ec2b4cecbc76933f6c63344bca73557900280dd89a9b1ef4 not found: ID does not exist" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.261532 4772 scope.go:117] "RemoveContainer" containerID="abb84f069b7ba6556a04c96fbef42abc5bac570c75f402b32a5f9f20ac96046c" Jan 27 15:31:31 crc kubenswrapper[4772]: E0127 15:31:31.262109 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abb84f069b7ba6556a04c96fbef42abc5bac570c75f402b32a5f9f20ac96046c\": container with ID starting with abb84f069b7ba6556a04c96fbef42abc5bac570c75f402b32a5f9f20ac96046c not found: ID does not exist" containerID="abb84f069b7ba6556a04c96fbef42abc5bac570c75f402b32a5f9f20ac96046c" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.262218 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abb84f069b7ba6556a04c96fbef42abc5bac570c75f402b32a5f9f20ac96046c"} err="failed to get container status \"abb84f069b7ba6556a04c96fbef42abc5bac570c75f402b32a5f9f20ac96046c\": rpc error: code = NotFound desc = could not find container \"abb84f069b7ba6556a04c96fbef42abc5bac570c75f402b32a5f9f20ac96046c\": container with ID starting with abb84f069b7ba6556a04c96fbef42abc5bac570c75f402b32a5f9f20ac96046c not found: ID does not exist" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.262312 4772 scope.go:117] "RemoveContainer" containerID="afc8ab10fea0840566de64c53bc97d22454ee25e120ead660e5999b0da009daf" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.274895 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e69643a-e8c2-4057-a993-d5506ceeec1b-kube-api-access-85ttx" (OuterVolumeSpecName: "kube-api-access-85ttx") pod "5e69643a-e8c2-4057-a993-d5506ceeec1b" (UID: "5e69643a-e8c2-4057-a993-d5506ceeec1b"). InnerVolumeSpecName "kube-api-access-85ttx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.290129 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/710edaa6-ba83-4b1f-a49a-769ca1911c9b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "710edaa6-ba83-4b1f-a49a-769ca1911c9b" (UID: "710edaa6-ba83-4b1f-a49a-769ca1911c9b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.294293 4772 scope.go:117] "RemoveContainer" containerID="b93ad84c922746d427d3e2a2deb04a875a239fcafbecb5146ae05b1b11e36a09" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.318711 4772 scope.go:117] "RemoveContainer" containerID="0c37dc673e475cc4ca1e8b831b0543b26650ceedc799dace964e07fb4c7c7ffa" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.332206 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.343332 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/710edaa6-ba83-4b1f-a49a-769ca1911c9b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.343366 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85ttx\" (UniqueName: \"kubernetes.io/projected/5e69643a-e8c2-4057-a993-d5506ceeec1b-kube-api-access-85ttx\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.385328 4772 scope.go:117] "RemoveContainer" containerID="0c37dc673e475cc4ca1e8b831b0543b26650ceedc799dace964e07fb4c7c7ffa" Jan 27 15:31:31 crc kubenswrapper[4772]: E0127 15:31:31.386732 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c37dc673e475cc4ca1e8b831b0543b26650ceedc799dace964e07fb4c7c7ffa\": container with ID starting with 0c37dc673e475cc4ca1e8b831b0543b26650ceedc799dace964e07fb4c7c7ffa not found: ID does not exist" containerID="0c37dc673e475cc4ca1e8b831b0543b26650ceedc799dace964e07fb4c7c7ffa" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.387007 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c37dc673e475cc4ca1e8b831b0543b26650ceedc799dace964e07fb4c7c7ffa"} err="failed to get container status \"0c37dc673e475cc4ca1e8b831b0543b26650ceedc799dace964e07fb4c7c7ffa\": rpc error: code = NotFound desc = could not find container \"0c37dc673e475cc4ca1e8b831b0543b26650ceedc799dace964e07fb4c7c7ffa\": container with ID starting with 0c37dc673e475cc4ca1e8b831b0543b26650ceedc799dace964e07fb4c7c7ffa not found: ID does not exist" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.387040 4772 scope.go:117] "RemoveContainer" containerID="a063d80b4cd5f0199157f5e139c54f744514f0203001f47e4bce93805443a546" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.407473 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-556764fb84-r628x"] Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.410224 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf619242-7348-4de4-a37e-8ebdc4ca54d7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf619242-7348-4de4-a37e-8ebdc4ca54d7" (UID: "cf619242-7348-4de4-a37e-8ebdc4ca54d7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.412245 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-556764fb84-r628x"] Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.412332 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-97c3-account-create-update-bvlvs" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.421262 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/710edaa6-ba83-4b1f-a49a-769ca1911c9b-config-data" (OuterVolumeSpecName: "config-data") pod "710edaa6-ba83-4b1f-a49a-769ca1911c9b" (UID: "710edaa6-ba83-4b1f-a49a-769ca1911c9b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.428638 4772 scope.go:117] "RemoveContainer" containerID="b1ca77abcb5dfa41040a6625bbdc220ae80143a0714b3ff9a856057794a6d02a" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.439259 4772 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.444010 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c16a29a0-7238-4a5e-b892-8f5195a1f486-log-httpd\") pod \"c16a29a0-7238-4a5e-b892-8f5195a1f486\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.444157 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c16a29a0-7238-4a5e-b892-8f5195a1f486-etc-swift\") pod \"c16a29a0-7238-4a5e-b892-8f5195a1f486\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.444257 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c16a29a0-7238-4a5e-b892-8f5195a1f486-config-data\") pod \"c16a29a0-7238-4a5e-b892-8f5195a1f486\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.444303 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c16a29a0-7238-4a5e-b892-8f5195a1f486-combined-ca-bundle\") pod \"c16a29a0-7238-4a5e-b892-8f5195a1f486\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.444342 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c16a29a0-7238-4a5e-b892-8f5195a1f486-run-httpd\") pod \"c16a29a0-7238-4a5e-b892-8f5195a1f486\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.444386 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c16a29a0-7238-4a5e-b892-8f5195a1f486-internal-tls-certs\") pod \"c16a29a0-7238-4a5e-b892-8f5195a1f486\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.444470 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfzf2\" (UniqueName: \"kubernetes.io/projected/c16a29a0-7238-4a5e-b892-8f5195a1f486-kube-api-access-pfzf2\") pod \"c16a29a0-7238-4a5e-b892-8f5195a1f486\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.444504 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c16a29a0-7238-4a5e-b892-8f5195a1f486-public-tls-certs\") pod \"c16a29a0-7238-4a5e-b892-8f5195a1f486\" (UID: \"c16a29a0-7238-4a5e-b892-8f5195a1f486\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.444928 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf619242-7348-4de4-a37e-8ebdc4ca54d7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.444948 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/710edaa6-ba83-4b1f-a49a-769ca1911c9b-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.444960 4772 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.446497 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf619242-7348-4de4-a37e-8ebdc4ca54d7-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "cf619242-7348-4de4-a37e-8ebdc4ca54d7" (UID: "cf619242-7348-4de4-a37e-8ebdc4ca54d7"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.446944 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c16a29a0-7238-4a5e-b892-8f5195a1f486-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c16a29a0-7238-4a5e-b892-8f5195a1f486" (UID: "c16a29a0-7238-4a5e-b892-8f5195a1f486"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.447507 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c16a29a0-7238-4a5e-b892-8f5195a1f486-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c16a29a0-7238-4a5e-b892-8f5195a1f486" (UID: "c16a29a0-7238-4a5e-b892-8f5195a1f486"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.448121 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e69643a-e8c2-4057-a993-d5506ceeec1b-config-data" (OuterVolumeSpecName: "config-data") pod "5e69643a-e8c2-4057-a993-d5506ceeec1b" (UID: "5e69643a-e8c2-4057-a993-d5506ceeec1b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.462695 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e69643a-e8c2-4057-a993-d5506ceeec1b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e69643a-e8c2-4057-a993-d5506ceeec1b" (UID: "5e69643a-e8c2-4057-a993-d5506ceeec1b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.468665 4772 scope.go:117] "RemoveContainer" containerID="a063d80b4cd5f0199157f5e139c54f744514f0203001f47e4bce93805443a546" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.468835 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c16a29a0-7238-4a5e-b892-8f5195a1f486-kube-api-access-pfzf2" (OuterVolumeSpecName: "kube-api-access-pfzf2") pod "c16a29a0-7238-4a5e-b892-8f5195a1f486" (UID: "c16a29a0-7238-4a5e-b892-8f5195a1f486"). InnerVolumeSpecName "kube-api-access-pfzf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.469117 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c16a29a0-7238-4a5e-b892-8f5195a1f486-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "c16a29a0-7238-4a5e-b892-8f5195a1f486" (UID: "c16a29a0-7238-4a5e-b892-8f5195a1f486"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: E0127 15:31:31.469319 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a063d80b4cd5f0199157f5e139c54f744514f0203001f47e4bce93805443a546\": container with ID starting with a063d80b4cd5f0199157f5e139c54f744514f0203001f47e4bce93805443a546 not found: ID does not exist" containerID="a063d80b4cd5f0199157f5e139c54f744514f0203001f47e4bce93805443a546" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.469354 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a063d80b4cd5f0199157f5e139c54f744514f0203001f47e4bce93805443a546"} err="failed to get container status \"a063d80b4cd5f0199157f5e139c54f744514f0203001f47e4bce93805443a546\": rpc error: code = NotFound desc = could not find container \"a063d80b4cd5f0199157f5e139c54f744514f0203001f47e4bce93805443a546\": container with ID starting with a063d80b4cd5f0199157f5e139c54f744514f0203001f47e4bce93805443a546 not found: ID does not exist" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.469373 4772 scope.go:117] "RemoveContainer" containerID="b1ca77abcb5dfa41040a6625bbdc220ae80143a0714b3ff9a856057794a6d02a" Jan 27 15:31:31 crc kubenswrapper[4772]: E0127 15:31:31.469871 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1ca77abcb5dfa41040a6625bbdc220ae80143a0714b3ff9a856057794a6d02a\": container with ID starting with b1ca77abcb5dfa41040a6625bbdc220ae80143a0714b3ff9a856057794a6d02a not found: ID does not exist" containerID="b1ca77abcb5dfa41040a6625bbdc220ae80143a0714b3ff9a856057794a6d02a" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.469892 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1ca77abcb5dfa41040a6625bbdc220ae80143a0714b3ff9a856057794a6d02a"} err="failed to get container status \"b1ca77abcb5dfa41040a6625bbdc220ae80143a0714b3ff9a856057794a6d02a\": rpc error: code = NotFound desc = could not find container \"b1ca77abcb5dfa41040a6625bbdc220ae80143a0714b3ff9a856057794a6d02a\": container with ID starting with b1ca77abcb5dfa41040a6625bbdc220ae80143a0714b3ff9a856057794a6d02a not found: ID does not exist" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.469905 4772 scope.go:117] "RemoveContainer" containerID="aa76ea75f91196a6ccffd5d7e7d149b5efe900bfae2e86e19fa1ec88171321cc" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.524816 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c16a29a0-7238-4a5e-b892-8f5195a1f486-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c16a29a0-7238-4a5e-b892-8f5195a1f486" (UID: "c16a29a0-7238-4a5e-b892-8f5195a1f486"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.532697 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e69643a-e8c2-4057-a993-d5506ceeec1b-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "5e69643a-e8c2-4057-a993-d5506ceeec1b" (UID: "5e69643a-e8c2-4057-a993-d5506ceeec1b"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.535939 4772 scope.go:117] "RemoveContainer" containerID="f068099f2f85afe1f1db1c1b4191de3b3198e413724471d516ae5586de30eb8f" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.545796 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2j94f\" (UniqueName: \"kubernetes.io/projected/ef060591-3809-4f0b-974f-0785261db9b9-kube-api-access-2j94f\") pod \"ef060591-3809-4f0b-974f-0785261db9b9\" (UID: \"ef060591-3809-4f0b-974f-0785261db9b9\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.548396 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e69643a-e8c2-4057-a993-d5506ceeec1b-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "5e69643a-e8c2-4057-a993-d5506ceeec1b" (UID: "5e69643a-e8c2-4057-a993-d5506ceeec1b"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.549363 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef060591-3809-4f0b-974f-0785261db9b9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ef060591-3809-4f0b-974f-0785261db9b9" (UID: "ef060591-3809-4f0b-974f-0785261db9b9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.550013 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef060591-3809-4f0b-974f-0785261db9b9-operator-scripts\") pod \"ef060591-3809-4f0b-974f-0785261db9b9\" (UID: \"ef060591-3809-4f0b-974f-0785261db9b9\") " Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.550038 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef060591-3809-4f0b-974f-0785261db9b9-kube-api-access-2j94f" (OuterVolumeSpecName: "kube-api-access-2j94f") pod "ef060591-3809-4f0b-974f-0785261db9b9" (UID: "ef060591-3809-4f0b-974f-0785261db9b9"). InnerVolumeSpecName "kube-api-access-2j94f". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.559906 4772 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf619242-7348-4de4-a37e-8ebdc4ca54d7-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.560021 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e69643a-e8c2-4057-a993-d5506ceeec1b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.560044 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfzf2\" (UniqueName: \"kubernetes.io/projected/c16a29a0-7238-4a5e-b892-8f5195a1f486-kube-api-access-pfzf2\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.560061 4772 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c16a29a0-7238-4a5e-b892-8f5195a1f486-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.560074 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2j94f\" (UniqueName: \"kubernetes.io/projected/ef060591-3809-4f0b-974f-0785261db9b9-kube-api-access-2j94f\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.560085 4772 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c16a29a0-7238-4a5e-b892-8f5195a1f486-etc-swift\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.560098 4772 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e69643a-e8c2-4057-a993-d5506ceeec1b-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.560110 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e69643a-e8c2-4057-a993-d5506ceeec1b-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.560120 4772 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c16a29a0-7238-4a5e-b892-8f5195a1f486-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.560135 4772 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e69643a-e8c2-4057-a993-d5506ceeec1b-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.560146 4772 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c16a29a0-7238-4a5e-b892-8f5195a1f486-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.560158 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef060591-3809-4f0b-974f-0785261db9b9-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: E0127 15:31:31.560262 4772 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Jan 27 15:31:31 crc kubenswrapper[4772]: E0127 15:31:31.560327 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-config-data podName:508c3d5b-212a-46da-9a55-de3f35d7019b nodeName:}" failed. No retries permitted until 2026-01-27 15:31:35.56030754 +0000 UTC m=+1481.540916638 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-config-data") pod "rabbitmq-server-0" (UID: "508c3d5b-212a-46da-9a55-de3f35d7019b") : configmap "rabbitmq-config-data" not found Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.564369 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c16a29a0-7238-4a5e-b892-8f5195a1f486-config-data" (OuterVolumeSpecName: "config-data") pod "c16a29a0-7238-4a5e-b892-8f5195a1f486" (UID: "c16a29a0-7238-4a5e-b892-8f5195a1f486"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.580848 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c16a29a0-7238-4a5e-b892-8f5195a1f486-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c16a29a0-7238-4a5e-b892-8f5195a1f486" (UID: "c16a29a0-7238-4a5e-b892-8f5195a1f486"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.583726 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c16a29a0-7238-4a5e-b892-8f5195a1f486-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c16a29a0-7238-4a5e-b892-8f5195a1f486" (UID: "c16a29a0-7238-4a5e-b892-8f5195a1f486"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.668000 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c16a29a0-7238-4a5e-b892-8f5195a1f486-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.668027 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c16a29a0-7238-4a5e-b892-8f5195a1f486-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.668036 4772 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c16a29a0-7238-4a5e-b892-8f5195a1f486-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.738266 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/placement-597699949b-q6msx" podUID="4205dfea-7dc7-496a-9745-fc5e3d0a418a" containerName="placement-log" probeResult="failure" output="Get \"https://10.217.0.152:8778/\": read tcp 10.217.0.2:56924->10.217.0.152:8778: read: connection reset by peer" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.738296 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/placement-597699949b-q6msx" podUID="4205dfea-7dc7-496a-9745-fc5e3d0a418a" containerName="placement-api" probeResult="failure" output="Get \"https://10.217.0.152:8778/\": read tcp 10.217.0.2:56916->10.217.0.152:8778: read: connection reset by peer" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.779009 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-6748df9c8c-zk7zp"] Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.783464 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-6748df9c8c-zk7zp"] Jan 27 15:31:31 crc kubenswrapper[4772]: E0127 15:31:31.786484 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e74d40bce110215c2607c9a9b716bf9d9db61e446fe99bb897518adbdc86d83 is running failed: container process not found" containerID="2e74d40bce110215c2607c9a9b716bf9d9db61e446fe99bb897518adbdc86d83" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.786685 4772 scope.go:117] "RemoveContainer" containerID="aa76ea75f91196a6ccffd5d7e7d149b5efe900bfae2e86e19fa1ec88171321cc" Jan 27 15:31:31 crc kubenswrapper[4772]: E0127 15:31:31.789362 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa76ea75f91196a6ccffd5d7e7d149b5efe900bfae2e86e19fa1ec88171321cc\": container with ID starting with aa76ea75f91196a6ccffd5d7e7d149b5efe900bfae2e86e19fa1ec88171321cc not found: ID does not exist" containerID="aa76ea75f91196a6ccffd5d7e7d149b5efe900bfae2e86e19fa1ec88171321cc" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.789405 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa76ea75f91196a6ccffd5d7e7d149b5efe900bfae2e86e19fa1ec88171321cc"} err="failed to get container status \"aa76ea75f91196a6ccffd5d7e7d149b5efe900bfae2e86e19fa1ec88171321cc\": rpc error: code = NotFound desc = could not find container \"aa76ea75f91196a6ccffd5d7e7d149b5efe900bfae2e86e19fa1ec88171321cc\": container with ID starting with aa76ea75f91196a6ccffd5d7e7d149b5efe900bfae2e86e19fa1ec88171321cc not found: ID does not exist" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.789431 4772 scope.go:117] "RemoveContainer" containerID="f068099f2f85afe1f1db1c1b4191de3b3198e413724471d516ae5586de30eb8f" Jan 27 15:31:31 crc kubenswrapper[4772]: E0127 15:31:31.789517 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e74d40bce110215c2607c9a9b716bf9d9db61e446fe99bb897518adbdc86d83 is running failed: container process not found" containerID="2e74d40bce110215c2607c9a9b716bf9d9db61e446fe99bb897518adbdc86d83" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Jan 27 15:31:31 crc kubenswrapper[4772]: E0127 15:31:31.789851 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f068099f2f85afe1f1db1c1b4191de3b3198e413724471d516ae5586de30eb8f\": container with ID starting with f068099f2f85afe1f1db1c1b4191de3b3198e413724471d516ae5586de30eb8f not found: ID does not exist" containerID="f068099f2f85afe1f1db1c1b4191de3b3198e413724471d516ae5586de30eb8f" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.789869 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f068099f2f85afe1f1db1c1b4191de3b3198e413724471d516ae5586de30eb8f"} err="failed to get container status \"f068099f2f85afe1f1db1c1b4191de3b3198e413724471d516ae5586de30eb8f\": rpc error: code = NotFound desc = could not find container \"f068099f2f85afe1f1db1c1b4191de3b3198e413724471d516ae5586de30eb8f\": container with ID starting with f068099f2f85afe1f1db1c1b4191de3b3198e413724471d516ae5586de30eb8f not found: ID does not exist" Jan 27 15:31:31 crc kubenswrapper[4772]: E0127 15:31:31.789883 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e74d40bce110215c2607c9a9b716bf9d9db61e446fe99bb897518adbdc86d83 is running failed: container process not found" containerID="2e74d40bce110215c2607c9a9b716bf9d9db61e446fe99bb897518adbdc86d83" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Jan 27 15:31:31 crc kubenswrapper[4772]: E0127 15:31:31.789905 4772 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e74d40bce110215c2607c9a9b716bf9d9db61e446fe99bb897518adbdc86d83 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="b20b9215-5398-4100-bac4-763daa5ed222" containerName="nova-cell0-conductor-conductor" Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.901026 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.901383 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aea5ee43-36e3-437d-8aca-b2faedd87c5b" containerName="ceilometer-central-agent" containerID="cri-o://a4293d3cbd138216987430f5dab62fa26e55c56743eee0b42dd4fc7797a52afd" gracePeriod=30 Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.901864 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aea5ee43-36e3-437d-8aca-b2faedd87c5b" containerName="proxy-httpd" containerID="cri-o://0447c2ea1d147e4cee27fce146e4edc38d746774dc492452f5da3c48df7973bb" gracePeriod=30 Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.901995 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aea5ee43-36e3-437d-8aca-b2faedd87c5b" containerName="sg-core" containerID="cri-o://7b0085db2ce3021657d7773e88196b66b6759beeca3bff2b51fc3fdf5d6b4bd2" gracePeriod=30 Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.902012 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aea5ee43-36e3-437d-8aca-b2faedd87c5b" containerName="ceilometer-notification-agent" containerID="cri-o://81bb10c06283521cef14702be02bc4e89a7f82e4ae6c7d56b76d0d05f92797d0" gracePeriod=30 Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.917880 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 27 15:31:31 crc kubenswrapper[4772]: I0127 15:31:31.918076 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="21f54218-5889-4ae9-a7a1-7ed4895ad63c" containerName="kube-state-metrics" containerID="cri-o://670d5287e2a9882bc2137122191964eb76c57b36df9c904f50db621c1141ab98" gracePeriod=30 Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.074722 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-4c0e-account-create-update-w9dkg"] Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.104215 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-647c88bb6f-wzf82" podUID="6cf131c4-a5bd-452b-8598-42312c3a0270" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.167:9696/\": dial tcp 10.217.0.167:9696: connect: connection refused" Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.115484 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b is running failed: container process not found" containerID="4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.116414 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6579efc0c6a14eb40a1349e6b5e9e288881435286dc3fff811374b436b48c91" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.127411 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b is running failed: container process not found" containerID="4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.131263 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b is running failed: container process not found" containerID="4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.131332 4772 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-cqx7r" podUID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerName="ovsdb-server" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.155484 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-4c0e-account-create-update-w9dkg"] Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.163875 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6579efc0c6a14eb40a1349e6b5e9e288881435286dc3fff811374b436b48c91" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.164148 4772 generic.go:334] "Generic (PLEG): container finished" podID="21f54218-5889-4ae9-a7a1-7ed4895ad63c" containerID="670d5287e2a9882bc2137122191964eb76c57b36df9c904f50db621c1141ab98" exitCode=2 Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.164216 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"21f54218-5889-4ae9-a7a1-7ed4895ad63c","Type":"ContainerDied","Data":"670d5287e2a9882bc2137122191964eb76c57b36df9c904f50db621c1141ab98"} Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.182942 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6579efc0c6a14eb40a1349e6b5e9e288881435286dc3fff811374b436b48c91" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.182998 4772 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-cqx7r" podUID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerName="ovs-vswitchd" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.218837 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="f63bf600-ff03-43a3-92b4-fe8ac68a9bb7" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.207:8775/\": read tcp 10.217.0.2:33664->10.217.0.207:8775: read: connection reset by peer" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.218844 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="f63bf600-ff03-43a3-92b4-fe8ac68a9bb7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.207:8775/\": read tcp 10.217.0.2:33648->10.217.0.207:8775: read: connection reset by peer" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.219254 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.219437 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66" containerName="memcached" containerID="cri-o://faf687181014b14838de86572705cbe5952bdabca1b3fad7e35afc3ce6238c0f" gracePeriod=30 Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.231566 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-4c0e-account-create-update-wlbfm"] Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.232111 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a490a71b-c33d-4c94-9592-f97d1d315e81" containerName="openstack-network-exporter" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.232197 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a490a71b-c33d-4c94-9592-f97d1d315e81" containerName="openstack-network-exporter" Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.232392 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="710edaa6-ba83-4b1f-a49a-769ca1911c9b" containerName="barbican-worker" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.232459 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="710edaa6-ba83-4b1f-a49a-769ca1911c9b" containerName="barbican-worker" Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.232523 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c16a29a0-7238-4a5e-b892-8f5195a1f486" containerName="proxy-server" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.232572 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="c16a29a0-7238-4a5e-b892-8f5195a1f486" containerName="proxy-server" Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.232649 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2" containerName="openstack-network-exporter" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.232723 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2" containerName="openstack-network-exporter" Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.232797 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e69643a-e8c2-4057-a993-d5506ceeec1b" containerName="nova-cell1-novncproxy-novncproxy" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.232865 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e69643a-e8c2-4057-a993-d5506ceeec1b" containerName="nova-cell1-novncproxy-novncproxy" Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.233003 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2" containerName="ovsdbserver-nb" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.233070 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2" containerName="ovsdbserver-nb" Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.233146 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c16a29a0-7238-4a5e-b892-8f5195a1f486" containerName="proxy-httpd" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.233212 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="c16a29a0-7238-4a5e-b892-8f5195a1f486" containerName="proxy-httpd" Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.233593 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ce27714-673f-47de-acc3-b6902b534bdd" containerName="barbican-keystone-listener" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.240509 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ce27714-673f-47de-acc3-b6902b534bdd" containerName="barbican-keystone-listener" Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.240714 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fddd5e59-3124-4a05-aafd-92d6aea05f7e" containerName="dnsmasq-dns" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.240800 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="fddd5e59-3124-4a05-aafd-92d6aea05f7e" containerName="dnsmasq-dns" Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.240875 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="710edaa6-ba83-4b1f-a49a-769ca1911c9b" containerName="barbican-worker-log" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.240927 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="710edaa6-ba83-4b1f-a49a-769ca1911c9b" containerName="barbican-worker-log" Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.241014 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fddd5e59-3124-4a05-aafd-92d6aea05f7e" containerName="init" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.241062 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="fddd5e59-3124-4a05-aafd-92d6aea05f7e" containerName="init" Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.241114 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="220011f2-8778-4a14-82d4-33a07bd33379" containerName="ovn-controller" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.241160 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="220011f2-8778-4a14-82d4-33a07bd33379" containerName="ovn-controller" Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.241237 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf619242-7348-4de4-a37e-8ebdc4ca54d7" containerName="galera" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.241286 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf619242-7348-4de4-a37e-8ebdc4ca54d7" containerName="galera" Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.241349 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf619242-7348-4de4-a37e-8ebdc4ca54d7" containerName="mysql-bootstrap" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.241395 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf619242-7348-4de4-a37e-8ebdc4ca54d7" containerName="mysql-bootstrap" Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.241447 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ce27714-673f-47de-acc3-b6902b534bdd" containerName="barbican-keystone-listener-log" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.241493 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ce27714-673f-47de-acc3-b6902b534bdd" containerName="barbican-keystone-listener-log" Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.241553 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc34a3a4-ad0b-4154-82c9-728227b19732" containerName="openstack-network-exporter" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.241604 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc34a3a4-ad0b-4154-82c9-728227b19732" containerName="openstack-network-exporter" Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.241655 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc34a3a4-ad0b-4154-82c9-728227b19732" containerName="ovsdbserver-sb" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.241702 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc34a3a4-ad0b-4154-82c9-728227b19732" containerName="ovsdbserver-sb" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.234387 4772 generic.go:334] "Generic (PLEG): container finished" podID="aea5ee43-36e3-437d-8aca-b2faedd87c5b" containerID="7b0085db2ce3021657d7773e88196b66b6759beeca3bff2b51fc3fdf5d6b4bd2" exitCode=2 Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.242057 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc34a3a4-ad0b-4154-82c9-728227b19732" containerName="ovsdbserver-sb" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.243382 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc34a3a4-ad0b-4154-82c9-728227b19732" containerName="openstack-network-exporter" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.243460 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="fddd5e59-3124-4a05-aafd-92d6aea05f7e" containerName="dnsmasq-dns" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.243511 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="710edaa6-ba83-4b1f-a49a-769ca1911c9b" containerName="barbican-worker-log" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.243559 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2" containerName="ovsdbserver-nb" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.243607 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e69643a-e8c2-4057-a993-d5506ceeec1b" containerName="nova-cell1-novncproxy-novncproxy" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.243660 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ce27714-673f-47de-acc3-b6902b534bdd" containerName="barbican-keystone-listener-log" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.243720 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="c16a29a0-7238-4a5e-b892-8f5195a1f486" containerName="proxy-httpd" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.243768 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="c16a29a0-7238-4a5e-b892-8f5195a1f486" containerName="proxy-server" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.243818 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf619242-7348-4de4-a37e-8ebdc4ca54d7" containerName="galera" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.243871 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2" containerName="openstack-network-exporter" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.243918 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ce27714-673f-47de-acc3-b6902b534bdd" containerName="barbican-keystone-listener" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.243965 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="710edaa6-ba83-4b1f-a49a-769ca1911c9b" containerName="barbican-worker" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.244013 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="a490a71b-c33d-4c94-9592-f97d1d315e81" containerName="openstack-network-exporter" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.244072 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="220011f2-8778-4a14-82d4-33a07bd33379" containerName="ovn-controller" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.247477 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-4c0e-account-create-update-wlbfm"] Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.247534 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aea5ee43-36e3-437d-8aca-b2faedd87c5b","Type":"ContainerDied","Data":"7b0085db2ce3021657d7773e88196b66b6759beeca3bff2b51fc3fdf5d6b4bd2"} Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.247578 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4c0e-account-create-update-wlbfm" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.249651 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.250791 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.250876 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5e69643a-e8c2-4057-a993-d5506ceeec1b","Type":"ContainerDied","Data":"6444b6c25763e568fb1ce306052e7e5dc898559abb4fc82fb393de7a9f4a2b66"} Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.250951 4772 scope.go:117] "RemoveContainer" containerID="7a1429ee13edd2169e8a683ea45dcb648c58812d36d48307ba37a8f39d0a67ce" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.253978 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-97c3-account-create-update-bvlvs" event={"ID":"ef060591-3809-4f0b-974f-0785261db9b9","Type":"ContainerDied","Data":"818d39110910f066b35d697f63a51f4883012a3f77256ac3126d09653c3a60e2"} Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.254030 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-97c3-account-create-update-bvlvs" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.264328 4772 generic.go:334] "Generic (PLEG): container finished" podID="4205dfea-7dc7-496a-9745-fc5e3d0a418a" containerID="ad26ca4835a223df0b0aa3065e02d9e54b67030d2b6d0436f1f1a0dd7bf06415" exitCode=0 Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.264444 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-597699949b-q6msx" event={"ID":"4205dfea-7dc7-496a-9745-fc5e3d0a418a","Type":"ContainerDied","Data":"ad26ca4835a223df0b0aa3065e02d9e54b67030d2b6d0436f1f1a0dd7bf06415"} Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.274713 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="be772158-a71c-448d-8972-014f0d3a9ab8" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.163:8776/healthcheck\": dial tcp 10.217.0.163:8776: connect: connection refused" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.280322 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-fl4nt"] Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.286785 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-d86f6cfbc-cwfmc" event={"ID":"c16a29a0-7238-4a5e-b892-8f5195a1f486","Type":"ContainerDied","Data":"92e9170b2797b87fe5816f61d1944a7f0cca88f2e0e21f7420f27a5ed25d4005"} Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.286900 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-d86f6cfbc-cwfmc" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.334205 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-kvb25"] Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.341287 4772 scope.go:117] "RemoveContainer" containerID="47a1d8c4913044388b407e6a5c05783d2d3731216d7862873425d28265a5fe05" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.342586 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-659485ddbb-5bnzg" podUID="766c2a26-46ea-41b2-ba0c-2101ec9477d5" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.162:9311/healthcheck\": read tcp 10.217.0.2:44230->10.217.0.162:9311: read: connection reset by peer" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.342859 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-659485ddbb-5bnzg" podUID="766c2a26-46ea-41b2-ba0c-2101ec9477d5" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.162:9311/healthcheck\": read tcp 10.217.0.2:44232->10.217.0.162:9311: read: connection reset by peer" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.349872 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4llv\" (UniqueName: \"kubernetes.io/projected/05445a27-d839-4a60-8338-5ee5f2c3f9d7-kube-api-access-z4llv\") pod \"keystone-4c0e-account-create-update-wlbfm\" (UID: \"05445a27-d839-4a60-8338-5ee5f2c3f9d7\") " pod="openstack/keystone-4c0e-account-create-update-wlbfm" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.350105 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05445a27-d839-4a60-8338-5ee5f2c3f9d7-operator-scripts\") pod \"keystone-4c0e-account-create-update-wlbfm\" (UID: \"05445a27-d839-4a60-8338-5ee5f2c3f9d7\") " pod="openstack/keystone-4c0e-account-create-update-wlbfm" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.351202 4772 generic.go:334] "Generic (PLEG): container finished" podID="4cbf7469-816d-4e54-a7ad-b5b76d0d59d6" containerID="9005da10eaad68221a5ab75b0d10da02a46a7bd38d46bece0339dd56d8e2fc51" exitCode=1 Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.351290 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-qmppl" event={"ID":"4cbf7469-816d-4e54-a7ad-b5b76d0d59d6","Type":"ContainerDied","Data":"9005da10eaad68221a5ab75b0d10da02a46a7bd38d46bece0339dd56d8e2fc51"} Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.351789 4772 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/root-account-create-update-qmppl" secret="" err="secret \"galera-openstack-dockercfg-4dfv4\" not found" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.351828 4772 scope.go:117] "RemoveContainer" containerID="9005da10eaad68221a5ab75b0d10da02a46a7bd38d46bece0339dd56d8e2fc51" Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.352097 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CrashLoopBackOff: \"back-off 10s restarting failed container=mariadb-account-create-update pod=root-account-create-update-qmppl_openstack(4cbf7469-816d-4e54-a7ad-b5b76d0d59d6)\"" pod="openstack/root-account-create-update-qmppl" podUID="4cbf7469-816d-4e54-a7ad-b5b76d0d59d6" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.356853 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-kvb25"] Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.392500 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-fl4nt"] Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.393592 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.393789 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"cf619242-7348-4de4-a37e-8ebdc4ca54d7","Type":"ContainerDied","Data":"9858c0fc9167c8fdb9fe56212a74207375b7ea71449891249cf75618c47eff4b"} Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.405357 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.451753 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4llv\" (UniqueName: \"kubernetes.io/projected/05445a27-d839-4a60-8338-5ee5f2c3f9d7-kube-api-access-z4llv\") pod \"keystone-4c0e-account-create-update-wlbfm\" (UID: \"05445a27-d839-4a60-8338-5ee5f2c3f9d7\") " pod="openstack/keystone-4c0e-account-create-update-wlbfm" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.463729 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05445a27-d839-4a60-8338-5ee5f2c3f9d7-operator-scripts\") pod \"keystone-4c0e-account-create-update-wlbfm\" (UID: \"05445a27-d839-4a60-8338-5ee5f2c3f9d7\") " pod="openstack/keystone-4c0e-account-create-update-wlbfm" Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.458374 4772 projected.go:194] Error preparing data for projected volume kube-api-access-z4llv for pod openstack/keystone-4c0e-account-create-update-wlbfm: failed to fetch token: serviceaccounts "galera-openstack" not found Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.464113 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/05445a27-d839-4a60-8338-5ee5f2c3f9d7-kube-api-access-z4llv podName:05445a27-d839-4a60-8338-5ee5f2c3f9d7 nodeName:}" failed. No retries permitted until 2026-01-27 15:31:32.964098289 +0000 UTC m=+1478.944707387 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-z4llv" (UniqueName: "kubernetes.io/projected/05445a27-d839-4a60-8338-5ee5f2c3f9d7-kube-api-access-z4llv") pod "keystone-4c0e-account-create-update-wlbfm" (UID: "05445a27-d839-4a60-8338-5ee5f2c3f9d7") : failed to fetch token: serviceaccounts "galera-openstack" not found Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.461445 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-677fb7d6fc-djjsx"] Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.465116 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-677fb7d6fc-djjsx" podUID="6e790127-8223-4b0c-8a5d-21e1bb15fa30" containerName="keystone-api" containerID="cri-o://468321c234874e808e21c356adbece5162a3e84011f0215b573e541258fb76bc" gracePeriod=30 Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.464326 4772 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.467645 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/05445a27-d839-4a60-8338-5ee5f2c3f9d7-operator-scripts podName:05445a27-d839-4a60-8338-5ee5f2c3f9d7 nodeName:}" failed. No retries permitted until 2026-01-27 15:31:32.967631071 +0000 UTC m=+1478.948240169 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/05445a27-d839-4a60-8338-5ee5f2c3f9d7-operator-scripts") pod "keystone-4c0e-account-create-update-wlbfm" (UID: "05445a27-d839-4a60-8338-5ee5f2c3f9d7") : configmap "openstack-scripts" not found Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.464016 4772 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.467769 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4cbf7469-816d-4e54-a7ad-b5b76d0d59d6-operator-scripts podName:4cbf7469-816d-4e54-a7ad-b5b76d0d59d6 nodeName:}" failed. No retries permitted until 2026-01-27 15:31:32.967762404 +0000 UTC m=+1478.948371502 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4cbf7469-816d-4e54-a7ad-b5b76d0d59d6-operator-scripts") pod "root-account-create-update-qmppl" (UID: "4cbf7469-816d-4e54-a7ad-b5b76d0d59d6") : configmap "openstack-scripts" not found Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.469593 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-nmvpf"] Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.505825 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-4c0e-account-create-update-wlbfm"] Jan 27 15:31:32 crc kubenswrapper[4772]: E0127 15:31:32.506453 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-z4llv operator-scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/keystone-4c0e-account-create-update-wlbfm" podUID="05445a27-d839-4a60-8338-5ee5f2c3f9d7" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.516946 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-nmvpf"] Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.589872 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.590329 4772 generic.go:334] "Generic (PLEG): container finished" podID="be772158-a71c-448d-8972-014f0d3a9ab8" containerID="c47159ab0aee5087f5a44073988d2ad8d6aaaa0e47ba7702dc2a03eab229b375" exitCode=0 Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.590386 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"be772158-a71c-448d-8972-014f0d3a9ab8","Type":"ContainerDied","Data":"c47159ab0aee5087f5a44073988d2ad8d6aaaa0e47ba7702dc2a03eab229b375"} Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.601234 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-qmppl"] Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.626597 4772 scope.go:117] "RemoveContainer" containerID="a476d84a3741734575b073569a645d9d973c5cdbb39812aa454a7257859db22b" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.635199 4772 generic.go:334] "Generic (PLEG): container finished" podID="9a02b617-28a7-4262-a110-f1c71763ad19" containerID="3114715e24bc63a93ce31ec7ec2cc2fdeaad0a6c7647de22f23d06ac45e3d864" exitCode=0 Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.635297 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9a02b617-28a7-4262-a110-f1c71763ad19","Type":"ContainerDied","Data":"3114715e24bc63a93ce31ec7ec2cc2fdeaad0a6c7647de22f23d06ac45e3d864"} Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.669215 4772 generic.go:334] "Generic (PLEG): container finished" podID="e7e78641-77e6-4c89-b5c9-0d6f3c9a9343" containerID="6481b50eed7f8997cc197c4b50a1b5d1b9aa395b3745aa30ff2d6ee451d23215" exitCode=0 Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.670278 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b20b9215-5398-4100-bac4-763daa5ed222-combined-ca-bundle\") pod \"b20b9215-5398-4100-bac4-763daa5ed222\" (UID: \"b20b9215-5398-4100-bac4-763daa5ed222\") " Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.674704 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b20b9215-5398-4100-bac4-763daa5ed222-config-data\") pod \"b20b9215-5398-4100-bac4-763daa5ed222\" (UID: \"b20b9215-5398-4100-bac4-763daa5ed222\") " Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.675808 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kbc8\" (UniqueName: \"kubernetes.io/projected/b20b9215-5398-4100-bac4-763daa5ed222-kube-api-access-8kbc8\") pod \"b20b9215-5398-4100-bac4-763daa5ed222\" (UID: \"b20b9215-5398-4100-bac4-763daa5ed222\") " Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.683565 4772 generic.go:334] "Generic (PLEG): container finished" podID="b20b9215-5398-4100-bac4-763daa5ed222" containerID="2e74d40bce110215c2607c9a9b716bf9d9db61e446fe99bb897518adbdc86d83" exitCode=0 Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.683653 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.687550 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="220011f2-8778-4a14-82d4-33a07bd33379" path="/var/lib/kubelet/pods/220011f2-8778-4a14-82d4-33a07bd33379/volumes" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.688943 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2" path="/var/lib/kubelet/pods/4270ab9b-f4a9-4d48-9cc2-f25152ee5fb2/volumes" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.689532 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ce27714-673f-47de-acc3-b6902b534bdd" path="/var/lib/kubelet/pods/4ce27714-673f-47de-acc3-b6902b534bdd/volumes" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.690687 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a1d71f-3b00-42c0-92c4-a29fb3d4518c" path="/var/lib/kubelet/pods/57a1d71f-3b00-42c0-92c4-a29fb3d4518c/volumes" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.691242 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="710edaa6-ba83-4b1f-a49a-769ca1911c9b" path="/var/lib/kubelet/pods/710edaa6-ba83-4b1f-a49a-769ca1911c9b/volumes" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.691781 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8322baad-60c1-4d0b-96e3-51038f2e447a" path="/var/lib/kubelet/pods/8322baad-60c1-4d0b-96e3-51038f2e447a/volumes" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.692901 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a490a71b-c33d-4c94-9592-f97d1d315e81" path="/var/lib/kubelet/pods/a490a71b-c33d-4c94-9592-f97d1d315e81/volumes" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.700057 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbad3a30-e11d-4ae8-9c42-e06b6382c6de" path="/var/lib/kubelet/pods/bbad3a30-e11d-4ae8-9c42-e06b6382c6de/volumes" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.700728 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc34a3a4-ad0b-4154-82c9-728227b19732" path="/var/lib/kubelet/pods/dc34a3a4-ad0b-4154-82c9-728227b19732/volumes" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.701284 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef900211-2a44-498c-adb6-fec1abcba5ec" path="/var/lib/kubelet/pods/ef900211-2a44-498c-adb6-fec1abcba5ec/volumes" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.702307 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fddd5e59-3124-4a05-aafd-92d6aea05f7e" path="/var/lib/kubelet/pods/fddd5e59-3124-4a05-aafd-92d6aea05f7e/volumes" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.705396 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b20b9215-5398-4100-bac4-763daa5ed222-kube-api-access-8kbc8" (OuterVolumeSpecName: "kube-api-access-8kbc8") pod "b20b9215-5398-4100-bac4-763daa5ed222" (UID: "b20b9215-5398-4100-bac4-763daa5ed222"). InnerVolumeSpecName "kube-api-access-8kbc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.716824 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b20b9215-5398-4100-bac4-763daa5ed222-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b20b9215-5398-4100-bac4-763daa5ed222" (UID: "b20b9215-5398-4100-bac4-763daa5ed222"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.734219 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b20b9215-5398-4100-bac4-763daa5ed222-config-data" (OuterVolumeSpecName: "config-data") pod "b20b9215-5398-4100-bac4-763daa5ed222" (UID: "b20b9215-5398-4100-bac4-763daa5ed222"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.755459 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="b1515626-5d79-408d-abc1-cb92abd58f3f" containerName="galera" containerID="cri-o://21613f2614f5809a9a792371fe2e685753a1a2fc6ea2f8fa7dcc2390d4bafda0" gracePeriod=30 Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.802821 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kbc8\" (UniqueName: \"kubernetes.io/projected/b20b9215-5398-4100-bac4-763daa5ed222-kube-api-access-8kbc8\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.802853 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b20b9215-5398-4100-bac4-763daa5ed222-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.802866 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b20b9215-5398-4100-bac4-763daa5ed222-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.805938 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.805967 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343","Type":"ContainerDied","Data":"6481b50eed7f8997cc197c4b50a1b5d1b9aa395b3745aa30ff2d6ee451d23215"} Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.805987 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"b20b9215-5398-4100-bac4-763daa5ed222","Type":"ContainerDied","Data":"2e74d40bce110215c2607c9a9b716bf9d9db61e446fe99bb897518adbdc86d83"} Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.806002 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.806019 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-97c3-account-create-update-bvlvs"] Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.806029 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-97c3-account-create-update-bvlvs"] Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.806043 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-d86f6cfbc-cwfmc"] Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.806056 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-d86f6cfbc-cwfmc"] Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.806070 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.806080 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.817978 4772 scope.go:117] "RemoveContainer" containerID="3cdc8204c9c28616053d96ae2843e1dddf8646f9b546be00bd78d90869086025" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.916090 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="508c3d5b-212a-46da-9a55-de3f35d7019b" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Jan 27 15:31:32 crc kubenswrapper[4772]: I0127 15:31:32.966828 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.004823 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.005576 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4llv\" (UniqueName: \"kubernetes.io/projected/05445a27-d839-4a60-8338-5ee5f2c3f9d7-kube-api-access-z4llv\") pod \"keystone-4c0e-account-create-update-wlbfm\" (UID: \"05445a27-d839-4a60-8338-5ee5f2c3f9d7\") " pod="openstack/keystone-4c0e-account-create-update-wlbfm" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.005680 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05445a27-d839-4a60-8338-5ee5f2c3f9d7-operator-scripts\") pod \"keystone-4c0e-account-create-update-wlbfm\" (UID: \"05445a27-d839-4a60-8338-5ee5f2c3f9d7\") " pod="openstack/keystone-4c0e-account-create-update-wlbfm" Jan 27 15:31:33 crc kubenswrapper[4772]: E0127 15:31:33.005864 4772 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Jan 27 15:31:33 crc kubenswrapper[4772]: E0127 15:31:33.005928 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4cbf7469-816d-4e54-a7ad-b5b76d0d59d6-operator-scripts podName:4cbf7469-816d-4e54-a7ad-b5b76d0d59d6 nodeName:}" failed. No retries permitted until 2026-01-27 15:31:34.005910956 +0000 UTC m=+1479.986520054 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4cbf7469-816d-4e54-a7ad-b5b76d0d59d6-operator-scripts") pod "root-account-create-update-qmppl" (UID: "4cbf7469-816d-4e54-a7ad-b5b76d0d59d6") : configmap "openstack-scripts" not found Jan 27 15:31:33 crc kubenswrapper[4772]: E0127 15:31:33.008329 4772 projected.go:194] Error preparing data for projected volume kube-api-access-z4llv for pod openstack/keystone-4c0e-account-create-update-wlbfm: failed to fetch token: serviceaccounts "galera-openstack" not found Jan 27 15:31:33 crc kubenswrapper[4772]: E0127 15:31:33.008465 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/05445a27-d839-4a60-8338-5ee5f2c3f9d7-kube-api-access-z4llv podName:05445a27-d839-4a60-8338-5ee5f2c3f9d7 nodeName:}" failed. No retries permitted until 2026-01-27 15:31:34.00845163 +0000 UTC m=+1479.989060718 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-z4llv" (UniqueName: "kubernetes.io/projected/05445a27-d839-4a60-8338-5ee5f2c3f9d7-kube-api-access-z4llv") pod "keystone-4c0e-account-create-update-wlbfm" (UID: "05445a27-d839-4a60-8338-5ee5f2c3f9d7") : failed to fetch token: serviceaccounts "galera-openstack" not found Jan 27 15:31:33 crc kubenswrapper[4772]: E0127 15:31:33.008907 4772 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Jan 27 15:31:33 crc kubenswrapper[4772]: E0127 15:31:33.008976 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/05445a27-d839-4a60-8338-5ee5f2c3f9d7-operator-scripts podName:05445a27-d839-4a60-8338-5ee5f2c3f9d7 nodeName:}" failed. No retries permitted until 2026-01-27 15:31:34.008958374 +0000 UTC m=+1479.989567502 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/05445a27-d839-4a60-8338-5ee5f2c3f9d7-operator-scripts") pod "keystone-4c0e-account-create-update-wlbfm" (UID: "05445a27-d839-4a60-8338-5ee5f2c3f9d7") : configmap "openstack-scripts" not found Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.022572 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.029011 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.107501 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nblgh\" (UniqueName: \"kubernetes.io/projected/21f54218-5889-4ae9-a7a1-7ed4895ad63c-kube-api-access-nblgh\") pod \"21f54218-5889-4ae9-a7a1-7ed4895ad63c\" (UID: \"21f54218-5889-4ae9-a7a1-7ed4895ad63c\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.108325 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21f54218-5889-4ae9-a7a1-7ed4895ad63c-combined-ca-bundle\") pod \"21f54218-5889-4ae9-a7a1-7ed4895ad63c\" (UID: \"21f54218-5889-4ae9-a7a1-7ed4895ad63c\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.108487 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmnhr\" (UniqueName: \"kubernetes.io/projected/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-kube-api-access-pmnhr\") pod \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\" (UID: \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.108964 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-config-data\") pod \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\" (UID: \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.109099 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-logs\") pod \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\" (UID: \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.109232 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-combined-ca-bundle\") pod \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\" (UID: \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.109302 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/21f54218-5889-4ae9-a7a1-7ed4895ad63c-kube-state-metrics-tls-config\") pod \"21f54218-5889-4ae9-a7a1-7ed4895ad63c\" (UID: \"21f54218-5889-4ae9-a7a1-7ed4895ad63c\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.109378 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-internal-tls-certs\") pod \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\" (UID: \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.109503 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/21f54218-5889-4ae9-a7a1-7ed4895ad63c-kube-state-metrics-tls-certs\") pod \"21f54218-5889-4ae9-a7a1-7ed4895ad63c\" (UID: \"21f54218-5889-4ae9-a7a1-7ed4895ad63c\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.109593 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-public-tls-certs\") pod \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\" (UID: \"93c8f9a4-c6ef-42b8-8543-ff8b5347977e\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.112408 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-kube-api-access-pmnhr" (OuterVolumeSpecName: "kube-api-access-pmnhr") pod "93c8f9a4-c6ef-42b8-8543-ff8b5347977e" (UID: "93c8f9a4-c6ef-42b8-8543-ff8b5347977e"). InnerVolumeSpecName "kube-api-access-pmnhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.114705 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21f54218-5889-4ae9-a7a1-7ed4895ad63c-kube-api-access-nblgh" (OuterVolumeSpecName: "kube-api-access-nblgh") pod "21f54218-5889-4ae9-a7a1-7ed4895ad63c" (UID: "21f54218-5889-4ae9-a7a1-7ed4895ad63c"). InnerVolumeSpecName "kube-api-access-nblgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.115538 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-logs" (OuterVolumeSpecName: "logs") pod "93c8f9a4-c6ef-42b8-8543-ff8b5347977e" (UID: "93c8f9a4-c6ef-42b8-8543-ff8b5347977e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.183319 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-config-data" (OuterVolumeSpecName: "config-data") pod "93c8f9a4-c6ef-42b8-8543-ff8b5347977e" (UID: "93c8f9a4-c6ef-42b8-8543-ff8b5347977e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.183537 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21f54218-5889-4ae9-a7a1-7ed4895ad63c-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "21f54218-5889-4ae9-a7a1-7ed4895ad63c" (UID: "21f54218-5889-4ae9-a7a1-7ed4895ad63c"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.187745 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21f54218-5889-4ae9-a7a1-7ed4895ad63c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "21f54218-5889-4ae9-a7a1-7ed4895ad63c" (UID: "21f54218-5889-4ae9-a7a1-7ed4895ad63c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.208950 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "93c8f9a4-c6ef-42b8-8543-ff8b5347977e" (UID: "93c8f9a4-c6ef-42b8-8543-ff8b5347977e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.213094 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nblgh\" (UniqueName: \"kubernetes.io/projected/21f54218-5889-4ae9-a7a1-7ed4895ad63c-kube-api-access-nblgh\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.213268 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21f54218-5889-4ae9-a7a1-7ed4895ad63c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.213393 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmnhr\" (UniqueName: \"kubernetes.io/projected/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-kube-api-access-pmnhr\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.213471 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.213546 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-logs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.213602 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.213654 4772 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/21f54218-5889-4ae9-a7a1-7ed4895ad63c-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.233367 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.241528 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "93c8f9a4-c6ef-42b8-8543-ff8b5347977e" (UID: "93c8f9a4-c6ef-42b8-8543-ff8b5347977e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.245117 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "93c8f9a4-c6ef-42b8-8543-ff8b5347977e" (UID: "93c8f9a4-c6ef-42b8-8543-ff8b5347977e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.247263 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.265179 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-597699949b-q6msx" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.270675 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.287068 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.289609 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21f54218-5889-4ae9-a7a1-7ed4895ad63c-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "21f54218-5889-4ae9-a7a1-7ed4895ad63c" (UID: "21f54218-5889-4ae9-a7a1-7ed4895ad63c"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.315465 4772 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/21f54218-5889-4ae9-a7a1-7ed4895ad63c-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.315494 4772 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.315503 4772 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/93c8f9a4-c6ef-42b8-8543-ff8b5347977e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.417353 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-public-tls-certs\") pod \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.417808 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/766c2a26-46ea-41b2-ba0c-2101ec9477d5-logs\") pod \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.417846 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be772158-a71c-448d-8972-014f0d3a9ab8-logs\") pod \"be772158-a71c-448d-8972-014f0d3a9ab8\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.417886 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-config-data\") pod \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.417915 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-public-tls-certs\") pod \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.417964 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-combined-ca-bundle\") pod \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418004 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-logs\") pod \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418037 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghzrk\" (UniqueName: \"kubernetes.io/projected/9a02b617-28a7-4262-a110-f1c71763ad19-kube-api-access-ghzrk\") pod \"9a02b617-28a7-4262-a110-f1c71763ad19\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418064 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-config-data\") pod \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418088 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpm5h\" (UniqueName: \"kubernetes.io/projected/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-kube-api-access-rpm5h\") pod \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418117 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418149 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-combined-ca-bundle\") pod \"be772158-a71c-448d-8972-014f0d3a9ab8\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418194 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-combined-ca-bundle\") pod \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418219 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-config-data-custom\") pod \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418238 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"9a02b617-28a7-4262-a110-f1c71763ad19\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418265 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-internal-tls-certs\") pod \"be772158-a71c-448d-8972-014f0d3a9ab8\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418295 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a02b617-28a7-4262-a110-f1c71763ad19-logs\") pod \"9a02b617-28a7-4262-a110-f1c71763ad19\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418326 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-internal-tls-certs\") pod \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418361 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-httpd-run\") pod \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418383 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/be772158-a71c-448d-8972-014f0d3a9ab8-etc-machine-id\") pod \"be772158-a71c-448d-8972-014f0d3a9ab8\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418419 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4205dfea-7dc7-496a-9745-fc5e3d0a418a-logs\") pod \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418450 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-config-data\") pod \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418464 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be772158-a71c-448d-8972-014f0d3a9ab8-logs" (OuterVolumeSpecName: "logs") pod "be772158-a71c-448d-8972-014f0d3a9ab8" (UID: "be772158-a71c-448d-8972-014f0d3a9ab8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418482 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-combined-ca-bundle\") pod \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418565 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a02b617-28a7-4262-a110-f1c71763ad19-httpd-run\") pod \"9a02b617-28a7-4262-a110-f1c71763ad19\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418602 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-internal-tls-certs\") pod \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418620 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-internal-tls-certs\") pod \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418638 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-config-data-custom\") pod \"be772158-a71c-448d-8972-014f0d3a9ab8\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418665 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a02b617-28a7-4262-a110-f1c71763ad19-config-data\") pod \"9a02b617-28a7-4262-a110-f1c71763ad19\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418687 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a02b617-28a7-4262-a110-f1c71763ad19-public-tls-certs\") pod \"9a02b617-28a7-4262-a110-f1c71763ad19\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418710 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-scripts\") pod \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418743 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqskw\" (UniqueName: \"kubernetes.io/projected/766c2a26-46ea-41b2-ba0c-2101ec9477d5-kube-api-access-sqskw\") pod \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\" (UID: \"766c2a26-46ea-41b2-ba0c-2101ec9477d5\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418771 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-config-data\") pod \"be772158-a71c-448d-8972-014f0d3a9ab8\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418794 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vclj\" (UniqueName: \"kubernetes.io/projected/be772158-a71c-448d-8972-014f0d3a9ab8-kube-api-access-2vclj\") pod \"be772158-a71c-448d-8972-014f0d3a9ab8\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418812 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-scripts\") pod \"be772158-a71c-448d-8972-014f0d3a9ab8\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418837 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-scripts\") pod \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\" (UID: \"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418864 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fknks\" (UniqueName: \"kubernetes.io/projected/4205dfea-7dc7-496a-9745-fc5e3d0a418a-kube-api-access-fknks\") pod \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\" (UID: \"4205dfea-7dc7-496a-9745-fc5e3d0a418a\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418882 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-public-tls-certs\") pod \"be772158-a71c-448d-8972-014f0d3a9ab8\" (UID: \"be772158-a71c-448d-8972-014f0d3a9ab8\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418897 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a02b617-28a7-4262-a110-f1c71763ad19-combined-ca-bundle\") pod \"9a02b617-28a7-4262-a110-f1c71763ad19\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.418916 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a02b617-28a7-4262-a110-f1c71763ad19-scripts\") pod \"9a02b617-28a7-4262-a110-f1c71763ad19\" (UID: \"9a02b617-28a7-4262-a110-f1c71763ad19\") " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.433390 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be772158-a71c-448d-8972-014f0d3a9ab8-logs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.438297 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-logs" (OuterVolumeSpecName: "logs") pod "e7e78641-77e6-4c89-b5c9-0d6f3c9a9343" (UID: "e7e78641-77e6-4c89-b5c9-0d6f3c9a9343"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.442697 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a02b617-28a7-4262-a110-f1c71763ad19-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9a02b617-28a7-4262-a110-f1c71763ad19" (UID: "9a02b617-28a7-4262-a110-f1c71763ad19"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.444449 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a02b617-28a7-4262-a110-f1c71763ad19-kube-api-access-ghzrk" (OuterVolumeSpecName: "kube-api-access-ghzrk") pod "9a02b617-28a7-4262-a110-f1c71763ad19" (UID: "9a02b617-28a7-4262-a110-f1c71763ad19"). InnerVolumeSpecName "kube-api-access-ghzrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.467157 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "be772158-a71c-448d-8972-014f0d3a9ab8" (UID: "be772158-a71c-448d-8972-014f0d3a9ab8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.475498 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a02b617-28a7-4262-a110-f1c71763ad19-scripts" (OuterVolumeSpecName: "scripts") pod "9a02b617-28a7-4262-a110-f1c71763ad19" (UID: "9a02b617-28a7-4262-a110-f1c71763ad19"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.477549 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/766c2a26-46ea-41b2-ba0c-2101ec9477d5-kube-api-access-sqskw" (OuterVolumeSpecName: "kube-api-access-sqskw") pod "766c2a26-46ea-41b2-ba0c-2101ec9477d5" (UID: "766c2a26-46ea-41b2-ba0c-2101ec9477d5"). InnerVolumeSpecName "kube-api-access-sqskw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.477833 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-scripts" (OuterVolumeSpecName: "scripts") pod "be772158-a71c-448d-8972-014f0d3a9ab8" (UID: "be772158-a71c-448d-8972-014f0d3a9ab8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.477912 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-scripts" (OuterVolumeSpecName: "scripts") pod "4205dfea-7dc7-496a-9745-fc5e3d0a418a" (UID: "4205dfea-7dc7-496a-9745-fc5e3d0a418a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.478401 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a02b617-28a7-4262-a110-f1c71763ad19-logs" (OuterVolumeSpecName: "logs") pod "9a02b617-28a7-4262-a110-f1c71763ad19" (UID: "9a02b617-28a7-4262-a110-f1c71763ad19"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.480192 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be772158-a71c-448d-8972-014f0d3a9ab8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "be772158-a71c-448d-8972-014f0d3a9ab8" (UID: "be772158-a71c-448d-8972-014f0d3a9ab8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.482486 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4205dfea-7dc7-496a-9745-fc5e3d0a418a-logs" (OuterVolumeSpecName: "logs") pod "4205dfea-7dc7-496a-9745-fc5e3d0a418a" (UID: "4205dfea-7dc7-496a-9745-fc5e3d0a418a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.483155 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e7e78641-77e6-4c89-b5c9-0d6f3c9a9343" (UID: "e7e78641-77e6-4c89-b5c9-0d6f3c9a9343"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.484247 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "e7e78641-77e6-4c89-b5c9-0d6f3c9a9343" (UID: "e7e78641-77e6-4c89-b5c9-0d6f3c9a9343"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.484417 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be772158-a71c-448d-8972-014f0d3a9ab8-kube-api-access-2vclj" (OuterVolumeSpecName: "kube-api-access-2vclj") pod "be772158-a71c-448d-8972-014f0d3a9ab8" (UID: "be772158-a71c-448d-8972-014f0d3a9ab8"). InnerVolumeSpecName "kube-api-access-2vclj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.484486 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-scripts" (OuterVolumeSpecName: "scripts") pod "e7e78641-77e6-4c89-b5c9-0d6f3c9a9343" (UID: "e7e78641-77e6-4c89-b5c9-0d6f3c9a9343"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.484561 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4205dfea-7dc7-496a-9745-fc5e3d0a418a-kube-api-access-fknks" (OuterVolumeSpecName: "kube-api-access-fknks") pod "4205dfea-7dc7-496a-9745-fc5e3d0a418a" (UID: "4205dfea-7dc7-496a-9745-fc5e3d0a418a"). InnerVolumeSpecName "kube-api-access-fknks". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.485278 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "766c2a26-46ea-41b2-ba0c-2101ec9477d5" (UID: "766c2a26-46ea-41b2-ba0c-2101ec9477d5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.485658 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "9a02b617-28a7-4262-a110-f1c71763ad19" (UID: "9a02b617-28a7-4262-a110-f1c71763ad19"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.491021 4772 scope.go:117] "RemoveContainer" containerID="2e743dfaa62b788cb68a4d553d64cf9affaf8ef6e4da1308fddf4dc259167b69" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.491093 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-kube-api-access-rpm5h" (OuterVolumeSpecName: "kube-api-access-rpm5h") pod "e7e78641-77e6-4c89-b5c9-0d6f3c9a9343" (UID: "e7e78641-77e6-4c89-b5c9-0d6f3c9a9343"). InnerVolumeSpecName "kube-api-access-rpm5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.506113 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e7e78641-77e6-4c89-b5c9-0d6f3c9a9343" (UID: "e7e78641-77e6-4c89-b5c9-0d6f3c9a9343"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.537048 4772 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.537085 4772 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.537094 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a02b617-28a7-4262-a110-f1c71763ad19-logs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.537103 4772 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.537112 4772 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/be772158-a71c-448d-8972-014f0d3a9ab8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.537120 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4205dfea-7dc7-496a-9745-fc5e3d0a418a-logs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.537128 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.537136 4772 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a02b617-28a7-4262-a110-f1c71763ad19-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.537145 4772 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.537161 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.537182 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqskw\" (UniqueName: \"kubernetes.io/projected/766c2a26-46ea-41b2-ba0c-2101ec9477d5-kube-api-access-sqskw\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.537192 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vclj\" (UniqueName: \"kubernetes.io/projected/be772158-a71c-448d-8972-014f0d3a9ab8-kube-api-access-2vclj\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.537200 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.537208 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.537216 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fknks\" (UniqueName: \"kubernetes.io/projected/4205dfea-7dc7-496a-9745-fc5e3d0a418a-kube-api-access-fknks\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.537224 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a02b617-28a7-4262-a110-f1c71763ad19-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.537231 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-logs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.537241 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghzrk\" (UniqueName: \"kubernetes.io/projected/9a02b617-28a7-4262-a110-f1c71763ad19-kube-api-access-ghzrk\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.537250 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpm5h\" (UniqueName: \"kubernetes.io/projected/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-kube-api-access-rpm5h\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.537266 4772 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.565520 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "766c2a26-46ea-41b2-ba0c-2101ec9477d5" (UID: "766c2a26-46ea-41b2-ba0c-2101ec9477d5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.567384 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "766c2a26-46ea-41b2-ba0c-2101ec9477d5" (UID: "766c2a26-46ea-41b2-ba0c-2101ec9477d5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.584257 4772 scope.go:117] "RemoveContainer" containerID="0d7ac15f647607d8d8b9ab55f639b5ec78749485b0e54cbc048e0727ed5dbce0" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.585410 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/766c2a26-46ea-41b2-ba0c-2101ec9477d5-logs" (OuterVolumeSpecName: "logs") pod "766c2a26-46ea-41b2-ba0c-2101ec9477d5" (UID: "766c2a26-46ea-41b2-ba0c-2101ec9477d5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.591141 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a02b617-28a7-4262-a110-f1c71763ad19-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9a02b617-28a7-4262-a110-f1c71763ad19" (UID: "9a02b617-28a7-4262-a110-f1c71763ad19"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.594434 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a02b617-28a7-4262-a110-f1c71763ad19-config-data" (OuterVolumeSpecName: "config-data") pod "9a02b617-28a7-4262-a110-f1c71763ad19" (UID: "9a02b617-28a7-4262-a110-f1c71763ad19"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.606566 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "be772158-a71c-448d-8972-014f0d3a9ab8" (UID: "be772158-a71c-448d-8972-014f0d3a9ab8"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.609214 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-config-data" (OuterVolumeSpecName: "config-data") pod "766c2a26-46ea-41b2-ba0c-2101ec9477d5" (UID: "766c2a26-46ea-41b2-ba0c-2101ec9477d5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.616159 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a02b617-28a7-4262-a110-f1c71763ad19-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a02b617-28a7-4262-a110-f1c71763ad19" (UID: "9a02b617-28a7-4262-a110-f1c71763ad19"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.622588 4772 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.624199 4772 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.639581 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a02b617-28a7-4262-a110-f1c71763ad19-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.639779 4772 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a02b617-28a7-4262-a110-f1c71763ad19-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.639895 4772 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.640228 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a02b617-28a7-4262-a110-f1c71763ad19-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.640327 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/766c2a26-46ea-41b2-ba0c-2101ec9477d5-logs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.640431 4772 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.640496 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.640561 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.640626 4772 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.640688 4772 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.643782 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-config-data" (OuterVolumeSpecName: "config-data") pod "be772158-a71c-448d-8972-014f0d3a9ab8" (UID: "be772158-a71c-448d-8972-014f0d3a9ab8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.691105 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-config-data" (OuterVolumeSpecName: "config-data") pod "4205dfea-7dc7-496a-9745-fc5e3d0a418a" (UID: "4205dfea-7dc7-496a-9745-fc5e3d0a418a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.696317 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.697285 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e7e78641-77e6-4c89-b5c9-0d6f3c9a9343","Type":"ContainerDied","Data":"64eb2d8855af54c245dc9d145df3ac0064c424271a5cf4af6c9815a1aa8bc16e"} Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.701511 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "be772158-a71c-448d-8972-014f0d3a9ab8" (UID: "be772158-a71c-448d-8972-014f0d3a9ab8"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.705912 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4205dfea-7dc7-496a-9745-fc5e3d0a418a" (UID: "4205dfea-7dc7-496a-9745-fc5e3d0a418a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.706030 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be772158-a71c-448d-8972-014f0d3a9ab8" (UID: "be772158-a71c-448d-8972-014f0d3a9ab8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.706067 4772 generic.go:334] "Generic (PLEG): container finished" podID="f63bf600-ff03-43a3-92b4-fe8ac68a9bb7" containerID="db38347574e8ea3471da74617b5c2b8fd8e23430f530dbd434f5aba2a153f9bb" exitCode=0 Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.706437 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7","Type":"ContainerDied","Data":"db38347574e8ea3471da74617b5c2b8fd8e23430f530dbd434f5aba2a153f9bb"} Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.706829 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e7e78641-77e6-4c89-b5c9-0d6f3c9a9343" (UID: "e7e78641-77e6-4c89-b5c9-0d6f3c9a9343"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.708960 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4205dfea-7dc7-496a-9745-fc5e3d0a418a" (UID: "4205dfea-7dc7-496a-9745-fc5e3d0a418a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.711685 4772 generic.go:334] "Generic (PLEG): container finished" podID="aea5ee43-36e3-437d-8aca-b2faedd87c5b" containerID="0447c2ea1d147e4cee27fce146e4edc38d746774dc492452f5da3c48df7973bb" exitCode=0 Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.711728 4772 generic.go:334] "Generic (PLEG): container finished" podID="aea5ee43-36e3-437d-8aca-b2faedd87c5b" containerID="81bb10c06283521cef14702be02bc4e89a7f82e4ae6c7d56b76d0d05f92797d0" exitCode=0 Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.711739 4772 generic.go:334] "Generic (PLEG): container finished" podID="aea5ee43-36e3-437d-8aca-b2faedd87c5b" containerID="a4293d3cbd138216987430f5dab62fa26e55c56743eee0b42dd4fc7797a52afd" exitCode=0 Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.711781 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aea5ee43-36e3-437d-8aca-b2faedd87c5b","Type":"ContainerDied","Data":"0447c2ea1d147e4cee27fce146e4edc38d746774dc492452f5da3c48df7973bb"} Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.711880 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aea5ee43-36e3-437d-8aca-b2faedd87c5b","Type":"ContainerDied","Data":"81bb10c06283521cef14702be02bc4e89a7f82e4ae6c7d56b76d0d05f92797d0"} Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.711899 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aea5ee43-36e3-437d-8aca-b2faedd87c5b","Type":"ContainerDied","Data":"a4293d3cbd138216987430f5dab62fa26e55c56743eee0b42dd4fc7797a52afd"} Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.714473 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"be772158-a71c-448d-8972-014f0d3a9ab8","Type":"ContainerDied","Data":"2e68d940e0eebbc1216da3357187ae70827b7d508fb0a26f0e91d9593aac8852"} Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.714614 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.718744 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-597699949b-q6msx" event={"ID":"4205dfea-7dc7-496a-9745-fc5e3d0a418a","Type":"ContainerDied","Data":"77708c49aaa66488bf09da947ac24b469a4cd3c49071689cbd09cfa6aa9b79b5"} Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.726332 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-597699949b-q6msx" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.731685 4772 generic.go:334] "Generic (PLEG): container finished" podID="93c8f9a4-c6ef-42b8-8543-ff8b5347977e" containerID="abfb528c89657cd0985ff90de17dace11a1be4c50ae49dc95a4a7ec03d093246" exitCode=0 Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.731772 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"93c8f9a4-c6ef-42b8-8543-ff8b5347977e","Type":"ContainerDied","Data":"abfb528c89657cd0985ff90de17dace11a1be4c50ae49dc95a4a7ec03d093246"} Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.731810 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"93c8f9a4-c6ef-42b8-8543-ff8b5347977e","Type":"ContainerDied","Data":"dfddffa6f559c177ea99d7f7fef5a8fb81a5dc7c7f2005faaf77278166e23279"} Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.731912 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.748981 4772 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.749015 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.749027 4772 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.749038 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.749047 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.749057 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.749067 4772 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be772158-a71c-448d-8972-014f0d3a9ab8-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.750753 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"21f54218-5889-4ae9-a7a1-7ed4895ad63c","Type":"ContainerDied","Data":"cf3bc864ff0528c25cfa09a147802c26a644517c099a85fc5bafd7c4da9534c3"} Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.750880 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.757719 4772 generic.go:334] "Generic (PLEG): container finished" podID="a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66" containerID="faf687181014b14838de86572705cbe5952bdabca1b3fad7e35afc3ce6238c0f" exitCode=0 Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.757796 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66","Type":"ContainerDied","Data":"faf687181014b14838de86572705cbe5952bdabca1b3fad7e35afc3ce6238c0f"} Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.763320 4772 generic.go:334] "Generic (PLEG): container finished" podID="683f458e-44e9-49ea-a66b-4ac91a3f2bc1" containerID="3e806373a2604b5465de7a3913d6865c82f0689bac61f26c430950d7d4efb948" exitCode=0 Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.763385 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"683f458e-44e9-49ea-a66b-4ac91a3f2bc1","Type":"ContainerDied","Data":"3e806373a2604b5465de7a3913d6865c82f0689bac61f26c430950d7d4efb948"} Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.767917 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="76fdbdb1-d48a-4cd1-8372-78887671dce8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: connect: connection refused" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.768427 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9a02b617-28a7-4262-a110-f1c71763ad19","Type":"ContainerDied","Data":"1536a68238e83bb2c89cfe9a0fce1841bc4d60d2a518fdc49dc1b005d27a6470"} Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.768551 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.774620 4772 generic.go:334] "Generic (PLEG): container finished" podID="766c2a26-46ea-41b2-ba0c-2101ec9477d5" containerID="ffbe05081a83d720881627c45e3d405aaf574d1db8fa63481da9c229023c0874" exitCode=0 Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.774727 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-659485ddbb-5bnzg" event={"ID":"766c2a26-46ea-41b2-ba0c-2101ec9477d5","Type":"ContainerDied","Data":"ffbe05081a83d720881627c45e3d405aaf574d1db8fa63481da9c229023c0874"} Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.774756 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-659485ddbb-5bnzg" event={"ID":"766c2a26-46ea-41b2-ba0c-2101ec9477d5","Type":"ContainerDied","Data":"181aa9237802812b703a88787d1d6892177f6147a0214d407241520c82b45857"} Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.774729 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-659485ddbb-5bnzg" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.779272 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "766c2a26-46ea-41b2-ba0c-2101ec9477d5" (UID: "766c2a26-46ea-41b2-ba0c-2101ec9477d5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.782159 4772 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/root-account-create-update-qmppl" secret="" err="secret \"galera-openstack-dockercfg-4dfv4\" not found" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.782221 4772 scope.go:117] "RemoveContainer" containerID="9005da10eaad68221a5ab75b0d10da02a46a7bd38d46bece0339dd56d8e2fc51" Jan 27 15:31:33 crc kubenswrapper[4772]: E0127 15:31:33.782433 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CrashLoopBackOff: \"back-off 10s restarting failed container=mariadb-account-create-update pod=root-account-create-update-qmppl_openstack(4cbf7469-816d-4e54-a7ad-b5b76d0d59d6)\"" pod="openstack/root-account-create-update-qmppl" podUID="4cbf7469-816d-4e54-a7ad-b5b76d0d59d6" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.796592 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4c0e-account-create-update-wlbfm" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.799993 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4205dfea-7dc7-496a-9745-fc5e3d0a418a" (UID: "4205dfea-7dc7-496a-9745-fc5e3d0a418a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.828616 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-config-data" (OuterVolumeSpecName: "config-data") pod "e7e78641-77e6-4c89-b5c9-0d6f3c9a9343" (UID: "e7e78641-77e6-4c89-b5c9-0d6f3c9a9343"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:33 crc kubenswrapper[4772]: E0127 15:31:33.842580 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f351431c9793a13f48f307e65178046dd4ccdc52ebd7ba269a580599ff0da01a" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Jan 27 15:31:33 crc kubenswrapper[4772]: E0127 15:31:33.843761 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f351431c9793a13f48f307e65178046dd4ccdc52ebd7ba269a580599ff0da01a" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Jan 27 15:31:33 crc kubenswrapper[4772]: E0127 15:31:33.847567 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f351431c9793a13f48f307e65178046dd4ccdc52ebd7ba269a580599ff0da01a" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Jan 27 15:31:33 crc kubenswrapper[4772]: E0127 15:31:33.847642 4772 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="8b8101bc-2ddf-48ed-9b92-e8f9e5e71938" containerName="ovn-northd" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.851817 4772 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4205dfea-7dc7-496a-9745-fc5e3d0a418a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.851853 4772 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/766c2a26-46ea-41b2-ba0c-2101ec9477d5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.851868 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.946217 4772 scope.go:117] "RemoveContainer" containerID="2e74d40bce110215c2607c9a9b716bf9d9db61e446fe99bb897518adbdc86d83" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.969188 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4c0e-account-create-update-wlbfm" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.983621 4772 scope.go:117] "RemoveContainer" containerID="6481b50eed7f8997cc197c4b50a1b5d1b9aa395b3745aa30ff2d6ee451d23215" Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.990681 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 27 15:31:33 crc kubenswrapper[4772]: I0127 15:31:33.995623 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.007641 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.013456 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.026348 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.049791 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.052405 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.057182 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4llv\" (UniqueName: \"kubernetes.io/projected/05445a27-d839-4a60-8338-5ee5f2c3f9d7-kube-api-access-z4llv\") pod \"keystone-4c0e-account-create-update-wlbfm\" (UID: \"05445a27-d839-4a60-8338-5ee5f2c3f9d7\") " pod="openstack/keystone-4c0e-account-create-update-wlbfm" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.057408 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05445a27-d839-4a60-8338-5ee5f2c3f9d7-operator-scripts\") pod \"keystone-4c0e-account-create-update-wlbfm\" (UID: \"05445a27-d839-4a60-8338-5ee5f2c3f9d7\") " pod="openstack/keystone-4c0e-account-create-update-wlbfm" Jan 27 15:31:34 crc kubenswrapper[4772]: E0127 15:31:34.057801 4772 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Jan 27 15:31:34 crc kubenswrapper[4772]: E0127 15:31:34.057874 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/05445a27-d839-4a60-8338-5ee5f2c3f9d7-operator-scripts podName:05445a27-d839-4a60-8338-5ee5f2c3f9d7 nodeName:}" failed. No retries permitted until 2026-01-27 15:31:36.057855095 +0000 UTC m=+1482.038464193 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/05445a27-d839-4a60-8338-5ee5f2c3f9d7-operator-scripts") pod "keystone-4c0e-account-create-update-wlbfm" (UID: "05445a27-d839-4a60-8338-5ee5f2c3f9d7") : configmap "openstack-scripts" not found Jan 27 15:31:34 crc kubenswrapper[4772]: E0127 15:31:34.058260 4772 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Jan 27 15:31:34 crc kubenswrapper[4772]: E0127 15:31:34.058292 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4cbf7469-816d-4e54-a7ad-b5b76d0d59d6-operator-scripts podName:4cbf7469-816d-4e54-a7ad-b5b76d0d59d6 nodeName:}" failed. No retries permitted until 2026-01-27 15:31:36.058281808 +0000 UTC m=+1482.038890906 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4cbf7469-816d-4e54-a7ad-b5b76d0d59d6-operator-scripts") pod "root-account-create-update-qmppl" (UID: "4cbf7469-816d-4e54-a7ad-b5b76d0d59d6") : configmap "openstack-scripts" not found Jan 27 15:31:34 crc kubenswrapper[4772]: E0127 15:31:34.061570 4772 projected.go:194] Error preparing data for projected volume kube-api-access-z4llv for pod openstack/keystone-4c0e-account-create-update-wlbfm: failed to fetch token: serviceaccounts "galera-openstack" not found Jan 27 15:31:34 crc kubenswrapper[4772]: E0127 15:31:34.061637 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/05445a27-d839-4a60-8338-5ee5f2c3f9d7-kube-api-access-z4llv podName:05445a27-d839-4a60-8338-5ee5f2c3f9d7 nodeName:}" failed. No retries permitted until 2026-01-27 15:31:36.061617904 +0000 UTC m=+1482.042227002 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-z4llv" (UniqueName: "kubernetes.io/projected/05445a27-d839-4a60-8338-5ee5f2c3f9d7-kube-api-access-z4llv") pod "keystone-4c0e-account-create-update-wlbfm" (UID: "05445a27-d839-4a60-8338-5ee5f2c3f9d7") : failed to fetch token: serviceaccounts "galera-openstack" not found Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.061827 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.067874 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.082520 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.088723 4772 scope.go:117] "RemoveContainer" containerID="3454f9899adaff309b52934e71697924735c1f269fb473444cba03b5baf4e1e5" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.114869 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-597699949b-q6msx"] Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.116341 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-597699949b-q6msx"] Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.120903 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.130887 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.141106 4772 scope.go:117] "RemoveContainer" containerID="c47159ab0aee5087f5a44073988d2ad8d6aaaa0e47ba7702dc2a03eab229b375" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.152049 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-659485ddbb-5bnzg"] Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.158102 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-659485ddbb-5bnzg"] Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.158646 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-kolla-config\") pod \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\" (UID: \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\") " Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.158688 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-nova-metadata-tls-certs\") pod \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\" (UID: \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\") " Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.158741 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-config-data\") pod \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\" (UID: \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\") " Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.158810 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6b44p\" (UniqueName: \"kubernetes.io/projected/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-kube-api-access-6b44p\") pod \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\" (UID: \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\") " Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.158836 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-logs\") pod \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\" (UID: \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\") " Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.158861 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-combined-ca-bundle\") pod \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\" (UID: \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\") " Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.158899 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r75qc\" (UniqueName: \"kubernetes.io/projected/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-kube-api-access-r75qc\") pod \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\" (UID: \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\") " Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.158952 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-memcached-tls-certs\") pod \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\" (UID: \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\") " Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.159000 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-config-data\") pod \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\" (UID: \"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7\") " Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.159033 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-combined-ca-bundle\") pod \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\" (UID: \"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66\") " Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.159230 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-logs" (OuterVolumeSpecName: "logs") pod "f63bf600-ff03-43a3-92b4-fe8ac68a9bb7" (UID: "f63bf600-ff03-43a3-92b4-fe8ac68a9bb7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.159408 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-logs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.159672 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-config-data" (OuterVolumeSpecName: "config-data") pod "a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66" (UID: "a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.160357 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66" (UID: "a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.164051 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-kube-api-access-r75qc" (OuterVolumeSpecName: "kube-api-access-r75qc") pod "a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66" (UID: "a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66"). InnerVolumeSpecName "kube-api-access-r75qc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.181471 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-kube-api-access-6b44p" (OuterVolumeSpecName: "kube-api-access-6b44p") pod "f63bf600-ff03-43a3-92b4-fe8ac68a9bb7" (UID: "f63bf600-ff03-43a3-92b4-fe8ac68a9bb7"). InnerVolumeSpecName "kube-api-access-6b44p". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.202410 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-config-data" (OuterVolumeSpecName: "config-data") pod "f63bf600-ff03-43a3-92b4-fe8ac68a9bb7" (UID: "f63bf600-ff03-43a3-92b4-fe8ac68a9bb7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.208512 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f63bf600-ff03-43a3-92b4-fe8ac68a9bb7" (UID: "f63bf600-ff03-43a3-92b4-fe8ac68a9bb7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.209279 4772 scope.go:117] "RemoveContainer" containerID="26cc6d1f580535edc969fb0f7d0d2e7d716fa8450f944ca1657554f90801529b" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.209659 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66" (UID: "a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.226271 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66" (UID: "a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.239001 4772 scope.go:117] "RemoveContainer" containerID="ad26ca4835a223df0b0aa3065e02d9e54b67030d2b6d0436f1f1a0dd7bf06415" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.250346 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "f63bf600-ff03-43a3-92b4-fe8ac68a9bb7" (UID: "f63bf600-ff03-43a3-92b4-fe8ac68a9bb7"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.260573 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.260612 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6b44p\" (UniqueName: \"kubernetes.io/projected/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-kube-api-access-6b44p\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.260626 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.260639 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r75qc\" (UniqueName: \"kubernetes.io/projected/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-kube-api-access-r75qc\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.260650 4772 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.260662 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.260672 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.260682 4772 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66-kolla-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.260691 4772 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.275274 4772 scope.go:117] "RemoveContainer" containerID="f10ed54f4ea68e56be83b8d8387a9768612b5c035b1fc42928132066af5bd689" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.300675 4772 scope.go:117] "RemoveContainer" containerID="abfb528c89657cd0985ff90de17dace11a1be4c50ae49dc95a4a7ec03d093246" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.352640 4772 scope.go:117] "RemoveContainer" containerID="db0ed28f713318389578164d5ba7364ff5e0ca569d4bd32de0483eb615fe7597" Jan 27 15:31:34 crc kubenswrapper[4772]: E0127 15:31:34.373286 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="788384a3ae6b89b3eeabbb3fe7578f4cb514172f7c7e0c341ec2b75ed4d75a29" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Jan 27 15:31:34 crc kubenswrapper[4772]: E0127 15:31:34.383669 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="788384a3ae6b89b3eeabbb3fe7578f4cb514172f7c7e0c341ec2b75ed4d75a29" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Jan 27 15:31:34 crc kubenswrapper[4772]: E0127 15:31:34.384989 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="788384a3ae6b89b3eeabbb3fe7578f4cb514172f7c7e0c341ec2b75ed4d75a29" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Jan 27 15:31:34 crc kubenswrapper[4772]: E0127 15:31:34.385014 4772 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="dbbd3c83-3fde-4b11-8ef0-add837d393ce" containerName="nova-cell1-conductor-conductor" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.400871 4772 scope.go:117] "RemoveContainer" containerID="abfb528c89657cd0985ff90de17dace11a1be4c50ae49dc95a4a7ec03d093246" Jan 27 15:31:34 crc kubenswrapper[4772]: E0127 15:31:34.401297 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abfb528c89657cd0985ff90de17dace11a1be4c50ae49dc95a4a7ec03d093246\": container with ID starting with abfb528c89657cd0985ff90de17dace11a1be4c50ae49dc95a4a7ec03d093246 not found: ID does not exist" containerID="abfb528c89657cd0985ff90de17dace11a1be4c50ae49dc95a4a7ec03d093246" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.401322 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abfb528c89657cd0985ff90de17dace11a1be4c50ae49dc95a4a7ec03d093246"} err="failed to get container status \"abfb528c89657cd0985ff90de17dace11a1be4c50ae49dc95a4a7ec03d093246\": rpc error: code = NotFound desc = could not find container \"abfb528c89657cd0985ff90de17dace11a1be4c50ae49dc95a4a7ec03d093246\": container with ID starting with abfb528c89657cd0985ff90de17dace11a1be4c50ae49dc95a4a7ec03d093246 not found: ID does not exist" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.401346 4772 scope.go:117] "RemoveContainer" containerID="db0ed28f713318389578164d5ba7364ff5e0ca569d4bd32de0483eb615fe7597" Jan 27 15:31:34 crc kubenswrapper[4772]: E0127 15:31:34.401583 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db0ed28f713318389578164d5ba7364ff5e0ca569d4bd32de0483eb615fe7597\": container with ID starting with db0ed28f713318389578164d5ba7364ff5e0ca569d4bd32de0483eb615fe7597 not found: ID does not exist" containerID="db0ed28f713318389578164d5ba7364ff5e0ca569d4bd32de0483eb615fe7597" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.401601 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db0ed28f713318389578164d5ba7364ff5e0ca569d4bd32de0483eb615fe7597"} err="failed to get container status \"db0ed28f713318389578164d5ba7364ff5e0ca569d4bd32de0483eb615fe7597\": rpc error: code = NotFound desc = could not find container \"db0ed28f713318389578164d5ba7364ff5e0ca569d4bd32de0483eb615fe7597\": container with ID starting with db0ed28f713318389578164d5ba7364ff5e0ca569d4bd32de0483eb615fe7597 not found: ID does not exist" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.401615 4772 scope.go:117] "RemoveContainer" containerID="670d5287e2a9882bc2137122191964eb76c57b36df9c904f50db621c1141ab98" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.434723 4772 scope.go:117] "RemoveContainer" containerID="3114715e24bc63a93ce31ec7ec2cc2fdeaad0a6c7647de22f23d06ac45e3d864" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.486213 4772 scope.go:117] "RemoveContainer" containerID="d767e789b4befb7b8caac693075691222c00bb6ae1189417345706dad41621f9" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.534211 4772 scope.go:117] "RemoveContainer" containerID="ffbe05081a83d720881627c45e3d405aaf574d1db8fa63481da9c229023c0874" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.613143 4772 scope.go:117] "RemoveContainer" containerID="23481794981b6875427087c492230b72f248918903ae28fa47bb73190cfa8573" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.680064 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21f54218-5889-4ae9-a7a1-7ed4895ad63c" path="/var/lib/kubelet/pods/21f54218-5889-4ae9-a7a1-7ed4895ad63c/volumes" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.680828 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4205dfea-7dc7-496a-9745-fc5e3d0a418a" path="/var/lib/kubelet/pods/4205dfea-7dc7-496a-9745-fc5e3d0a418a/volumes" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.681543 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e69643a-e8c2-4057-a993-d5506ceeec1b" path="/var/lib/kubelet/pods/5e69643a-e8c2-4057-a993-d5506ceeec1b/volumes" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.682823 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="766c2a26-46ea-41b2-ba0c-2101ec9477d5" path="/var/lib/kubelet/pods/766c2a26-46ea-41b2-ba0c-2101ec9477d5/volumes" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.683612 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93c8f9a4-c6ef-42b8-8543-ff8b5347977e" path="/var/lib/kubelet/pods/93c8f9a4-c6ef-42b8-8543-ff8b5347977e/volumes" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.684288 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a02b617-28a7-4262-a110-f1c71763ad19" path="/var/lib/kubelet/pods/9a02b617-28a7-4262-a110-f1c71763ad19/volumes" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.685651 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b20b9215-5398-4100-bac4-763daa5ed222" path="/var/lib/kubelet/pods/b20b9215-5398-4100-bac4-763daa5ed222/volumes" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.686135 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be772158-a71c-448d-8972-014f0d3a9ab8" path="/var/lib/kubelet/pods/be772158-a71c-448d-8972-014f0d3a9ab8/volumes" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.686765 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c16a29a0-7238-4a5e-b892-8f5195a1f486" path="/var/lib/kubelet/pods/c16a29a0-7238-4a5e-b892-8f5195a1f486/volumes" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.687890 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf619242-7348-4de4-a37e-8ebdc4ca54d7" path="/var/lib/kubelet/pods/cf619242-7348-4de4-a37e-8ebdc4ca54d7/volumes" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.688695 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e78641-77e6-4c89-b5c9-0d6f3c9a9343" path="/var/lib/kubelet/pods/e7e78641-77e6-4c89-b5c9-0d6f3c9a9343/volumes" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.690082 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef060591-3809-4f0b-974f-0785261db9b9" path="/var/lib/kubelet/pods/ef060591-3809-4f0b-974f-0785261db9b9/volumes" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.714610 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.721420 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.730843 4772 scope.go:117] "RemoveContainer" containerID="ffbe05081a83d720881627c45e3d405aaf574d1db8fa63481da9c229023c0874" Jan 27 15:31:34 crc kubenswrapper[4772]: E0127 15:31:34.731701 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffbe05081a83d720881627c45e3d405aaf574d1db8fa63481da9c229023c0874\": container with ID starting with ffbe05081a83d720881627c45e3d405aaf574d1db8fa63481da9c229023c0874 not found: ID does not exist" containerID="ffbe05081a83d720881627c45e3d405aaf574d1db8fa63481da9c229023c0874" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.731759 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffbe05081a83d720881627c45e3d405aaf574d1db8fa63481da9c229023c0874"} err="failed to get container status \"ffbe05081a83d720881627c45e3d405aaf574d1db8fa63481da9c229023c0874\": rpc error: code = NotFound desc = could not find container \"ffbe05081a83d720881627c45e3d405aaf574d1db8fa63481da9c229023c0874\": container with ID starting with ffbe05081a83d720881627c45e3d405aaf574d1db8fa63481da9c229023c0874 not found: ID does not exist" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.731779 4772 scope.go:117] "RemoveContainer" containerID="23481794981b6875427087c492230b72f248918903ae28fa47bb73190cfa8573" Jan 27 15:31:34 crc kubenswrapper[4772]: E0127 15:31:34.733409 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23481794981b6875427087c492230b72f248918903ae28fa47bb73190cfa8573\": container with ID starting with 23481794981b6875427087c492230b72f248918903ae28fa47bb73190cfa8573 not found: ID does not exist" containerID="23481794981b6875427087c492230b72f248918903ae28fa47bb73190cfa8573" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.733428 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23481794981b6875427087c492230b72f248918903ae28fa47bb73190cfa8573"} err="failed to get container status \"23481794981b6875427087c492230b72f248918903ae28fa47bb73190cfa8573\": rpc error: code = NotFound desc = could not find container \"23481794981b6875427087c492230b72f248918903ae28fa47bb73190cfa8573\": container with ID starting with 23481794981b6875427087c492230b72f248918903ae28fa47bb73190cfa8573 not found: ID does not exist" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.733470 4772 scope.go:117] "RemoveContainer" containerID="2e74d40bce110215c2607c9a9b716bf9d9db61e446fe99bb897518adbdc86d83" Jan 27 15:31:34 crc kubenswrapper[4772]: E0127 15:31:34.733749 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e74d40bce110215c2607c9a9b716bf9d9db61e446fe99bb897518adbdc86d83\": container with ID starting with 2e74d40bce110215c2607c9a9b716bf9d9db61e446fe99bb897518adbdc86d83 not found: ID does not exist" containerID="2e74d40bce110215c2607c9a9b716bf9d9db61e446fe99bb897518adbdc86d83" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.733794 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e74d40bce110215c2607c9a9b716bf9d9db61e446fe99bb897518adbdc86d83"} err="failed to get container status \"2e74d40bce110215c2607c9a9b716bf9d9db61e446fe99bb897518adbdc86d83\": rpc error: code = NotFound desc = could not find container \"2e74d40bce110215c2607c9a9b716bf9d9db61e446fe99bb897518adbdc86d83\": container with ID starting with 2e74d40bce110215c2607c9a9b716bf9d9db61e446fe99bb897518adbdc86d83 not found: ID does not exist" Jan 27 15:31:34 crc kubenswrapper[4772]: E0127 15:31:34.774958 4772 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Jan 27 15:31:34 crc kubenswrapper[4772]: E0127 15:31:34.775051 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-config-data podName:76fdbdb1-d48a-4cd1-8372-78887671dce8 nodeName:}" failed. No retries permitted until 2026-01-27 15:31:42.775031751 +0000 UTC m=+1488.755640859 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-config-data") pod "rabbitmq-cell1-server-0" (UID: "76fdbdb1-d48a-4cd1-8372-78887671dce8") : configmap "rabbitmq-cell1-config-data" not found Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.832815 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"683f458e-44e9-49ea-a66b-4ac91a3f2bc1","Type":"ContainerDied","Data":"bb200c044803c6c5491d60dc192f271f4cdf0adcf18a5f0f12ab40acb77fdf72"} Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.832864 4772 scope.go:117] "RemoveContainer" containerID="112ddc6068b3694383f83c1ffece42788a7623920d1c02ff9f46202f7c8c0d7e" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.832994 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.837653 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8b8101bc-2ddf-48ed-9b92-e8f9e5e71938/ovn-northd/0.log" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.837688 4772 generic.go:334] "Generic (PLEG): container finished" podID="8b8101bc-2ddf-48ed-9b92-e8f9e5e71938" containerID="f351431c9793a13f48f307e65178046dd4ccdc52ebd7ba269a580599ff0da01a" exitCode=139 Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.837729 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938","Type":"ContainerDied","Data":"f351431c9793a13f48f307e65178046dd4ccdc52ebd7ba269a580599ff0da01a"} Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.841708 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.841722 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f63bf600-ff03-43a3-92b4-fe8ac68a9bb7","Type":"ContainerDied","Data":"a940184dde4998665ff3925c8d268f050f05912d1265137578965cd151d251c3"} Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.857069 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66","Type":"ContainerDied","Data":"022c0f29ec3ec9ea31194094e372dfed87fe074f880cb471419a54885eeba246"} Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.857300 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.864930 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.865451 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aea5ee43-36e3-437d-8aca-b2faedd87c5b","Type":"ContainerDied","Data":"0d9e1d64ee2212bcbce9b483a76517d64478f416567bc79c87cd9fc874d3b4e1"} Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.865622 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-4c0e-account-create-update-wlbfm" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.879593 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72jk7\" (UniqueName: \"kubernetes.io/projected/aea5ee43-36e3-437d-8aca-b2faedd87c5b-kube-api-access-72jk7\") pod \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.879848 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-config-data\") pod \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\" (UID: \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\") " Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.879935 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-scripts\") pod \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.880016 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aea5ee43-36e3-437d-8aca-b2faedd87c5b-run-httpd\") pod \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.880109 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-config-data\") pod \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.880343 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aea5ee43-36e3-437d-8aca-b2faedd87c5b-log-httpd\") pod \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.880456 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-sg-core-conf-yaml\") pod \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.880653 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-config-data-custom\") pod \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\" (UID: \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\") " Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.880803 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-etc-machine-id\") pod \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\" (UID: \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\") " Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.880895 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-ceilometer-tls-certs\") pod \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.880984 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-combined-ca-bundle\") pod \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\" (UID: \"aea5ee43-36e3-437d-8aca-b2faedd87c5b\") " Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.881063 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrz6x\" (UniqueName: \"kubernetes.io/projected/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-kube-api-access-rrz6x\") pod \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\" (UID: \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\") " Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.881133 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-scripts\") pod \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\" (UID: \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\") " Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.881237 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-combined-ca-bundle\") pod \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\" (UID: \"683f458e-44e9-49ea-a66b-4ac91a3f2bc1\") " Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.885304 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aea5ee43-36e3-437d-8aca-b2faedd87c5b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "aea5ee43-36e3-437d-8aca-b2faedd87c5b" (UID: "aea5ee43-36e3-437d-8aca-b2faedd87c5b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.886411 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "683f458e-44e9-49ea-a66b-4ac91a3f2bc1" (UID: "683f458e-44e9-49ea-a66b-4ac91a3f2bc1"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.891017 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aea5ee43-36e3-437d-8aca-b2faedd87c5b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "aea5ee43-36e3-437d-8aca-b2faedd87c5b" (UID: "aea5ee43-36e3-437d-8aca-b2faedd87c5b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.894859 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aea5ee43-36e3-437d-8aca-b2faedd87c5b-kube-api-access-72jk7" (OuterVolumeSpecName: "kube-api-access-72jk7") pod "aea5ee43-36e3-437d-8aca-b2faedd87c5b" (UID: "aea5ee43-36e3-437d-8aca-b2faedd87c5b"). InnerVolumeSpecName "kube-api-access-72jk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.894985 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-scripts" (OuterVolumeSpecName: "scripts") pod "683f458e-44e9-49ea-a66b-4ac91a3f2bc1" (UID: "683f458e-44e9-49ea-a66b-4ac91a3f2bc1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.895292 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "683f458e-44e9-49ea-a66b-4ac91a3f2bc1" (UID: "683f458e-44e9-49ea-a66b-4ac91a3f2bc1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.895516 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-kube-api-access-rrz6x" (OuterVolumeSpecName: "kube-api-access-rrz6x") pod "683f458e-44e9-49ea-a66b-4ac91a3f2bc1" (UID: "683f458e-44e9-49ea-a66b-4ac91a3f2bc1"). InnerVolumeSpecName "kube-api-access-rrz6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.897217 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-scripts" (OuterVolumeSpecName: "scripts") pod "aea5ee43-36e3-437d-8aca-b2faedd87c5b" (UID: "aea5ee43-36e3-437d-8aca-b2faedd87c5b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.908641 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8b8101bc-2ddf-48ed-9b92-e8f9e5e71938/ovn-northd/0.log" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.908714 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.948973 4772 scope.go:117] "RemoveContainer" containerID="3e806373a2604b5465de7a3913d6865c82f0689bac61f26c430950d7d4efb948" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.976495 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-4c0e-account-create-update-wlbfm"] Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.976622 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "683f458e-44e9-49ea-a66b-4ac91a3f2bc1" (UID: "683f458e-44e9-49ea-a66b-4ac91a3f2bc1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.987007 4772 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.987035 4772 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.987044 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrz6x\" (UniqueName: \"kubernetes.io/projected/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-kube-api-access-rrz6x\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.987053 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.987061 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.987087 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72jk7\" (UniqueName: \"kubernetes.io/projected/aea5ee43-36e3-437d-8aca-b2faedd87c5b-kube-api-access-72jk7\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.987097 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.987105 4772 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aea5ee43-36e3-437d-8aca-b2faedd87c5b-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.987113 4772 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aea5ee43-36e3-437d-8aca-b2faedd87c5b-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.987664 4772 scope.go:117] "RemoveContainer" containerID="db38347574e8ea3471da74617b5c2b8fd8e23430f530dbd434f5aba2a153f9bb" Jan 27 15:31:34 crc kubenswrapper[4772]: I0127 15:31:34.995685 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-4c0e-account-create-update-wlbfm"] Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.002386 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-config-data" (OuterVolumeSpecName: "config-data") pod "aea5ee43-36e3-437d-8aca-b2faedd87c5b" (UID: "aea5ee43-36e3-437d-8aca-b2faedd87c5b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.006229 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "aea5ee43-36e3-437d-8aca-b2faedd87c5b" (UID: "aea5ee43-36e3-437d-8aca-b2faedd87c5b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.008324 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.013868 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.019010 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.024818 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.027004 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "aea5ee43-36e3-437d-8aca-b2faedd87c5b" (UID: "aea5ee43-36e3-437d-8aca-b2faedd87c5b"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.028779 4772 scope.go:117] "RemoveContainer" containerID="7343cd6a2a5cf705b558b4cc862d749d392235682218489d0106143cb8a5d4bc" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.032099 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-config-data" (OuterVolumeSpecName: "config-data") pod "683f458e-44e9-49ea-a66b-4ac91a3f2bc1" (UID: "683f458e-44e9-49ea-a66b-4ac91a3f2bc1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:35 crc kubenswrapper[4772]: E0127 15:31:35.044662 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e92037085b98ccc46bbd64416f98018d2426a17d0883dd17b830d5574a8a0f4e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.047144 4772 scope.go:117] "RemoveContainer" containerID="faf687181014b14838de86572705cbe5952bdabca1b3fad7e35afc3ce6238c0f" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.066377 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aea5ee43-36e3-437d-8aca-b2faedd87c5b" (UID: "aea5ee43-36e3-437d-8aca-b2faedd87c5b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:35 crc kubenswrapper[4772]: E0127 15:31:35.071325 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e92037085b98ccc46bbd64416f98018d2426a17d0883dd17b830d5574a8a0f4e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 27 15:31:35 crc kubenswrapper[4772]: E0127 15:31:35.077893 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e92037085b98ccc46bbd64416f98018d2426a17d0883dd17b830d5574a8a0f4e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 27 15:31:35 crc kubenswrapper[4772]: E0127 15:31:35.077959 4772 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="b83f7578-8113-46c8-be24-5968aa0ca563" containerName="nova-scheduler-scheduler" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.087735 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-config\") pod \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.087781 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-429qb\" (UniqueName: \"kubernetes.io/projected/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-kube-api-access-429qb\") pod \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.087843 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-metrics-certs-tls-certs\") pod \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.087880 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-combined-ca-bundle\") pod \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.087911 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-ovn-rundir\") pod \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.088012 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-ovn-northd-tls-certs\") pod \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.088037 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-scripts\") pod \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\" (UID: \"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938\") " Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.088371 4772 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.088383 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.088394 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/683f458e-44e9-49ea-a66b-4ac91a3f2bc1-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.088403 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.088413 4772 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aea5ee43-36e3-437d-8aca-b2faedd87c5b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.088805 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-config" (OuterVolumeSpecName: "config") pod "8b8101bc-2ddf-48ed-9b92-e8f9e5e71938" (UID: "8b8101bc-2ddf-48ed-9b92-e8f9e5e71938"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.088990 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "8b8101bc-2ddf-48ed-9b92-e8f9e5e71938" (UID: "8b8101bc-2ddf-48ed-9b92-e8f9e5e71938"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.089020 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-scripts" (OuterVolumeSpecName: "scripts") pod "8b8101bc-2ddf-48ed-9b92-e8f9e5e71938" (UID: "8b8101bc-2ddf-48ed-9b92-e8f9e5e71938"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.090561 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-kube-api-access-429qb" (OuterVolumeSpecName: "kube-api-access-429qb") pod "8b8101bc-2ddf-48ed-9b92-e8f9e5e71938" (UID: "8b8101bc-2ddf-48ed-9b92-e8f9e5e71938"). InnerVolumeSpecName "kube-api-access-429qb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.092787 4772 scope.go:117] "RemoveContainer" containerID="0447c2ea1d147e4cee27fce146e4edc38d746774dc492452f5da3c48df7973bb" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.115386 4772 scope.go:117] "RemoveContainer" containerID="7b0085db2ce3021657d7773e88196b66b6759beeca3bff2b51fc3fdf5d6b4bd2" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.115859 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8b8101bc-2ddf-48ed-9b92-e8f9e5e71938" (UID: "8b8101bc-2ddf-48ed-9b92-e8f9e5e71938"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:35 crc kubenswrapper[4772]: E0127 15:31:35.122115 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="21613f2614f5809a9a792371fe2e685753a1a2fc6ea2f8fa7dcc2390d4bafda0" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Jan 27 15:31:35 crc kubenswrapper[4772]: E0127 15:31:35.124421 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="21613f2614f5809a9a792371fe2e685753a1a2fc6ea2f8fa7dcc2390d4bafda0" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Jan 27 15:31:35 crc kubenswrapper[4772]: E0127 15:31:35.128261 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="21613f2614f5809a9a792371fe2e685753a1a2fc6ea2f8fa7dcc2390d4bafda0" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Jan 27 15:31:35 crc kubenswrapper[4772]: E0127 15:31:35.128385 4772 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="b1515626-5d79-408d-abc1-cb92abd58f3f" containerName="galera" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.141784 4772 scope.go:117] "RemoveContainer" containerID="81bb10c06283521cef14702be02bc4e89a7f82e4ae6c7d56b76d0d05f92797d0" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.166299 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "8b8101bc-2ddf-48ed-9b92-e8f9e5e71938" (UID: "8b8101bc-2ddf-48ed-9b92-e8f9e5e71938"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.168541 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "8b8101bc-2ddf-48ed-9b92-e8f9e5e71938" (UID: "8b8101bc-2ddf-48ed-9b92-e8f9e5e71938"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.169474 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.184448 4772 scope.go:117] "RemoveContainer" containerID="a4293d3cbd138216987430f5dab62fa26e55c56743eee0b42dd4fc7797a52afd" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.188051 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.193674 4772 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.193715 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.193725 4772 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-ovn-rundir\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.193737 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05445a27-d839-4a60-8338-5ee5f2c3f9d7-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.193746 4772 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.193754 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.193762 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.193769 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4llv\" (UniqueName: \"kubernetes.io/projected/05445a27-d839-4a60-8338-5ee5f2c3f9d7-kube-api-access-z4llv\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.193779 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-429qb\" (UniqueName: \"kubernetes.io/projected/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938-kube-api-access-429qb\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.198938 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.205176 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.245065 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-qmppl" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.396472 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thp4z\" (UniqueName: \"kubernetes.io/projected/4cbf7469-816d-4e54-a7ad-b5b76d0d59d6-kube-api-access-thp4z\") pod \"4cbf7469-816d-4e54-a7ad-b5b76d0d59d6\" (UID: \"4cbf7469-816d-4e54-a7ad-b5b76d0d59d6\") " Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.396627 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cbf7469-816d-4e54-a7ad-b5b76d0d59d6-operator-scripts\") pod \"4cbf7469-816d-4e54-a7ad-b5b76d0d59d6\" (UID: \"4cbf7469-816d-4e54-a7ad-b5b76d0d59d6\") " Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.397422 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cbf7469-816d-4e54-a7ad-b5b76d0d59d6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4cbf7469-816d-4e54-a7ad-b5b76d0d59d6" (UID: "4cbf7469-816d-4e54-a7ad-b5b76d0d59d6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.401213 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cbf7469-816d-4e54-a7ad-b5b76d0d59d6-kube-api-access-thp4z" (OuterVolumeSpecName: "kube-api-access-thp4z") pod "4cbf7469-816d-4e54-a7ad-b5b76d0d59d6" (UID: "4cbf7469-816d-4e54-a7ad-b5b76d0d59d6"). InnerVolumeSpecName "kube-api-access-thp4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.498577 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thp4z\" (UniqueName: \"kubernetes.io/projected/4cbf7469-816d-4e54-a7ad-b5b76d0d59d6-kube-api-access-thp4z\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.498619 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cbf7469-816d-4e54-a7ad-b5b76d0d59d6-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:35 crc kubenswrapper[4772]: E0127 15:31:35.605452 4772 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Jan 27 15:31:35 crc kubenswrapper[4772]: E0127 15:31:35.606900 4772 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-config-data podName:508c3d5b-212a-46da-9a55-de3f35d7019b nodeName:}" failed. No retries permitted until 2026-01-27 15:31:43.60686792 +0000 UTC m=+1489.587477018 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-config-data") pod "rabbitmq-server-0" (UID: "508c3d5b-212a-46da-9a55-de3f35d7019b") : configmap "rabbitmq-config-data" not found Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.788213 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.895564 4772 generic.go:334] "Generic (PLEG): container finished" podID="76fdbdb1-d48a-4cd1-8372-78887671dce8" containerID="d8699d4e2fb6bcbb97c43048a20aeda8d17be226258e1acddb0364ab41c23e4a" exitCode=0 Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.895614 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"76fdbdb1-d48a-4cd1-8372-78887671dce8","Type":"ContainerDied","Data":"d8699d4e2fb6bcbb97c43048a20aeda8d17be226258e1acddb0364ab41c23e4a"} Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.895646 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.895688 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"76fdbdb1-d48a-4cd1-8372-78887671dce8","Type":"ContainerDied","Data":"09e6c8b66552c99b1f924df5f88d4156d8a5bb2bf8b6bbb8e0fc50cdfa96e1ad"} Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.895709 4772 scope.go:117] "RemoveContainer" containerID="d8699d4e2fb6bcbb97c43048a20aeda8d17be226258e1acddb0364ab41c23e4a" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.899123 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8b8101bc-2ddf-48ed-9b92-e8f9e5e71938/ovn-northd/0.log" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.899211 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"8b8101bc-2ddf-48ed-9b92-e8f9e5e71938","Type":"ContainerDied","Data":"383ce19f4879446a46975b1e3757ca75d5dbab13e103b56af11750ee3019f6bc"} Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.899297 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.909526 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-qmppl" event={"ID":"4cbf7469-816d-4e54-a7ad-b5b76d0d59d6","Type":"ContainerDied","Data":"d5deb0f3cfb55cc15d206b2ff6d6a2e4b5ccc8b9efe8608e2073fe3df0f8d559"} Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.909611 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-qmppl" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.915798 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/76fdbdb1-d48a-4cd1-8372-78887671dce8-pod-info\") pod \"76fdbdb1-d48a-4cd1-8372-78887671dce8\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.915914 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-config-data\") pod \"76fdbdb1-d48a-4cd1-8372-78887671dce8\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.915953 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-plugins\") pod \"76fdbdb1-d48a-4cd1-8372-78887671dce8\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.916009 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/76fdbdb1-d48a-4cd1-8372-78887671dce8-erlang-cookie-secret\") pod \"76fdbdb1-d48a-4cd1-8372-78887671dce8\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.916044 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-tls\") pod \"76fdbdb1-d48a-4cd1-8372-78887671dce8\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.916085 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-confd\") pod \"76fdbdb1-d48a-4cd1-8372-78887671dce8\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.916107 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-erlang-cookie\") pod \"76fdbdb1-d48a-4cd1-8372-78887671dce8\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.916126 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"76fdbdb1-d48a-4cd1-8372-78887671dce8\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.916146 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gbh9\" (UniqueName: \"kubernetes.io/projected/76fdbdb1-d48a-4cd1-8372-78887671dce8-kube-api-access-9gbh9\") pod \"76fdbdb1-d48a-4cd1-8372-78887671dce8\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.916200 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-server-conf\") pod \"76fdbdb1-d48a-4cd1-8372-78887671dce8\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.916217 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-plugins-conf\") pod \"76fdbdb1-d48a-4cd1-8372-78887671dce8\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.916902 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "76fdbdb1-d48a-4cd1-8372-78887671dce8" (UID: "76fdbdb1-d48a-4cd1-8372-78887671dce8"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.918387 4772 generic.go:334] "Generic (PLEG): container finished" podID="508c3d5b-212a-46da-9a55-de3f35d7019b" containerID="f002759dea4443f7600e0f76f24481c1604449a5ee31bd8aa53171a2121ec4b2" exitCode=0 Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.918466 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"508c3d5b-212a-46da-9a55-de3f35d7019b","Type":"ContainerDied","Data":"f002759dea4443f7600e0f76f24481c1604449a5ee31bd8aa53171a2121ec4b2"} Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.918538 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "76fdbdb1-d48a-4cd1-8372-78887671dce8" (UID: "76fdbdb1-d48a-4cd1-8372-78887671dce8"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.918911 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "76fdbdb1-d48a-4cd1-8372-78887671dce8" (UID: "76fdbdb1-d48a-4cd1-8372-78887671dce8"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.922268 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "76fdbdb1-d48a-4cd1-8372-78887671dce8" (UID: "76fdbdb1-d48a-4cd1-8372-78887671dce8"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.922367 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/76fdbdb1-d48a-4cd1-8372-78887671dce8-pod-info" (OuterVolumeSpecName: "pod-info") pod "76fdbdb1-d48a-4cd1-8372-78887671dce8" (UID: "76fdbdb1-d48a-4cd1-8372-78887671dce8"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.923033 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "76fdbdb1-d48a-4cd1-8372-78887671dce8" (UID: "76fdbdb1-d48a-4cd1-8372-78887671dce8"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.925251 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76fdbdb1-d48a-4cd1-8372-78887671dce8-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "76fdbdb1-d48a-4cd1-8372-78887671dce8" (UID: "76fdbdb1-d48a-4cd1-8372-78887671dce8"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.926304 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76fdbdb1-d48a-4cd1-8372-78887671dce8-kube-api-access-9gbh9" (OuterVolumeSpecName: "kube-api-access-9gbh9") pod "76fdbdb1-d48a-4cd1-8372-78887671dce8" (UID: "76fdbdb1-d48a-4cd1-8372-78887671dce8"). InnerVolumeSpecName "kube-api-access-9gbh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.939452 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-config-data" (OuterVolumeSpecName: "config-data") pod "76fdbdb1-d48a-4cd1-8372-78887671dce8" (UID: "76fdbdb1-d48a-4cd1-8372-78887671dce8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.961832 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-server-conf" (OuterVolumeSpecName: "server-conf") pod "76fdbdb1-d48a-4cd1-8372-78887671dce8" (UID: "76fdbdb1-d48a-4cd1-8372-78887671dce8"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.978134 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 27 15:31:35 crc kubenswrapper[4772]: I0127 15:31:35.986678 4772 scope.go:117] "RemoveContainer" containerID="d53d0dfba4b0af64ac6186cc8eb8efb21a1ec89a66c075c5a53ab1db9987e594" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.017611 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "76fdbdb1-d48a-4cd1-8372-78887671dce8" (UID: "76fdbdb1-d48a-4cd1-8372-78887671dce8"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.017683 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/508c3d5b-212a-46da-9a55-de3f35d7019b-rabbitmq-confd\") pod \"508c3d5b-212a-46da-9a55-de3f35d7019b\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.017727 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/508c3d5b-212a-46da-9a55-de3f35d7019b-rabbitmq-plugins\") pod \"508c3d5b-212a-46da-9a55-de3f35d7019b\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.017759 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/508c3d5b-212a-46da-9a55-de3f35d7019b-rabbitmq-erlang-cookie\") pod \"508c3d5b-212a-46da-9a55-de3f35d7019b\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.017787 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-server-conf\") pod \"508c3d5b-212a-46da-9a55-de3f35d7019b\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.017816 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-config-data\") pod \"508c3d5b-212a-46da-9a55-de3f35d7019b\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.017839 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8h8d\" (UniqueName: \"kubernetes.io/projected/508c3d5b-212a-46da-9a55-de3f35d7019b-kube-api-access-l8h8d\") pod \"508c3d5b-212a-46da-9a55-de3f35d7019b\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.017883 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/508c3d5b-212a-46da-9a55-de3f35d7019b-pod-info\") pod \"508c3d5b-212a-46da-9a55-de3f35d7019b\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.017912 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"508c3d5b-212a-46da-9a55-de3f35d7019b\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.017936 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-plugins-conf\") pod \"508c3d5b-212a-46da-9a55-de3f35d7019b\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.017974 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/508c3d5b-212a-46da-9a55-de3f35d7019b-erlang-cookie-secret\") pod \"508c3d5b-212a-46da-9a55-de3f35d7019b\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.018056 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-confd\") pod \"76fdbdb1-d48a-4cd1-8372-78887671dce8\" (UID: \"76fdbdb1-d48a-4cd1-8372-78887671dce8\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.018076 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/508c3d5b-212a-46da-9a55-de3f35d7019b-rabbitmq-tls\") pod \"508c3d5b-212a-46da-9a55-de3f35d7019b\" (UID: \"508c3d5b-212a-46da-9a55-de3f35d7019b\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.019791 4772 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.019825 4772 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/76fdbdb1-d48a-4cd1-8372-78887671dce8-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.019837 4772 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.019850 4772 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.019877 4772 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.019891 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gbh9\" (UniqueName: \"kubernetes.io/projected/76fdbdb1-d48a-4cd1-8372-78887671dce8-kube-api-access-9gbh9\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.019902 4772 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-server-conf\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.019913 4772 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-plugins-conf\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.019924 4772 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/76fdbdb1-d48a-4cd1-8372-78887671dce8-pod-info\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.019936 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76fdbdb1-d48a-4cd1-8372-78887671dce8-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.023402 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/508c3d5b-212a-46da-9a55-de3f35d7019b-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "508c3d5b-212a-46da-9a55-de3f35d7019b" (UID: "508c3d5b-212a-46da-9a55-de3f35d7019b"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.023885 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "508c3d5b-212a-46da-9a55-de3f35d7019b" (UID: "508c3d5b-212a-46da-9a55-de3f35d7019b"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.024867 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/508c3d5b-212a-46da-9a55-de3f35d7019b-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "508c3d5b-212a-46da-9a55-de3f35d7019b" (UID: "508c3d5b-212a-46da-9a55-de3f35d7019b"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: W0127 15:31:36.024943 4772 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/76fdbdb1-d48a-4cd1-8372-78887671dce8/volumes/kubernetes.io~projected/rabbitmq-confd Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.025000 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "76fdbdb1-d48a-4cd1-8372-78887671dce8" (UID: "76fdbdb1-d48a-4cd1-8372-78887671dce8"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.028391 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/508c3d5b-212a-46da-9a55-de3f35d7019b-kube-api-access-l8h8d" (OuterVolumeSpecName: "kube-api-access-l8h8d") pod "508c3d5b-212a-46da-9a55-de3f35d7019b" (UID: "508c3d5b-212a-46da-9a55-de3f35d7019b"). InnerVolumeSpecName "kube-api-access-l8h8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.028683 4772 scope.go:117] "RemoveContainer" containerID="d8699d4e2fb6bcbb97c43048a20aeda8d17be226258e1acddb0364ab41c23e4a" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.028475 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "persistence") pod "508c3d5b-212a-46da-9a55-de3f35d7019b" (UID: "508c3d5b-212a-46da-9a55-de3f35d7019b"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: E0127 15:31:36.029924 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8699d4e2fb6bcbb97c43048a20aeda8d17be226258e1acddb0364ab41c23e4a\": container with ID starting with d8699d4e2fb6bcbb97c43048a20aeda8d17be226258e1acddb0364ab41c23e4a not found: ID does not exist" containerID="d8699d4e2fb6bcbb97c43048a20aeda8d17be226258e1acddb0364ab41c23e4a" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.029987 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8699d4e2fb6bcbb97c43048a20aeda8d17be226258e1acddb0364ab41c23e4a"} err="failed to get container status \"d8699d4e2fb6bcbb97c43048a20aeda8d17be226258e1acddb0364ab41c23e4a\": rpc error: code = NotFound desc = could not find container \"d8699d4e2fb6bcbb97c43048a20aeda8d17be226258e1acddb0364ab41c23e4a\": container with ID starting with d8699d4e2fb6bcbb97c43048a20aeda8d17be226258e1acddb0364ab41c23e4a not found: ID does not exist" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.030014 4772 scope.go:117] "RemoveContainer" containerID="d53d0dfba4b0af64ac6186cc8eb8efb21a1ec89a66c075c5a53ab1db9987e594" Jan 27 15:31:36 crc kubenswrapper[4772]: E0127 15:31:36.030622 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d53d0dfba4b0af64ac6186cc8eb8efb21a1ec89a66c075c5a53ab1db9987e594\": container with ID starting with d53d0dfba4b0af64ac6186cc8eb8efb21a1ec89a66c075c5a53ab1db9987e594 not found: ID does not exist" containerID="d53d0dfba4b0af64ac6186cc8eb8efb21a1ec89a66c075c5a53ab1db9987e594" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.030652 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d53d0dfba4b0af64ac6186cc8eb8efb21a1ec89a66c075c5a53ab1db9987e594"} err="failed to get container status \"d53d0dfba4b0af64ac6186cc8eb8efb21a1ec89a66c075c5a53ab1db9987e594\": rpc error: code = NotFound desc = could not find container \"d53d0dfba4b0af64ac6186cc8eb8efb21a1ec89a66c075c5a53ab1db9987e594\": container with ID starting with d53d0dfba4b0af64ac6186cc8eb8efb21a1ec89a66c075c5a53ab1db9987e594 not found: ID does not exist" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.030680 4772 scope.go:117] "RemoveContainer" containerID="b1542ba131aec1cffd5520f2969b843d3aa12fe7b4cd60022addce3e73977b99" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.031857 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/508c3d5b-212a-46da-9a55-de3f35d7019b-pod-info" (OuterVolumeSpecName: "pod-info") pod "508c3d5b-212a-46da-9a55-de3f35d7019b" (UID: "508c3d5b-212a-46da-9a55-de3f35d7019b"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.033343 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/508c3d5b-212a-46da-9a55-de3f35d7019b-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "508c3d5b-212a-46da-9a55-de3f35d7019b" (UID: "508c3d5b-212a-46da-9a55-de3f35d7019b"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.033402 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/508c3d5b-212a-46da-9a55-de3f35d7019b-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "508c3d5b-212a-46da-9a55-de3f35d7019b" (UID: "508c3d5b-212a-46da-9a55-de3f35d7019b"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.040063 4772 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.044327 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-qmppl"] Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.049364 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-qmppl"] Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.049978 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-config-data" (OuterVolumeSpecName: "config-data") pod "508c3d5b-212a-46da-9a55-de3f35d7019b" (UID: "508c3d5b-212a-46da-9a55-de3f35d7019b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.054409 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.060007 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.070039 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-server-conf" (OuterVolumeSpecName: "server-conf") pod "508c3d5b-212a-46da-9a55-de3f35d7019b" (UID: "508c3d5b-212a-46da-9a55-de3f35d7019b"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.092062 4772 scope.go:117] "RemoveContainer" containerID="f351431c9793a13f48f307e65178046dd4ccdc52ebd7ba269a580599ff0da01a" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.115587 4772 scope.go:117] "RemoveContainer" containerID="9005da10eaad68221a5ab75b0d10da02a46a7bd38d46bece0339dd56d8e2fc51" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.116438 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/508c3d5b-212a-46da-9a55-de3f35d7019b-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "508c3d5b-212a-46da-9a55-de3f35d7019b" (UID: "508c3d5b-212a-46da-9a55-de3f35d7019b"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.121709 4772 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/508c3d5b-212a-46da-9a55-de3f35d7019b-pod-info\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.121767 4772 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.121781 4772 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-plugins-conf\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.121797 4772 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/508c3d5b-212a-46da-9a55-de3f35d7019b-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.121813 4772 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/76fdbdb1-d48a-4cd1-8372-78887671dce8-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.121825 4772 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/508c3d5b-212a-46da-9a55-de3f35d7019b-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.121901 4772 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.121962 4772 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/508c3d5b-212a-46da-9a55-de3f35d7019b-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.121990 4772 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/508c3d5b-212a-46da-9a55-de3f35d7019b-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.123883 4772 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/508c3d5b-212a-46da-9a55-de3f35d7019b-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.123896 4772 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-server-conf\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.123907 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/508c3d5b-212a-46da-9a55-de3f35d7019b-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.123919 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8h8d\" (UniqueName: \"kubernetes.io/projected/508c3d5b-212a-46da-9a55-de3f35d7019b-kube-api-access-l8h8d\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.139868 4772 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.225394 4772 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.244888 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.261742 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.421650 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.531718 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-config-data\") pod \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.531789 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksdpm\" (UniqueName: \"kubernetes.io/projected/6e790127-8223-4b0c-8a5d-21e1bb15fa30-kube-api-access-ksdpm\") pod \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.531839 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-credential-keys\") pod \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.531865 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-fernet-keys\") pod \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.531886 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-combined-ca-bundle\") pod \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.532680 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-public-tls-certs\") pod \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.532777 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-internal-tls-certs\") pod \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.532847 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-scripts\") pod \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\" (UID: \"6e790127-8223-4b0c-8a5d-21e1bb15fa30\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.538312 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-scripts" (OuterVolumeSpecName: "scripts") pod "6e790127-8223-4b0c-8a5d-21e1bb15fa30" (UID: "6e790127-8223-4b0c-8a5d-21e1bb15fa30"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.538541 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e790127-8223-4b0c-8a5d-21e1bb15fa30-kube-api-access-ksdpm" (OuterVolumeSpecName: "kube-api-access-ksdpm") pod "6e790127-8223-4b0c-8a5d-21e1bb15fa30" (UID: "6e790127-8223-4b0c-8a5d-21e1bb15fa30"). InnerVolumeSpecName "kube-api-access-ksdpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.554453 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "6e790127-8223-4b0c-8a5d-21e1bb15fa30" (UID: "6e790127-8223-4b0c-8a5d-21e1bb15fa30"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.558406 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e790127-8223-4b0c-8a5d-21e1bb15fa30" (UID: "6e790127-8223-4b0c-8a5d-21e1bb15fa30"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.560845 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-config-data" (OuterVolumeSpecName: "config-data") pod "6e790127-8223-4b0c-8a5d-21e1bb15fa30" (UID: "6e790127-8223-4b0c-8a5d-21e1bb15fa30"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.563532 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6e790127-8223-4b0c-8a5d-21e1bb15fa30" (UID: "6e790127-8223-4b0c-8a5d-21e1bb15fa30"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.580900 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6e790127-8223-4b0c-8a5d-21e1bb15fa30" (UID: "6e790127-8223-4b0c-8a5d-21e1bb15fa30"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.583691 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6e790127-8223-4b0c-8a5d-21e1bb15fa30" (UID: "6e790127-8223-4b0c-8a5d-21e1bb15fa30"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.634805 4772 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.634855 4772 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.634868 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.634877 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.634887 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksdpm\" (UniqueName: \"kubernetes.io/projected/6e790127-8223-4b0c-8a5d-21e1bb15fa30-kube-api-access-ksdpm\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.634896 4772 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.634904 4772 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.634912 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e790127-8223-4b0c-8a5d-21e1bb15fa30-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.685269 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05445a27-d839-4a60-8338-5ee5f2c3f9d7" path="/var/lib/kubelet/pods/05445a27-d839-4a60-8338-5ee5f2c3f9d7/volumes" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.685654 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cbf7469-816d-4e54-a7ad-b5b76d0d59d6" path="/var/lib/kubelet/pods/4cbf7469-816d-4e54-a7ad-b5b76d0d59d6/volumes" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.686354 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="683f458e-44e9-49ea-a66b-4ac91a3f2bc1" path="/var/lib/kubelet/pods/683f458e-44e9-49ea-a66b-4ac91a3f2bc1/volumes" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.687512 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76fdbdb1-d48a-4cd1-8372-78887671dce8" path="/var/lib/kubelet/pods/76fdbdb1-d48a-4cd1-8372-78887671dce8/volumes" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.688891 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b8101bc-2ddf-48ed-9b92-e8f9e5e71938" path="/var/lib/kubelet/pods/8b8101bc-2ddf-48ed-9b92-e8f9e5e71938/volumes" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.690370 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66" path="/var/lib/kubelet/pods/a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66/volumes" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.691668 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aea5ee43-36e3-437d-8aca-b2faedd87c5b" path="/var/lib/kubelet/pods/aea5ee43-36e3-437d-8aca-b2faedd87c5b/volumes" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.692875 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f63bf600-ff03-43a3-92b4-fe8ac68a9bb7" path="/var/lib/kubelet/pods/f63bf600-ff03-43a3-92b4-fe8ac68a9bb7/volumes" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.829847 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.934143 4772 generic.go:334] "Generic (PLEG): container finished" podID="6e790127-8223-4b0c-8a5d-21e1bb15fa30" containerID="468321c234874e808e21c356adbece5162a3e84011f0215b573e541258fb76bc" exitCode=0 Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.934214 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-677fb7d6fc-djjsx" event={"ID":"6e790127-8223-4b0c-8a5d-21e1bb15fa30","Type":"ContainerDied","Data":"468321c234874e808e21c356adbece5162a3e84011f0215b573e541258fb76bc"} Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.934518 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-677fb7d6fc-djjsx" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.934593 4772 scope.go:117] "RemoveContainer" containerID="468321c234874e808e21c356adbece5162a3e84011f0215b573e541258fb76bc" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.935201 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-677fb7d6fc-djjsx" event={"ID":"6e790127-8223-4b0c-8a5d-21e1bb15fa30","Type":"ContainerDied","Data":"8b30129bf5b3504ae600edeaafe66f62d6f0c11b788461d423310f03199da3c5"} Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.944773 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1515626-5d79-408d-abc1-cb92abd58f3f-galera-tls-certs\") pod \"b1515626-5d79-408d-abc1-cb92abd58f3f\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.944853 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rxbj\" (UniqueName: \"kubernetes.io/projected/b1515626-5d79-408d-abc1-cb92abd58f3f-kube-api-access-4rxbj\") pod \"b1515626-5d79-408d-abc1-cb92abd58f3f\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.944965 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b1515626-5d79-408d-abc1-cb92abd58f3f-config-data-generated\") pod \"b1515626-5d79-408d-abc1-cb92abd58f3f\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.944990 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"b1515626-5d79-408d-abc1-cb92abd58f3f\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.945139 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b1515626-5d79-408d-abc1-cb92abd58f3f-config-data-default\") pod \"b1515626-5d79-408d-abc1-cb92abd58f3f\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.945212 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1515626-5d79-408d-abc1-cb92abd58f3f-combined-ca-bundle\") pod \"b1515626-5d79-408d-abc1-cb92abd58f3f\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.945263 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1515626-5d79-408d-abc1-cb92abd58f3f-operator-scripts\") pod \"b1515626-5d79-408d-abc1-cb92abd58f3f\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.945300 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b1515626-5d79-408d-abc1-cb92abd58f3f-kolla-config\") pod \"b1515626-5d79-408d-abc1-cb92abd58f3f\" (UID: \"b1515626-5d79-408d-abc1-cb92abd58f3f\") " Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.946299 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1515626-5d79-408d-abc1-cb92abd58f3f-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "b1515626-5d79-408d-abc1-cb92abd58f3f" (UID: "b1515626-5d79-408d-abc1-cb92abd58f3f"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.946394 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1515626-5d79-408d-abc1-cb92abd58f3f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b1515626-5d79-408d-abc1-cb92abd58f3f" (UID: "b1515626-5d79-408d-abc1-cb92abd58f3f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.947079 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1515626-5d79-408d-abc1-cb92abd58f3f-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "b1515626-5d79-408d-abc1-cb92abd58f3f" (UID: "b1515626-5d79-408d-abc1-cb92abd58f3f"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.947198 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1515626-5d79-408d-abc1-cb92abd58f3f-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "b1515626-5d79-408d-abc1-cb92abd58f3f" (UID: "b1515626-5d79-408d-abc1-cb92abd58f3f"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.947644 4772 generic.go:334] "Generic (PLEG): container finished" podID="dbbd3c83-3fde-4b11-8ef0-add837d393ce" containerID="788384a3ae6b89b3eeabbb3fe7578f4cb514172f7c7e0c341ec2b75ed4d75a29" exitCode=0 Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.947702 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"dbbd3c83-3fde-4b11-8ef0-add837d393ce","Type":"ContainerDied","Data":"788384a3ae6b89b3eeabbb3fe7578f4cb514172f7c7e0c341ec2b75ed4d75a29"} Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.949179 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1515626-5d79-408d-abc1-cb92abd58f3f-kube-api-access-4rxbj" (OuterVolumeSpecName: "kube-api-access-4rxbj") pod "b1515626-5d79-408d-abc1-cb92abd58f3f" (UID: "b1515626-5d79-408d-abc1-cb92abd58f3f"). InnerVolumeSpecName "kube-api-access-4rxbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.951632 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"508c3d5b-212a-46da-9a55-de3f35d7019b","Type":"ContainerDied","Data":"044e360ab5ed48dba1c044f12dafd0e510d6847bb09f3238ce3b8c8d2130f226"} Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.951752 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.955193 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "mysql-db") pod "b1515626-5d79-408d-abc1-cb92abd58f3f" (UID: "b1515626-5d79-408d-abc1-cb92abd58f3f"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.957651 4772 generic.go:334] "Generic (PLEG): container finished" podID="b1515626-5d79-408d-abc1-cb92abd58f3f" containerID="21613f2614f5809a9a792371fe2e685753a1a2fc6ea2f8fa7dcc2390d4bafda0" exitCode=0 Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.957691 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b1515626-5d79-408d-abc1-cb92abd58f3f","Type":"ContainerDied","Data":"21613f2614f5809a9a792371fe2e685753a1a2fc6ea2f8fa7dcc2390d4bafda0"} Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.957711 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b1515626-5d79-408d-abc1-cb92abd58f3f","Type":"ContainerDied","Data":"e699d423eedfd6502021873114f8ac6157951b5b24e3387e2b6a5c652a5f6465"} Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.957971 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.973037 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1515626-5d79-408d-abc1-cb92abd58f3f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b1515626-5d79-408d-abc1-cb92abd58f3f" (UID: "b1515626-5d79-408d-abc1-cb92abd58f3f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:36 crc kubenswrapper[4772]: I0127 15:31:36.989380 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1515626-5d79-408d-abc1-cb92abd58f3f-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "b1515626-5d79-408d-abc1-cb92abd58f3f" (UID: "b1515626-5d79-408d-abc1-cb92abd58f3f"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.003305 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.020806 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.028253 4772 scope.go:117] "RemoveContainer" containerID="468321c234874e808e21c356adbece5162a3e84011f0215b573e541258fb76bc" Jan 27 15:31:37 crc kubenswrapper[4772]: E0127 15:31:37.028989 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"468321c234874e808e21c356adbece5162a3e84011f0215b573e541258fb76bc\": container with ID starting with 468321c234874e808e21c356adbece5162a3e84011f0215b573e541258fb76bc not found: ID does not exist" containerID="468321c234874e808e21c356adbece5162a3e84011f0215b573e541258fb76bc" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.029038 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"468321c234874e808e21c356adbece5162a3e84011f0215b573e541258fb76bc"} err="failed to get container status \"468321c234874e808e21c356adbece5162a3e84011f0215b573e541258fb76bc\": rpc error: code = NotFound desc = could not find container \"468321c234874e808e21c356adbece5162a3e84011f0215b573e541258fb76bc\": container with ID starting with 468321c234874e808e21c356adbece5162a3e84011f0215b573e541258fb76bc not found: ID does not exist" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.029071 4772 scope.go:117] "RemoveContainer" containerID="f002759dea4443f7600e0f76f24481c1604449a5ee31bd8aa53171a2121ec4b2" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.032349 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.046025 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vh2cp\" (UniqueName: \"kubernetes.io/projected/dbbd3c83-3fde-4b11-8ef0-add837d393ce-kube-api-access-vh2cp\") pod \"dbbd3c83-3fde-4b11-8ef0-add837d393ce\" (UID: \"dbbd3c83-3fde-4b11-8ef0-add837d393ce\") " Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.046130 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbbd3c83-3fde-4b11-8ef0-add837d393ce-config-data\") pod \"dbbd3c83-3fde-4b11-8ef0-add837d393ce\" (UID: \"dbbd3c83-3fde-4b11-8ef0-add837d393ce\") " Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.046215 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbbd3c83-3fde-4b11-8ef0-add837d393ce-combined-ca-bundle\") pod \"dbbd3c83-3fde-4b11-8ef0-add837d393ce\" (UID: \"dbbd3c83-3fde-4b11-8ef0-add837d393ce\") " Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.046691 4772 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1515626-5d79-408d-abc1-cb92abd58f3f-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.046720 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rxbj\" (UniqueName: \"kubernetes.io/projected/b1515626-5d79-408d-abc1-cb92abd58f3f-kube-api-access-4rxbj\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.046740 4772 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b1515626-5d79-408d-abc1-cb92abd58f3f-config-data-generated\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.046763 4772 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.046777 4772 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b1515626-5d79-408d-abc1-cb92abd58f3f-config-data-default\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.046790 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1515626-5d79-408d-abc1-cb92abd58f3f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.046801 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1515626-5d79-408d-abc1-cb92abd58f3f-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.046811 4772 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b1515626-5d79-408d-abc1-cb92abd58f3f-kolla-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.049304 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbbd3c83-3fde-4b11-8ef0-add837d393ce-kube-api-access-vh2cp" (OuterVolumeSpecName: "kube-api-access-vh2cp") pod "dbbd3c83-3fde-4b11-8ef0-add837d393ce" (UID: "dbbd3c83-3fde-4b11-8ef0-add837d393ce"). InnerVolumeSpecName "kube-api-access-vh2cp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.065816 4772 scope.go:117] "RemoveContainer" containerID="900401625caff4c2d87fe06884c7dcba7f46fdc58e9213b1a6cc2cf36d383e52" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.068972 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-677fb7d6fc-djjsx"] Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.079483 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbbd3c83-3fde-4b11-8ef0-add837d393ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dbbd3c83-3fde-4b11-8ef0-add837d393ce" (UID: "dbbd3c83-3fde-4b11-8ef0-add837d393ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.079897 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-677fb7d6fc-djjsx"] Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.080817 4772 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.090927 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbbd3c83-3fde-4b11-8ef0-add837d393ce-config-data" (OuterVolumeSpecName: "config-data") pod "dbbd3c83-3fde-4b11-8ef0-add837d393ce" (UID: "dbbd3c83-3fde-4b11-8ef0-add837d393ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.096532 4772 scope.go:117] "RemoveContainer" containerID="21613f2614f5809a9a792371fe2e685753a1a2fc6ea2f8fa7dcc2390d4bafda0" Jan 27 15:31:37 crc kubenswrapper[4772]: E0127 15:31:37.109780 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b is running failed: container process not found" containerID="4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Jan 27 15:31:37 crc kubenswrapper[4772]: E0127 15:31:37.110042 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b is running failed: container process not found" containerID="4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Jan 27 15:31:37 crc kubenswrapper[4772]: E0127 15:31:37.110249 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b is running failed: container process not found" containerID="4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Jan 27 15:31:37 crc kubenswrapper[4772]: E0127 15:31:37.110270 4772 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-cqx7r" podUID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerName="ovsdb-server" Jan 27 15:31:37 crc kubenswrapper[4772]: E0127 15:31:37.118922 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6579efc0c6a14eb40a1349e6b5e9e288881435286dc3fff811374b436b48c91" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.120245 4772 scope.go:117] "RemoveContainer" containerID="3f38ceb1ab131833479b9e418df05230599249523d60e8e11929add232b03e39" Jan 27 15:31:37 crc kubenswrapper[4772]: E0127 15:31:37.125428 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6579efc0c6a14eb40a1349e6b5e9e288881435286dc3fff811374b436b48c91" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Jan 27 15:31:37 crc kubenswrapper[4772]: E0127 15:31:37.126644 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6579efc0c6a14eb40a1349e6b5e9e288881435286dc3fff811374b436b48c91" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Jan 27 15:31:37 crc kubenswrapper[4772]: E0127 15:31:37.126677 4772 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-cqx7r" podUID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerName="ovs-vswitchd" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.148321 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vh2cp\" (UniqueName: \"kubernetes.io/projected/dbbd3c83-3fde-4b11-8ef0-add837d393ce-kube-api-access-vh2cp\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.148663 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbbd3c83-3fde-4b11-8ef0-add837d393ce-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.148678 4772 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.148692 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbbd3c83-3fde-4b11-8ef0-add837d393ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.148925 4772 scope.go:117] "RemoveContainer" containerID="21613f2614f5809a9a792371fe2e685753a1a2fc6ea2f8fa7dcc2390d4bafda0" Jan 27 15:31:37 crc kubenswrapper[4772]: E0127 15:31:37.149605 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21613f2614f5809a9a792371fe2e685753a1a2fc6ea2f8fa7dcc2390d4bafda0\": container with ID starting with 21613f2614f5809a9a792371fe2e685753a1a2fc6ea2f8fa7dcc2390d4bafda0 not found: ID does not exist" containerID="21613f2614f5809a9a792371fe2e685753a1a2fc6ea2f8fa7dcc2390d4bafda0" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.149653 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21613f2614f5809a9a792371fe2e685753a1a2fc6ea2f8fa7dcc2390d4bafda0"} err="failed to get container status \"21613f2614f5809a9a792371fe2e685753a1a2fc6ea2f8fa7dcc2390d4bafda0\": rpc error: code = NotFound desc = could not find container \"21613f2614f5809a9a792371fe2e685753a1a2fc6ea2f8fa7dcc2390d4bafda0\": container with ID starting with 21613f2614f5809a9a792371fe2e685753a1a2fc6ea2f8fa7dcc2390d4bafda0 not found: ID does not exist" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.149680 4772 scope.go:117] "RemoveContainer" containerID="3f38ceb1ab131833479b9e418df05230599249523d60e8e11929add232b03e39" Jan 27 15:31:37 crc kubenswrapper[4772]: E0127 15:31:37.149955 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f38ceb1ab131833479b9e418df05230599249523d60e8e11929add232b03e39\": container with ID starting with 3f38ceb1ab131833479b9e418df05230599249523d60e8e11929add232b03e39 not found: ID does not exist" containerID="3f38ceb1ab131833479b9e418df05230599249523d60e8e11929add232b03e39" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.150732 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f38ceb1ab131833479b9e418df05230599249523d60e8e11929add232b03e39"} err="failed to get container status \"3f38ceb1ab131833479b9e418df05230599249523d60e8e11929add232b03e39\": rpc error: code = NotFound desc = could not find container \"3f38ceb1ab131833479b9e418df05230599249523d60e8e11929add232b03e39\": container with ID starting with 3f38ceb1ab131833479b9e418df05230599249523d60e8e11929add232b03e39 not found: ID does not exist" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.290053 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.296116 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.320596 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.350124 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b83f7578-8113-46c8-be24-5968aa0ca563-combined-ca-bundle\") pod \"b83f7578-8113-46c8-be24-5968aa0ca563\" (UID: \"b83f7578-8113-46c8-be24-5968aa0ca563\") " Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.350269 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hq4lp\" (UniqueName: \"kubernetes.io/projected/b83f7578-8113-46c8-be24-5968aa0ca563-kube-api-access-hq4lp\") pod \"b83f7578-8113-46c8-be24-5968aa0ca563\" (UID: \"b83f7578-8113-46c8-be24-5968aa0ca563\") " Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.350692 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b83f7578-8113-46c8-be24-5968aa0ca563-config-data\") pod \"b83f7578-8113-46c8-be24-5968aa0ca563\" (UID: \"b83f7578-8113-46c8-be24-5968aa0ca563\") " Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.353297 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b83f7578-8113-46c8-be24-5968aa0ca563-kube-api-access-hq4lp" (OuterVolumeSpecName: "kube-api-access-hq4lp") pod "b83f7578-8113-46c8-be24-5968aa0ca563" (UID: "b83f7578-8113-46c8-be24-5968aa0ca563"). InnerVolumeSpecName "kube-api-access-hq4lp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.367383 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b83f7578-8113-46c8-be24-5968aa0ca563-config-data" (OuterVolumeSpecName: "config-data") pod "b83f7578-8113-46c8-be24-5968aa0ca563" (UID: "b83f7578-8113-46c8-be24-5968aa0ca563"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.379767 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b83f7578-8113-46c8-be24-5968aa0ca563-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b83f7578-8113-46c8-be24-5968aa0ca563" (UID: "b83f7578-8113-46c8-be24-5968aa0ca563"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.452460 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b83f7578-8113-46c8-be24-5968aa0ca563-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.452490 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b83f7578-8113-46c8-be24-5968aa0ca563-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.452500 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hq4lp\" (UniqueName: \"kubernetes.io/projected/b83f7578-8113-46c8-be24-5968aa0ca563-kube-api-access-hq4lp\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.971438 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"dbbd3c83-3fde-4b11-8ef0-add837d393ce","Type":"ContainerDied","Data":"4ac8efb7b8696b151d1bdc121a58850ad086edea13390d08276f3048f0eea493"} Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.971493 4772 scope.go:117] "RemoveContainer" containerID="788384a3ae6b89b3eeabbb3fe7578f4cb514172f7c7e0c341ec2b75ed4d75a29" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.971507 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.973492 4772 generic.go:334] "Generic (PLEG): container finished" podID="b83f7578-8113-46c8-be24-5968aa0ca563" containerID="e92037085b98ccc46bbd64416f98018d2426a17d0883dd17b830d5574a8a0f4e" exitCode=0 Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.973542 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.973553 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b83f7578-8113-46c8-be24-5968aa0ca563","Type":"ContainerDied","Data":"e92037085b98ccc46bbd64416f98018d2426a17d0883dd17b830d5574a8a0f4e"} Jan 27 15:31:37 crc kubenswrapper[4772]: I0127 15:31:37.973578 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b83f7578-8113-46c8-be24-5968aa0ca563","Type":"ContainerDied","Data":"ee6d65efb4f3df3d96335b7b6b58d4ee20a12c71c0ca644b8c8c4208300d2710"} Jan 27 15:31:38 crc kubenswrapper[4772]: I0127 15:31:38.008515 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 15:31:38 crc kubenswrapper[4772]: I0127 15:31:38.019381 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 15:31:38 crc kubenswrapper[4772]: I0127 15:31:38.025472 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 27 15:31:38 crc kubenswrapper[4772]: I0127 15:31:38.030009 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 27 15:31:38 crc kubenswrapper[4772]: I0127 15:31:38.046097 4772 scope.go:117] "RemoveContainer" containerID="e92037085b98ccc46bbd64416f98018d2426a17d0883dd17b830d5574a8a0f4e" Jan 27 15:31:38 crc kubenswrapper[4772]: I0127 15:31:38.073915 4772 scope.go:117] "RemoveContainer" containerID="e92037085b98ccc46bbd64416f98018d2426a17d0883dd17b830d5574a8a0f4e" Jan 27 15:31:38 crc kubenswrapper[4772]: E0127 15:31:38.074473 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e92037085b98ccc46bbd64416f98018d2426a17d0883dd17b830d5574a8a0f4e\": container with ID starting with e92037085b98ccc46bbd64416f98018d2426a17d0883dd17b830d5574a8a0f4e not found: ID does not exist" containerID="e92037085b98ccc46bbd64416f98018d2426a17d0883dd17b830d5574a8a0f4e" Jan 27 15:31:38 crc kubenswrapper[4772]: I0127 15:31:38.074509 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e92037085b98ccc46bbd64416f98018d2426a17d0883dd17b830d5574a8a0f4e"} err="failed to get container status \"e92037085b98ccc46bbd64416f98018d2426a17d0883dd17b830d5574a8a0f4e\": rpc error: code = NotFound desc = could not find container \"e92037085b98ccc46bbd64416f98018d2426a17d0883dd17b830d5574a8a0f4e\": container with ID starting with e92037085b98ccc46bbd64416f98018d2426a17d0883dd17b830d5574a8a0f4e not found: ID does not exist" Jan 27 15:31:38 crc kubenswrapper[4772]: I0127 15:31:38.676119 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="508c3d5b-212a-46da-9a55-de3f35d7019b" path="/var/lib/kubelet/pods/508c3d5b-212a-46da-9a55-de3f35d7019b/volumes" Jan 27 15:31:38 crc kubenswrapper[4772]: I0127 15:31:38.678279 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e790127-8223-4b0c-8a5d-21e1bb15fa30" path="/var/lib/kubelet/pods/6e790127-8223-4b0c-8a5d-21e1bb15fa30/volumes" Jan 27 15:31:38 crc kubenswrapper[4772]: I0127 15:31:38.679870 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1515626-5d79-408d-abc1-cb92abd58f3f" path="/var/lib/kubelet/pods/b1515626-5d79-408d-abc1-cb92abd58f3f/volumes" Jan 27 15:31:38 crc kubenswrapper[4772]: I0127 15:31:38.682656 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b83f7578-8113-46c8-be24-5968aa0ca563" path="/var/lib/kubelet/pods/b83f7578-8113-46c8-be24-5968aa0ca563/volumes" Jan 27 15:31:38 crc kubenswrapper[4772]: I0127 15:31:38.684010 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbbd3c83-3fde-4b11-8ef0-add837d393ce" path="/var/lib/kubelet/pods/dbbd3c83-3fde-4b11-8ef0-add837d393ce/volumes" Jan 27 15:31:42 crc kubenswrapper[4772]: I0127 15:31:42.058674 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:31:42 crc kubenswrapper[4772]: I0127 15:31:42.059290 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:31:42 crc kubenswrapper[4772]: E0127 15:31:42.109314 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b is running failed: container process not found" containerID="4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Jan 27 15:31:42 crc kubenswrapper[4772]: E0127 15:31:42.109814 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b is running failed: container process not found" containerID="4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Jan 27 15:31:42 crc kubenswrapper[4772]: E0127 15:31:42.110193 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b is running failed: container process not found" containerID="4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Jan 27 15:31:42 crc kubenswrapper[4772]: E0127 15:31:42.110239 4772 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-cqx7r" podUID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerName="ovsdb-server" Jan 27 15:31:42 crc kubenswrapper[4772]: E0127 15:31:42.110917 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6579efc0c6a14eb40a1349e6b5e9e288881435286dc3fff811374b436b48c91" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Jan 27 15:31:42 crc kubenswrapper[4772]: E0127 15:31:42.112578 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6579efc0c6a14eb40a1349e6b5e9e288881435286dc3fff811374b436b48c91" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Jan 27 15:31:42 crc kubenswrapper[4772]: E0127 15:31:42.113940 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6579efc0c6a14eb40a1349e6b5e9e288881435286dc3fff811374b436b48c91" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Jan 27 15:31:42 crc kubenswrapper[4772]: E0127 15:31:42.113994 4772 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-cqx7r" podUID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerName="ovs-vswitchd" Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.022364 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.065471 4772 generic.go:334] "Generic (PLEG): container finished" podID="6cf131c4-a5bd-452b-8598-42312c3a0270" containerID="72824ad39b806a2254b462f1a46f766a404dd5dd1e5172059745c7930bd54b37" exitCode=0 Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.065510 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-647c88bb6f-wzf82" Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.065516 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-647c88bb6f-wzf82" event={"ID":"6cf131c4-a5bd-452b-8598-42312c3a0270","Type":"ContainerDied","Data":"72824ad39b806a2254b462f1a46f766a404dd5dd1e5172059745c7930bd54b37"} Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.065545 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-647c88bb6f-wzf82" event={"ID":"6cf131c4-a5bd-452b-8598-42312c3a0270","Type":"ContainerDied","Data":"eb3fc136e47d75ea92171b2a25f1728b294a61ff0f248fa056a324eadfc98f00"} Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.065567 4772 scope.go:117] "RemoveContainer" containerID="b96f34157cbed4eef2143feeb0fd51ea8ff8193f4fa6d28ad6a9487061aba8b2" Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.089374 4772 scope.go:117] "RemoveContainer" containerID="72824ad39b806a2254b462f1a46f766a404dd5dd1e5172059745c7930bd54b37" Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.106633 4772 scope.go:117] "RemoveContainer" containerID="b96f34157cbed4eef2143feeb0fd51ea8ff8193f4fa6d28ad6a9487061aba8b2" Jan 27 15:31:47 crc kubenswrapper[4772]: E0127 15:31:47.107029 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b96f34157cbed4eef2143feeb0fd51ea8ff8193f4fa6d28ad6a9487061aba8b2\": container with ID starting with b96f34157cbed4eef2143feeb0fd51ea8ff8193f4fa6d28ad6a9487061aba8b2 not found: ID does not exist" containerID="b96f34157cbed4eef2143feeb0fd51ea8ff8193f4fa6d28ad6a9487061aba8b2" Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.107069 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b96f34157cbed4eef2143feeb0fd51ea8ff8193f4fa6d28ad6a9487061aba8b2"} err="failed to get container status \"b96f34157cbed4eef2143feeb0fd51ea8ff8193f4fa6d28ad6a9487061aba8b2\": rpc error: code = NotFound desc = could not find container \"b96f34157cbed4eef2143feeb0fd51ea8ff8193f4fa6d28ad6a9487061aba8b2\": container with ID starting with b96f34157cbed4eef2143feeb0fd51ea8ff8193f4fa6d28ad6a9487061aba8b2 not found: ID does not exist" Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.107092 4772 scope.go:117] "RemoveContainer" containerID="72824ad39b806a2254b462f1a46f766a404dd5dd1e5172059745c7930bd54b37" Jan 27 15:31:47 crc kubenswrapper[4772]: E0127 15:31:47.107543 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72824ad39b806a2254b462f1a46f766a404dd5dd1e5172059745c7930bd54b37\": container with ID starting with 72824ad39b806a2254b462f1a46f766a404dd5dd1e5172059745c7930bd54b37 not found: ID does not exist" containerID="72824ad39b806a2254b462f1a46f766a404dd5dd1e5172059745c7930bd54b37" Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.107583 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72824ad39b806a2254b462f1a46f766a404dd5dd1e5172059745c7930bd54b37"} err="failed to get container status \"72824ad39b806a2254b462f1a46f766a404dd5dd1e5172059745c7930bd54b37\": rpc error: code = NotFound desc = could not find container \"72824ad39b806a2254b462f1a46f766a404dd5dd1e5172059745c7930bd54b37\": container with ID starting with 72824ad39b806a2254b462f1a46f766a404dd5dd1e5172059745c7930bd54b37 not found: ID does not exist" Jan 27 15:31:47 crc kubenswrapper[4772]: E0127 15:31:47.109030 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b is running failed: container process not found" containerID="4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Jan 27 15:31:47 crc kubenswrapper[4772]: E0127 15:31:47.109423 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b is running failed: container process not found" containerID="4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Jan 27 15:31:47 crc kubenswrapper[4772]: E0127 15:31:47.109672 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b is running failed: container process not found" containerID="4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Jan 27 15:31:47 crc kubenswrapper[4772]: E0127 15:31:47.109710 4772 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-cqx7r" podUID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerName="ovsdb-server" Jan 27 15:31:47 crc kubenswrapper[4772]: E0127 15:31:47.110212 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6579efc0c6a14eb40a1349e6b5e9e288881435286dc3fff811374b436b48c91" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Jan 27 15:31:47 crc kubenswrapper[4772]: E0127 15:31:47.111778 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6579efc0c6a14eb40a1349e6b5e9e288881435286dc3fff811374b436b48c91" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Jan 27 15:31:47 crc kubenswrapper[4772]: E0127 15:31:47.113161 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6579efc0c6a14eb40a1349e6b5e9e288881435286dc3fff811374b436b48c91" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Jan 27 15:31:47 crc kubenswrapper[4772]: E0127 15:31:47.113293 4772 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-cqx7r" podUID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerName="ovs-vswitchd" Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.153375 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-httpd-config\") pod \"6cf131c4-a5bd-452b-8598-42312c3a0270\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.153467 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-internal-tls-certs\") pod \"6cf131c4-a5bd-452b-8598-42312c3a0270\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.153511 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-public-tls-certs\") pod \"6cf131c4-a5bd-452b-8598-42312c3a0270\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.153553 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brxbx\" (UniqueName: \"kubernetes.io/projected/6cf131c4-a5bd-452b-8598-42312c3a0270-kube-api-access-brxbx\") pod \"6cf131c4-a5bd-452b-8598-42312c3a0270\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.153618 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-combined-ca-bundle\") pod \"6cf131c4-a5bd-452b-8598-42312c3a0270\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.153700 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-config\") pod \"6cf131c4-a5bd-452b-8598-42312c3a0270\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.153736 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-ovndb-tls-certs\") pod \"6cf131c4-a5bd-452b-8598-42312c3a0270\" (UID: \"6cf131c4-a5bd-452b-8598-42312c3a0270\") " Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.158993 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cf131c4-a5bd-452b-8598-42312c3a0270-kube-api-access-brxbx" (OuterVolumeSpecName: "kube-api-access-brxbx") pod "6cf131c4-a5bd-452b-8598-42312c3a0270" (UID: "6cf131c4-a5bd-452b-8598-42312c3a0270"). InnerVolumeSpecName "kube-api-access-brxbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.160878 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "6cf131c4-a5bd-452b-8598-42312c3a0270" (UID: "6cf131c4-a5bd-452b-8598-42312c3a0270"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.193245 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6cf131c4-a5bd-452b-8598-42312c3a0270" (UID: "6cf131c4-a5bd-452b-8598-42312c3a0270"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.195897 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6cf131c4-a5bd-452b-8598-42312c3a0270" (UID: "6cf131c4-a5bd-452b-8598-42312c3a0270"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.202036 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-config" (OuterVolumeSpecName: "config") pod "6cf131c4-a5bd-452b-8598-42312c3a0270" (UID: "6cf131c4-a5bd-452b-8598-42312c3a0270"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.208127 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6cf131c4-a5bd-452b-8598-42312c3a0270" (UID: "6cf131c4-a5bd-452b-8598-42312c3a0270"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.228932 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "6cf131c4-a5bd-452b-8598-42312c3a0270" (UID: "6cf131c4-a5bd-452b-8598-42312c3a0270"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.255314 4772 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.255353 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brxbx\" (UniqueName: \"kubernetes.io/projected/6cf131c4-a5bd-452b-8598-42312c3a0270-kube-api-access-brxbx\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.255389 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.255401 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.255412 4772 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.255422 4772 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-httpd-config\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.255431 4772 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cf131c4-a5bd-452b-8598-42312c3a0270-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.395150 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-647c88bb6f-wzf82"] Jan 27 15:31:47 crc kubenswrapper[4772]: I0127 15:31:47.399748 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-647c88bb6f-wzf82"] Jan 27 15:31:48 crc kubenswrapper[4772]: I0127 15:31:48.675960 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cf131c4-a5bd-452b-8598-42312c3a0270" path="/var/lib/kubelet/pods/6cf131c4-a5bd-452b-8598-42312c3a0270/volumes" Jan 27 15:31:52 crc kubenswrapper[4772]: E0127 15:31:52.111408 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6579efc0c6a14eb40a1349e6b5e9e288881435286dc3fff811374b436b48c91" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Jan 27 15:31:52 crc kubenswrapper[4772]: E0127 15:31:52.111819 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b is running failed: container process not found" containerID="4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Jan 27 15:31:52 crc kubenswrapper[4772]: E0127 15:31:52.332030 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b is running failed: container process not found" containerID="4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Jan 27 15:31:52 crc kubenswrapper[4772]: E0127 15:31:52.333369 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b is running failed: container process not found" containerID="4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Jan 27 15:31:52 crc kubenswrapper[4772]: E0127 15:31:52.333409 4772 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-cqx7r" podUID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerName="ovsdb-server" Jan 27 15:31:52 crc kubenswrapper[4772]: E0127 15:31:52.334331 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6579efc0c6a14eb40a1349e6b5e9e288881435286dc3fff811374b436b48c91" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Jan 27 15:31:52 crc kubenswrapper[4772]: E0127 15:31:52.339690 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6579efc0c6a14eb40a1349e6b5e9e288881435286dc3fff811374b436b48c91" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Jan 27 15:31:52 crc kubenswrapper[4772]: E0127 15:31:52.339745 4772 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-cqx7r" podUID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerName="ovs-vswitchd" Jan 27 15:31:57 crc kubenswrapper[4772]: E0127 15:31:57.109793 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b is running failed: container process not found" containerID="4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Jan 27 15:31:57 crc kubenswrapper[4772]: E0127 15:31:57.111573 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6579efc0c6a14eb40a1349e6b5e9e288881435286dc3fff811374b436b48c91" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Jan 27 15:31:57 crc kubenswrapper[4772]: E0127 15:31:57.111771 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b is running failed: container process not found" containerID="4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Jan 27 15:31:57 crc kubenswrapper[4772]: E0127 15:31:57.112102 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b is running failed: container process not found" containerID="4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Jan 27 15:31:57 crc kubenswrapper[4772]: E0127 15:31:57.112141 4772 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-cqx7r" podUID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerName="ovsdb-server" Jan 27 15:31:57 crc kubenswrapper[4772]: E0127 15:31:57.113260 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6579efc0c6a14eb40a1349e6b5e9e288881435286dc3fff811374b436b48c91" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Jan 27 15:31:57 crc kubenswrapper[4772]: E0127 15:31:57.115641 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d6579efc0c6a14eb40a1349e6b5e9e288881435286dc3fff811374b436b48c91" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Jan 27 15:31:57 crc kubenswrapper[4772]: E0127 15:31:57.115703 4772 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-cqx7r" podUID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerName="ovs-vswitchd" Jan 27 15:31:57 crc kubenswrapper[4772]: I0127 15:31:57.384255 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-cqx7r_38ebd422-35c5-4682-8a4d-ca9073728d7c/ovs-vswitchd/0.log" Jan 27 15:31:57 crc kubenswrapper[4772]: I0127 15:31:57.385114 4772 generic.go:334] "Generic (PLEG): container finished" podID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerID="d6579efc0c6a14eb40a1349e6b5e9e288881435286dc3fff811374b436b48c91" exitCode=137 Jan 27 15:31:57 crc kubenswrapper[4772]: I0127 15:31:57.385181 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-cqx7r" event={"ID":"38ebd422-35c5-4682-8a4d-ca9073728d7c","Type":"ContainerDied","Data":"d6579efc0c6a14eb40a1349e6b5e9e288881435286dc3fff811374b436b48c91"} Jan 27 15:31:57 crc kubenswrapper[4772]: I0127 15:31:57.862059 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-cqx7r_38ebd422-35c5-4682-8a4d-ca9073728d7c/ovs-vswitchd/0.log" Jan 27 15:31:57 crc kubenswrapper[4772]: I0127 15:31:57.863340 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:31:57 crc kubenswrapper[4772]: I0127 15:31:57.906535 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvchv\" (UniqueName: \"kubernetes.io/projected/38ebd422-35c5-4682-8a4d-ca9073728d7c-kube-api-access-zvchv\") pod \"38ebd422-35c5-4682-8a4d-ca9073728d7c\" (UID: \"38ebd422-35c5-4682-8a4d-ca9073728d7c\") " Jan 27 15:31:57 crc kubenswrapper[4772]: I0127 15:31:57.907197 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/38ebd422-35c5-4682-8a4d-ca9073728d7c-var-lib\") pod \"38ebd422-35c5-4682-8a4d-ca9073728d7c\" (UID: \"38ebd422-35c5-4682-8a4d-ca9073728d7c\") " Jan 27 15:31:57 crc kubenswrapper[4772]: I0127 15:31:57.907398 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/38ebd422-35c5-4682-8a4d-ca9073728d7c-scripts\") pod \"38ebd422-35c5-4682-8a4d-ca9073728d7c\" (UID: \"38ebd422-35c5-4682-8a4d-ca9073728d7c\") " Jan 27 15:31:57 crc kubenswrapper[4772]: I0127 15:31:57.907534 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/38ebd422-35c5-4682-8a4d-ca9073728d7c-etc-ovs\") pod \"38ebd422-35c5-4682-8a4d-ca9073728d7c\" (UID: \"38ebd422-35c5-4682-8a4d-ca9073728d7c\") " Jan 27 15:31:57 crc kubenswrapper[4772]: I0127 15:31:57.907643 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/38ebd422-35c5-4682-8a4d-ca9073728d7c-var-log\") pod \"38ebd422-35c5-4682-8a4d-ca9073728d7c\" (UID: \"38ebd422-35c5-4682-8a4d-ca9073728d7c\") " Jan 27 15:31:57 crc kubenswrapper[4772]: I0127 15:31:57.908258 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/38ebd422-35c5-4682-8a4d-ca9073728d7c-var-log" (OuterVolumeSpecName: "var-log") pod "38ebd422-35c5-4682-8a4d-ca9073728d7c" (UID: "38ebd422-35c5-4682-8a4d-ca9073728d7c"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:31:57 crc kubenswrapper[4772]: I0127 15:31:57.908376 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/38ebd422-35c5-4682-8a4d-ca9073728d7c-var-lib" (OuterVolumeSpecName: "var-lib") pod "38ebd422-35c5-4682-8a4d-ca9073728d7c" (UID: "38ebd422-35c5-4682-8a4d-ca9073728d7c"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:31:57 crc kubenswrapper[4772]: I0127 15:31:57.908453 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/38ebd422-35c5-4682-8a4d-ca9073728d7c-var-run\") pod \"38ebd422-35c5-4682-8a4d-ca9073728d7c\" (UID: \"38ebd422-35c5-4682-8a4d-ca9073728d7c\") " Jan 27 15:31:57 crc kubenswrapper[4772]: I0127 15:31:57.908504 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/38ebd422-35c5-4682-8a4d-ca9073728d7c-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "38ebd422-35c5-4682-8a4d-ca9073728d7c" (UID: "38ebd422-35c5-4682-8a4d-ca9073728d7c"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:31:57 crc kubenswrapper[4772]: I0127 15:31:57.908877 4772 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/38ebd422-35c5-4682-8a4d-ca9073728d7c-var-lib\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:57 crc kubenswrapper[4772]: I0127 15:31:57.908896 4772 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/38ebd422-35c5-4682-8a4d-ca9073728d7c-etc-ovs\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:57 crc kubenswrapper[4772]: I0127 15:31:57.908906 4772 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/38ebd422-35c5-4682-8a4d-ca9073728d7c-var-log\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:57 crc kubenswrapper[4772]: I0127 15:31:57.908931 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/38ebd422-35c5-4682-8a4d-ca9073728d7c-var-run" (OuterVolumeSpecName: "var-run") pod "38ebd422-35c5-4682-8a4d-ca9073728d7c" (UID: "38ebd422-35c5-4682-8a4d-ca9073728d7c"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 15:31:57 crc kubenswrapper[4772]: I0127 15:31:57.909724 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38ebd422-35c5-4682-8a4d-ca9073728d7c-scripts" (OuterVolumeSpecName: "scripts") pod "38ebd422-35c5-4682-8a4d-ca9073728d7c" (UID: "38ebd422-35c5-4682-8a4d-ca9073728d7c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:31:57 crc kubenswrapper[4772]: I0127 15:31:57.912935 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38ebd422-35c5-4682-8a4d-ca9073728d7c-kube-api-access-zvchv" (OuterVolumeSpecName: "kube-api-access-zvchv") pod "38ebd422-35c5-4682-8a4d-ca9073728d7c" (UID: "38ebd422-35c5-4682-8a4d-ca9073728d7c"). InnerVolumeSpecName "kube-api-access-zvchv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.010347 4772 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/38ebd422-35c5-4682-8a4d-ca9073728d7c-var-run\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.010386 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvchv\" (UniqueName: \"kubernetes.io/projected/38ebd422-35c5-4682-8a4d-ca9073728d7c-kube-api-access-zvchv\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.010395 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/38ebd422-35c5-4682-8a4d-ca9073728d7c-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.396314 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-cqx7r_38ebd422-35c5-4682-8a4d-ca9073728d7c/ovs-vswitchd/0.log" Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.397151 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-cqx7r" Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.397147 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-cqx7r" event={"ID":"38ebd422-35c5-4682-8a4d-ca9073728d7c","Type":"ContainerDied","Data":"b4ae3e61c086f91c9c3a7442484ecc85a4bdf545d39601e45239a3351393b9ff"} Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.397241 4772 scope.go:117] "RemoveContainer" containerID="d6579efc0c6a14eb40a1349e6b5e9e288881435286dc3fff811374b436b48c91" Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.410546 4772 generic.go:334] "Generic (PLEG): container finished" podID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerID="0b50101071feccad5793667a8f4849d22482c6d522fac228c249d69d6d557cdf" exitCode=137 Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.410594 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerDied","Data":"0b50101071feccad5793667a8f4849d22482c6d522fac228c249d69d6d557cdf"} Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.437448 4772 scope.go:117] "RemoveContainer" containerID="4b64614c7f3007f9118f8ed226ede92035da74f66d831e70ce26d6d3d8e9f47b" Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.437975 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-cqx7r"] Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.442679 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-cqx7r"] Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.466359 4772 scope.go:117] "RemoveContainer" containerID="c2ec2d9ef51a12150ebe6df637e29030ff2b622c19a7ada45c6cd396c44b8636" Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.585222 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.673575 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38ebd422-35c5-4682-8a4d-ca9073728d7c" path="/var/lib/kubelet/pods/38ebd422-35c5-4682-8a4d-ca9073728d7c/volumes" Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.719483 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/3ef68955-b80c-4732-9e87-0bec53d0b3a0-lock\") pod \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.719591 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/3ef68955-b80c-4732-9e87-0bec53d0b3a0-cache\") pod \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.719644 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-etc-swift\") pod \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.719666 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxlv4\" (UniqueName: \"kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-kube-api-access-mxlv4\") pod \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.719703 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ef68955-b80c-4732-9e87-0bec53d0b3a0-combined-ca-bundle\") pod \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.719786 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\" (UID: \"3ef68955-b80c-4732-9e87-0bec53d0b3a0\") " Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.720125 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ef68955-b80c-4732-9e87-0bec53d0b3a0-lock" (OuterVolumeSpecName: "lock") pod "3ef68955-b80c-4732-9e87-0bec53d0b3a0" (UID: "3ef68955-b80c-4732-9e87-0bec53d0b3a0"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.720221 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ef68955-b80c-4732-9e87-0bec53d0b3a0-cache" (OuterVolumeSpecName: "cache") pod "3ef68955-b80c-4732-9e87-0bec53d0b3a0" (UID: "3ef68955-b80c-4732-9e87-0bec53d0b3a0"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.723007 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3ef68955-b80c-4732-9e87-0bec53d0b3a0" (UID: "3ef68955-b80c-4732-9e87-0bec53d0b3a0"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.723405 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-kube-api-access-mxlv4" (OuterVolumeSpecName: "kube-api-access-mxlv4") pod "3ef68955-b80c-4732-9e87-0bec53d0b3a0" (UID: "3ef68955-b80c-4732-9e87-0bec53d0b3a0"). InnerVolumeSpecName "kube-api-access-mxlv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.726325 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "swift") pod "3ef68955-b80c-4732-9e87-0bec53d0b3a0" (UID: "3ef68955-b80c-4732-9e87-0bec53d0b3a0"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.821119 4772 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/3ef68955-b80c-4732-9e87-0bec53d0b3a0-cache\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.821156 4772 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-etc-swift\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.821199 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxlv4\" (UniqueName: \"kubernetes.io/projected/3ef68955-b80c-4732-9e87-0bec53d0b3a0-kube-api-access-mxlv4\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.821236 4772 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.821248 4772 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/3ef68955-b80c-4732-9e87-0bec53d0b3a0-lock\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.835175 4772 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.923343 4772 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:58 crc kubenswrapper[4772]: I0127 15:31:58.972160 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ef68955-b80c-4732-9e87-0bec53d0b3a0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ef68955-b80c-4732-9e87-0bec53d0b3a0" (UID: "3ef68955-b80c-4732-9e87-0bec53d0b3a0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:31:59 crc kubenswrapper[4772]: I0127 15:31:59.024325 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ef68955-b80c-4732-9e87-0bec53d0b3a0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 15:31:59 crc kubenswrapper[4772]: I0127 15:31:59.427497 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"3ef68955-b80c-4732-9e87-0bec53d0b3a0","Type":"ContainerDied","Data":"7e2686f92b31392fd2420828f9959abe37458794a1d13beae3bf48377776f704"} Jan 27 15:31:59 crc kubenswrapper[4772]: I0127 15:31:59.427559 4772 scope.go:117] "RemoveContainer" containerID="0b50101071feccad5793667a8f4849d22482c6d522fac228c249d69d6d557cdf" Jan 27 15:31:59 crc kubenswrapper[4772]: I0127 15:31:59.427596 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Jan 27 15:31:59 crc kubenswrapper[4772]: I0127 15:31:59.472346 4772 scope.go:117] "RemoveContainer" containerID="8d889567d10b3e8868d76680ff442da2a14216919aae766c356918ec9960b9a4" Jan 27 15:31:59 crc kubenswrapper[4772]: I0127 15:31:59.473086 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Jan 27 15:31:59 crc kubenswrapper[4772]: I0127 15:31:59.480115 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Jan 27 15:31:59 crc kubenswrapper[4772]: I0127 15:31:59.531239 4772 scope.go:117] "RemoveContainer" containerID="c1cf3012e8501ba3a809e028a1ab49c960d95fb090a04b4dbca6cd01d2de9524" Jan 27 15:31:59 crc kubenswrapper[4772]: I0127 15:31:59.548883 4772 scope.go:117] "RemoveContainer" containerID="b0a7c137687a720a7d8c3f84cc586f4b9d3bde7c9bc9e2e0c83a325c2ae23322" Jan 27 15:31:59 crc kubenswrapper[4772]: I0127 15:31:59.575274 4772 scope.go:117] "RemoveContainer" containerID="8bbb31c1be222187b0e9b27f07c1ac0fe66d8ad583df4ff6b26fec62ab98cf87" Jan 27 15:31:59 crc kubenswrapper[4772]: I0127 15:31:59.597563 4772 scope.go:117] "RemoveContainer" containerID="71b4242b9081be055bfb8bd2db6959d32259cd0c3ee2b95ddde1c1d2154be74b" Jan 27 15:31:59 crc kubenswrapper[4772]: I0127 15:31:59.623096 4772 scope.go:117] "RemoveContainer" containerID="bc57f117c387fb10832190ea21f63cdb319308d9390292395fb515e28966d217" Jan 27 15:31:59 crc kubenswrapper[4772]: I0127 15:31:59.646979 4772 scope.go:117] "RemoveContainer" containerID="99c9f47c0720632dfecbfc5e9152885ab96d751677b561767c79f0a032ca5cf5" Jan 27 15:31:59 crc kubenswrapper[4772]: I0127 15:31:59.671976 4772 scope.go:117] "RemoveContainer" containerID="0c6f6ecf89a4947c23560538762ca73dfe5e13c4acb04e206d91772a3cfc9c49" Jan 27 15:31:59 crc kubenswrapper[4772]: I0127 15:31:59.693791 4772 scope.go:117] "RemoveContainer" containerID="94e4c588a745acb16ce919a52f7150cf54119c1c41e94c9e658206e6b58958ed" Jan 27 15:31:59 crc kubenswrapper[4772]: I0127 15:31:59.739295 4772 scope.go:117] "RemoveContainer" containerID="494d3ebaeddb756bf375d2bc394a4b4086ee3e25d9a76747552d41c1f40a9737" Jan 27 15:31:59 crc kubenswrapper[4772]: I0127 15:31:59.755397 4772 scope.go:117] "RemoveContainer" containerID="ac32767b3784713a66fbfe32a337398a7461aa8ffad58bbfea7ccf6e3c4ee19d" Jan 27 15:31:59 crc kubenswrapper[4772]: I0127 15:31:59.770425 4772 scope.go:117] "RemoveContainer" containerID="c3f602f5b8fe5f978c40989adc1d0130c6aaae0dce0fc13d5e34bbe819e8eccb" Jan 27 15:31:59 crc kubenswrapper[4772]: I0127 15:31:59.788185 4772 scope.go:117] "RemoveContainer" containerID="5f271cd2dcb6b658cde722402c5b2945c28f4d7486cab8c56e064081779416a1" Jan 27 15:31:59 crc kubenswrapper[4772]: I0127 15:31:59.807224 4772 scope.go:117] "RemoveContainer" containerID="d35aa807e61d39133b8319305719556fcfa6889495c80253864eaf2dc48a450b" Jan 27 15:32:00 crc kubenswrapper[4772]: I0127 15:32:00.673437 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" path="/var/lib/kubelet/pods/3ef68955-b80c-4732-9e87-0bec53d0b3a0/volumes" Jan 27 15:32:02 crc kubenswrapper[4772]: I0127 15:32:02.553598 4772 scope.go:117] "RemoveContainer" containerID="e068687fbbe8ba2bc884327a323113a2f9b397134b3783cc71145217f0aced72" Jan 27 15:32:02 crc kubenswrapper[4772]: I0127 15:32:02.575988 4772 scope.go:117] "RemoveContainer" containerID="a76c09aaadcee4723d7ef767396afbe7396ff3e3af040a33171b3953859d1cba" Jan 27 15:32:02 crc kubenswrapper[4772]: I0127 15:32:02.658106 4772 scope.go:117] "RemoveContainer" containerID="6aa60721dd7c09b05e3a663482308f5a6da188370cc19651da9a73a40e00696f" Jan 27 15:32:02 crc kubenswrapper[4772]: I0127 15:32:02.706010 4772 scope.go:117] "RemoveContainer" containerID="a5599751ce46331dd2a224ba692cd6619979f4eb0205e3a54352eb587e777c31" Jan 27 15:32:02 crc kubenswrapper[4772]: I0127 15:32:02.732710 4772 scope.go:117] "RemoveContainer" containerID="666a2855e8df449d0b2a9f22d64efe41fc16e80a56e57924cea7c6f56eb00af0" Jan 27 15:32:02 crc kubenswrapper[4772]: I0127 15:32:02.762941 4772 scope.go:117] "RemoveContainer" containerID="317ff691da5e191e31778e1d02f29484703e057687e372739fcbc9dd6f8088d2" Jan 27 15:32:02 crc kubenswrapper[4772]: I0127 15:32:02.809110 4772 scope.go:117] "RemoveContainer" containerID="d1b5117c10f9331477f591f10a624b08ae6968087cc1bb15580ee055f80a719c" Jan 27 15:32:02 crc kubenswrapper[4772]: I0127 15:32:02.831424 4772 scope.go:117] "RemoveContainer" containerID="7ce8beebc480cca9e2ff0700b901cda6f6e2d53f77d8edbfd7e337a2359ae80a" Jan 27 15:32:02 crc kubenswrapper[4772]: I0127 15:32:02.858418 4772 scope.go:117] "RemoveContainer" containerID="fcb62876ceaa2921dde5172c985a61fd201c04281f9d06dbb383e8128d91c935" Jan 27 15:32:02 crc kubenswrapper[4772]: I0127 15:32:02.876386 4772 scope.go:117] "RemoveContainer" containerID="37cb21cfa353006443b3a1e31571db32c636cbf5e0c7a880cb766a2b91769826" Jan 27 15:32:02 crc kubenswrapper[4772]: I0127 15:32:02.913310 4772 scope.go:117] "RemoveContainer" containerID="e93f9f446173d4fd985d40db28827a7f313c9dbe0522a2d3003fa93c8ac7de5e" Jan 27 15:32:02 crc kubenswrapper[4772]: I0127 15:32:02.962939 4772 scope.go:117] "RemoveContainer" containerID="d22be9ecfb9cc0389dd0f2e64dbdb2f980e40813484563a0a652ad657fd8f5b7" Jan 27 15:32:02 crc kubenswrapper[4772]: I0127 15:32:02.984044 4772 scope.go:117] "RemoveContainer" containerID="6d6c94667c0ae61eab0c4931fc95c11f862c674ae06fd177d824e395ced6b9a6" Jan 27 15:32:03 crc kubenswrapper[4772]: I0127 15:32:03.000592 4772 scope.go:117] "RemoveContainer" containerID="d569280ad66a5087c9e0aa7b8abe04a7d97361bee2ca7b7c30646e77734ba51d" Jan 27 15:32:03 crc kubenswrapper[4772]: I0127 15:32:03.024428 4772 scope.go:117] "RemoveContainer" containerID="31d9e486da5aa706768e022e398d969ef41f15c9db5b579c83d50ae160db05a7" Jan 27 15:32:12 crc kubenswrapper[4772]: I0127 15:32:12.058898 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:32:12 crc kubenswrapper[4772]: I0127 15:32:12.059294 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:32:42 crc kubenswrapper[4772]: I0127 15:32:42.058036 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:32:42 crc kubenswrapper[4772]: I0127 15:32:42.058663 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:32:42 crc kubenswrapper[4772]: I0127 15:32:42.058708 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:32:42 crc kubenswrapper[4772]: I0127 15:32:42.059306 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 15:32:42 crc kubenswrapper[4772]: I0127 15:32:42.059377 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" gracePeriod=600 Jan 27 15:32:42 crc kubenswrapper[4772]: E0127 15:32:42.378529 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:32:42 crc kubenswrapper[4772]: I0127 15:32:42.778799 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" exitCode=0 Jan 27 15:32:42 crc kubenswrapper[4772]: I0127 15:32:42.778855 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2"} Jan 27 15:32:42 crc kubenswrapper[4772]: I0127 15:32:42.778972 4772 scope.go:117] "RemoveContainer" containerID="1d1c45659af37dbb5fcad6152d119ca4f804c58006a54555795ff000f3b7aea9" Jan 27 15:32:42 crc kubenswrapper[4772]: I0127 15:32:42.779486 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:32:42 crc kubenswrapper[4772]: E0127 15:32:42.781088 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:32:54 crc kubenswrapper[4772]: I0127 15:32:54.667753 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:32:54 crc kubenswrapper[4772]: E0127 15:32:54.668758 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:33:03 crc kubenswrapper[4772]: I0127 15:33:03.738727 4772 scope.go:117] "RemoveContainer" containerID="c454404cb2dabeb6539bab075b0096e5a7ba9d3726f1b7a2ce5d55b30cc778e8" Jan 27 15:33:03 crc kubenswrapper[4772]: I0127 15:33:03.762959 4772 scope.go:117] "RemoveContainer" containerID="c63a10e019701dbe41c4487398c76cb4acdd6a0eda99f6edb9df7d6273b71a27" Jan 27 15:33:03 crc kubenswrapper[4772]: I0127 15:33:03.813607 4772 scope.go:117] "RemoveContainer" containerID="e1a2cafeb608c7919a88b50bf39a141cb90ef87745db78d4f8f6a94522bb8d2e" Jan 27 15:33:03 crc kubenswrapper[4772]: I0127 15:33:03.838875 4772 scope.go:117] "RemoveContainer" containerID="a5ffbaeea04257a22f38554ccc4304785fadfe22ac90bb6e3544b162aab10857" Jan 27 15:33:03 crc kubenswrapper[4772]: I0127 15:33:03.854811 4772 scope.go:117] "RemoveContainer" containerID="d265eb93689c326c68ce844d36ec8e13845ff3f6cfb1ed7e88273d0cf4e91cbd" Jan 27 15:33:03 crc kubenswrapper[4772]: I0127 15:33:03.890120 4772 scope.go:117] "RemoveContainer" containerID="597636ff183f237bb3b639ea5c67c6b5f6f29f40e362b71df3d4ec02eaa6036b" Jan 27 15:33:03 crc kubenswrapper[4772]: I0127 15:33:03.912247 4772 scope.go:117] "RemoveContainer" containerID="478f9eb73f50cba542d4259825587e98caddfe9513876ed4823af8b00681f571" Jan 27 15:33:03 crc kubenswrapper[4772]: I0127 15:33:03.932956 4772 scope.go:117] "RemoveContainer" containerID="7cb0416a54334bdd5699afd4b64397c193035c399e5586172a360ff52cd674f9" Jan 27 15:33:03 crc kubenswrapper[4772]: I0127 15:33:03.949915 4772 scope.go:117] "RemoveContainer" containerID="d2de8b3a1c27ebd01b5c3393c6dcb85d202fe549eef0c41d0f9f318c3b15d219" Jan 27 15:33:06 crc kubenswrapper[4772]: I0127 15:33:06.663746 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:33:06 crc kubenswrapper[4772]: E0127 15:33:06.664031 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:33:20 crc kubenswrapper[4772]: I0127 15:33:20.663949 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:33:20 crc kubenswrapper[4772]: E0127 15:33:20.664624 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:33:31 crc kubenswrapper[4772]: I0127 15:33:31.662999 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:33:31 crc kubenswrapper[4772]: E0127 15:33:31.663800 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:33:42 crc kubenswrapper[4772]: I0127 15:33:42.663719 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:33:42 crc kubenswrapper[4772]: E0127 15:33:42.664982 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:33:55 crc kubenswrapper[4772]: I0127 15:33:55.662801 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:33:55 crc kubenswrapper[4772]: E0127 15:33:55.663662 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:34:04 crc kubenswrapper[4772]: I0127 15:34:04.119733 4772 scope.go:117] "RemoveContainer" containerID="6ba95c7bf22c812cf8d7d855d86c702f5f7f90db05ec7fc2281ddec549f7d67b" Jan 27 15:34:04 crc kubenswrapper[4772]: I0127 15:34:04.153733 4772 scope.go:117] "RemoveContainer" containerID="03f8da2d80772e659c36db9a1b10a6be24dc704eb86ce89c04a5a14351b7726d" Jan 27 15:34:04 crc kubenswrapper[4772]: I0127 15:34:04.178426 4772 scope.go:117] "RemoveContainer" containerID="f76b5eae8b9d1fd746edffe9a9f5a02ca0ad4ea09665e63c5dbeacff4753fa40" Jan 27 15:34:04 crc kubenswrapper[4772]: I0127 15:34:04.195231 4772 scope.go:117] "RemoveContainer" containerID="ed51d0aa4ae1c7166bbf0464f2b405f79a0faa50f99c4244c9717d1a1fd81db2" Jan 27 15:34:04 crc kubenswrapper[4772]: I0127 15:34:04.214759 4772 scope.go:117] "RemoveContainer" containerID="d2b29cba9bcd684a9fa3005c73cbd809102e0bb6c21ef6ed5d53662bb4cdcdaa" Jan 27 15:34:04 crc kubenswrapper[4772]: I0127 15:34:04.237904 4772 scope.go:117] "RemoveContainer" containerID="5de6bd74908b324e47419d9f37b784b689e01e1c833ca0e1c7d7483a1e19037c" Jan 27 15:34:04 crc kubenswrapper[4772]: I0127 15:34:04.277077 4772 scope.go:117] "RemoveContainer" containerID="f581dd644d182efa5f740dc0b5a2f4adfb865bef3f027972802161889179f1d4" Jan 27 15:34:04 crc kubenswrapper[4772]: I0127 15:34:04.310794 4772 scope.go:117] "RemoveContainer" containerID="b87da5e7b978350e6830e0f65fce50644eee1e1665a4ebcd45d4d0010f0f31d7" Jan 27 15:34:09 crc kubenswrapper[4772]: I0127 15:34:09.662759 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:34:09 crc kubenswrapper[4772]: E0127 15:34:09.663327 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.593053 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jm8p7"] Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.593775 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1515626-5d79-408d-abc1-cb92abd58f3f" containerName="galera" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.593786 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1515626-5d79-408d-abc1-cb92abd58f3f" containerName="galera" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.593807 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="swift-recon-cron" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.593813 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="swift-recon-cron" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.593820 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be772158-a71c-448d-8972-014f0d3a9ab8" containerName="cinder-api-log" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.593828 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="be772158-a71c-448d-8972-014f0d3a9ab8" containerName="cinder-api-log" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.593836 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aea5ee43-36e3-437d-8aca-b2faedd87c5b" containerName="ceilometer-notification-agent" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.593842 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="aea5ee43-36e3-437d-8aca-b2faedd87c5b" containerName="ceilometer-notification-agent" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.593853 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="766c2a26-46ea-41b2-ba0c-2101ec9477d5" containerName="barbican-api-log" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.593859 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="766c2a26-46ea-41b2-ba0c-2101ec9477d5" containerName="barbican-api-log" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.593870 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f63bf600-ff03-43a3-92b4-fe8ac68a9bb7" containerName="nova-metadata-metadata" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.593875 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f63bf600-ff03-43a3-92b4-fe8ac68a9bb7" containerName="nova-metadata-metadata" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.593882 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="766c2a26-46ea-41b2-ba0c-2101ec9477d5" containerName="barbican-api" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.593888 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="766c2a26-46ea-41b2-ba0c-2101ec9477d5" containerName="barbican-api" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.593894 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b8101bc-2ddf-48ed-9b92-e8f9e5e71938" containerName="openstack-network-exporter" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.593900 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b8101bc-2ddf-48ed-9b92-e8f9e5e71938" containerName="openstack-network-exporter" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.593908 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76fdbdb1-d48a-4cd1-8372-78887671dce8" containerName="setup-container" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.593914 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="76fdbdb1-d48a-4cd1-8372-78887671dce8" containerName="setup-container" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.593920 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="account-server" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.593926 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="account-server" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.593937 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="container-server" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.593943 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="container-server" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.593948 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="object-expirer" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.593953 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="object-expirer" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.593962 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93c8f9a4-c6ef-42b8-8543-ff8b5347977e" containerName="nova-api-log" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.593967 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="93c8f9a4-c6ef-42b8-8543-ff8b5347977e" containerName="nova-api-log" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.593974 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aea5ee43-36e3-437d-8aca-b2faedd87c5b" containerName="proxy-httpd" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.593980 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="aea5ee43-36e3-437d-8aca-b2faedd87c5b" containerName="proxy-httpd" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.593988 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cbf7469-816d-4e54-a7ad-b5b76d0d59d6" containerName="mariadb-account-create-update" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.593994 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cbf7469-816d-4e54-a7ad-b5b76d0d59d6" containerName="mariadb-account-create-update" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594003 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76fdbdb1-d48a-4cd1-8372-78887671dce8" containerName="rabbitmq" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594009 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="76fdbdb1-d48a-4cd1-8372-78887671dce8" containerName="rabbitmq" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594026 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cf131c4-a5bd-452b-8598-42312c3a0270" containerName="neutron-api" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594032 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cf131c4-a5bd-452b-8598-42312c3a0270" containerName="neutron-api" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594040 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66" containerName="memcached" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594045 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66" containerName="memcached" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594055 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="508c3d5b-212a-46da-9a55-de3f35d7019b" containerName="setup-container" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594061 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="508c3d5b-212a-46da-9a55-de3f35d7019b" containerName="setup-container" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594066 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4205dfea-7dc7-496a-9745-fc5e3d0a418a" containerName="placement-log" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594072 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4205dfea-7dc7-496a-9745-fc5e3d0a418a" containerName="placement-log" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594080 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7e78641-77e6-4c89-b5c9-0d6f3c9a9343" containerName="glance-httpd" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594086 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7e78641-77e6-4c89-b5c9-0d6f3c9a9343" containerName="glance-httpd" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594094 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1515626-5d79-408d-abc1-cb92abd58f3f" containerName="mysql-bootstrap" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594099 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1515626-5d79-408d-abc1-cb92abd58f3f" containerName="mysql-bootstrap" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594106 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="account-replicator" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594112 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="account-replicator" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594123 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="object-updater" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594129 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="object-updater" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594136 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="account-auditor" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594142 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="account-auditor" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594151 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="object-replicator" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594157 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="object-replicator" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594178 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be772158-a71c-448d-8972-014f0d3a9ab8" containerName="cinder-api" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594185 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="be772158-a71c-448d-8972-014f0d3a9ab8" containerName="cinder-api" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594193 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21f54218-5889-4ae9-a7a1-7ed4895ad63c" containerName="kube-state-metrics" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594198 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="21f54218-5889-4ae9-a7a1-7ed4895ad63c" containerName="kube-state-metrics" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594208 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerName="ovsdb-server-init" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594214 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerName="ovsdb-server-init" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594222 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="683f458e-44e9-49ea-a66b-4ac91a3f2bc1" containerName="probe" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594227 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="683f458e-44e9-49ea-a66b-4ac91a3f2bc1" containerName="probe" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594239 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cbf7469-816d-4e54-a7ad-b5b76d0d59d6" containerName="mariadb-account-create-update" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594244 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cbf7469-816d-4e54-a7ad-b5b76d0d59d6" containerName="mariadb-account-create-update" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594252 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f63bf600-ff03-43a3-92b4-fe8ac68a9bb7" containerName="nova-metadata-log" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594258 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f63bf600-ff03-43a3-92b4-fe8ac68a9bb7" containerName="nova-metadata-log" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594267 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="account-reaper" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594273 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="account-reaper" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594281 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="object-server" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594287 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="object-server" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594294 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7e78641-77e6-4c89-b5c9-0d6f3c9a9343" containerName="glance-log" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594300 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7e78641-77e6-4c89-b5c9-0d6f3c9a9343" containerName="glance-log" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594309 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="object-auditor" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594315 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="object-auditor" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594321 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e790127-8223-4b0c-8a5d-21e1bb15fa30" containerName="keystone-api" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594326 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e790127-8223-4b0c-8a5d-21e1bb15fa30" containerName="keystone-api" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594334 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b20b9215-5398-4100-bac4-763daa5ed222" containerName="nova-cell0-conductor-conductor" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594339 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="b20b9215-5398-4100-bac4-763daa5ed222" containerName="nova-cell0-conductor-conductor" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594349 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="rsync" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594354 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="rsync" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594363 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a02b617-28a7-4262-a110-f1c71763ad19" containerName="glance-log" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594369 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a02b617-28a7-4262-a110-f1c71763ad19" containerName="glance-log" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594375 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4205dfea-7dc7-496a-9745-fc5e3d0a418a" containerName="placement-api" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594380 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4205dfea-7dc7-496a-9745-fc5e3d0a418a" containerName="placement-api" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594390 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="683f458e-44e9-49ea-a66b-4ac91a3f2bc1" containerName="cinder-scheduler" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594396 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="683f458e-44e9-49ea-a66b-4ac91a3f2bc1" containerName="cinder-scheduler" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594404 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b8101bc-2ddf-48ed-9b92-e8f9e5e71938" containerName="ovn-northd" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594409 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b8101bc-2ddf-48ed-9b92-e8f9e5e71938" containerName="ovn-northd" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594416 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aea5ee43-36e3-437d-8aca-b2faedd87c5b" containerName="ceilometer-central-agent" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594422 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="aea5ee43-36e3-437d-8aca-b2faedd87c5b" containerName="ceilometer-central-agent" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594431 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="container-replicator" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594437 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="container-replicator" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594443 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aea5ee43-36e3-437d-8aca-b2faedd87c5b" containerName="sg-core" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594448 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="aea5ee43-36e3-437d-8aca-b2faedd87c5b" containerName="sg-core" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594455 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a02b617-28a7-4262-a110-f1c71763ad19" containerName="glance-httpd" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594460 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a02b617-28a7-4262-a110-f1c71763ad19" containerName="glance-httpd" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594470 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b83f7578-8113-46c8-be24-5968aa0ca563" containerName="nova-scheduler-scheduler" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594476 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="b83f7578-8113-46c8-be24-5968aa0ca563" containerName="nova-scheduler-scheduler" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594483 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerName="ovs-vswitchd" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594488 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerName="ovs-vswitchd" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594495 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbbd3c83-3fde-4b11-8ef0-add837d393ce" containerName="nova-cell1-conductor-conductor" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594502 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbbd3c83-3fde-4b11-8ef0-add837d393ce" containerName="nova-cell1-conductor-conductor" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594513 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93c8f9a4-c6ef-42b8-8543-ff8b5347977e" containerName="nova-api-api" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594518 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="93c8f9a4-c6ef-42b8-8543-ff8b5347977e" containerName="nova-api-api" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594529 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="508c3d5b-212a-46da-9a55-de3f35d7019b" containerName="rabbitmq" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594534 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="508c3d5b-212a-46da-9a55-de3f35d7019b" containerName="rabbitmq" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594543 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cf131c4-a5bd-452b-8598-42312c3a0270" containerName="neutron-httpd" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594549 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cf131c4-a5bd-452b-8598-42312c3a0270" containerName="neutron-httpd" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594557 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="container-auditor" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594562 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="container-auditor" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594571 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerName="ovsdb-server" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594577 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerName="ovsdb-server" Jan 27 15:34:14 crc kubenswrapper[4772]: E0127 15:34:14.594583 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="container-updater" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594588 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="container-updater" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594704 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="4205dfea-7dc7-496a-9745-fc5e3d0a418a" containerName="placement-log" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594714 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b8101bc-2ddf-48ed-9b92-e8f9e5e71938" containerName="openstack-network-exporter" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594725 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a02b617-28a7-4262-a110-f1c71763ad19" containerName="glance-httpd" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594733 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="container-replicator" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594742 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1515626-5d79-408d-abc1-cb92abd58f3f" containerName="galera" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594750 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="container-server" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594756 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b8101bc-2ddf-48ed-9b92-e8f9e5e71938" containerName="ovn-northd" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594763 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="container-auditor" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594774 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="swift-recon-cron" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594784 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="683f458e-44e9-49ea-a66b-4ac91a3f2bc1" containerName="cinder-scheduler" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594793 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4ba01b3-fadf-4bc3-9bc3-7c647cfe7e66" containerName="memcached" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594801 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7e78641-77e6-4c89-b5c9-0d6f3c9a9343" containerName="glance-log" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594807 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="object-server" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594816 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="rsync" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594825 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="93c8f9a4-c6ef-42b8-8543-ff8b5347977e" containerName="nova-api-log" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594830 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a02b617-28a7-4262-a110-f1c71763ad19" containerName="glance-log" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594839 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="93c8f9a4-c6ef-42b8-8543-ff8b5347977e" containerName="nova-api-api" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594845 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7e78641-77e6-4c89-b5c9-0d6f3c9a9343" containerName="glance-httpd" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594853 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="account-replicator" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594862 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="be772158-a71c-448d-8972-014f0d3a9ab8" containerName="cinder-api" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594870 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbbd3c83-3fde-4b11-8ef0-add837d393ce" containerName="nova-cell1-conductor-conductor" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594880 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="account-server" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594889 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="aea5ee43-36e3-437d-8aca-b2faedd87c5b" containerName="proxy-httpd" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594898 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="683f458e-44e9-49ea-a66b-4ac91a3f2bc1" containerName="probe" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594908 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="f63bf600-ff03-43a3-92b4-fe8ac68a9bb7" containerName="nova-metadata-log" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594917 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="account-auditor" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594923 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="508c3d5b-212a-46da-9a55-de3f35d7019b" containerName="rabbitmq" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594931 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="object-replicator" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594939 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="object-auditor" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594945 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="aea5ee43-36e3-437d-8aca-b2faedd87c5b" containerName="ceilometer-central-agent" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594953 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="object-updater" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594962 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cf131c4-a5bd-452b-8598-42312c3a0270" containerName="neutron-api" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594970 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerName="ovs-vswitchd" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594977 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="object-expirer" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594984 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="f63bf600-ff03-43a3-92b4-fe8ac68a9bb7" containerName="nova-metadata-metadata" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.594992 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cbf7469-816d-4e54-a7ad-b5b76d0d59d6" containerName="mariadb-account-create-update" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.595000 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cbf7469-816d-4e54-a7ad-b5b76d0d59d6" containerName="mariadb-account-create-update" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.595007 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="b20b9215-5398-4100-bac4-763daa5ed222" containerName="nova-cell0-conductor-conductor" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.595015 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e790127-8223-4b0c-8a5d-21e1bb15fa30" containerName="keystone-api" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.595024 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="4205dfea-7dc7-496a-9745-fc5e3d0a418a" containerName="placement-api" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.595031 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="b83f7578-8113-46c8-be24-5968aa0ca563" containerName="nova-scheduler-scheduler" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.595040 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cf131c4-a5bd-452b-8598-42312c3a0270" containerName="neutron-httpd" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.595048 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="aea5ee43-36e3-437d-8aca-b2faedd87c5b" containerName="sg-core" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.595055 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="container-updater" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.595061 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="766c2a26-46ea-41b2-ba0c-2101ec9477d5" containerName="barbican-api" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.595069 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="76fdbdb1-d48a-4cd1-8372-78887671dce8" containerName="rabbitmq" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.595078 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="21f54218-5889-4ae9-a7a1-7ed4895ad63c" containerName="kube-state-metrics" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.595085 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="766c2a26-46ea-41b2-ba0c-2101ec9477d5" containerName="barbican-api-log" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.595092 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="aea5ee43-36e3-437d-8aca-b2faedd87c5b" containerName="ceilometer-notification-agent" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.595100 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="be772158-a71c-448d-8972-014f0d3a9ab8" containerName="cinder-api-log" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.595108 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="38ebd422-35c5-4682-8a4d-ca9073728d7c" containerName="ovsdb-server" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.595115 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef68955-b80c-4732-9e87-0bec53d0b3a0" containerName="account-reaper" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.596105 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jm8p7" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.604870 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jm8p7"] Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.639515 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72fd7bb8-3c20-4a0a-b0eb-94f4e9059078-catalog-content\") pod \"certified-operators-jm8p7\" (UID: \"72fd7bb8-3c20-4a0a-b0eb-94f4e9059078\") " pod="openshift-marketplace/certified-operators-jm8p7" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.639605 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqkw2\" (UniqueName: \"kubernetes.io/projected/72fd7bb8-3c20-4a0a-b0eb-94f4e9059078-kube-api-access-fqkw2\") pod \"certified-operators-jm8p7\" (UID: \"72fd7bb8-3c20-4a0a-b0eb-94f4e9059078\") " pod="openshift-marketplace/certified-operators-jm8p7" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.639628 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72fd7bb8-3c20-4a0a-b0eb-94f4e9059078-utilities\") pod \"certified-operators-jm8p7\" (UID: \"72fd7bb8-3c20-4a0a-b0eb-94f4e9059078\") " pod="openshift-marketplace/certified-operators-jm8p7" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.741475 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72fd7bb8-3c20-4a0a-b0eb-94f4e9059078-catalog-content\") pod \"certified-operators-jm8p7\" (UID: \"72fd7bb8-3c20-4a0a-b0eb-94f4e9059078\") " pod="openshift-marketplace/certified-operators-jm8p7" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.741554 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqkw2\" (UniqueName: \"kubernetes.io/projected/72fd7bb8-3c20-4a0a-b0eb-94f4e9059078-kube-api-access-fqkw2\") pod \"certified-operators-jm8p7\" (UID: \"72fd7bb8-3c20-4a0a-b0eb-94f4e9059078\") " pod="openshift-marketplace/certified-operators-jm8p7" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.741576 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72fd7bb8-3c20-4a0a-b0eb-94f4e9059078-utilities\") pod \"certified-operators-jm8p7\" (UID: \"72fd7bb8-3c20-4a0a-b0eb-94f4e9059078\") " pod="openshift-marketplace/certified-operators-jm8p7" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.742124 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72fd7bb8-3c20-4a0a-b0eb-94f4e9059078-catalog-content\") pod \"certified-operators-jm8p7\" (UID: \"72fd7bb8-3c20-4a0a-b0eb-94f4e9059078\") " pod="openshift-marketplace/certified-operators-jm8p7" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.742183 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72fd7bb8-3c20-4a0a-b0eb-94f4e9059078-utilities\") pod \"certified-operators-jm8p7\" (UID: \"72fd7bb8-3c20-4a0a-b0eb-94f4e9059078\") " pod="openshift-marketplace/certified-operators-jm8p7" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.762583 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqkw2\" (UniqueName: \"kubernetes.io/projected/72fd7bb8-3c20-4a0a-b0eb-94f4e9059078-kube-api-access-fqkw2\") pod \"certified-operators-jm8p7\" (UID: \"72fd7bb8-3c20-4a0a-b0eb-94f4e9059078\") " pod="openshift-marketplace/certified-operators-jm8p7" Jan 27 15:34:14 crc kubenswrapper[4772]: I0127 15:34:14.929147 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jm8p7" Jan 27 15:34:15 crc kubenswrapper[4772]: I0127 15:34:15.420314 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jm8p7"] Jan 27 15:34:15 crc kubenswrapper[4772]: I0127 15:34:15.517320 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jm8p7" event={"ID":"72fd7bb8-3c20-4a0a-b0eb-94f4e9059078","Type":"ContainerStarted","Data":"8a2e691ba04205978b1b1471c0a34857459b9169f2b1389dbdaf100526709579"} Jan 27 15:34:16 crc kubenswrapper[4772]: I0127 15:34:16.531095 4772 generic.go:334] "Generic (PLEG): container finished" podID="72fd7bb8-3c20-4a0a-b0eb-94f4e9059078" containerID="53008a4df13b2b0f0aaa84c7676410cc85c7c9299c329127da8a6c0db94363b8" exitCode=0 Jan 27 15:34:16 crc kubenswrapper[4772]: I0127 15:34:16.531496 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jm8p7" event={"ID":"72fd7bb8-3c20-4a0a-b0eb-94f4e9059078","Type":"ContainerDied","Data":"53008a4df13b2b0f0aaa84c7676410cc85c7c9299c329127da8a6c0db94363b8"} Jan 27 15:34:19 crc kubenswrapper[4772]: I0127 15:34:19.555778 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jm8p7" event={"ID":"72fd7bb8-3c20-4a0a-b0eb-94f4e9059078","Type":"ContainerStarted","Data":"a0cad38dce3bede1d6985650b216e3e0c1590ff7d27f9064db990057561ca0d9"} Jan 27 15:34:20 crc kubenswrapper[4772]: I0127 15:34:20.570791 4772 generic.go:334] "Generic (PLEG): container finished" podID="72fd7bb8-3c20-4a0a-b0eb-94f4e9059078" containerID="a0cad38dce3bede1d6985650b216e3e0c1590ff7d27f9064db990057561ca0d9" exitCode=0 Jan 27 15:34:20 crc kubenswrapper[4772]: I0127 15:34:20.570911 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jm8p7" event={"ID":"72fd7bb8-3c20-4a0a-b0eb-94f4e9059078","Type":"ContainerDied","Data":"a0cad38dce3bede1d6985650b216e3e0c1590ff7d27f9064db990057561ca0d9"} Jan 27 15:34:21 crc kubenswrapper[4772]: I0127 15:34:21.663060 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:34:21 crc kubenswrapper[4772]: E0127 15:34:21.664917 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:34:22 crc kubenswrapper[4772]: I0127 15:34:22.588625 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jm8p7" event={"ID":"72fd7bb8-3c20-4a0a-b0eb-94f4e9059078","Type":"ContainerStarted","Data":"ec057535eadf435a5c1fbd29551134ac66e1d9025e2bb32c9f6a0f8140dd95ad"} Jan 27 15:34:22 crc kubenswrapper[4772]: I0127 15:34:22.620562 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jm8p7" podStartSLOduration=3.441936655 podStartE2EDuration="8.620525328s" podCreationTimestamp="2026-01-27 15:34:14 +0000 UTC" firstStartedPulling="2026-01-27 15:34:16.533407582 +0000 UTC m=+1642.514016720" lastFinishedPulling="2026-01-27 15:34:21.711996295 +0000 UTC m=+1647.692605393" observedRunningTime="2026-01-27 15:34:22.614740101 +0000 UTC m=+1648.595349239" watchObservedRunningTime="2026-01-27 15:34:22.620525328 +0000 UTC m=+1648.601134466" Jan 27 15:34:24 crc kubenswrapper[4772]: I0127 15:34:24.929894 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jm8p7" Jan 27 15:34:24 crc kubenswrapper[4772]: I0127 15:34:24.929958 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jm8p7" Jan 27 15:34:24 crc kubenswrapper[4772]: I0127 15:34:24.976463 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jm8p7" Jan 27 15:34:34 crc kubenswrapper[4772]: I0127 15:34:34.979241 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jm8p7" Jan 27 15:34:35 crc kubenswrapper[4772]: I0127 15:34:35.040142 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jm8p7"] Jan 27 15:34:35 crc kubenswrapper[4772]: I0127 15:34:35.687462 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jm8p7" podUID="72fd7bb8-3c20-4a0a-b0eb-94f4e9059078" containerName="registry-server" containerID="cri-o://ec057535eadf435a5c1fbd29551134ac66e1d9025e2bb32c9f6a0f8140dd95ad" gracePeriod=2 Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.090198 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jm8p7" Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.115325 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqkw2\" (UniqueName: \"kubernetes.io/projected/72fd7bb8-3c20-4a0a-b0eb-94f4e9059078-kube-api-access-fqkw2\") pod \"72fd7bb8-3c20-4a0a-b0eb-94f4e9059078\" (UID: \"72fd7bb8-3c20-4a0a-b0eb-94f4e9059078\") " Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.115513 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72fd7bb8-3c20-4a0a-b0eb-94f4e9059078-utilities\") pod \"72fd7bb8-3c20-4a0a-b0eb-94f4e9059078\" (UID: \"72fd7bb8-3c20-4a0a-b0eb-94f4e9059078\") " Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.115588 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72fd7bb8-3c20-4a0a-b0eb-94f4e9059078-catalog-content\") pod \"72fd7bb8-3c20-4a0a-b0eb-94f4e9059078\" (UID: \"72fd7bb8-3c20-4a0a-b0eb-94f4e9059078\") " Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.118515 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72fd7bb8-3c20-4a0a-b0eb-94f4e9059078-utilities" (OuterVolumeSpecName: "utilities") pod "72fd7bb8-3c20-4a0a-b0eb-94f4e9059078" (UID: "72fd7bb8-3c20-4a0a-b0eb-94f4e9059078"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.123537 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72fd7bb8-3c20-4a0a-b0eb-94f4e9059078-kube-api-access-fqkw2" (OuterVolumeSpecName: "kube-api-access-fqkw2") pod "72fd7bb8-3c20-4a0a-b0eb-94f4e9059078" (UID: "72fd7bb8-3c20-4a0a-b0eb-94f4e9059078"). InnerVolumeSpecName "kube-api-access-fqkw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.183998 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72fd7bb8-3c20-4a0a-b0eb-94f4e9059078-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "72fd7bb8-3c20-4a0a-b0eb-94f4e9059078" (UID: "72fd7bb8-3c20-4a0a-b0eb-94f4e9059078"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.217299 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72fd7bb8-3c20-4a0a-b0eb-94f4e9059078-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.217617 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72fd7bb8-3c20-4a0a-b0eb-94f4e9059078-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.217688 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqkw2\" (UniqueName: \"kubernetes.io/projected/72fd7bb8-3c20-4a0a-b0eb-94f4e9059078-kube-api-access-fqkw2\") on node \"crc\" DevicePath \"\"" Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.663425 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:34:36 crc kubenswrapper[4772]: E0127 15:34:36.664222 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.698905 4772 generic.go:334] "Generic (PLEG): container finished" podID="72fd7bb8-3c20-4a0a-b0eb-94f4e9059078" containerID="ec057535eadf435a5c1fbd29551134ac66e1d9025e2bb32c9f6a0f8140dd95ad" exitCode=0 Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.698956 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jm8p7" event={"ID":"72fd7bb8-3c20-4a0a-b0eb-94f4e9059078","Type":"ContainerDied","Data":"ec057535eadf435a5c1fbd29551134ac66e1d9025e2bb32c9f6a0f8140dd95ad"} Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.698988 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jm8p7" event={"ID":"72fd7bb8-3c20-4a0a-b0eb-94f4e9059078","Type":"ContainerDied","Data":"8a2e691ba04205978b1b1471c0a34857459b9169f2b1389dbdaf100526709579"} Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.699008 4772 scope.go:117] "RemoveContainer" containerID="ec057535eadf435a5c1fbd29551134ac66e1d9025e2bb32c9f6a0f8140dd95ad" Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.699005 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jm8p7" Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.742339 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jm8p7"] Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.744592 4772 scope.go:117] "RemoveContainer" containerID="a0cad38dce3bede1d6985650b216e3e0c1590ff7d27f9064db990057561ca0d9" Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.748597 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jm8p7"] Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.774533 4772 scope.go:117] "RemoveContainer" containerID="53008a4df13b2b0f0aaa84c7676410cc85c7c9299c329127da8a6c0db94363b8" Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.801493 4772 scope.go:117] "RemoveContainer" containerID="ec057535eadf435a5c1fbd29551134ac66e1d9025e2bb32c9f6a0f8140dd95ad" Jan 27 15:34:36 crc kubenswrapper[4772]: E0127 15:34:36.802296 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec057535eadf435a5c1fbd29551134ac66e1d9025e2bb32c9f6a0f8140dd95ad\": container with ID starting with ec057535eadf435a5c1fbd29551134ac66e1d9025e2bb32c9f6a0f8140dd95ad not found: ID does not exist" containerID="ec057535eadf435a5c1fbd29551134ac66e1d9025e2bb32c9f6a0f8140dd95ad" Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.802365 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec057535eadf435a5c1fbd29551134ac66e1d9025e2bb32c9f6a0f8140dd95ad"} err="failed to get container status \"ec057535eadf435a5c1fbd29551134ac66e1d9025e2bb32c9f6a0f8140dd95ad\": rpc error: code = NotFound desc = could not find container \"ec057535eadf435a5c1fbd29551134ac66e1d9025e2bb32c9f6a0f8140dd95ad\": container with ID starting with ec057535eadf435a5c1fbd29551134ac66e1d9025e2bb32c9f6a0f8140dd95ad not found: ID does not exist" Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.802408 4772 scope.go:117] "RemoveContainer" containerID="a0cad38dce3bede1d6985650b216e3e0c1590ff7d27f9064db990057561ca0d9" Jan 27 15:34:36 crc kubenswrapper[4772]: E0127 15:34:36.803213 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0cad38dce3bede1d6985650b216e3e0c1590ff7d27f9064db990057561ca0d9\": container with ID starting with a0cad38dce3bede1d6985650b216e3e0c1590ff7d27f9064db990057561ca0d9 not found: ID does not exist" containerID="a0cad38dce3bede1d6985650b216e3e0c1590ff7d27f9064db990057561ca0d9" Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.803258 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0cad38dce3bede1d6985650b216e3e0c1590ff7d27f9064db990057561ca0d9"} err="failed to get container status \"a0cad38dce3bede1d6985650b216e3e0c1590ff7d27f9064db990057561ca0d9\": rpc error: code = NotFound desc = could not find container \"a0cad38dce3bede1d6985650b216e3e0c1590ff7d27f9064db990057561ca0d9\": container with ID starting with a0cad38dce3bede1d6985650b216e3e0c1590ff7d27f9064db990057561ca0d9 not found: ID does not exist" Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.803286 4772 scope.go:117] "RemoveContainer" containerID="53008a4df13b2b0f0aaa84c7676410cc85c7c9299c329127da8a6c0db94363b8" Jan 27 15:34:36 crc kubenswrapper[4772]: E0127 15:34:36.804630 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53008a4df13b2b0f0aaa84c7676410cc85c7c9299c329127da8a6c0db94363b8\": container with ID starting with 53008a4df13b2b0f0aaa84c7676410cc85c7c9299c329127da8a6c0db94363b8 not found: ID does not exist" containerID="53008a4df13b2b0f0aaa84c7676410cc85c7c9299c329127da8a6c0db94363b8" Jan 27 15:34:36 crc kubenswrapper[4772]: I0127 15:34:36.804687 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53008a4df13b2b0f0aaa84c7676410cc85c7c9299c329127da8a6c0db94363b8"} err="failed to get container status \"53008a4df13b2b0f0aaa84c7676410cc85c7c9299c329127da8a6c0db94363b8\": rpc error: code = NotFound desc = could not find container \"53008a4df13b2b0f0aaa84c7676410cc85c7c9299c329127da8a6c0db94363b8\": container with ID starting with 53008a4df13b2b0f0aaa84c7676410cc85c7c9299c329127da8a6c0db94363b8 not found: ID does not exist" Jan 27 15:34:38 crc kubenswrapper[4772]: I0127 15:34:38.671443 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72fd7bb8-3c20-4a0a-b0eb-94f4e9059078" path="/var/lib/kubelet/pods/72fd7bb8-3c20-4a0a-b0eb-94f4e9059078/volumes" Jan 27 15:34:51 crc kubenswrapper[4772]: I0127 15:34:51.662799 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:34:51 crc kubenswrapper[4772]: E0127 15:34:51.663565 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:35:04 crc kubenswrapper[4772]: I0127 15:35:04.401908 4772 scope.go:117] "RemoveContainer" containerID="9627fca4ce2bbd20c54de88fa2250d98bc1976636644d325a8225826fd2e9ef2" Jan 27 15:35:04 crc kubenswrapper[4772]: I0127 15:35:04.436621 4772 scope.go:117] "RemoveContainer" containerID="f28ff63f10f8899bc8cd8fd5a42bd4249a187e430ea97934ef6b489554310751" Jan 27 15:35:04 crc kubenswrapper[4772]: I0127 15:35:04.472505 4772 scope.go:117] "RemoveContainer" containerID="4da9288c82c7401f434d2a53ff336e0d653eb3932d204eafc2869a5860cee4bc" Jan 27 15:35:04 crc kubenswrapper[4772]: I0127 15:35:04.537250 4772 scope.go:117] "RemoveContainer" containerID="9853bf54eae9ce0f1c3b8ddee31101fe10bc44f0b0f41d495f936c0ac3cc7ec8" Jan 27 15:35:04 crc kubenswrapper[4772]: I0127 15:35:04.570812 4772 scope.go:117] "RemoveContainer" containerID="48911a4a107b6bf266b45bdb20df360ce0efcf35791daa4bc1413cb966d28fb0" Jan 27 15:35:04 crc kubenswrapper[4772]: I0127 15:35:04.601930 4772 scope.go:117] "RemoveContainer" containerID="46996df047d6fd10b3034c52a93ce3634cebbfdcb4bf44854f66da5e6d342110" Jan 27 15:35:04 crc kubenswrapper[4772]: I0127 15:35:04.631962 4772 scope.go:117] "RemoveContainer" containerID="4c6da56f01306accbad60e3ba02a91f4cc6ed8bb905bd9286671fd7f32153ed5" Jan 27 15:35:04 crc kubenswrapper[4772]: I0127 15:35:04.662147 4772 scope.go:117] "RemoveContainer" containerID="dce84557790ce392eba68b822eea435ede1d05fd9a392c9bd393123a9c7bf467" Jan 27 15:35:04 crc kubenswrapper[4772]: I0127 15:35:04.715545 4772 scope.go:117] "RemoveContainer" containerID="e995550ae720943eacfd405b30c920c20d450c9bc6c2389b27261b188859406e" Jan 27 15:35:04 crc kubenswrapper[4772]: I0127 15:35:04.735953 4772 scope.go:117] "RemoveContainer" containerID="e1b312b7631d415f567909a3003da4cdfd7208b6894d1397aa7da34098746b5a" Jan 27 15:35:04 crc kubenswrapper[4772]: I0127 15:35:04.760654 4772 scope.go:117] "RemoveContainer" containerID="35964dfe2e497930630aeb0996d17bf7bbe0e9d5e7bfb1d7efca05167ac578fc" Jan 27 15:35:06 crc kubenswrapper[4772]: I0127 15:35:06.438506 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mcqtt"] Jan 27 15:35:06 crc kubenswrapper[4772]: E0127 15:35:06.439357 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72fd7bb8-3c20-4a0a-b0eb-94f4e9059078" containerName="extract-content" Jan 27 15:35:06 crc kubenswrapper[4772]: I0127 15:35:06.439379 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="72fd7bb8-3c20-4a0a-b0eb-94f4e9059078" containerName="extract-content" Jan 27 15:35:06 crc kubenswrapper[4772]: E0127 15:35:06.439414 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72fd7bb8-3c20-4a0a-b0eb-94f4e9059078" containerName="registry-server" Jan 27 15:35:06 crc kubenswrapper[4772]: I0127 15:35:06.439426 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="72fd7bb8-3c20-4a0a-b0eb-94f4e9059078" containerName="registry-server" Jan 27 15:35:06 crc kubenswrapper[4772]: E0127 15:35:06.439453 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72fd7bb8-3c20-4a0a-b0eb-94f4e9059078" containerName="extract-utilities" Jan 27 15:35:06 crc kubenswrapper[4772]: I0127 15:35:06.439466 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="72fd7bb8-3c20-4a0a-b0eb-94f4e9059078" containerName="extract-utilities" Jan 27 15:35:06 crc kubenswrapper[4772]: I0127 15:35:06.439732 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="72fd7bb8-3c20-4a0a-b0eb-94f4e9059078" containerName="registry-server" Jan 27 15:35:06 crc kubenswrapper[4772]: I0127 15:35:06.441469 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mcqtt" Jan 27 15:35:06 crc kubenswrapper[4772]: I0127 15:35:06.442454 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mcqtt"] Jan 27 15:35:06 crc kubenswrapper[4772]: I0127 15:35:06.539318 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a-catalog-content\") pod \"community-operators-mcqtt\" (UID: \"ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a\") " pod="openshift-marketplace/community-operators-mcqtt" Jan 27 15:35:06 crc kubenswrapper[4772]: I0127 15:35:06.539376 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a-utilities\") pod \"community-operators-mcqtt\" (UID: \"ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a\") " pod="openshift-marketplace/community-operators-mcqtt" Jan 27 15:35:06 crc kubenswrapper[4772]: I0127 15:35:06.539407 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxnbq\" (UniqueName: \"kubernetes.io/projected/ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a-kube-api-access-dxnbq\") pod \"community-operators-mcqtt\" (UID: \"ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a\") " pod="openshift-marketplace/community-operators-mcqtt" Jan 27 15:35:06 crc kubenswrapper[4772]: I0127 15:35:06.640323 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a-catalog-content\") pod \"community-operators-mcqtt\" (UID: \"ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a\") " pod="openshift-marketplace/community-operators-mcqtt" Jan 27 15:35:06 crc kubenswrapper[4772]: I0127 15:35:06.640377 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a-utilities\") pod \"community-operators-mcqtt\" (UID: \"ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a\") " pod="openshift-marketplace/community-operators-mcqtt" Jan 27 15:35:06 crc kubenswrapper[4772]: I0127 15:35:06.640404 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxnbq\" (UniqueName: \"kubernetes.io/projected/ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a-kube-api-access-dxnbq\") pod \"community-operators-mcqtt\" (UID: \"ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a\") " pod="openshift-marketplace/community-operators-mcqtt" Jan 27 15:35:06 crc kubenswrapper[4772]: I0127 15:35:06.640918 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a-catalog-content\") pod \"community-operators-mcqtt\" (UID: \"ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a\") " pod="openshift-marketplace/community-operators-mcqtt" Jan 27 15:35:06 crc kubenswrapper[4772]: I0127 15:35:06.640973 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a-utilities\") pod \"community-operators-mcqtt\" (UID: \"ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a\") " pod="openshift-marketplace/community-operators-mcqtt" Jan 27 15:35:06 crc kubenswrapper[4772]: I0127 15:35:06.663253 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:35:06 crc kubenswrapper[4772]: E0127 15:35:06.663488 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:35:06 crc kubenswrapper[4772]: I0127 15:35:06.672035 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxnbq\" (UniqueName: \"kubernetes.io/projected/ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a-kube-api-access-dxnbq\") pod \"community-operators-mcqtt\" (UID: \"ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a\") " pod="openshift-marketplace/community-operators-mcqtt" Jan 27 15:35:06 crc kubenswrapper[4772]: I0127 15:35:06.783741 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mcqtt" Jan 27 15:35:07 crc kubenswrapper[4772]: I0127 15:35:07.261264 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mcqtt"] Jan 27 15:35:07 crc kubenswrapper[4772]: I0127 15:35:07.978992 4772 generic.go:334] "Generic (PLEG): container finished" podID="ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a" containerID="91ae6cb76970d4cc89de236593aee41408991154309c0dac8ac8245c02925dfb" exitCode=0 Jan 27 15:35:07 crc kubenswrapper[4772]: I0127 15:35:07.979047 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mcqtt" event={"ID":"ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a","Type":"ContainerDied","Data":"91ae6cb76970d4cc89de236593aee41408991154309c0dac8ac8245c02925dfb"} Jan 27 15:35:07 crc kubenswrapper[4772]: I0127 15:35:07.979113 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mcqtt" event={"ID":"ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a","Type":"ContainerStarted","Data":"72909f838bde105e1566f87719019a2bb56b2533e276fbb1c841e04408b58a81"} Jan 27 15:35:07 crc kubenswrapper[4772]: I0127 15:35:07.981223 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 15:35:08 crc kubenswrapper[4772]: I0127 15:35:08.996990 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mcqtt" event={"ID":"ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a","Type":"ContainerStarted","Data":"4d6928d9c333649d0fd45228aacd272e5472427e08d477b556c71fa0f31fd418"} Jan 27 15:35:10 crc kubenswrapper[4772]: I0127 15:35:10.005659 4772 generic.go:334] "Generic (PLEG): container finished" podID="ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a" containerID="4d6928d9c333649d0fd45228aacd272e5472427e08d477b556c71fa0f31fd418" exitCode=0 Jan 27 15:35:10 crc kubenswrapper[4772]: I0127 15:35:10.005709 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mcqtt" event={"ID":"ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a","Type":"ContainerDied","Data":"4d6928d9c333649d0fd45228aacd272e5472427e08d477b556c71fa0f31fd418"} Jan 27 15:35:11 crc kubenswrapper[4772]: I0127 15:35:11.031534 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mcqtt" event={"ID":"ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a","Type":"ContainerStarted","Data":"9768bde0eb8c6061a919ebcea3ee20bb86649ff6c879aaef71e49fd52c6736fe"} Jan 27 15:35:11 crc kubenswrapper[4772]: I0127 15:35:11.057282 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mcqtt" podStartSLOduration=2.443619788 podStartE2EDuration="5.057262192s" podCreationTimestamp="2026-01-27 15:35:06 +0000 UTC" firstStartedPulling="2026-01-27 15:35:07.980799129 +0000 UTC m=+1693.961408237" lastFinishedPulling="2026-01-27 15:35:10.594441533 +0000 UTC m=+1696.575050641" observedRunningTime="2026-01-27 15:35:11.051768214 +0000 UTC m=+1697.032377332" watchObservedRunningTime="2026-01-27 15:35:11.057262192 +0000 UTC m=+1697.037871290" Jan 27 15:35:16 crc kubenswrapper[4772]: I0127 15:35:16.784211 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mcqtt" Jan 27 15:35:16 crc kubenswrapper[4772]: I0127 15:35:16.784854 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mcqtt" Jan 27 15:35:16 crc kubenswrapper[4772]: I0127 15:35:16.855607 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mcqtt" Jan 27 15:35:17 crc kubenswrapper[4772]: I0127 15:35:17.147770 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mcqtt" Jan 27 15:35:17 crc kubenswrapper[4772]: I0127 15:35:17.207863 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mcqtt"] Jan 27 15:35:19 crc kubenswrapper[4772]: I0127 15:35:19.097963 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mcqtt" podUID="ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a" containerName="registry-server" containerID="cri-o://9768bde0eb8c6061a919ebcea3ee20bb86649ff6c879aaef71e49fd52c6736fe" gracePeriod=2 Jan 27 15:35:20 crc kubenswrapper[4772]: I0127 15:35:20.112432 4772 generic.go:334] "Generic (PLEG): container finished" podID="ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a" containerID="9768bde0eb8c6061a919ebcea3ee20bb86649ff6c879aaef71e49fd52c6736fe" exitCode=0 Jan 27 15:35:20 crc kubenswrapper[4772]: I0127 15:35:20.112537 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mcqtt" event={"ID":"ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a","Type":"ContainerDied","Data":"9768bde0eb8c6061a919ebcea3ee20bb86649ff6c879aaef71e49fd52c6736fe"} Jan 27 15:35:20 crc kubenswrapper[4772]: I0127 15:35:20.663651 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:35:20 crc kubenswrapper[4772]: E0127 15:35:20.664082 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:35:20 crc kubenswrapper[4772]: I0127 15:35:20.871407 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mcqtt" Jan 27 15:35:20 crc kubenswrapper[4772]: I0127 15:35:20.973442 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a-catalog-content\") pod \"ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a\" (UID: \"ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a\") " Jan 27 15:35:20 crc kubenswrapper[4772]: I0127 15:35:20.973863 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a-utilities\") pod \"ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a\" (UID: \"ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a\") " Jan 27 15:35:20 crc kubenswrapper[4772]: I0127 15:35:20.974245 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxnbq\" (UniqueName: \"kubernetes.io/projected/ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a-kube-api-access-dxnbq\") pod \"ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a\" (UID: \"ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a\") " Jan 27 15:35:20 crc kubenswrapper[4772]: I0127 15:35:20.978277 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a-utilities" (OuterVolumeSpecName: "utilities") pod "ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a" (UID: "ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:35:20 crc kubenswrapper[4772]: I0127 15:35:20.998557 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a-kube-api-access-dxnbq" (OuterVolumeSpecName: "kube-api-access-dxnbq") pod "ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a" (UID: "ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a"). InnerVolumeSpecName "kube-api-access-dxnbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:35:21 crc kubenswrapper[4772]: I0127 15:35:21.077140 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxnbq\" (UniqueName: \"kubernetes.io/projected/ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a-kube-api-access-dxnbq\") on node \"crc\" DevicePath \"\"" Jan 27 15:35:21 crc kubenswrapper[4772]: I0127 15:35:21.077201 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 15:35:21 crc kubenswrapper[4772]: I0127 15:35:21.127357 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mcqtt" event={"ID":"ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a","Type":"ContainerDied","Data":"72909f838bde105e1566f87719019a2bb56b2533e276fbb1c841e04408b58a81"} Jan 27 15:35:21 crc kubenswrapper[4772]: I0127 15:35:21.128299 4772 scope.go:117] "RemoveContainer" containerID="9768bde0eb8c6061a919ebcea3ee20bb86649ff6c879aaef71e49fd52c6736fe" Jan 27 15:35:21 crc kubenswrapper[4772]: I0127 15:35:21.128530 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mcqtt" Jan 27 15:35:21 crc kubenswrapper[4772]: I0127 15:35:21.150010 4772 scope.go:117] "RemoveContainer" containerID="4d6928d9c333649d0fd45228aacd272e5472427e08d477b556c71fa0f31fd418" Jan 27 15:35:21 crc kubenswrapper[4772]: I0127 15:35:21.168681 4772 scope.go:117] "RemoveContainer" containerID="91ae6cb76970d4cc89de236593aee41408991154309c0dac8ac8245c02925dfb" Jan 27 15:35:21 crc kubenswrapper[4772]: I0127 15:35:21.300030 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a" (UID: "ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:35:21 crc kubenswrapper[4772]: I0127 15:35:21.380734 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 15:35:21 crc kubenswrapper[4772]: I0127 15:35:21.460249 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mcqtt"] Jan 27 15:35:21 crc kubenswrapper[4772]: I0127 15:35:21.467031 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mcqtt"] Jan 27 15:35:22 crc kubenswrapper[4772]: I0127 15:35:22.674791 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a" path="/var/lib/kubelet/pods/ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a/volumes" Jan 27 15:35:31 crc kubenswrapper[4772]: I0127 15:35:31.663062 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:35:31 crc kubenswrapper[4772]: E0127 15:35:31.663904 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:35:44 crc kubenswrapper[4772]: I0127 15:35:44.667734 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:35:44 crc kubenswrapper[4772]: E0127 15:35:44.668669 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:35:58 crc kubenswrapper[4772]: I0127 15:35:58.664505 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:35:58 crc kubenswrapper[4772]: E0127 15:35:58.665833 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:36:04 crc kubenswrapper[4772]: I0127 15:36:04.985567 4772 scope.go:117] "RemoveContainer" containerID="e1df482be0829e766abad6c9eb6842ba0e9d9f6fb517127a47f819ce8b296c7d" Jan 27 15:36:05 crc kubenswrapper[4772]: I0127 15:36:05.053817 4772 scope.go:117] "RemoveContainer" containerID="dfac9725f6b1a542e430af600747ad2b7e4c5c445357868cef0cc0fe2f4dae49" Jan 27 15:36:05 crc kubenswrapper[4772]: I0127 15:36:05.075364 4772 scope.go:117] "RemoveContainer" containerID="bea9ecc5c8bd7f22996f379a16987a5468d25478afcbfdd986751cd73382ded7" Jan 27 15:36:09 crc kubenswrapper[4772]: I0127 15:36:09.663784 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:36:09 crc kubenswrapper[4772]: E0127 15:36:09.664349 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:36:24 crc kubenswrapper[4772]: I0127 15:36:24.668697 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:36:24 crc kubenswrapper[4772]: E0127 15:36:24.669523 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:36:36 crc kubenswrapper[4772]: I0127 15:36:36.663599 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:36:36 crc kubenswrapper[4772]: E0127 15:36:36.664896 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:36:51 crc kubenswrapper[4772]: I0127 15:36:51.662871 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:36:51 crc kubenswrapper[4772]: E0127 15:36:51.663663 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:37:05 crc kubenswrapper[4772]: I0127 15:37:05.214871 4772 scope.go:117] "RemoveContainer" containerID="7c4a200cbf0e299c55e6a175696503b17a34325960db8f2addd09db07bdebe00" Jan 27 15:37:06 crc kubenswrapper[4772]: I0127 15:37:06.663714 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:37:06 crc kubenswrapper[4772]: E0127 15:37:06.664312 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:37:19 crc kubenswrapper[4772]: I0127 15:37:19.664155 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:37:19 crc kubenswrapper[4772]: E0127 15:37:19.665773 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:37:31 crc kubenswrapper[4772]: I0127 15:37:31.662984 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:37:31 crc kubenswrapper[4772]: E0127 15:37:31.663771 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:37:44 crc kubenswrapper[4772]: I0127 15:37:44.669816 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:37:45 crc kubenswrapper[4772]: I0127 15:37:45.204604 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"7b761d6bc884bc3f2c2a56d23c68b1cd3740e77326d275aa98b5b71fcaad6f52"} Jan 27 15:39:28 crc kubenswrapper[4772]: I0127 15:39:28.377734 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7f4lg"] Jan 27 15:39:28 crc kubenswrapper[4772]: E0127 15:39:28.378731 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a" containerName="extract-utilities" Jan 27 15:39:28 crc kubenswrapper[4772]: I0127 15:39:28.378747 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a" containerName="extract-utilities" Jan 27 15:39:28 crc kubenswrapper[4772]: E0127 15:39:28.378761 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a" containerName="registry-server" Jan 27 15:39:28 crc kubenswrapper[4772]: I0127 15:39:28.378771 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a" containerName="registry-server" Jan 27 15:39:28 crc kubenswrapper[4772]: E0127 15:39:28.378783 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a" containerName="extract-content" Jan 27 15:39:28 crc kubenswrapper[4772]: I0127 15:39:28.378791 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a" containerName="extract-content" Jan 27 15:39:28 crc kubenswrapper[4772]: I0127 15:39:28.378969 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae1aaeb5-4ac9-42fc-8a59-a47ae8baf54a" containerName="registry-server" Jan 27 15:39:28 crc kubenswrapper[4772]: I0127 15:39:28.380257 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7f4lg" Jan 27 15:39:28 crc kubenswrapper[4772]: I0127 15:39:28.403987 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7f4lg"] Jan 27 15:39:28 crc kubenswrapper[4772]: I0127 15:39:28.457783 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3e87a75-cf2d-4ee1-b89b-75cbce1a3041-utilities\") pod \"redhat-operators-7f4lg\" (UID: \"f3e87a75-cf2d-4ee1-b89b-75cbce1a3041\") " pod="openshift-marketplace/redhat-operators-7f4lg" Jan 27 15:39:28 crc kubenswrapper[4772]: I0127 15:39:28.457849 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3e87a75-cf2d-4ee1-b89b-75cbce1a3041-catalog-content\") pod \"redhat-operators-7f4lg\" (UID: \"f3e87a75-cf2d-4ee1-b89b-75cbce1a3041\") " pod="openshift-marketplace/redhat-operators-7f4lg" Jan 27 15:39:28 crc kubenswrapper[4772]: I0127 15:39:28.457871 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsj5g\" (UniqueName: \"kubernetes.io/projected/f3e87a75-cf2d-4ee1-b89b-75cbce1a3041-kube-api-access-vsj5g\") pod \"redhat-operators-7f4lg\" (UID: \"f3e87a75-cf2d-4ee1-b89b-75cbce1a3041\") " pod="openshift-marketplace/redhat-operators-7f4lg" Jan 27 15:39:28 crc kubenswrapper[4772]: I0127 15:39:28.559341 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3e87a75-cf2d-4ee1-b89b-75cbce1a3041-utilities\") pod \"redhat-operators-7f4lg\" (UID: \"f3e87a75-cf2d-4ee1-b89b-75cbce1a3041\") " pod="openshift-marketplace/redhat-operators-7f4lg" Jan 27 15:39:28 crc kubenswrapper[4772]: I0127 15:39:28.559405 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3e87a75-cf2d-4ee1-b89b-75cbce1a3041-catalog-content\") pod \"redhat-operators-7f4lg\" (UID: \"f3e87a75-cf2d-4ee1-b89b-75cbce1a3041\") " pod="openshift-marketplace/redhat-operators-7f4lg" Jan 27 15:39:28 crc kubenswrapper[4772]: I0127 15:39:28.559428 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsj5g\" (UniqueName: \"kubernetes.io/projected/f3e87a75-cf2d-4ee1-b89b-75cbce1a3041-kube-api-access-vsj5g\") pod \"redhat-operators-7f4lg\" (UID: \"f3e87a75-cf2d-4ee1-b89b-75cbce1a3041\") " pod="openshift-marketplace/redhat-operators-7f4lg" Jan 27 15:39:28 crc kubenswrapper[4772]: I0127 15:39:28.560069 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3e87a75-cf2d-4ee1-b89b-75cbce1a3041-catalog-content\") pod \"redhat-operators-7f4lg\" (UID: \"f3e87a75-cf2d-4ee1-b89b-75cbce1a3041\") " pod="openshift-marketplace/redhat-operators-7f4lg" Jan 27 15:39:28 crc kubenswrapper[4772]: I0127 15:39:28.560409 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3e87a75-cf2d-4ee1-b89b-75cbce1a3041-utilities\") pod \"redhat-operators-7f4lg\" (UID: \"f3e87a75-cf2d-4ee1-b89b-75cbce1a3041\") " pod="openshift-marketplace/redhat-operators-7f4lg" Jan 27 15:39:28 crc kubenswrapper[4772]: I0127 15:39:28.583821 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsj5g\" (UniqueName: \"kubernetes.io/projected/f3e87a75-cf2d-4ee1-b89b-75cbce1a3041-kube-api-access-vsj5g\") pod \"redhat-operators-7f4lg\" (UID: \"f3e87a75-cf2d-4ee1-b89b-75cbce1a3041\") " pod="openshift-marketplace/redhat-operators-7f4lg" Jan 27 15:39:28 crc kubenswrapper[4772]: I0127 15:39:28.702231 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7f4lg" Jan 27 15:39:29 crc kubenswrapper[4772]: I0127 15:39:29.150487 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7f4lg"] Jan 27 15:39:29 crc kubenswrapper[4772]: I0127 15:39:29.958888 4772 generic.go:334] "Generic (PLEG): container finished" podID="f3e87a75-cf2d-4ee1-b89b-75cbce1a3041" containerID="096fb8c1b8d3fbe84b8c8656f135ae1fef6771b626847166ee451c948bc66ce7" exitCode=0 Jan 27 15:39:29 crc kubenswrapper[4772]: I0127 15:39:29.959037 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f4lg" event={"ID":"f3e87a75-cf2d-4ee1-b89b-75cbce1a3041","Type":"ContainerDied","Data":"096fb8c1b8d3fbe84b8c8656f135ae1fef6771b626847166ee451c948bc66ce7"} Jan 27 15:39:29 crc kubenswrapper[4772]: I0127 15:39:29.959512 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f4lg" event={"ID":"f3e87a75-cf2d-4ee1-b89b-75cbce1a3041","Type":"ContainerStarted","Data":"55f822061f7b79771202dd86ea827bfce4a99cda83a00c362905aef55a31aa17"} Jan 27 15:39:30 crc kubenswrapper[4772]: I0127 15:39:30.969876 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f4lg" event={"ID":"f3e87a75-cf2d-4ee1-b89b-75cbce1a3041","Type":"ContainerStarted","Data":"cef5d3c9dde46ed19fedc1688b6a3a0442d90768aa13b5b1c866559b791cff2c"} Jan 27 15:39:31 crc kubenswrapper[4772]: I0127 15:39:31.978484 4772 generic.go:334] "Generic (PLEG): container finished" podID="f3e87a75-cf2d-4ee1-b89b-75cbce1a3041" containerID="cef5d3c9dde46ed19fedc1688b6a3a0442d90768aa13b5b1c866559b791cff2c" exitCode=0 Jan 27 15:39:31 crc kubenswrapper[4772]: I0127 15:39:31.978546 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f4lg" event={"ID":"f3e87a75-cf2d-4ee1-b89b-75cbce1a3041","Type":"ContainerDied","Data":"cef5d3c9dde46ed19fedc1688b6a3a0442d90768aa13b5b1c866559b791cff2c"} Jan 27 15:39:32 crc kubenswrapper[4772]: I0127 15:39:32.989432 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f4lg" event={"ID":"f3e87a75-cf2d-4ee1-b89b-75cbce1a3041","Type":"ContainerStarted","Data":"098817128bd0191e6d2a586cabdbaf5ba95e52947cc3cb768cf31df336960f75"} Jan 27 15:39:33 crc kubenswrapper[4772]: I0127 15:39:33.009714 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7f4lg" podStartSLOduration=2.462876389 podStartE2EDuration="5.009694064s" podCreationTimestamp="2026-01-27 15:39:28 +0000 UTC" firstStartedPulling="2026-01-27 15:39:29.962101945 +0000 UTC m=+1955.942711053" lastFinishedPulling="2026-01-27 15:39:32.50891963 +0000 UTC m=+1958.489528728" observedRunningTime="2026-01-27 15:39:33.008686394 +0000 UTC m=+1958.989295492" watchObservedRunningTime="2026-01-27 15:39:33.009694064 +0000 UTC m=+1958.990303162" Jan 27 15:39:38 crc kubenswrapper[4772]: I0127 15:39:38.703550 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7f4lg" Jan 27 15:39:38 crc kubenswrapper[4772]: I0127 15:39:38.704211 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7f4lg" Jan 27 15:39:38 crc kubenswrapper[4772]: I0127 15:39:38.763075 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7f4lg" Jan 27 15:39:39 crc kubenswrapper[4772]: I0127 15:39:39.077790 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7f4lg" Jan 27 15:39:39 crc kubenswrapper[4772]: I0127 15:39:39.122522 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7f4lg"] Jan 27 15:39:41 crc kubenswrapper[4772]: I0127 15:39:41.047291 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7f4lg" podUID="f3e87a75-cf2d-4ee1-b89b-75cbce1a3041" containerName="registry-server" containerID="cri-o://098817128bd0191e6d2a586cabdbaf5ba95e52947cc3cb768cf31df336960f75" gracePeriod=2 Jan 27 15:39:41 crc kubenswrapper[4772]: I0127 15:39:41.474541 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7f4lg" Jan 27 15:39:41 crc kubenswrapper[4772]: I0127 15:39:41.551826 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsj5g\" (UniqueName: \"kubernetes.io/projected/f3e87a75-cf2d-4ee1-b89b-75cbce1a3041-kube-api-access-vsj5g\") pod \"f3e87a75-cf2d-4ee1-b89b-75cbce1a3041\" (UID: \"f3e87a75-cf2d-4ee1-b89b-75cbce1a3041\") " Jan 27 15:39:41 crc kubenswrapper[4772]: I0127 15:39:41.551893 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3e87a75-cf2d-4ee1-b89b-75cbce1a3041-utilities\") pod \"f3e87a75-cf2d-4ee1-b89b-75cbce1a3041\" (UID: \"f3e87a75-cf2d-4ee1-b89b-75cbce1a3041\") " Jan 27 15:39:41 crc kubenswrapper[4772]: I0127 15:39:41.551939 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3e87a75-cf2d-4ee1-b89b-75cbce1a3041-catalog-content\") pod \"f3e87a75-cf2d-4ee1-b89b-75cbce1a3041\" (UID: \"f3e87a75-cf2d-4ee1-b89b-75cbce1a3041\") " Jan 27 15:39:41 crc kubenswrapper[4772]: I0127 15:39:41.552970 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3e87a75-cf2d-4ee1-b89b-75cbce1a3041-utilities" (OuterVolumeSpecName: "utilities") pod "f3e87a75-cf2d-4ee1-b89b-75cbce1a3041" (UID: "f3e87a75-cf2d-4ee1-b89b-75cbce1a3041"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:39:41 crc kubenswrapper[4772]: I0127 15:39:41.557063 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3e87a75-cf2d-4ee1-b89b-75cbce1a3041-kube-api-access-vsj5g" (OuterVolumeSpecName: "kube-api-access-vsj5g") pod "f3e87a75-cf2d-4ee1-b89b-75cbce1a3041" (UID: "f3e87a75-cf2d-4ee1-b89b-75cbce1a3041"). InnerVolumeSpecName "kube-api-access-vsj5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:39:41 crc kubenswrapper[4772]: I0127 15:39:41.653727 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsj5g\" (UniqueName: \"kubernetes.io/projected/f3e87a75-cf2d-4ee1-b89b-75cbce1a3041-kube-api-access-vsj5g\") on node \"crc\" DevicePath \"\"" Jan 27 15:39:41 crc kubenswrapper[4772]: I0127 15:39:41.653767 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3e87a75-cf2d-4ee1-b89b-75cbce1a3041-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 15:39:41 crc kubenswrapper[4772]: I0127 15:39:41.675835 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3e87a75-cf2d-4ee1-b89b-75cbce1a3041-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f3e87a75-cf2d-4ee1-b89b-75cbce1a3041" (UID: "f3e87a75-cf2d-4ee1-b89b-75cbce1a3041"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:39:41 crc kubenswrapper[4772]: I0127 15:39:41.760368 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3e87a75-cf2d-4ee1-b89b-75cbce1a3041-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 15:39:42 crc kubenswrapper[4772]: I0127 15:39:42.059444 4772 generic.go:334] "Generic (PLEG): container finished" podID="f3e87a75-cf2d-4ee1-b89b-75cbce1a3041" containerID="098817128bd0191e6d2a586cabdbaf5ba95e52947cc3cb768cf31df336960f75" exitCode=0 Jan 27 15:39:42 crc kubenswrapper[4772]: I0127 15:39:42.059506 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f4lg" event={"ID":"f3e87a75-cf2d-4ee1-b89b-75cbce1a3041","Type":"ContainerDied","Data":"098817128bd0191e6d2a586cabdbaf5ba95e52947cc3cb768cf31df336960f75"} Jan 27 15:39:42 crc kubenswrapper[4772]: I0127 15:39:42.059544 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7f4lg" event={"ID":"f3e87a75-cf2d-4ee1-b89b-75cbce1a3041","Type":"ContainerDied","Data":"55f822061f7b79771202dd86ea827bfce4a99cda83a00c362905aef55a31aa17"} Jan 27 15:39:42 crc kubenswrapper[4772]: I0127 15:39:42.059570 4772 scope.go:117] "RemoveContainer" containerID="098817128bd0191e6d2a586cabdbaf5ba95e52947cc3cb768cf31df336960f75" Jan 27 15:39:42 crc kubenswrapper[4772]: I0127 15:39:42.059613 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7f4lg" Jan 27 15:39:42 crc kubenswrapper[4772]: I0127 15:39:42.086840 4772 scope.go:117] "RemoveContainer" containerID="cef5d3c9dde46ed19fedc1688b6a3a0442d90768aa13b5b1c866559b791cff2c" Jan 27 15:39:42 crc kubenswrapper[4772]: I0127 15:39:42.114296 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7f4lg"] Jan 27 15:39:42 crc kubenswrapper[4772]: I0127 15:39:42.120580 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7f4lg"] Jan 27 15:39:42 crc kubenswrapper[4772]: I0127 15:39:42.138051 4772 scope.go:117] "RemoveContainer" containerID="096fb8c1b8d3fbe84b8c8656f135ae1fef6771b626847166ee451c948bc66ce7" Jan 27 15:39:42 crc kubenswrapper[4772]: I0127 15:39:42.162100 4772 scope.go:117] "RemoveContainer" containerID="098817128bd0191e6d2a586cabdbaf5ba95e52947cc3cb768cf31df336960f75" Jan 27 15:39:42 crc kubenswrapper[4772]: E0127 15:39:42.162839 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"098817128bd0191e6d2a586cabdbaf5ba95e52947cc3cb768cf31df336960f75\": container with ID starting with 098817128bd0191e6d2a586cabdbaf5ba95e52947cc3cb768cf31df336960f75 not found: ID does not exist" containerID="098817128bd0191e6d2a586cabdbaf5ba95e52947cc3cb768cf31df336960f75" Jan 27 15:39:42 crc kubenswrapper[4772]: I0127 15:39:42.162914 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"098817128bd0191e6d2a586cabdbaf5ba95e52947cc3cb768cf31df336960f75"} err="failed to get container status \"098817128bd0191e6d2a586cabdbaf5ba95e52947cc3cb768cf31df336960f75\": rpc error: code = NotFound desc = could not find container \"098817128bd0191e6d2a586cabdbaf5ba95e52947cc3cb768cf31df336960f75\": container with ID starting with 098817128bd0191e6d2a586cabdbaf5ba95e52947cc3cb768cf31df336960f75 not found: ID does not exist" Jan 27 15:39:42 crc kubenswrapper[4772]: I0127 15:39:42.162973 4772 scope.go:117] "RemoveContainer" containerID="cef5d3c9dde46ed19fedc1688b6a3a0442d90768aa13b5b1c866559b791cff2c" Jan 27 15:39:42 crc kubenswrapper[4772]: E0127 15:39:42.163618 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cef5d3c9dde46ed19fedc1688b6a3a0442d90768aa13b5b1c866559b791cff2c\": container with ID starting with cef5d3c9dde46ed19fedc1688b6a3a0442d90768aa13b5b1c866559b791cff2c not found: ID does not exist" containerID="cef5d3c9dde46ed19fedc1688b6a3a0442d90768aa13b5b1c866559b791cff2c" Jan 27 15:39:42 crc kubenswrapper[4772]: I0127 15:39:42.163659 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cef5d3c9dde46ed19fedc1688b6a3a0442d90768aa13b5b1c866559b791cff2c"} err="failed to get container status \"cef5d3c9dde46ed19fedc1688b6a3a0442d90768aa13b5b1c866559b791cff2c\": rpc error: code = NotFound desc = could not find container \"cef5d3c9dde46ed19fedc1688b6a3a0442d90768aa13b5b1c866559b791cff2c\": container with ID starting with cef5d3c9dde46ed19fedc1688b6a3a0442d90768aa13b5b1c866559b791cff2c not found: ID does not exist" Jan 27 15:39:42 crc kubenswrapper[4772]: I0127 15:39:42.163690 4772 scope.go:117] "RemoveContainer" containerID="096fb8c1b8d3fbe84b8c8656f135ae1fef6771b626847166ee451c948bc66ce7" Jan 27 15:39:42 crc kubenswrapper[4772]: E0127 15:39:42.164270 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"096fb8c1b8d3fbe84b8c8656f135ae1fef6771b626847166ee451c948bc66ce7\": container with ID starting with 096fb8c1b8d3fbe84b8c8656f135ae1fef6771b626847166ee451c948bc66ce7 not found: ID does not exist" containerID="096fb8c1b8d3fbe84b8c8656f135ae1fef6771b626847166ee451c948bc66ce7" Jan 27 15:39:42 crc kubenswrapper[4772]: I0127 15:39:42.164342 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"096fb8c1b8d3fbe84b8c8656f135ae1fef6771b626847166ee451c948bc66ce7"} err="failed to get container status \"096fb8c1b8d3fbe84b8c8656f135ae1fef6771b626847166ee451c948bc66ce7\": rpc error: code = NotFound desc = could not find container \"096fb8c1b8d3fbe84b8c8656f135ae1fef6771b626847166ee451c948bc66ce7\": container with ID starting with 096fb8c1b8d3fbe84b8c8656f135ae1fef6771b626847166ee451c948bc66ce7 not found: ID does not exist" Jan 27 15:39:42 crc kubenswrapper[4772]: I0127 15:39:42.672317 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3e87a75-cf2d-4ee1-b89b-75cbce1a3041" path="/var/lib/kubelet/pods/f3e87a75-cf2d-4ee1-b89b-75cbce1a3041/volumes" Jan 27 15:40:12 crc kubenswrapper[4772]: I0127 15:40:12.058288 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:40:12 crc kubenswrapper[4772]: I0127 15:40:12.059039 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:40:42 crc kubenswrapper[4772]: I0127 15:40:42.058864 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:40:42 crc kubenswrapper[4772]: I0127 15:40:42.059801 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:41:12 crc kubenswrapper[4772]: I0127 15:41:12.058249 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:41:12 crc kubenswrapper[4772]: I0127 15:41:12.059235 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:41:12 crc kubenswrapper[4772]: I0127 15:41:12.059319 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:41:12 crc kubenswrapper[4772]: I0127 15:41:12.060468 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7b761d6bc884bc3f2c2a56d23c68b1cd3740e77326d275aa98b5b71fcaad6f52"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 15:41:12 crc kubenswrapper[4772]: I0127 15:41:12.060553 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://7b761d6bc884bc3f2c2a56d23c68b1cd3740e77326d275aa98b5b71fcaad6f52" gracePeriod=600 Jan 27 15:41:12 crc kubenswrapper[4772]: E0127 15:41:12.181074 4772 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67794a44_d793_4fd7_9e54_e40437f67c0b.slice/crio-7b761d6bc884bc3f2c2a56d23c68b1cd3740e77326d275aa98b5b71fcaad6f52.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67794a44_d793_4fd7_9e54_e40437f67c0b.slice/crio-conmon-7b761d6bc884bc3f2c2a56d23c68b1cd3740e77326d275aa98b5b71fcaad6f52.scope\": RecentStats: unable to find data in memory cache]" Jan 27 15:41:12 crc kubenswrapper[4772]: I0127 15:41:12.788022 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"7b761d6bc884bc3f2c2a56d23c68b1cd3740e77326d275aa98b5b71fcaad6f52"} Jan 27 15:41:12 crc kubenswrapper[4772]: I0127 15:41:12.787953 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="7b761d6bc884bc3f2c2a56d23c68b1cd3740e77326d275aa98b5b71fcaad6f52" exitCode=0 Jan 27 15:41:12 crc kubenswrapper[4772]: I0127 15:41:12.788134 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d"} Jan 27 15:41:12 crc kubenswrapper[4772]: I0127 15:41:12.788109 4772 scope.go:117] "RemoveContainer" containerID="b0ae39c80720edbba923270ddb9a5ec4d4548e971f6133e3594454030be573c2" Jan 27 15:41:17 crc kubenswrapper[4772]: I0127 15:41:17.467919 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dkq6v"] Jan 27 15:41:17 crc kubenswrapper[4772]: E0127 15:41:17.468970 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3e87a75-cf2d-4ee1-b89b-75cbce1a3041" containerName="registry-server" Jan 27 15:41:17 crc kubenswrapper[4772]: I0127 15:41:17.468989 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3e87a75-cf2d-4ee1-b89b-75cbce1a3041" containerName="registry-server" Jan 27 15:41:17 crc kubenswrapper[4772]: E0127 15:41:17.469006 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3e87a75-cf2d-4ee1-b89b-75cbce1a3041" containerName="extract-utilities" Jan 27 15:41:17 crc kubenswrapper[4772]: I0127 15:41:17.469014 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3e87a75-cf2d-4ee1-b89b-75cbce1a3041" containerName="extract-utilities" Jan 27 15:41:17 crc kubenswrapper[4772]: E0127 15:41:17.469033 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3e87a75-cf2d-4ee1-b89b-75cbce1a3041" containerName="extract-content" Jan 27 15:41:17 crc kubenswrapper[4772]: I0127 15:41:17.469043 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3e87a75-cf2d-4ee1-b89b-75cbce1a3041" containerName="extract-content" Jan 27 15:41:17 crc kubenswrapper[4772]: I0127 15:41:17.469215 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3e87a75-cf2d-4ee1-b89b-75cbce1a3041" containerName="registry-server" Jan 27 15:41:17 crc kubenswrapper[4772]: I0127 15:41:17.470525 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dkq6v" Jan 27 15:41:17 crc kubenswrapper[4772]: I0127 15:41:17.482536 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dkq6v"] Jan 27 15:41:17 crc kubenswrapper[4772]: I0127 15:41:17.586646 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb350c6d-0ecc-4fa7-8270-36a7589aa9a8-catalog-content\") pod \"redhat-marketplace-dkq6v\" (UID: \"fb350c6d-0ecc-4fa7-8270-36a7589aa9a8\") " pod="openshift-marketplace/redhat-marketplace-dkq6v" Jan 27 15:41:17 crc kubenswrapper[4772]: I0127 15:41:17.586709 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb350c6d-0ecc-4fa7-8270-36a7589aa9a8-utilities\") pod \"redhat-marketplace-dkq6v\" (UID: \"fb350c6d-0ecc-4fa7-8270-36a7589aa9a8\") " pod="openshift-marketplace/redhat-marketplace-dkq6v" Jan 27 15:41:17 crc kubenswrapper[4772]: I0127 15:41:17.586815 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95fz9\" (UniqueName: \"kubernetes.io/projected/fb350c6d-0ecc-4fa7-8270-36a7589aa9a8-kube-api-access-95fz9\") pod \"redhat-marketplace-dkq6v\" (UID: \"fb350c6d-0ecc-4fa7-8270-36a7589aa9a8\") " pod="openshift-marketplace/redhat-marketplace-dkq6v" Jan 27 15:41:17 crc kubenswrapper[4772]: I0127 15:41:17.688042 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb350c6d-0ecc-4fa7-8270-36a7589aa9a8-catalog-content\") pod \"redhat-marketplace-dkq6v\" (UID: \"fb350c6d-0ecc-4fa7-8270-36a7589aa9a8\") " pod="openshift-marketplace/redhat-marketplace-dkq6v" Jan 27 15:41:17 crc kubenswrapper[4772]: I0127 15:41:17.688126 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb350c6d-0ecc-4fa7-8270-36a7589aa9a8-utilities\") pod \"redhat-marketplace-dkq6v\" (UID: \"fb350c6d-0ecc-4fa7-8270-36a7589aa9a8\") " pod="openshift-marketplace/redhat-marketplace-dkq6v" Jan 27 15:41:17 crc kubenswrapper[4772]: I0127 15:41:17.688158 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95fz9\" (UniqueName: \"kubernetes.io/projected/fb350c6d-0ecc-4fa7-8270-36a7589aa9a8-kube-api-access-95fz9\") pod \"redhat-marketplace-dkq6v\" (UID: \"fb350c6d-0ecc-4fa7-8270-36a7589aa9a8\") " pod="openshift-marketplace/redhat-marketplace-dkq6v" Jan 27 15:41:17 crc kubenswrapper[4772]: I0127 15:41:17.688769 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb350c6d-0ecc-4fa7-8270-36a7589aa9a8-utilities\") pod \"redhat-marketplace-dkq6v\" (UID: \"fb350c6d-0ecc-4fa7-8270-36a7589aa9a8\") " pod="openshift-marketplace/redhat-marketplace-dkq6v" Jan 27 15:41:17 crc kubenswrapper[4772]: I0127 15:41:17.688769 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb350c6d-0ecc-4fa7-8270-36a7589aa9a8-catalog-content\") pod \"redhat-marketplace-dkq6v\" (UID: \"fb350c6d-0ecc-4fa7-8270-36a7589aa9a8\") " pod="openshift-marketplace/redhat-marketplace-dkq6v" Jan 27 15:41:17 crc kubenswrapper[4772]: I0127 15:41:17.711397 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95fz9\" (UniqueName: \"kubernetes.io/projected/fb350c6d-0ecc-4fa7-8270-36a7589aa9a8-kube-api-access-95fz9\") pod \"redhat-marketplace-dkq6v\" (UID: \"fb350c6d-0ecc-4fa7-8270-36a7589aa9a8\") " pod="openshift-marketplace/redhat-marketplace-dkq6v" Jan 27 15:41:17 crc kubenswrapper[4772]: I0127 15:41:17.790386 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dkq6v" Jan 27 15:41:18 crc kubenswrapper[4772]: I0127 15:41:18.239076 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dkq6v"] Jan 27 15:41:18 crc kubenswrapper[4772]: I0127 15:41:18.839927 4772 generic.go:334] "Generic (PLEG): container finished" podID="fb350c6d-0ecc-4fa7-8270-36a7589aa9a8" containerID="2734d7c6099c2b586983d9f79bc0e7eba2fc0c635ae80160a19b4cfe53a27194" exitCode=0 Jan 27 15:41:18 crc kubenswrapper[4772]: I0127 15:41:18.840008 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkq6v" event={"ID":"fb350c6d-0ecc-4fa7-8270-36a7589aa9a8","Type":"ContainerDied","Data":"2734d7c6099c2b586983d9f79bc0e7eba2fc0c635ae80160a19b4cfe53a27194"} Jan 27 15:41:18 crc kubenswrapper[4772]: I0127 15:41:18.840241 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkq6v" event={"ID":"fb350c6d-0ecc-4fa7-8270-36a7589aa9a8","Type":"ContainerStarted","Data":"c1a4e1a540a2765632567eb226c9e18c795fcbbcddf9a726b1cc0c6e091e2a13"} Jan 27 15:41:18 crc kubenswrapper[4772]: I0127 15:41:18.843828 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 15:41:20 crc kubenswrapper[4772]: I0127 15:41:20.875038 4772 generic.go:334] "Generic (PLEG): container finished" podID="fb350c6d-0ecc-4fa7-8270-36a7589aa9a8" containerID="ae87e00dd314123fd500107d9147be5ee94d0ef022bbaadd406d05a97d92ce2f" exitCode=0 Jan 27 15:41:20 crc kubenswrapper[4772]: I0127 15:41:20.875391 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkq6v" event={"ID":"fb350c6d-0ecc-4fa7-8270-36a7589aa9a8","Type":"ContainerDied","Data":"ae87e00dd314123fd500107d9147be5ee94d0ef022bbaadd406d05a97d92ce2f"} Jan 27 15:41:21 crc kubenswrapper[4772]: I0127 15:41:21.889377 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkq6v" event={"ID":"fb350c6d-0ecc-4fa7-8270-36a7589aa9a8","Type":"ContainerStarted","Data":"1109392b5c2e2da59eacb68ca1039f67700720c68032965782bdee89b041c0f5"} Jan 27 15:41:21 crc kubenswrapper[4772]: I0127 15:41:21.912258 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dkq6v" podStartSLOduration=2.132035986 podStartE2EDuration="4.91223516s" podCreationTimestamp="2026-01-27 15:41:17 +0000 UTC" firstStartedPulling="2026-01-27 15:41:18.843567656 +0000 UTC m=+2064.824176754" lastFinishedPulling="2026-01-27 15:41:21.62376683 +0000 UTC m=+2067.604375928" observedRunningTime="2026-01-27 15:41:21.908134252 +0000 UTC m=+2067.888743350" watchObservedRunningTime="2026-01-27 15:41:21.91223516 +0000 UTC m=+2067.892844268" Jan 27 15:41:27 crc kubenswrapper[4772]: I0127 15:41:27.790771 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dkq6v" Jan 27 15:41:27 crc kubenswrapper[4772]: I0127 15:41:27.791381 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dkq6v" Jan 27 15:41:27 crc kubenswrapper[4772]: I0127 15:41:27.833023 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dkq6v" Jan 27 15:41:27 crc kubenswrapper[4772]: I0127 15:41:27.968968 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dkq6v" Jan 27 15:41:28 crc kubenswrapper[4772]: I0127 15:41:28.066303 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dkq6v"] Jan 27 15:41:29 crc kubenswrapper[4772]: I0127 15:41:29.943295 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dkq6v" podUID="fb350c6d-0ecc-4fa7-8270-36a7589aa9a8" containerName="registry-server" containerID="cri-o://1109392b5c2e2da59eacb68ca1039f67700720c68032965782bdee89b041c0f5" gracePeriod=2 Jan 27 15:41:30 crc kubenswrapper[4772]: I0127 15:41:30.437063 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dkq6v" Jan 27 15:41:30 crc kubenswrapper[4772]: I0127 15:41:30.576626 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb350c6d-0ecc-4fa7-8270-36a7589aa9a8-catalog-content\") pod \"fb350c6d-0ecc-4fa7-8270-36a7589aa9a8\" (UID: \"fb350c6d-0ecc-4fa7-8270-36a7589aa9a8\") " Jan 27 15:41:30 crc kubenswrapper[4772]: I0127 15:41:30.576798 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95fz9\" (UniqueName: \"kubernetes.io/projected/fb350c6d-0ecc-4fa7-8270-36a7589aa9a8-kube-api-access-95fz9\") pod \"fb350c6d-0ecc-4fa7-8270-36a7589aa9a8\" (UID: \"fb350c6d-0ecc-4fa7-8270-36a7589aa9a8\") " Jan 27 15:41:30 crc kubenswrapper[4772]: I0127 15:41:30.576972 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb350c6d-0ecc-4fa7-8270-36a7589aa9a8-utilities\") pod \"fb350c6d-0ecc-4fa7-8270-36a7589aa9a8\" (UID: \"fb350c6d-0ecc-4fa7-8270-36a7589aa9a8\") " Jan 27 15:41:30 crc kubenswrapper[4772]: I0127 15:41:30.578455 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb350c6d-0ecc-4fa7-8270-36a7589aa9a8-utilities" (OuterVolumeSpecName: "utilities") pod "fb350c6d-0ecc-4fa7-8270-36a7589aa9a8" (UID: "fb350c6d-0ecc-4fa7-8270-36a7589aa9a8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:41:30 crc kubenswrapper[4772]: I0127 15:41:30.588650 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb350c6d-0ecc-4fa7-8270-36a7589aa9a8-kube-api-access-95fz9" (OuterVolumeSpecName: "kube-api-access-95fz9") pod "fb350c6d-0ecc-4fa7-8270-36a7589aa9a8" (UID: "fb350c6d-0ecc-4fa7-8270-36a7589aa9a8"). InnerVolumeSpecName "kube-api-access-95fz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:41:30 crc kubenswrapper[4772]: I0127 15:41:30.604078 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb350c6d-0ecc-4fa7-8270-36a7589aa9a8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fb350c6d-0ecc-4fa7-8270-36a7589aa9a8" (UID: "fb350c6d-0ecc-4fa7-8270-36a7589aa9a8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:41:30 crc kubenswrapper[4772]: I0127 15:41:30.678946 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb350c6d-0ecc-4fa7-8270-36a7589aa9a8-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 15:41:30 crc kubenswrapper[4772]: I0127 15:41:30.678993 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb350c6d-0ecc-4fa7-8270-36a7589aa9a8-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 15:41:30 crc kubenswrapper[4772]: I0127 15:41:30.679006 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95fz9\" (UniqueName: \"kubernetes.io/projected/fb350c6d-0ecc-4fa7-8270-36a7589aa9a8-kube-api-access-95fz9\") on node \"crc\" DevicePath \"\"" Jan 27 15:41:30 crc kubenswrapper[4772]: I0127 15:41:30.952968 4772 generic.go:334] "Generic (PLEG): container finished" podID="fb350c6d-0ecc-4fa7-8270-36a7589aa9a8" containerID="1109392b5c2e2da59eacb68ca1039f67700720c68032965782bdee89b041c0f5" exitCode=0 Jan 27 15:41:30 crc kubenswrapper[4772]: I0127 15:41:30.953015 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkq6v" event={"ID":"fb350c6d-0ecc-4fa7-8270-36a7589aa9a8","Type":"ContainerDied","Data":"1109392b5c2e2da59eacb68ca1039f67700720c68032965782bdee89b041c0f5"} Jan 27 15:41:30 crc kubenswrapper[4772]: I0127 15:41:30.953048 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkq6v" event={"ID":"fb350c6d-0ecc-4fa7-8270-36a7589aa9a8","Type":"ContainerDied","Data":"c1a4e1a540a2765632567eb226c9e18c795fcbbcddf9a726b1cc0c6e091e2a13"} Jan 27 15:41:30 crc kubenswrapper[4772]: I0127 15:41:30.953066 4772 scope.go:117] "RemoveContainer" containerID="1109392b5c2e2da59eacb68ca1039f67700720c68032965782bdee89b041c0f5" Jan 27 15:41:30 crc kubenswrapper[4772]: I0127 15:41:30.953081 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dkq6v" Jan 27 15:41:30 crc kubenswrapper[4772]: I0127 15:41:30.977839 4772 scope.go:117] "RemoveContainer" containerID="ae87e00dd314123fd500107d9147be5ee94d0ef022bbaadd406d05a97d92ce2f" Jan 27 15:41:30 crc kubenswrapper[4772]: I0127 15:41:30.993410 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dkq6v"] Jan 27 15:41:31 crc kubenswrapper[4772]: I0127 15:41:31.001197 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dkq6v"] Jan 27 15:41:31 crc kubenswrapper[4772]: I0127 15:41:31.017483 4772 scope.go:117] "RemoveContainer" containerID="2734d7c6099c2b586983d9f79bc0e7eba2fc0c635ae80160a19b4cfe53a27194" Jan 27 15:41:31 crc kubenswrapper[4772]: I0127 15:41:31.039903 4772 scope.go:117] "RemoveContainer" containerID="1109392b5c2e2da59eacb68ca1039f67700720c68032965782bdee89b041c0f5" Jan 27 15:41:31 crc kubenswrapper[4772]: E0127 15:41:31.040423 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1109392b5c2e2da59eacb68ca1039f67700720c68032965782bdee89b041c0f5\": container with ID starting with 1109392b5c2e2da59eacb68ca1039f67700720c68032965782bdee89b041c0f5 not found: ID does not exist" containerID="1109392b5c2e2da59eacb68ca1039f67700720c68032965782bdee89b041c0f5" Jan 27 15:41:31 crc kubenswrapper[4772]: I0127 15:41:31.040471 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1109392b5c2e2da59eacb68ca1039f67700720c68032965782bdee89b041c0f5"} err="failed to get container status \"1109392b5c2e2da59eacb68ca1039f67700720c68032965782bdee89b041c0f5\": rpc error: code = NotFound desc = could not find container \"1109392b5c2e2da59eacb68ca1039f67700720c68032965782bdee89b041c0f5\": container with ID starting with 1109392b5c2e2da59eacb68ca1039f67700720c68032965782bdee89b041c0f5 not found: ID does not exist" Jan 27 15:41:31 crc kubenswrapper[4772]: I0127 15:41:31.040502 4772 scope.go:117] "RemoveContainer" containerID="ae87e00dd314123fd500107d9147be5ee94d0ef022bbaadd406d05a97d92ce2f" Jan 27 15:41:31 crc kubenswrapper[4772]: E0127 15:41:31.040914 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae87e00dd314123fd500107d9147be5ee94d0ef022bbaadd406d05a97d92ce2f\": container with ID starting with ae87e00dd314123fd500107d9147be5ee94d0ef022bbaadd406d05a97d92ce2f not found: ID does not exist" containerID="ae87e00dd314123fd500107d9147be5ee94d0ef022bbaadd406d05a97d92ce2f" Jan 27 15:41:31 crc kubenswrapper[4772]: I0127 15:41:31.040946 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae87e00dd314123fd500107d9147be5ee94d0ef022bbaadd406d05a97d92ce2f"} err="failed to get container status \"ae87e00dd314123fd500107d9147be5ee94d0ef022bbaadd406d05a97d92ce2f\": rpc error: code = NotFound desc = could not find container \"ae87e00dd314123fd500107d9147be5ee94d0ef022bbaadd406d05a97d92ce2f\": container with ID starting with ae87e00dd314123fd500107d9147be5ee94d0ef022bbaadd406d05a97d92ce2f not found: ID does not exist" Jan 27 15:41:31 crc kubenswrapper[4772]: I0127 15:41:31.040970 4772 scope.go:117] "RemoveContainer" containerID="2734d7c6099c2b586983d9f79bc0e7eba2fc0c635ae80160a19b4cfe53a27194" Jan 27 15:41:31 crc kubenswrapper[4772]: E0127 15:41:31.041274 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2734d7c6099c2b586983d9f79bc0e7eba2fc0c635ae80160a19b4cfe53a27194\": container with ID starting with 2734d7c6099c2b586983d9f79bc0e7eba2fc0c635ae80160a19b4cfe53a27194 not found: ID does not exist" containerID="2734d7c6099c2b586983d9f79bc0e7eba2fc0c635ae80160a19b4cfe53a27194" Jan 27 15:41:31 crc kubenswrapper[4772]: I0127 15:41:31.041298 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2734d7c6099c2b586983d9f79bc0e7eba2fc0c635ae80160a19b4cfe53a27194"} err="failed to get container status \"2734d7c6099c2b586983d9f79bc0e7eba2fc0c635ae80160a19b4cfe53a27194\": rpc error: code = NotFound desc = could not find container \"2734d7c6099c2b586983d9f79bc0e7eba2fc0c635ae80160a19b4cfe53a27194\": container with ID starting with 2734d7c6099c2b586983d9f79bc0e7eba2fc0c635ae80160a19b4cfe53a27194 not found: ID does not exist" Jan 27 15:41:32 crc kubenswrapper[4772]: I0127 15:41:32.688657 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb350c6d-0ecc-4fa7-8270-36a7589aa9a8" path="/var/lib/kubelet/pods/fb350c6d-0ecc-4fa7-8270-36a7589aa9a8/volumes" Jan 27 15:43:12 crc kubenswrapper[4772]: I0127 15:43:12.058350 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:43:12 crc kubenswrapper[4772]: I0127 15:43:12.059130 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:43:42 crc kubenswrapper[4772]: I0127 15:43:42.058276 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:43:42 crc kubenswrapper[4772]: I0127 15:43:42.058883 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:44:12 crc kubenswrapper[4772]: I0127 15:44:12.058494 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:44:12 crc kubenswrapper[4772]: I0127 15:44:12.059088 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:44:12 crc kubenswrapper[4772]: I0127 15:44:12.059142 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:44:12 crc kubenswrapper[4772]: I0127 15:44:12.059836 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 15:44:12 crc kubenswrapper[4772]: I0127 15:44:12.059896 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" gracePeriod=600 Jan 27 15:44:12 crc kubenswrapper[4772]: E0127 15:44:12.199948 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:44:13 crc kubenswrapper[4772]: I0127 15:44:13.143935 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" exitCode=0 Jan 27 15:44:13 crc kubenswrapper[4772]: I0127 15:44:13.143983 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d"} Jan 27 15:44:13 crc kubenswrapper[4772]: I0127 15:44:13.144018 4772 scope.go:117] "RemoveContainer" containerID="7b761d6bc884bc3f2c2a56d23c68b1cd3740e77326d275aa98b5b71fcaad6f52" Jan 27 15:44:13 crc kubenswrapper[4772]: I0127 15:44:13.144613 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:44:13 crc kubenswrapper[4772]: E0127 15:44:13.144837 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:44:20 crc kubenswrapper[4772]: I0127 15:44:20.139291 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qsdqm"] Jan 27 15:44:20 crc kubenswrapper[4772]: E0127 15:44:20.140075 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb350c6d-0ecc-4fa7-8270-36a7589aa9a8" containerName="extract-utilities" Jan 27 15:44:20 crc kubenswrapper[4772]: I0127 15:44:20.140087 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb350c6d-0ecc-4fa7-8270-36a7589aa9a8" containerName="extract-utilities" Jan 27 15:44:20 crc kubenswrapper[4772]: E0127 15:44:20.140104 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb350c6d-0ecc-4fa7-8270-36a7589aa9a8" containerName="registry-server" Jan 27 15:44:20 crc kubenswrapper[4772]: I0127 15:44:20.140111 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb350c6d-0ecc-4fa7-8270-36a7589aa9a8" containerName="registry-server" Jan 27 15:44:20 crc kubenswrapper[4772]: E0127 15:44:20.140125 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb350c6d-0ecc-4fa7-8270-36a7589aa9a8" containerName="extract-content" Jan 27 15:44:20 crc kubenswrapper[4772]: I0127 15:44:20.140131 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb350c6d-0ecc-4fa7-8270-36a7589aa9a8" containerName="extract-content" Jan 27 15:44:20 crc kubenswrapper[4772]: I0127 15:44:20.140315 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb350c6d-0ecc-4fa7-8270-36a7589aa9a8" containerName="registry-server" Jan 27 15:44:20 crc kubenswrapper[4772]: I0127 15:44:20.141309 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qsdqm" Jan 27 15:44:20 crc kubenswrapper[4772]: I0127 15:44:20.148699 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qsdqm"] Jan 27 15:44:20 crc kubenswrapper[4772]: I0127 15:44:20.209530 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f233b51-cb2c-420c-8041-51f37d626af8-catalog-content\") pod \"certified-operators-qsdqm\" (UID: \"4f233b51-cb2c-420c-8041-51f37d626af8\") " pod="openshift-marketplace/certified-operators-qsdqm" Jan 27 15:44:20 crc kubenswrapper[4772]: I0127 15:44:20.209628 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f233b51-cb2c-420c-8041-51f37d626af8-utilities\") pod \"certified-operators-qsdqm\" (UID: \"4f233b51-cb2c-420c-8041-51f37d626af8\") " pod="openshift-marketplace/certified-operators-qsdqm" Jan 27 15:44:20 crc kubenswrapper[4772]: I0127 15:44:20.209800 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrknc\" (UniqueName: \"kubernetes.io/projected/4f233b51-cb2c-420c-8041-51f37d626af8-kube-api-access-hrknc\") pod \"certified-operators-qsdqm\" (UID: \"4f233b51-cb2c-420c-8041-51f37d626af8\") " pod="openshift-marketplace/certified-operators-qsdqm" Jan 27 15:44:20 crc kubenswrapper[4772]: I0127 15:44:20.310746 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f233b51-cb2c-420c-8041-51f37d626af8-utilities\") pod \"certified-operators-qsdqm\" (UID: \"4f233b51-cb2c-420c-8041-51f37d626af8\") " pod="openshift-marketplace/certified-operators-qsdqm" Jan 27 15:44:20 crc kubenswrapper[4772]: I0127 15:44:20.310876 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrknc\" (UniqueName: \"kubernetes.io/projected/4f233b51-cb2c-420c-8041-51f37d626af8-kube-api-access-hrknc\") pod \"certified-operators-qsdqm\" (UID: \"4f233b51-cb2c-420c-8041-51f37d626af8\") " pod="openshift-marketplace/certified-operators-qsdqm" Jan 27 15:44:20 crc kubenswrapper[4772]: I0127 15:44:20.310899 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f233b51-cb2c-420c-8041-51f37d626af8-catalog-content\") pod \"certified-operators-qsdqm\" (UID: \"4f233b51-cb2c-420c-8041-51f37d626af8\") " pod="openshift-marketplace/certified-operators-qsdqm" Jan 27 15:44:20 crc kubenswrapper[4772]: I0127 15:44:20.311412 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f233b51-cb2c-420c-8041-51f37d626af8-catalog-content\") pod \"certified-operators-qsdqm\" (UID: \"4f233b51-cb2c-420c-8041-51f37d626af8\") " pod="openshift-marketplace/certified-operators-qsdqm" Jan 27 15:44:20 crc kubenswrapper[4772]: I0127 15:44:20.312348 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f233b51-cb2c-420c-8041-51f37d626af8-utilities\") pod \"certified-operators-qsdqm\" (UID: \"4f233b51-cb2c-420c-8041-51f37d626af8\") " pod="openshift-marketplace/certified-operators-qsdqm" Jan 27 15:44:20 crc kubenswrapper[4772]: I0127 15:44:20.329502 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrknc\" (UniqueName: \"kubernetes.io/projected/4f233b51-cb2c-420c-8041-51f37d626af8-kube-api-access-hrknc\") pod \"certified-operators-qsdqm\" (UID: \"4f233b51-cb2c-420c-8041-51f37d626af8\") " pod="openshift-marketplace/certified-operators-qsdqm" Jan 27 15:44:20 crc kubenswrapper[4772]: I0127 15:44:20.467293 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qsdqm" Jan 27 15:44:20 crc kubenswrapper[4772]: I0127 15:44:20.728460 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qsdqm"] Jan 27 15:44:21 crc kubenswrapper[4772]: I0127 15:44:21.207136 4772 generic.go:334] "Generic (PLEG): container finished" podID="4f233b51-cb2c-420c-8041-51f37d626af8" containerID="efa341e571e74e65be29aaef4632650e18ccae6a662231c4759328298a83b725" exitCode=0 Jan 27 15:44:21 crc kubenswrapper[4772]: I0127 15:44:21.207200 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qsdqm" event={"ID":"4f233b51-cb2c-420c-8041-51f37d626af8","Type":"ContainerDied","Data":"efa341e571e74e65be29aaef4632650e18ccae6a662231c4759328298a83b725"} Jan 27 15:44:21 crc kubenswrapper[4772]: I0127 15:44:21.207248 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qsdqm" event={"ID":"4f233b51-cb2c-420c-8041-51f37d626af8","Type":"ContainerStarted","Data":"cf55e1a8d42200dea1c6edfd0f692b2f2dd7a9c0337a4be623c23a3c32885937"} Jan 27 15:44:22 crc kubenswrapper[4772]: I0127 15:44:22.214997 4772 generic.go:334] "Generic (PLEG): container finished" podID="4f233b51-cb2c-420c-8041-51f37d626af8" containerID="3b3133f4d862cd6fbf455830ef984b383c19cfdc69480ebb3bf4d964e5a3bf9b" exitCode=0 Jan 27 15:44:22 crc kubenswrapper[4772]: I0127 15:44:22.215041 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qsdqm" event={"ID":"4f233b51-cb2c-420c-8041-51f37d626af8","Type":"ContainerDied","Data":"3b3133f4d862cd6fbf455830ef984b383c19cfdc69480ebb3bf4d964e5a3bf9b"} Jan 27 15:44:23 crc kubenswrapper[4772]: I0127 15:44:23.224145 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qsdqm" event={"ID":"4f233b51-cb2c-420c-8041-51f37d626af8","Type":"ContainerStarted","Data":"4ce816182821da35d33be57a085da39c9beb9bdfd04fd5d55ee933fbdbd497da"} Jan 27 15:44:23 crc kubenswrapper[4772]: I0127 15:44:23.241569 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qsdqm" podStartSLOduration=1.748532371 podStartE2EDuration="3.241547665s" podCreationTimestamp="2026-01-27 15:44:20 +0000 UTC" firstStartedPulling="2026-01-27 15:44:21.208837046 +0000 UTC m=+2247.189446144" lastFinishedPulling="2026-01-27 15:44:22.70185234 +0000 UTC m=+2248.682461438" observedRunningTime="2026-01-27 15:44:23.239000083 +0000 UTC m=+2249.219609191" watchObservedRunningTime="2026-01-27 15:44:23.241547665 +0000 UTC m=+2249.222156764" Jan 27 15:44:26 crc kubenswrapper[4772]: I0127 15:44:26.663377 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:44:26 crc kubenswrapper[4772]: E0127 15:44:26.664099 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:44:30 crc kubenswrapper[4772]: I0127 15:44:30.468468 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qsdqm" Jan 27 15:44:30 crc kubenswrapper[4772]: I0127 15:44:30.468526 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qsdqm" Jan 27 15:44:30 crc kubenswrapper[4772]: I0127 15:44:30.507150 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qsdqm" Jan 27 15:44:31 crc kubenswrapper[4772]: I0127 15:44:31.320488 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qsdqm" Jan 27 15:44:31 crc kubenswrapper[4772]: I0127 15:44:31.371235 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qsdqm"] Jan 27 15:44:33 crc kubenswrapper[4772]: I0127 15:44:33.294016 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qsdqm" podUID="4f233b51-cb2c-420c-8041-51f37d626af8" containerName="registry-server" containerID="cri-o://4ce816182821da35d33be57a085da39c9beb9bdfd04fd5d55ee933fbdbd497da" gracePeriod=2 Jan 27 15:44:33 crc kubenswrapper[4772]: I0127 15:44:33.666848 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qsdqm" Jan 27 15:44:33 crc kubenswrapper[4772]: I0127 15:44:33.800201 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f233b51-cb2c-420c-8041-51f37d626af8-utilities\") pod \"4f233b51-cb2c-420c-8041-51f37d626af8\" (UID: \"4f233b51-cb2c-420c-8041-51f37d626af8\") " Jan 27 15:44:33 crc kubenswrapper[4772]: I0127 15:44:33.800240 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f233b51-cb2c-420c-8041-51f37d626af8-catalog-content\") pod \"4f233b51-cb2c-420c-8041-51f37d626af8\" (UID: \"4f233b51-cb2c-420c-8041-51f37d626af8\") " Jan 27 15:44:33 crc kubenswrapper[4772]: I0127 15:44:33.800333 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrknc\" (UniqueName: \"kubernetes.io/projected/4f233b51-cb2c-420c-8041-51f37d626af8-kube-api-access-hrknc\") pod \"4f233b51-cb2c-420c-8041-51f37d626af8\" (UID: \"4f233b51-cb2c-420c-8041-51f37d626af8\") " Jan 27 15:44:33 crc kubenswrapper[4772]: I0127 15:44:33.801538 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f233b51-cb2c-420c-8041-51f37d626af8-utilities" (OuterVolumeSpecName: "utilities") pod "4f233b51-cb2c-420c-8041-51f37d626af8" (UID: "4f233b51-cb2c-420c-8041-51f37d626af8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:44:33 crc kubenswrapper[4772]: I0127 15:44:33.802308 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f233b51-cb2c-420c-8041-51f37d626af8-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 15:44:33 crc kubenswrapper[4772]: I0127 15:44:33.805393 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f233b51-cb2c-420c-8041-51f37d626af8-kube-api-access-hrknc" (OuterVolumeSpecName: "kube-api-access-hrknc") pod "4f233b51-cb2c-420c-8041-51f37d626af8" (UID: "4f233b51-cb2c-420c-8041-51f37d626af8"). InnerVolumeSpecName "kube-api-access-hrknc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:44:33 crc kubenswrapper[4772]: I0127 15:44:33.849930 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f233b51-cb2c-420c-8041-51f37d626af8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4f233b51-cb2c-420c-8041-51f37d626af8" (UID: "4f233b51-cb2c-420c-8041-51f37d626af8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:44:33 crc kubenswrapper[4772]: I0127 15:44:33.903991 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f233b51-cb2c-420c-8041-51f37d626af8-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 15:44:33 crc kubenswrapper[4772]: I0127 15:44:33.904023 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrknc\" (UniqueName: \"kubernetes.io/projected/4f233b51-cb2c-420c-8041-51f37d626af8-kube-api-access-hrknc\") on node \"crc\" DevicePath \"\"" Jan 27 15:44:34 crc kubenswrapper[4772]: I0127 15:44:34.304829 4772 generic.go:334] "Generic (PLEG): container finished" podID="4f233b51-cb2c-420c-8041-51f37d626af8" containerID="4ce816182821da35d33be57a085da39c9beb9bdfd04fd5d55ee933fbdbd497da" exitCode=0 Jan 27 15:44:34 crc kubenswrapper[4772]: I0127 15:44:34.304875 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qsdqm" event={"ID":"4f233b51-cb2c-420c-8041-51f37d626af8","Type":"ContainerDied","Data":"4ce816182821da35d33be57a085da39c9beb9bdfd04fd5d55ee933fbdbd497da"} Jan 27 15:44:34 crc kubenswrapper[4772]: I0127 15:44:34.304903 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qsdqm" event={"ID":"4f233b51-cb2c-420c-8041-51f37d626af8","Type":"ContainerDied","Data":"cf55e1a8d42200dea1c6edfd0f692b2f2dd7a9c0337a4be623c23a3c32885937"} Jan 27 15:44:34 crc kubenswrapper[4772]: I0127 15:44:34.304922 4772 scope.go:117] "RemoveContainer" containerID="4ce816182821da35d33be57a085da39c9beb9bdfd04fd5d55ee933fbdbd497da" Jan 27 15:44:34 crc kubenswrapper[4772]: I0127 15:44:34.305049 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qsdqm" Jan 27 15:44:34 crc kubenswrapper[4772]: I0127 15:44:34.334011 4772 scope.go:117] "RemoveContainer" containerID="3b3133f4d862cd6fbf455830ef984b383c19cfdc69480ebb3bf4d964e5a3bf9b" Jan 27 15:44:34 crc kubenswrapper[4772]: I0127 15:44:34.347012 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qsdqm"] Jan 27 15:44:34 crc kubenswrapper[4772]: I0127 15:44:34.358496 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qsdqm"] Jan 27 15:44:34 crc kubenswrapper[4772]: I0127 15:44:34.374516 4772 scope.go:117] "RemoveContainer" containerID="efa341e571e74e65be29aaef4632650e18ccae6a662231c4759328298a83b725" Jan 27 15:44:34 crc kubenswrapper[4772]: I0127 15:44:34.392606 4772 scope.go:117] "RemoveContainer" containerID="4ce816182821da35d33be57a085da39c9beb9bdfd04fd5d55ee933fbdbd497da" Jan 27 15:44:34 crc kubenswrapper[4772]: E0127 15:44:34.393710 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ce816182821da35d33be57a085da39c9beb9bdfd04fd5d55ee933fbdbd497da\": container with ID starting with 4ce816182821da35d33be57a085da39c9beb9bdfd04fd5d55ee933fbdbd497da not found: ID does not exist" containerID="4ce816182821da35d33be57a085da39c9beb9bdfd04fd5d55ee933fbdbd497da" Jan 27 15:44:34 crc kubenswrapper[4772]: I0127 15:44:34.393747 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ce816182821da35d33be57a085da39c9beb9bdfd04fd5d55ee933fbdbd497da"} err="failed to get container status \"4ce816182821da35d33be57a085da39c9beb9bdfd04fd5d55ee933fbdbd497da\": rpc error: code = NotFound desc = could not find container \"4ce816182821da35d33be57a085da39c9beb9bdfd04fd5d55ee933fbdbd497da\": container with ID starting with 4ce816182821da35d33be57a085da39c9beb9bdfd04fd5d55ee933fbdbd497da not found: ID does not exist" Jan 27 15:44:34 crc kubenswrapper[4772]: I0127 15:44:34.393770 4772 scope.go:117] "RemoveContainer" containerID="3b3133f4d862cd6fbf455830ef984b383c19cfdc69480ebb3bf4d964e5a3bf9b" Jan 27 15:44:34 crc kubenswrapper[4772]: E0127 15:44:34.394156 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b3133f4d862cd6fbf455830ef984b383c19cfdc69480ebb3bf4d964e5a3bf9b\": container with ID starting with 3b3133f4d862cd6fbf455830ef984b383c19cfdc69480ebb3bf4d964e5a3bf9b not found: ID does not exist" containerID="3b3133f4d862cd6fbf455830ef984b383c19cfdc69480ebb3bf4d964e5a3bf9b" Jan 27 15:44:34 crc kubenswrapper[4772]: I0127 15:44:34.394189 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b3133f4d862cd6fbf455830ef984b383c19cfdc69480ebb3bf4d964e5a3bf9b"} err="failed to get container status \"3b3133f4d862cd6fbf455830ef984b383c19cfdc69480ebb3bf4d964e5a3bf9b\": rpc error: code = NotFound desc = could not find container \"3b3133f4d862cd6fbf455830ef984b383c19cfdc69480ebb3bf4d964e5a3bf9b\": container with ID starting with 3b3133f4d862cd6fbf455830ef984b383c19cfdc69480ebb3bf4d964e5a3bf9b not found: ID does not exist" Jan 27 15:44:34 crc kubenswrapper[4772]: I0127 15:44:34.394201 4772 scope.go:117] "RemoveContainer" containerID="efa341e571e74e65be29aaef4632650e18ccae6a662231c4759328298a83b725" Jan 27 15:44:34 crc kubenswrapper[4772]: E0127 15:44:34.394453 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efa341e571e74e65be29aaef4632650e18ccae6a662231c4759328298a83b725\": container with ID starting with efa341e571e74e65be29aaef4632650e18ccae6a662231c4759328298a83b725 not found: ID does not exist" containerID="efa341e571e74e65be29aaef4632650e18ccae6a662231c4759328298a83b725" Jan 27 15:44:34 crc kubenswrapper[4772]: I0127 15:44:34.394472 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efa341e571e74e65be29aaef4632650e18ccae6a662231c4759328298a83b725"} err="failed to get container status \"efa341e571e74e65be29aaef4632650e18ccae6a662231c4759328298a83b725\": rpc error: code = NotFound desc = could not find container \"efa341e571e74e65be29aaef4632650e18ccae6a662231c4759328298a83b725\": container with ID starting with efa341e571e74e65be29aaef4632650e18ccae6a662231c4759328298a83b725 not found: ID does not exist" Jan 27 15:44:34 crc kubenswrapper[4772]: I0127 15:44:34.679763 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f233b51-cb2c-420c-8041-51f37d626af8" path="/var/lib/kubelet/pods/4f233b51-cb2c-420c-8041-51f37d626af8/volumes" Jan 27 15:44:37 crc kubenswrapper[4772]: I0127 15:44:37.662866 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:44:37 crc kubenswrapper[4772]: E0127 15:44:37.663110 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:44:50 crc kubenswrapper[4772]: I0127 15:44:50.663449 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:44:50 crc kubenswrapper[4772]: E0127 15:44:50.664221 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:45:00 crc kubenswrapper[4772]: I0127 15:45:00.147728 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492145-st6dd"] Jan 27 15:45:00 crc kubenswrapper[4772]: E0127 15:45:00.148930 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f233b51-cb2c-420c-8041-51f37d626af8" containerName="extract-content" Jan 27 15:45:00 crc kubenswrapper[4772]: I0127 15:45:00.148948 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f233b51-cb2c-420c-8041-51f37d626af8" containerName="extract-content" Jan 27 15:45:00 crc kubenswrapper[4772]: E0127 15:45:00.148961 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f233b51-cb2c-420c-8041-51f37d626af8" containerName="registry-server" Jan 27 15:45:00 crc kubenswrapper[4772]: I0127 15:45:00.148972 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f233b51-cb2c-420c-8041-51f37d626af8" containerName="registry-server" Jan 27 15:45:00 crc kubenswrapper[4772]: E0127 15:45:00.148986 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f233b51-cb2c-420c-8041-51f37d626af8" containerName="extract-utilities" Jan 27 15:45:00 crc kubenswrapper[4772]: I0127 15:45:00.148993 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f233b51-cb2c-420c-8041-51f37d626af8" containerName="extract-utilities" Jan 27 15:45:00 crc kubenswrapper[4772]: I0127 15:45:00.149159 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f233b51-cb2c-420c-8041-51f37d626af8" containerName="registry-server" Jan 27 15:45:00 crc kubenswrapper[4772]: I0127 15:45:00.149779 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492145-st6dd" Jan 27 15:45:00 crc kubenswrapper[4772]: I0127 15:45:00.152873 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 27 15:45:00 crc kubenswrapper[4772]: I0127 15:45:00.152920 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 27 15:45:00 crc kubenswrapper[4772]: I0127 15:45:00.178779 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492145-st6dd"] Jan 27 15:45:00 crc kubenswrapper[4772]: I0127 15:45:00.292485 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d288b\" (UniqueName: \"kubernetes.io/projected/df82c0c4-9652-407e-b63d-17e2ccdb38aa-kube-api-access-d288b\") pod \"collect-profiles-29492145-st6dd\" (UID: \"df82c0c4-9652-407e-b63d-17e2ccdb38aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492145-st6dd" Jan 27 15:45:00 crc kubenswrapper[4772]: I0127 15:45:00.292812 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df82c0c4-9652-407e-b63d-17e2ccdb38aa-config-volume\") pod \"collect-profiles-29492145-st6dd\" (UID: \"df82c0c4-9652-407e-b63d-17e2ccdb38aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492145-st6dd" Jan 27 15:45:00 crc kubenswrapper[4772]: I0127 15:45:00.292929 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df82c0c4-9652-407e-b63d-17e2ccdb38aa-secret-volume\") pod \"collect-profiles-29492145-st6dd\" (UID: \"df82c0c4-9652-407e-b63d-17e2ccdb38aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492145-st6dd" Jan 27 15:45:00 crc kubenswrapper[4772]: I0127 15:45:00.394305 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d288b\" (UniqueName: \"kubernetes.io/projected/df82c0c4-9652-407e-b63d-17e2ccdb38aa-kube-api-access-d288b\") pod \"collect-profiles-29492145-st6dd\" (UID: \"df82c0c4-9652-407e-b63d-17e2ccdb38aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492145-st6dd" Jan 27 15:45:00 crc kubenswrapper[4772]: I0127 15:45:00.394634 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df82c0c4-9652-407e-b63d-17e2ccdb38aa-config-volume\") pod \"collect-profiles-29492145-st6dd\" (UID: \"df82c0c4-9652-407e-b63d-17e2ccdb38aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492145-st6dd" Jan 27 15:45:00 crc kubenswrapper[4772]: I0127 15:45:00.394776 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df82c0c4-9652-407e-b63d-17e2ccdb38aa-secret-volume\") pod \"collect-profiles-29492145-st6dd\" (UID: \"df82c0c4-9652-407e-b63d-17e2ccdb38aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492145-st6dd" Jan 27 15:45:00 crc kubenswrapper[4772]: I0127 15:45:00.395570 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df82c0c4-9652-407e-b63d-17e2ccdb38aa-config-volume\") pod \"collect-profiles-29492145-st6dd\" (UID: \"df82c0c4-9652-407e-b63d-17e2ccdb38aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492145-st6dd" Jan 27 15:45:00 crc kubenswrapper[4772]: I0127 15:45:00.407295 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df82c0c4-9652-407e-b63d-17e2ccdb38aa-secret-volume\") pod \"collect-profiles-29492145-st6dd\" (UID: \"df82c0c4-9652-407e-b63d-17e2ccdb38aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492145-st6dd" Jan 27 15:45:00 crc kubenswrapper[4772]: I0127 15:45:00.414589 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d288b\" (UniqueName: \"kubernetes.io/projected/df82c0c4-9652-407e-b63d-17e2ccdb38aa-kube-api-access-d288b\") pod \"collect-profiles-29492145-st6dd\" (UID: \"df82c0c4-9652-407e-b63d-17e2ccdb38aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492145-st6dd" Jan 27 15:45:00 crc kubenswrapper[4772]: I0127 15:45:00.522842 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492145-st6dd" Jan 27 15:45:00 crc kubenswrapper[4772]: I0127 15:45:00.760664 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492145-st6dd"] Jan 27 15:45:01 crc kubenswrapper[4772]: I0127 15:45:01.484823 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492145-st6dd" event={"ID":"df82c0c4-9652-407e-b63d-17e2ccdb38aa","Type":"ContainerStarted","Data":"72ea0a33955c0509b888997e5b6ca0dc68de786a608fe5aae9035bbbf84ae773"} Jan 27 15:45:01 crc kubenswrapper[4772]: I0127 15:45:01.485317 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492145-st6dd" event={"ID":"df82c0c4-9652-407e-b63d-17e2ccdb38aa","Type":"ContainerStarted","Data":"6e8b24753000f1a5886c84ae46b17b89b02042bde799c38ca5eda8c6b7c07dde"} Jan 27 15:45:01 crc kubenswrapper[4772]: I0127 15:45:01.507435 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29492145-st6dd" podStartSLOduration=1.507416612 podStartE2EDuration="1.507416612s" podCreationTimestamp="2026-01-27 15:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 15:45:01.50204835 +0000 UTC m=+2287.482657448" watchObservedRunningTime="2026-01-27 15:45:01.507416612 +0000 UTC m=+2287.488025710" Jan 27 15:45:02 crc kubenswrapper[4772]: I0127 15:45:02.493759 4772 generic.go:334] "Generic (PLEG): container finished" podID="df82c0c4-9652-407e-b63d-17e2ccdb38aa" containerID="72ea0a33955c0509b888997e5b6ca0dc68de786a608fe5aae9035bbbf84ae773" exitCode=0 Jan 27 15:45:02 crc kubenswrapper[4772]: I0127 15:45:02.493810 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492145-st6dd" event={"ID":"df82c0c4-9652-407e-b63d-17e2ccdb38aa","Type":"ContainerDied","Data":"72ea0a33955c0509b888997e5b6ca0dc68de786a608fe5aae9035bbbf84ae773"} Jan 27 15:45:02 crc kubenswrapper[4772]: I0127 15:45:02.663209 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:45:02 crc kubenswrapper[4772]: E0127 15:45:02.663496 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:45:03 crc kubenswrapper[4772]: I0127 15:45:03.767238 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492145-st6dd" Jan 27 15:45:03 crc kubenswrapper[4772]: I0127 15:45:03.846476 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df82c0c4-9652-407e-b63d-17e2ccdb38aa-config-volume\") pod \"df82c0c4-9652-407e-b63d-17e2ccdb38aa\" (UID: \"df82c0c4-9652-407e-b63d-17e2ccdb38aa\") " Jan 27 15:45:03 crc kubenswrapper[4772]: I0127 15:45:03.846661 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df82c0c4-9652-407e-b63d-17e2ccdb38aa-secret-volume\") pod \"df82c0c4-9652-407e-b63d-17e2ccdb38aa\" (UID: \"df82c0c4-9652-407e-b63d-17e2ccdb38aa\") " Jan 27 15:45:03 crc kubenswrapper[4772]: I0127 15:45:03.846755 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d288b\" (UniqueName: \"kubernetes.io/projected/df82c0c4-9652-407e-b63d-17e2ccdb38aa-kube-api-access-d288b\") pod \"df82c0c4-9652-407e-b63d-17e2ccdb38aa\" (UID: \"df82c0c4-9652-407e-b63d-17e2ccdb38aa\") " Jan 27 15:45:03 crc kubenswrapper[4772]: I0127 15:45:03.847943 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df82c0c4-9652-407e-b63d-17e2ccdb38aa-config-volume" (OuterVolumeSpecName: "config-volume") pod "df82c0c4-9652-407e-b63d-17e2ccdb38aa" (UID: "df82c0c4-9652-407e-b63d-17e2ccdb38aa"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 15:45:03 crc kubenswrapper[4772]: I0127 15:45:03.853479 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df82c0c4-9652-407e-b63d-17e2ccdb38aa-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "df82c0c4-9652-407e-b63d-17e2ccdb38aa" (UID: "df82c0c4-9652-407e-b63d-17e2ccdb38aa"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 15:45:03 crc kubenswrapper[4772]: I0127 15:45:03.853901 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df82c0c4-9652-407e-b63d-17e2ccdb38aa-kube-api-access-d288b" (OuterVolumeSpecName: "kube-api-access-d288b") pod "df82c0c4-9652-407e-b63d-17e2ccdb38aa" (UID: "df82c0c4-9652-407e-b63d-17e2ccdb38aa"). InnerVolumeSpecName "kube-api-access-d288b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:45:03 crc kubenswrapper[4772]: I0127 15:45:03.948815 4772 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/df82c0c4-9652-407e-b63d-17e2ccdb38aa-config-volume\") on node \"crc\" DevicePath \"\"" Jan 27 15:45:03 crc kubenswrapper[4772]: I0127 15:45:03.948863 4772 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/df82c0c4-9652-407e-b63d-17e2ccdb38aa-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 27 15:45:03 crc kubenswrapper[4772]: I0127 15:45:03.948878 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d288b\" (UniqueName: \"kubernetes.io/projected/df82c0c4-9652-407e-b63d-17e2ccdb38aa-kube-api-access-d288b\") on node \"crc\" DevicePath \"\"" Jan 27 15:45:04 crc kubenswrapper[4772]: I0127 15:45:04.509016 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492145-st6dd" event={"ID":"df82c0c4-9652-407e-b63d-17e2ccdb38aa","Type":"ContainerDied","Data":"6e8b24753000f1a5886c84ae46b17b89b02042bde799c38ca5eda8c6b7c07dde"} Jan 27 15:45:04 crc kubenswrapper[4772]: I0127 15:45:04.509075 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e8b24753000f1a5886c84ae46b17b89b02042bde799c38ca5eda8c6b7c07dde" Jan 27 15:45:04 crc kubenswrapper[4772]: I0127 15:45:04.509096 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492145-st6dd" Jan 27 15:45:04 crc kubenswrapper[4772]: I0127 15:45:04.583094 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492100-r2zj6"] Jan 27 15:45:04 crc kubenswrapper[4772]: I0127 15:45:04.589656 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492100-r2zj6"] Jan 27 15:45:04 crc kubenswrapper[4772]: I0127 15:45:04.672960 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6b54ae2-d365-4988-8e69-704574c7962a" path="/var/lib/kubelet/pods/c6b54ae2-d365-4988-8e69-704574c7962a/volumes" Jan 27 15:45:05 crc kubenswrapper[4772]: I0127 15:45:05.385539 4772 scope.go:117] "RemoveContainer" containerID="4f5ed02624877f82608d4a7a5fead892a80497d0b63bf729eaa6c0d56cf6aac6" Jan 27 15:45:16 crc kubenswrapper[4772]: I0127 15:45:16.663197 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:45:16 crc kubenswrapper[4772]: E0127 15:45:16.664056 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:45:28 crc kubenswrapper[4772]: I0127 15:45:28.664093 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:45:28 crc kubenswrapper[4772]: E0127 15:45:28.665785 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:45:41 crc kubenswrapper[4772]: I0127 15:45:41.663628 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:45:41 crc kubenswrapper[4772]: E0127 15:45:41.665618 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:45:53 crc kubenswrapper[4772]: I0127 15:45:53.663444 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:45:53 crc kubenswrapper[4772]: E0127 15:45:53.664104 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:46:04 crc kubenswrapper[4772]: I0127 15:46:04.667523 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:46:04 crc kubenswrapper[4772]: E0127 15:46:04.668285 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:46:16 crc kubenswrapper[4772]: I0127 15:46:16.664193 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:46:16 crc kubenswrapper[4772]: E0127 15:46:16.665283 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:46:29 crc kubenswrapper[4772]: I0127 15:46:29.664705 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:46:29 crc kubenswrapper[4772]: E0127 15:46:29.666004 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:46:31 crc kubenswrapper[4772]: I0127 15:46:31.772479 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n5mdv"] Jan 27 15:46:31 crc kubenswrapper[4772]: E0127 15:46:31.773215 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df82c0c4-9652-407e-b63d-17e2ccdb38aa" containerName="collect-profiles" Jan 27 15:46:31 crc kubenswrapper[4772]: I0127 15:46:31.773232 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="df82c0c4-9652-407e-b63d-17e2ccdb38aa" containerName="collect-profiles" Jan 27 15:46:31 crc kubenswrapper[4772]: I0127 15:46:31.773397 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="df82c0c4-9652-407e-b63d-17e2ccdb38aa" containerName="collect-profiles" Jan 27 15:46:31 crc kubenswrapper[4772]: I0127 15:46:31.774727 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n5mdv" Jan 27 15:46:31 crc kubenswrapper[4772]: I0127 15:46:31.786248 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n5mdv"] Jan 27 15:46:31 crc kubenswrapper[4772]: I0127 15:46:31.824126 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ced76afc-fd78-450c-b74e-dec420ed75db-catalog-content\") pod \"community-operators-n5mdv\" (UID: \"ced76afc-fd78-450c-b74e-dec420ed75db\") " pod="openshift-marketplace/community-operators-n5mdv" Jan 27 15:46:31 crc kubenswrapper[4772]: I0127 15:46:31.824233 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ced76afc-fd78-450c-b74e-dec420ed75db-utilities\") pod \"community-operators-n5mdv\" (UID: \"ced76afc-fd78-450c-b74e-dec420ed75db\") " pod="openshift-marketplace/community-operators-n5mdv" Jan 27 15:46:31 crc kubenswrapper[4772]: I0127 15:46:31.824269 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92bc9\" (UniqueName: \"kubernetes.io/projected/ced76afc-fd78-450c-b74e-dec420ed75db-kube-api-access-92bc9\") pod \"community-operators-n5mdv\" (UID: \"ced76afc-fd78-450c-b74e-dec420ed75db\") " pod="openshift-marketplace/community-operators-n5mdv" Jan 27 15:46:31 crc kubenswrapper[4772]: I0127 15:46:31.925254 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ced76afc-fd78-450c-b74e-dec420ed75db-utilities\") pod \"community-operators-n5mdv\" (UID: \"ced76afc-fd78-450c-b74e-dec420ed75db\") " pod="openshift-marketplace/community-operators-n5mdv" Jan 27 15:46:31 crc kubenswrapper[4772]: I0127 15:46:31.925312 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92bc9\" (UniqueName: \"kubernetes.io/projected/ced76afc-fd78-450c-b74e-dec420ed75db-kube-api-access-92bc9\") pod \"community-operators-n5mdv\" (UID: \"ced76afc-fd78-450c-b74e-dec420ed75db\") " pod="openshift-marketplace/community-operators-n5mdv" Jan 27 15:46:31 crc kubenswrapper[4772]: I0127 15:46:31.925345 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ced76afc-fd78-450c-b74e-dec420ed75db-catalog-content\") pod \"community-operators-n5mdv\" (UID: \"ced76afc-fd78-450c-b74e-dec420ed75db\") " pod="openshift-marketplace/community-operators-n5mdv" Jan 27 15:46:31 crc kubenswrapper[4772]: I0127 15:46:31.925757 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ced76afc-fd78-450c-b74e-dec420ed75db-utilities\") pod \"community-operators-n5mdv\" (UID: \"ced76afc-fd78-450c-b74e-dec420ed75db\") " pod="openshift-marketplace/community-operators-n5mdv" Jan 27 15:46:31 crc kubenswrapper[4772]: I0127 15:46:31.925799 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ced76afc-fd78-450c-b74e-dec420ed75db-catalog-content\") pod \"community-operators-n5mdv\" (UID: \"ced76afc-fd78-450c-b74e-dec420ed75db\") " pod="openshift-marketplace/community-operators-n5mdv" Jan 27 15:46:31 crc kubenswrapper[4772]: I0127 15:46:31.946266 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92bc9\" (UniqueName: \"kubernetes.io/projected/ced76afc-fd78-450c-b74e-dec420ed75db-kube-api-access-92bc9\") pod \"community-operators-n5mdv\" (UID: \"ced76afc-fd78-450c-b74e-dec420ed75db\") " pod="openshift-marketplace/community-operators-n5mdv" Jan 27 15:46:32 crc kubenswrapper[4772]: I0127 15:46:32.139690 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n5mdv" Jan 27 15:46:32 crc kubenswrapper[4772]: I0127 15:46:32.626626 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n5mdv"] Jan 27 15:46:33 crc kubenswrapper[4772]: I0127 15:46:33.149333 4772 generic.go:334] "Generic (PLEG): container finished" podID="ced76afc-fd78-450c-b74e-dec420ed75db" containerID="8ac0fc5d06479293a7d120d508c0b491eefe60dc43603e556115fb79adac9ffa" exitCode=0 Jan 27 15:46:33 crc kubenswrapper[4772]: I0127 15:46:33.149474 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n5mdv" event={"ID":"ced76afc-fd78-450c-b74e-dec420ed75db","Type":"ContainerDied","Data":"8ac0fc5d06479293a7d120d508c0b491eefe60dc43603e556115fb79adac9ffa"} Jan 27 15:46:33 crc kubenswrapper[4772]: I0127 15:46:33.149689 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n5mdv" event={"ID":"ced76afc-fd78-450c-b74e-dec420ed75db","Type":"ContainerStarted","Data":"b82ef5fd5075c8859afe639fb3cfeb3d073b576a10725fee36633021a1033fa0"} Jan 27 15:46:33 crc kubenswrapper[4772]: I0127 15:46:33.152222 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 15:46:35 crc kubenswrapper[4772]: I0127 15:46:35.165815 4772 generic.go:334] "Generic (PLEG): container finished" podID="ced76afc-fd78-450c-b74e-dec420ed75db" containerID="c4efac9ce3e6005b64b6af3bdc85f43a1a209171b2708c0685747465c7568cf9" exitCode=0 Jan 27 15:46:35 crc kubenswrapper[4772]: I0127 15:46:35.166016 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n5mdv" event={"ID":"ced76afc-fd78-450c-b74e-dec420ed75db","Type":"ContainerDied","Data":"c4efac9ce3e6005b64b6af3bdc85f43a1a209171b2708c0685747465c7568cf9"} Jan 27 15:46:36 crc kubenswrapper[4772]: I0127 15:46:36.176814 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n5mdv" event={"ID":"ced76afc-fd78-450c-b74e-dec420ed75db","Type":"ContainerStarted","Data":"c3b0df09bdf9b7355b490a4af154b0d2e84dbe10db8179c311fd237940d8cbc0"} Jan 27 15:46:36 crc kubenswrapper[4772]: I0127 15:46:36.194275 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n5mdv" podStartSLOduration=2.741323146 podStartE2EDuration="5.194255754s" podCreationTimestamp="2026-01-27 15:46:31 +0000 UTC" firstStartedPulling="2026-01-27 15:46:33.151908311 +0000 UTC m=+2379.132517409" lastFinishedPulling="2026-01-27 15:46:35.604840919 +0000 UTC m=+2381.585450017" observedRunningTime="2026-01-27 15:46:36.190824667 +0000 UTC m=+2382.171433765" watchObservedRunningTime="2026-01-27 15:46:36.194255754 +0000 UTC m=+2382.174864852" Jan 27 15:46:40 crc kubenswrapper[4772]: I0127 15:46:40.663374 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:46:40 crc kubenswrapper[4772]: E0127 15:46:40.664101 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:46:42 crc kubenswrapper[4772]: I0127 15:46:42.140785 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n5mdv" Jan 27 15:46:42 crc kubenswrapper[4772]: I0127 15:46:42.141103 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n5mdv" Jan 27 15:46:42 crc kubenswrapper[4772]: I0127 15:46:42.198857 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n5mdv" Jan 27 15:46:42 crc kubenswrapper[4772]: I0127 15:46:42.272105 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n5mdv" Jan 27 15:46:42 crc kubenswrapper[4772]: I0127 15:46:42.439832 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n5mdv"] Jan 27 15:46:44 crc kubenswrapper[4772]: I0127 15:46:44.237387 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n5mdv" podUID="ced76afc-fd78-450c-b74e-dec420ed75db" containerName="registry-server" containerID="cri-o://c3b0df09bdf9b7355b490a4af154b0d2e84dbe10db8179c311fd237940d8cbc0" gracePeriod=2 Jan 27 15:46:44 crc kubenswrapper[4772]: I0127 15:46:44.587963 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n5mdv" Jan 27 15:46:44 crc kubenswrapper[4772]: I0127 15:46:44.617028 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ced76afc-fd78-450c-b74e-dec420ed75db-utilities\") pod \"ced76afc-fd78-450c-b74e-dec420ed75db\" (UID: \"ced76afc-fd78-450c-b74e-dec420ed75db\") " Jan 27 15:46:44 crc kubenswrapper[4772]: I0127 15:46:44.617080 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92bc9\" (UniqueName: \"kubernetes.io/projected/ced76afc-fd78-450c-b74e-dec420ed75db-kube-api-access-92bc9\") pod \"ced76afc-fd78-450c-b74e-dec420ed75db\" (UID: \"ced76afc-fd78-450c-b74e-dec420ed75db\") " Jan 27 15:46:44 crc kubenswrapper[4772]: I0127 15:46:44.617150 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ced76afc-fd78-450c-b74e-dec420ed75db-catalog-content\") pod \"ced76afc-fd78-450c-b74e-dec420ed75db\" (UID: \"ced76afc-fd78-450c-b74e-dec420ed75db\") " Jan 27 15:46:44 crc kubenswrapper[4772]: I0127 15:46:44.619130 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ced76afc-fd78-450c-b74e-dec420ed75db-utilities" (OuterVolumeSpecName: "utilities") pod "ced76afc-fd78-450c-b74e-dec420ed75db" (UID: "ced76afc-fd78-450c-b74e-dec420ed75db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:46:44 crc kubenswrapper[4772]: I0127 15:46:44.627037 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ced76afc-fd78-450c-b74e-dec420ed75db-kube-api-access-92bc9" (OuterVolumeSpecName: "kube-api-access-92bc9") pod "ced76afc-fd78-450c-b74e-dec420ed75db" (UID: "ced76afc-fd78-450c-b74e-dec420ed75db"). InnerVolumeSpecName "kube-api-access-92bc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:46:44 crc kubenswrapper[4772]: I0127 15:46:44.682596 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ced76afc-fd78-450c-b74e-dec420ed75db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ced76afc-fd78-450c-b74e-dec420ed75db" (UID: "ced76afc-fd78-450c-b74e-dec420ed75db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:46:44 crc kubenswrapper[4772]: I0127 15:46:44.718575 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92bc9\" (UniqueName: \"kubernetes.io/projected/ced76afc-fd78-450c-b74e-dec420ed75db-kube-api-access-92bc9\") on node \"crc\" DevicePath \"\"" Jan 27 15:46:44 crc kubenswrapper[4772]: I0127 15:46:44.718604 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ced76afc-fd78-450c-b74e-dec420ed75db-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 15:46:44 crc kubenswrapper[4772]: I0127 15:46:44.718613 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ced76afc-fd78-450c-b74e-dec420ed75db-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 15:46:45 crc kubenswrapper[4772]: I0127 15:46:45.250368 4772 generic.go:334] "Generic (PLEG): container finished" podID="ced76afc-fd78-450c-b74e-dec420ed75db" containerID="c3b0df09bdf9b7355b490a4af154b0d2e84dbe10db8179c311fd237940d8cbc0" exitCode=0 Jan 27 15:46:45 crc kubenswrapper[4772]: I0127 15:46:45.250494 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n5mdv" event={"ID":"ced76afc-fd78-450c-b74e-dec420ed75db","Type":"ContainerDied","Data":"c3b0df09bdf9b7355b490a4af154b0d2e84dbe10db8179c311fd237940d8cbc0"} Jan 27 15:46:45 crc kubenswrapper[4772]: I0127 15:46:45.252038 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n5mdv" event={"ID":"ced76afc-fd78-450c-b74e-dec420ed75db","Type":"ContainerDied","Data":"b82ef5fd5075c8859afe639fb3cfeb3d073b576a10725fee36633021a1033fa0"} Jan 27 15:46:45 crc kubenswrapper[4772]: I0127 15:46:45.250653 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n5mdv" Jan 27 15:46:45 crc kubenswrapper[4772]: I0127 15:46:45.252078 4772 scope.go:117] "RemoveContainer" containerID="c3b0df09bdf9b7355b490a4af154b0d2e84dbe10db8179c311fd237940d8cbc0" Jan 27 15:46:45 crc kubenswrapper[4772]: I0127 15:46:45.276526 4772 scope.go:117] "RemoveContainer" containerID="c4efac9ce3e6005b64b6af3bdc85f43a1a209171b2708c0685747465c7568cf9" Jan 27 15:46:45 crc kubenswrapper[4772]: I0127 15:46:45.312398 4772 scope.go:117] "RemoveContainer" containerID="8ac0fc5d06479293a7d120d508c0b491eefe60dc43603e556115fb79adac9ffa" Jan 27 15:46:45 crc kubenswrapper[4772]: I0127 15:46:45.362485 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n5mdv"] Jan 27 15:46:45 crc kubenswrapper[4772]: I0127 15:46:45.371384 4772 scope.go:117] "RemoveContainer" containerID="c3b0df09bdf9b7355b490a4af154b0d2e84dbe10db8179c311fd237940d8cbc0" Jan 27 15:46:45 crc kubenswrapper[4772]: I0127 15:46:45.372229 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n5mdv"] Jan 27 15:46:45 crc kubenswrapper[4772]: E0127 15:46:45.372510 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3b0df09bdf9b7355b490a4af154b0d2e84dbe10db8179c311fd237940d8cbc0\": container with ID starting with c3b0df09bdf9b7355b490a4af154b0d2e84dbe10db8179c311fd237940d8cbc0 not found: ID does not exist" containerID="c3b0df09bdf9b7355b490a4af154b0d2e84dbe10db8179c311fd237940d8cbc0" Jan 27 15:46:45 crc kubenswrapper[4772]: I0127 15:46:45.372573 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3b0df09bdf9b7355b490a4af154b0d2e84dbe10db8179c311fd237940d8cbc0"} err="failed to get container status \"c3b0df09bdf9b7355b490a4af154b0d2e84dbe10db8179c311fd237940d8cbc0\": rpc error: code = NotFound desc = could not find container \"c3b0df09bdf9b7355b490a4af154b0d2e84dbe10db8179c311fd237940d8cbc0\": container with ID starting with c3b0df09bdf9b7355b490a4af154b0d2e84dbe10db8179c311fd237940d8cbc0 not found: ID does not exist" Jan 27 15:46:45 crc kubenswrapper[4772]: I0127 15:46:45.372606 4772 scope.go:117] "RemoveContainer" containerID="c4efac9ce3e6005b64b6af3bdc85f43a1a209171b2708c0685747465c7568cf9" Jan 27 15:46:45 crc kubenswrapper[4772]: E0127 15:46:45.373108 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4efac9ce3e6005b64b6af3bdc85f43a1a209171b2708c0685747465c7568cf9\": container with ID starting with c4efac9ce3e6005b64b6af3bdc85f43a1a209171b2708c0685747465c7568cf9 not found: ID does not exist" containerID="c4efac9ce3e6005b64b6af3bdc85f43a1a209171b2708c0685747465c7568cf9" Jan 27 15:46:45 crc kubenswrapper[4772]: I0127 15:46:45.373143 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4efac9ce3e6005b64b6af3bdc85f43a1a209171b2708c0685747465c7568cf9"} err="failed to get container status \"c4efac9ce3e6005b64b6af3bdc85f43a1a209171b2708c0685747465c7568cf9\": rpc error: code = NotFound desc = could not find container \"c4efac9ce3e6005b64b6af3bdc85f43a1a209171b2708c0685747465c7568cf9\": container with ID starting with c4efac9ce3e6005b64b6af3bdc85f43a1a209171b2708c0685747465c7568cf9 not found: ID does not exist" Jan 27 15:46:45 crc kubenswrapper[4772]: I0127 15:46:45.373187 4772 scope.go:117] "RemoveContainer" containerID="8ac0fc5d06479293a7d120d508c0b491eefe60dc43603e556115fb79adac9ffa" Jan 27 15:46:45 crc kubenswrapper[4772]: E0127 15:46:45.373511 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ac0fc5d06479293a7d120d508c0b491eefe60dc43603e556115fb79adac9ffa\": container with ID starting with 8ac0fc5d06479293a7d120d508c0b491eefe60dc43603e556115fb79adac9ffa not found: ID does not exist" containerID="8ac0fc5d06479293a7d120d508c0b491eefe60dc43603e556115fb79adac9ffa" Jan 27 15:46:45 crc kubenswrapper[4772]: I0127 15:46:45.373545 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ac0fc5d06479293a7d120d508c0b491eefe60dc43603e556115fb79adac9ffa"} err="failed to get container status \"8ac0fc5d06479293a7d120d508c0b491eefe60dc43603e556115fb79adac9ffa\": rpc error: code = NotFound desc = could not find container \"8ac0fc5d06479293a7d120d508c0b491eefe60dc43603e556115fb79adac9ffa\": container with ID starting with 8ac0fc5d06479293a7d120d508c0b491eefe60dc43603e556115fb79adac9ffa not found: ID does not exist" Jan 27 15:46:46 crc kubenswrapper[4772]: I0127 15:46:46.682444 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ced76afc-fd78-450c-b74e-dec420ed75db" path="/var/lib/kubelet/pods/ced76afc-fd78-450c-b74e-dec420ed75db/volumes" Jan 27 15:46:51 crc kubenswrapper[4772]: I0127 15:46:51.662970 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:46:51 crc kubenswrapper[4772]: E0127 15:46:51.663858 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:47:05 crc kubenswrapper[4772]: I0127 15:47:05.663271 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:47:05 crc kubenswrapper[4772]: E0127 15:47:05.663959 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:47:17 crc kubenswrapper[4772]: I0127 15:47:17.663519 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:47:17 crc kubenswrapper[4772]: E0127 15:47:17.664649 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:47:28 crc kubenswrapper[4772]: I0127 15:47:28.663603 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:47:28 crc kubenswrapper[4772]: E0127 15:47:28.664379 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:47:41 crc kubenswrapper[4772]: I0127 15:47:41.662842 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:47:41 crc kubenswrapper[4772]: E0127 15:47:41.663697 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:47:53 crc kubenswrapper[4772]: I0127 15:47:53.663910 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:47:53 crc kubenswrapper[4772]: E0127 15:47:53.664814 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:48:05 crc kubenswrapper[4772]: I0127 15:48:05.662624 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:48:05 crc kubenswrapper[4772]: E0127 15:48:05.663324 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:48:17 crc kubenswrapper[4772]: I0127 15:48:17.663551 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:48:17 crc kubenswrapper[4772]: E0127 15:48:17.664323 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:48:31 crc kubenswrapper[4772]: I0127 15:48:31.663144 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:48:31 crc kubenswrapper[4772]: E0127 15:48:31.663872 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:48:44 crc kubenswrapper[4772]: I0127 15:48:44.668706 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:48:44 crc kubenswrapper[4772]: E0127 15:48:44.669619 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:48:57 crc kubenswrapper[4772]: I0127 15:48:57.663606 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:48:57 crc kubenswrapper[4772]: E0127 15:48:57.664464 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:49:12 crc kubenswrapper[4772]: I0127 15:49:12.663515 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:49:12 crc kubenswrapper[4772]: I0127 15:49:12.980397 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"882ea0b40217e0829fd486d6dedd680ff982d35f68339f2340bf98c3fe9a0364"} Jan 27 15:50:19 crc kubenswrapper[4772]: I0127 15:50:19.420092 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-659gv"] Jan 27 15:50:19 crc kubenswrapper[4772]: E0127 15:50:19.421047 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ced76afc-fd78-450c-b74e-dec420ed75db" containerName="extract-utilities" Jan 27 15:50:19 crc kubenswrapper[4772]: I0127 15:50:19.421063 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="ced76afc-fd78-450c-b74e-dec420ed75db" containerName="extract-utilities" Jan 27 15:50:19 crc kubenswrapper[4772]: E0127 15:50:19.421083 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ced76afc-fd78-450c-b74e-dec420ed75db" containerName="extract-content" Jan 27 15:50:19 crc kubenswrapper[4772]: I0127 15:50:19.421092 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="ced76afc-fd78-450c-b74e-dec420ed75db" containerName="extract-content" Jan 27 15:50:19 crc kubenswrapper[4772]: E0127 15:50:19.421113 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ced76afc-fd78-450c-b74e-dec420ed75db" containerName="registry-server" Jan 27 15:50:19 crc kubenswrapper[4772]: I0127 15:50:19.421121 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="ced76afc-fd78-450c-b74e-dec420ed75db" containerName="registry-server" Jan 27 15:50:19 crc kubenswrapper[4772]: I0127 15:50:19.421330 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="ced76afc-fd78-450c-b74e-dec420ed75db" containerName="registry-server" Jan 27 15:50:19 crc kubenswrapper[4772]: I0127 15:50:19.422478 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-659gv" Jan 27 15:50:19 crc kubenswrapper[4772]: I0127 15:50:19.443747 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-659gv"] Jan 27 15:50:19 crc kubenswrapper[4772]: I0127 15:50:19.553187 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/226b848a-07cc-44c2-abc3-c60d88c569ca-catalog-content\") pod \"redhat-operators-659gv\" (UID: \"226b848a-07cc-44c2-abc3-c60d88c569ca\") " pod="openshift-marketplace/redhat-operators-659gv" Jan 27 15:50:19 crc kubenswrapper[4772]: I0127 15:50:19.553288 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-575hx\" (UniqueName: \"kubernetes.io/projected/226b848a-07cc-44c2-abc3-c60d88c569ca-kube-api-access-575hx\") pod \"redhat-operators-659gv\" (UID: \"226b848a-07cc-44c2-abc3-c60d88c569ca\") " pod="openshift-marketplace/redhat-operators-659gv" Jan 27 15:50:19 crc kubenswrapper[4772]: I0127 15:50:19.553343 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/226b848a-07cc-44c2-abc3-c60d88c569ca-utilities\") pod \"redhat-operators-659gv\" (UID: \"226b848a-07cc-44c2-abc3-c60d88c569ca\") " pod="openshift-marketplace/redhat-operators-659gv" Jan 27 15:50:19 crc kubenswrapper[4772]: I0127 15:50:19.654771 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/226b848a-07cc-44c2-abc3-c60d88c569ca-catalog-content\") pod \"redhat-operators-659gv\" (UID: \"226b848a-07cc-44c2-abc3-c60d88c569ca\") " pod="openshift-marketplace/redhat-operators-659gv" Jan 27 15:50:19 crc kubenswrapper[4772]: I0127 15:50:19.654891 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-575hx\" (UniqueName: \"kubernetes.io/projected/226b848a-07cc-44c2-abc3-c60d88c569ca-kube-api-access-575hx\") pod \"redhat-operators-659gv\" (UID: \"226b848a-07cc-44c2-abc3-c60d88c569ca\") " pod="openshift-marketplace/redhat-operators-659gv" Jan 27 15:50:19 crc kubenswrapper[4772]: I0127 15:50:19.654949 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/226b848a-07cc-44c2-abc3-c60d88c569ca-utilities\") pod \"redhat-operators-659gv\" (UID: \"226b848a-07cc-44c2-abc3-c60d88c569ca\") " pod="openshift-marketplace/redhat-operators-659gv" Jan 27 15:50:19 crc kubenswrapper[4772]: I0127 15:50:19.655554 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/226b848a-07cc-44c2-abc3-c60d88c569ca-catalog-content\") pod \"redhat-operators-659gv\" (UID: \"226b848a-07cc-44c2-abc3-c60d88c569ca\") " pod="openshift-marketplace/redhat-operators-659gv" Jan 27 15:50:19 crc kubenswrapper[4772]: I0127 15:50:19.655581 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/226b848a-07cc-44c2-abc3-c60d88c569ca-utilities\") pod \"redhat-operators-659gv\" (UID: \"226b848a-07cc-44c2-abc3-c60d88c569ca\") " pod="openshift-marketplace/redhat-operators-659gv" Jan 27 15:50:19 crc kubenswrapper[4772]: I0127 15:50:19.678162 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-575hx\" (UniqueName: \"kubernetes.io/projected/226b848a-07cc-44c2-abc3-c60d88c569ca-kube-api-access-575hx\") pod \"redhat-operators-659gv\" (UID: \"226b848a-07cc-44c2-abc3-c60d88c569ca\") " pod="openshift-marketplace/redhat-operators-659gv" Jan 27 15:50:19 crc kubenswrapper[4772]: I0127 15:50:19.745143 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-659gv" Jan 27 15:50:20 crc kubenswrapper[4772]: I0127 15:50:20.205386 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-659gv"] Jan 27 15:50:20 crc kubenswrapper[4772]: I0127 15:50:20.459702 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-659gv" event={"ID":"226b848a-07cc-44c2-abc3-c60d88c569ca","Type":"ContainerStarted","Data":"fce9617314a607b31399658547b245df6c440a62125f27475f93010305600c7d"} Jan 27 15:50:21 crc kubenswrapper[4772]: I0127 15:50:21.468693 4772 generic.go:334] "Generic (PLEG): container finished" podID="226b848a-07cc-44c2-abc3-c60d88c569ca" containerID="79a50a5aee22472dfcc6cefd94361dac06f6e4d3df5bd5b2bc7875a667c02ee9" exitCode=0 Jan 27 15:50:21 crc kubenswrapper[4772]: I0127 15:50:21.468740 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-659gv" event={"ID":"226b848a-07cc-44c2-abc3-c60d88c569ca","Type":"ContainerDied","Data":"79a50a5aee22472dfcc6cefd94361dac06f6e4d3df5bd5b2bc7875a667c02ee9"} Jan 27 15:50:24 crc kubenswrapper[4772]: I0127 15:50:24.520772 4772 generic.go:334] "Generic (PLEG): container finished" podID="226b848a-07cc-44c2-abc3-c60d88c569ca" containerID="585a2a943f1fd491f2e5864b367be29122b001bcdf8da8f664946a2b7086fbc4" exitCode=0 Jan 27 15:50:24 crc kubenswrapper[4772]: I0127 15:50:24.520860 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-659gv" event={"ID":"226b848a-07cc-44c2-abc3-c60d88c569ca","Type":"ContainerDied","Data":"585a2a943f1fd491f2e5864b367be29122b001bcdf8da8f664946a2b7086fbc4"} Jan 27 15:50:26 crc kubenswrapper[4772]: I0127 15:50:26.538152 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-659gv" event={"ID":"226b848a-07cc-44c2-abc3-c60d88c569ca","Type":"ContainerStarted","Data":"d19c7c888f9b5a4f918c07659870716735aa11ca8638eaee98485a627be9df3b"} Jan 27 15:50:29 crc kubenswrapper[4772]: I0127 15:50:29.746200 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-659gv" Jan 27 15:50:29 crc kubenswrapper[4772]: I0127 15:50:29.746559 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-659gv" Jan 27 15:50:30 crc kubenswrapper[4772]: I0127 15:50:30.797628 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-659gv" podUID="226b848a-07cc-44c2-abc3-c60d88c569ca" containerName="registry-server" probeResult="failure" output=< Jan 27 15:50:30 crc kubenswrapper[4772]: timeout: failed to connect service ":50051" within 1s Jan 27 15:50:30 crc kubenswrapper[4772]: > Jan 27 15:50:39 crc kubenswrapper[4772]: I0127 15:50:39.789153 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-659gv" Jan 27 15:50:39 crc kubenswrapper[4772]: I0127 15:50:39.814702 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-659gv" podStartSLOduration=16.034314127 podStartE2EDuration="20.814681841s" podCreationTimestamp="2026-01-27 15:50:19 +0000 UTC" firstStartedPulling="2026-01-27 15:50:21.472318495 +0000 UTC m=+2607.452927593" lastFinishedPulling="2026-01-27 15:50:26.252686209 +0000 UTC m=+2612.233295307" observedRunningTime="2026-01-27 15:50:26.56369759 +0000 UTC m=+2612.544306688" watchObservedRunningTime="2026-01-27 15:50:39.814681841 +0000 UTC m=+2625.795290939" Jan 27 15:50:39 crc kubenswrapper[4772]: I0127 15:50:39.832869 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-659gv" Jan 27 15:50:40 crc kubenswrapper[4772]: I0127 15:50:40.026881 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-659gv"] Jan 27 15:50:41 crc kubenswrapper[4772]: I0127 15:50:41.654322 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-659gv" podUID="226b848a-07cc-44c2-abc3-c60d88c569ca" containerName="registry-server" containerID="cri-o://d19c7c888f9b5a4f918c07659870716735aa11ca8638eaee98485a627be9df3b" gracePeriod=2 Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.153656 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-659gv" Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.278022 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-575hx\" (UniqueName: \"kubernetes.io/projected/226b848a-07cc-44c2-abc3-c60d88c569ca-kube-api-access-575hx\") pod \"226b848a-07cc-44c2-abc3-c60d88c569ca\" (UID: \"226b848a-07cc-44c2-abc3-c60d88c569ca\") " Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.278402 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/226b848a-07cc-44c2-abc3-c60d88c569ca-utilities\") pod \"226b848a-07cc-44c2-abc3-c60d88c569ca\" (UID: \"226b848a-07cc-44c2-abc3-c60d88c569ca\") " Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.278560 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/226b848a-07cc-44c2-abc3-c60d88c569ca-catalog-content\") pod \"226b848a-07cc-44c2-abc3-c60d88c569ca\" (UID: \"226b848a-07cc-44c2-abc3-c60d88c569ca\") " Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.279266 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/226b848a-07cc-44c2-abc3-c60d88c569ca-utilities" (OuterVolumeSpecName: "utilities") pod "226b848a-07cc-44c2-abc3-c60d88c569ca" (UID: "226b848a-07cc-44c2-abc3-c60d88c569ca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.288981 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/226b848a-07cc-44c2-abc3-c60d88c569ca-kube-api-access-575hx" (OuterVolumeSpecName: "kube-api-access-575hx") pod "226b848a-07cc-44c2-abc3-c60d88c569ca" (UID: "226b848a-07cc-44c2-abc3-c60d88c569ca"). InnerVolumeSpecName "kube-api-access-575hx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.380402 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-575hx\" (UniqueName: \"kubernetes.io/projected/226b848a-07cc-44c2-abc3-c60d88c569ca-kube-api-access-575hx\") on node \"crc\" DevicePath \"\"" Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.380459 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/226b848a-07cc-44c2-abc3-c60d88c569ca-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.416528 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/226b848a-07cc-44c2-abc3-c60d88c569ca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "226b848a-07cc-44c2-abc3-c60d88c569ca" (UID: "226b848a-07cc-44c2-abc3-c60d88c569ca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.481552 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/226b848a-07cc-44c2-abc3-c60d88c569ca-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.664015 4772 generic.go:334] "Generic (PLEG): container finished" podID="226b848a-07cc-44c2-abc3-c60d88c569ca" containerID="d19c7c888f9b5a4f918c07659870716735aa11ca8638eaee98485a627be9df3b" exitCode=0 Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.664183 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-659gv" Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.671298 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-659gv" event={"ID":"226b848a-07cc-44c2-abc3-c60d88c569ca","Type":"ContainerDied","Data":"d19c7c888f9b5a4f918c07659870716735aa11ca8638eaee98485a627be9df3b"} Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.671340 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-659gv" event={"ID":"226b848a-07cc-44c2-abc3-c60d88c569ca","Type":"ContainerDied","Data":"fce9617314a607b31399658547b245df6c440a62125f27475f93010305600c7d"} Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.671362 4772 scope.go:117] "RemoveContainer" containerID="d19c7c888f9b5a4f918c07659870716735aa11ca8638eaee98485a627be9df3b" Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.695625 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-659gv"] Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.698639 4772 scope.go:117] "RemoveContainer" containerID="585a2a943f1fd491f2e5864b367be29122b001bcdf8da8f664946a2b7086fbc4" Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.702407 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-659gv"] Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.715113 4772 scope.go:117] "RemoveContainer" containerID="79a50a5aee22472dfcc6cefd94361dac06f6e4d3df5bd5b2bc7875a667c02ee9" Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.743386 4772 scope.go:117] "RemoveContainer" containerID="d19c7c888f9b5a4f918c07659870716735aa11ca8638eaee98485a627be9df3b" Jan 27 15:50:42 crc kubenswrapper[4772]: E0127 15:50:42.745568 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d19c7c888f9b5a4f918c07659870716735aa11ca8638eaee98485a627be9df3b\": container with ID starting with d19c7c888f9b5a4f918c07659870716735aa11ca8638eaee98485a627be9df3b not found: ID does not exist" containerID="d19c7c888f9b5a4f918c07659870716735aa11ca8638eaee98485a627be9df3b" Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.745641 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d19c7c888f9b5a4f918c07659870716735aa11ca8638eaee98485a627be9df3b"} err="failed to get container status \"d19c7c888f9b5a4f918c07659870716735aa11ca8638eaee98485a627be9df3b\": rpc error: code = NotFound desc = could not find container \"d19c7c888f9b5a4f918c07659870716735aa11ca8638eaee98485a627be9df3b\": container with ID starting with d19c7c888f9b5a4f918c07659870716735aa11ca8638eaee98485a627be9df3b not found: ID does not exist" Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.745686 4772 scope.go:117] "RemoveContainer" containerID="585a2a943f1fd491f2e5864b367be29122b001bcdf8da8f664946a2b7086fbc4" Jan 27 15:50:42 crc kubenswrapper[4772]: E0127 15:50:42.746139 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"585a2a943f1fd491f2e5864b367be29122b001bcdf8da8f664946a2b7086fbc4\": container with ID starting with 585a2a943f1fd491f2e5864b367be29122b001bcdf8da8f664946a2b7086fbc4 not found: ID does not exist" containerID="585a2a943f1fd491f2e5864b367be29122b001bcdf8da8f664946a2b7086fbc4" Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.746200 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"585a2a943f1fd491f2e5864b367be29122b001bcdf8da8f664946a2b7086fbc4"} err="failed to get container status \"585a2a943f1fd491f2e5864b367be29122b001bcdf8da8f664946a2b7086fbc4\": rpc error: code = NotFound desc = could not find container \"585a2a943f1fd491f2e5864b367be29122b001bcdf8da8f664946a2b7086fbc4\": container with ID starting with 585a2a943f1fd491f2e5864b367be29122b001bcdf8da8f664946a2b7086fbc4 not found: ID does not exist" Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.746231 4772 scope.go:117] "RemoveContainer" containerID="79a50a5aee22472dfcc6cefd94361dac06f6e4d3df5bd5b2bc7875a667c02ee9" Jan 27 15:50:42 crc kubenswrapper[4772]: E0127 15:50:42.746626 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79a50a5aee22472dfcc6cefd94361dac06f6e4d3df5bd5b2bc7875a667c02ee9\": container with ID starting with 79a50a5aee22472dfcc6cefd94361dac06f6e4d3df5bd5b2bc7875a667c02ee9 not found: ID does not exist" containerID="79a50a5aee22472dfcc6cefd94361dac06f6e4d3df5bd5b2bc7875a667c02ee9" Jan 27 15:50:42 crc kubenswrapper[4772]: I0127 15:50:42.746666 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79a50a5aee22472dfcc6cefd94361dac06f6e4d3df5bd5b2bc7875a667c02ee9"} err="failed to get container status \"79a50a5aee22472dfcc6cefd94361dac06f6e4d3df5bd5b2bc7875a667c02ee9\": rpc error: code = NotFound desc = could not find container \"79a50a5aee22472dfcc6cefd94361dac06f6e4d3df5bd5b2bc7875a667c02ee9\": container with ID starting with 79a50a5aee22472dfcc6cefd94361dac06f6e4d3df5bd5b2bc7875a667c02ee9 not found: ID does not exist" Jan 27 15:50:44 crc kubenswrapper[4772]: I0127 15:50:44.670475 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="226b848a-07cc-44c2-abc3-c60d88c569ca" path="/var/lib/kubelet/pods/226b848a-07cc-44c2-abc3-c60d88c569ca/volumes" Jan 27 15:51:12 crc kubenswrapper[4772]: I0127 15:51:12.058766 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:51:12 crc kubenswrapper[4772]: I0127 15:51:12.059449 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:51:42 crc kubenswrapper[4772]: I0127 15:51:42.058661 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:51:42 crc kubenswrapper[4772]: I0127 15:51:42.059305 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:52:12 crc kubenswrapper[4772]: I0127 15:52:12.058991 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:52:12 crc kubenswrapper[4772]: I0127 15:52:12.059578 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:52:12 crc kubenswrapper[4772]: I0127 15:52:12.059631 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:52:12 crc kubenswrapper[4772]: I0127 15:52:12.060224 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"882ea0b40217e0829fd486d6dedd680ff982d35f68339f2340bf98c3fe9a0364"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 15:52:12 crc kubenswrapper[4772]: I0127 15:52:12.060276 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://882ea0b40217e0829fd486d6dedd680ff982d35f68339f2340bf98c3fe9a0364" gracePeriod=600 Jan 27 15:52:12 crc kubenswrapper[4772]: I0127 15:52:12.280036 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="882ea0b40217e0829fd486d6dedd680ff982d35f68339f2340bf98c3fe9a0364" exitCode=0 Jan 27 15:52:12 crc kubenswrapper[4772]: I0127 15:52:12.280098 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"882ea0b40217e0829fd486d6dedd680ff982d35f68339f2340bf98c3fe9a0364"} Jan 27 15:52:12 crc kubenswrapper[4772]: I0127 15:52:12.280681 4772 scope.go:117] "RemoveContainer" containerID="2d44d6ccc12cd5721067948851620a9f0611d13982269d5631689aef90c34d5d" Jan 27 15:52:13 crc kubenswrapper[4772]: I0127 15:52:13.289990 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51"} Jan 27 15:54:12 crc kubenswrapper[4772]: I0127 15:54:12.059119 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:54:12 crc kubenswrapper[4772]: I0127 15:54:12.059737 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:54:42 crc kubenswrapper[4772]: I0127 15:54:42.058350 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:54:42 crc kubenswrapper[4772]: I0127 15:54:42.059031 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:55:12 crc kubenswrapper[4772]: I0127 15:55:12.058749 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 15:55:12 crc kubenswrapper[4772]: I0127 15:55:12.060321 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 15:55:12 crc kubenswrapper[4772]: I0127 15:55:12.060425 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 15:55:12 crc kubenswrapper[4772]: I0127 15:55:12.061583 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 15:55:12 crc kubenswrapper[4772]: I0127 15:55:12.061678 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" gracePeriod=600 Jan 27 15:55:12 crc kubenswrapper[4772]: E0127 15:55:12.185053 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:55:12 crc kubenswrapper[4772]: I0127 15:55:12.695953 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" exitCode=0 Jan 27 15:55:12 crc kubenswrapper[4772]: I0127 15:55:12.696006 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51"} Jan 27 15:55:12 crc kubenswrapper[4772]: I0127 15:55:12.696050 4772 scope.go:117] "RemoveContainer" containerID="882ea0b40217e0829fd486d6dedd680ff982d35f68339f2340bf98c3fe9a0364" Jan 27 15:55:12 crc kubenswrapper[4772]: I0127 15:55:12.696762 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 15:55:12 crc kubenswrapper[4772]: E0127 15:55:12.697068 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:55:26 crc kubenswrapper[4772]: I0127 15:55:26.662872 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 15:55:26 crc kubenswrapper[4772]: E0127 15:55:26.663505 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:55:39 crc kubenswrapper[4772]: I0127 15:55:39.662539 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 15:55:39 crc kubenswrapper[4772]: E0127 15:55:39.663209 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:55:54 crc kubenswrapper[4772]: I0127 15:55:54.668622 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 15:55:54 crc kubenswrapper[4772]: E0127 15:55:54.669840 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:56:08 crc kubenswrapper[4772]: I0127 15:56:08.663472 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 15:56:08 crc kubenswrapper[4772]: E0127 15:56:08.664347 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:56:20 crc kubenswrapper[4772]: I0127 15:56:20.663332 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 15:56:20 crc kubenswrapper[4772]: E0127 15:56:20.664201 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:56:35 crc kubenswrapper[4772]: I0127 15:56:35.662392 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 15:56:35 crc kubenswrapper[4772]: E0127 15:56:35.663078 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:56:48 crc kubenswrapper[4772]: I0127 15:56:48.663153 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 15:56:48 crc kubenswrapper[4772]: E0127 15:56:48.663847 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:57:00 crc kubenswrapper[4772]: I0127 15:57:00.663559 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 15:57:00 crc kubenswrapper[4772]: E0127 15:57:00.664799 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:57:11 crc kubenswrapper[4772]: I0127 15:57:11.663128 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 15:57:11 crc kubenswrapper[4772]: E0127 15:57:11.663870 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:57:25 crc kubenswrapper[4772]: I0127 15:57:25.663256 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 15:57:25 crc kubenswrapper[4772]: E0127 15:57:25.664105 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:57:36 crc kubenswrapper[4772]: I0127 15:57:36.663329 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 15:57:36 crc kubenswrapper[4772]: E0127 15:57:36.665768 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:57:49 crc kubenswrapper[4772]: I0127 15:57:49.663145 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 15:57:49 crc kubenswrapper[4772]: E0127 15:57:49.664488 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:58:00 crc kubenswrapper[4772]: I0127 15:58:00.664141 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 15:58:00 crc kubenswrapper[4772]: E0127 15:58:00.664988 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:58:12 crc kubenswrapper[4772]: I0127 15:58:12.663060 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 15:58:12 crc kubenswrapper[4772]: E0127 15:58:12.664061 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:58:26 crc kubenswrapper[4772]: I0127 15:58:26.666797 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 15:58:26 crc kubenswrapper[4772]: E0127 15:58:26.667569 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:58:40 crc kubenswrapper[4772]: I0127 15:58:40.664033 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 15:58:40 crc kubenswrapper[4772]: E0127 15:58:40.664887 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:58:54 crc kubenswrapper[4772]: I0127 15:58:54.667067 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 15:58:54 crc kubenswrapper[4772]: E0127 15:58:54.667830 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:59:06 crc kubenswrapper[4772]: I0127 15:59:06.664190 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 15:59:06 crc kubenswrapper[4772]: E0127 15:59:06.664772 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:59:18 crc kubenswrapper[4772]: I0127 15:59:18.663951 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 15:59:18 crc kubenswrapper[4772]: E0127 15:59:18.664735 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:59:29 crc kubenswrapper[4772]: I0127 15:59:29.663483 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 15:59:29 crc kubenswrapper[4772]: E0127 15:59:29.664345 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:59:43 crc kubenswrapper[4772]: I0127 15:59:43.663495 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 15:59:43 crc kubenswrapper[4772]: E0127 15:59:43.664368 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 15:59:56 crc kubenswrapper[4772]: I0127 15:59:56.663801 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 15:59:56 crc kubenswrapper[4772]: E0127 15:59:56.665020 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:00:00 crc kubenswrapper[4772]: I0127 16:00:00.160359 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492160-4mmgr"] Jan 27 16:00:00 crc kubenswrapper[4772]: E0127 16:00:00.161199 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="226b848a-07cc-44c2-abc3-c60d88c569ca" containerName="extract-utilities" Jan 27 16:00:00 crc kubenswrapper[4772]: I0127 16:00:00.161218 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="226b848a-07cc-44c2-abc3-c60d88c569ca" containerName="extract-utilities" Jan 27 16:00:00 crc kubenswrapper[4772]: E0127 16:00:00.161236 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="226b848a-07cc-44c2-abc3-c60d88c569ca" containerName="registry-server" Jan 27 16:00:00 crc kubenswrapper[4772]: I0127 16:00:00.161243 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="226b848a-07cc-44c2-abc3-c60d88c569ca" containerName="registry-server" Jan 27 16:00:00 crc kubenswrapper[4772]: E0127 16:00:00.161267 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="226b848a-07cc-44c2-abc3-c60d88c569ca" containerName="extract-content" Jan 27 16:00:00 crc kubenswrapper[4772]: I0127 16:00:00.161276 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="226b848a-07cc-44c2-abc3-c60d88c569ca" containerName="extract-content" Jan 27 16:00:00 crc kubenswrapper[4772]: I0127 16:00:00.161436 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="226b848a-07cc-44c2-abc3-c60d88c569ca" containerName="registry-server" Jan 27 16:00:00 crc kubenswrapper[4772]: I0127 16:00:00.161979 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492160-4mmgr" Jan 27 16:00:00 crc kubenswrapper[4772]: I0127 16:00:00.166452 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 27 16:00:00 crc kubenswrapper[4772]: I0127 16:00:00.166456 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 27 16:00:00 crc kubenswrapper[4772]: I0127 16:00:00.170533 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492160-4mmgr"] Jan 27 16:00:00 crc kubenswrapper[4772]: I0127 16:00:00.355645 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2488\" (UniqueName: \"kubernetes.io/projected/9b4afc52-82aa-4768-9cc8-5e9236fc4330-kube-api-access-c2488\") pod \"collect-profiles-29492160-4mmgr\" (UID: \"9b4afc52-82aa-4768-9cc8-5e9236fc4330\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492160-4mmgr" Jan 27 16:00:00 crc kubenswrapper[4772]: I0127 16:00:00.356452 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9b4afc52-82aa-4768-9cc8-5e9236fc4330-secret-volume\") pod \"collect-profiles-29492160-4mmgr\" (UID: \"9b4afc52-82aa-4768-9cc8-5e9236fc4330\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492160-4mmgr" Jan 27 16:00:00 crc kubenswrapper[4772]: I0127 16:00:00.356527 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9b4afc52-82aa-4768-9cc8-5e9236fc4330-config-volume\") pod \"collect-profiles-29492160-4mmgr\" (UID: \"9b4afc52-82aa-4768-9cc8-5e9236fc4330\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492160-4mmgr" Jan 27 16:00:00 crc kubenswrapper[4772]: I0127 16:00:00.458155 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2488\" (UniqueName: \"kubernetes.io/projected/9b4afc52-82aa-4768-9cc8-5e9236fc4330-kube-api-access-c2488\") pod \"collect-profiles-29492160-4mmgr\" (UID: \"9b4afc52-82aa-4768-9cc8-5e9236fc4330\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492160-4mmgr" Jan 27 16:00:00 crc kubenswrapper[4772]: I0127 16:00:00.458252 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9b4afc52-82aa-4768-9cc8-5e9236fc4330-secret-volume\") pod \"collect-profiles-29492160-4mmgr\" (UID: \"9b4afc52-82aa-4768-9cc8-5e9236fc4330\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492160-4mmgr" Jan 27 16:00:00 crc kubenswrapper[4772]: I0127 16:00:00.458272 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9b4afc52-82aa-4768-9cc8-5e9236fc4330-config-volume\") pod \"collect-profiles-29492160-4mmgr\" (UID: \"9b4afc52-82aa-4768-9cc8-5e9236fc4330\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492160-4mmgr" Jan 27 16:00:00 crc kubenswrapper[4772]: I0127 16:00:00.459138 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9b4afc52-82aa-4768-9cc8-5e9236fc4330-config-volume\") pod \"collect-profiles-29492160-4mmgr\" (UID: \"9b4afc52-82aa-4768-9cc8-5e9236fc4330\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492160-4mmgr" Jan 27 16:00:00 crc kubenswrapper[4772]: I0127 16:00:00.467782 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9b4afc52-82aa-4768-9cc8-5e9236fc4330-secret-volume\") pod \"collect-profiles-29492160-4mmgr\" (UID: \"9b4afc52-82aa-4768-9cc8-5e9236fc4330\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492160-4mmgr" Jan 27 16:00:00 crc kubenswrapper[4772]: I0127 16:00:00.478372 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2488\" (UniqueName: \"kubernetes.io/projected/9b4afc52-82aa-4768-9cc8-5e9236fc4330-kube-api-access-c2488\") pod \"collect-profiles-29492160-4mmgr\" (UID: \"9b4afc52-82aa-4768-9cc8-5e9236fc4330\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492160-4mmgr" Jan 27 16:00:00 crc kubenswrapper[4772]: I0127 16:00:00.484756 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492160-4mmgr" Jan 27 16:00:00 crc kubenswrapper[4772]: I0127 16:00:00.918636 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492160-4mmgr"] Jan 27 16:00:01 crc kubenswrapper[4772]: I0127 16:00:01.102683 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492160-4mmgr" event={"ID":"9b4afc52-82aa-4768-9cc8-5e9236fc4330","Type":"ContainerStarted","Data":"8dd8add741d2c060daa432ff2f192c7a04c82b2eab3360197f22c851ca7bd6c0"} Jan 27 16:00:01 crc kubenswrapper[4772]: I0127 16:00:01.102732 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492160-4mmgr" event={"ID":"9b4afc52-82aa-4768-9cc8-5e9236fc4330","Type":"ContainerStarted","Data":"4e6de38fa85b71d772074ba8f95e475599369849f60252db8cf1c2b677f38292"} Jan 27 16:00:01 crc kubenswrapper[4772]: I0127 16:00:01.119108 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29492160-4mmgr" podStartSLOduration=1.119093843 podStartE2EDuration="1.119093843s" podCreationTimestamp="2026-01-27 16:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:00:01.116762686 +0000 UTC m=+3187.097371784" watchObservedRunningTime="2026-01-27 16:00:01.119093843 +0000 UTC m=+3187.099702931" Jan 27 16:00:02 crc kubenswrapper[4772]: I0127 16:00:02.113993 4772 generic.go:334] "Generic (PLEG): container finished" podID="9b4afc52-82aa-4768-9cc8-5e9236fc4330" containerID="8dd8add741d2c060daa432ff2f192c7a04c82b2eab3360197f22c851ca7bd6c0" exitCode=0 Jan 27 16:00:02 crc kubenswrapper[4772]: I0127 16:00:02.114115 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492160-4mmgr" event={"ID":"9b4afc52-82aa-4768-9cc8-5e9236fc4330","Type":"ContainerDied","Data":"8dd8add741d2c060daa432ff2f192c7a04c82b2eab3360197f22c851ca7bd6c0"} Jan 27 16:00:03 crc kubenswrapper[4772]: I0127 16:00:03.392433 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492160-4mmgr" Jan 27 16:00:03 crc kubenswrapper[4772]: I0127 16:00:03.517792 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2488\" (UniqueName: \"kubernetes.io/projected/9b4afc52-82aa-4768-9cc8-5e9236fc4330-kube-api-access-c2488\") pod \"9b4afc52-82aa-4768-9cc8-5e9236fc4330\" (UID: \"9b4afc52-82aa-4768-9cc8-5e9236fc4330\") " Jan 27 16:00:03 crc kubenswrapper[4772]: I0127 16:00:03.517859 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9b4afc52-82aa-4768-9cc8-5e9236fc4330-secret-volume\") pod \"9b4afc52-82aa-4768-9cc8-5e9236fc4330\" (UID: \"9b4afc52-82aa-4768-9cc8-5e9236fc4330\") " Jan 27 16:00:03 crc kubenswrapper[4772]: I0127 16:00:03.517883 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9b4afc52-82aa-4768-9cc8-5e9236fc4330-config-volume\") pod \"9b4afc52-82aa-4768-9cc8-5e9236fc4330\" (UID: \"9b4afc52-82aa-4768-9cc8-5e9236fc4330\") " Jan 27 16:00:03 crc kubenswrapper[4772]: I0127 16:00:03.518629 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b4afc52-82aa-4768-9cc8-5e9236fc4330-config-volume" (OuterVolumeSpecName: "config-volume") pod "9b4afc52-82aa-4768-9cc8-5e9236fc4330" (UID: "9b4afc52-82aa-4768-9cc8-5e9236fc4330"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:00:03 crc kubenswrapper[4772]: I0127 16:00:03.522663 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b4afc52-82aa-4768-9cc8-5e9236fc4330-kube-api-access-c2488" (OuterVolumeSpecName: "kube-api-access-c2488") pod "9b4afc52-82aa-4768-9cc8-5e9236fc4330" (UID: "9b4afc52-82aa-4768-9cc8-5e9236fc4330"). InnerVolumeSpecName "kube-api-access-c2488". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:00:03 crc kubenswrapper[4772]: I0127 16:00:03.522908 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b4afc52-82aa-4768-9cc8-5e9236fc4330-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9b4afc52-82aa-4768-9cc8-5e9236fc4330" (UID: "9b4afc52-82aa-4768-9cc8-5e9236fc4330"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:00:03 crc kubenswrapper[4772]: I0127 16:00:03.619480 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2488\" (UniqueName: \"kubernetes.io/projected/9b4afc52-82aa-4768-9cc8-5e9236fc4330-kube-api-access-c2488\") on node \"crc\" DevicePath \"\"" Jan 27 16:00:03 crc kubenswrapper[4772]: I0127 16:00:03.619534 4772 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9b4afc52-82aa-4768-9cc8-5e9236fc4330-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 27 16:00:03 crc kubenswrapper[4772]: I0127 16:00:03.619551 4772 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9b4afc52-82aa-4768-9cc8-5e9236fc4330-config-volume\") on node \"crc\" DevicePath \"\"" Jan 27 16:00:04 crc kubenswrapper[4772]: I0127 16:00:04.128034 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492160-4mmgr" event={"ID":"9b4afc52-82aa-4768-9cc8-5e9236fc4330","Type":"ContainerDied","Data":"4e6de38fa85b71d772074ba8f95e475599369849f60252db8cf1c2b677f38292"} Jan 27 16:00:04 crc kubenswrapper[4772]: I0127 16:00:04.128086 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e6de38fa85b71d772074ba8f95e475599369849f60252db8cf1c2b677f38292" Jan 27 16:00:04 crc kubenswrapper[4772]: I0127 16:00:04.128123 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492160-4mmgr" Jan 27 16:00:04 crc kubenswrapper[4772]: I0127 16:00:04.465278 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492115-hb89g"] Jan 27 16:00:04 crc kubenswrapper[4772]: I0127 16:00:04.469320 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492115-hb89g"] Jan 27 16:00:04 crc kubenswrapper[4772]: I0127 16:00:04.676415 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1" path="/var/lib/kubelet/pods/616d65ac-8b2b-4b0a-b7a1-ca3516ad7cf1/volumes" Jan 27 16:00:05 crc kubenswrapper[4772]: I0127 16:00:05.727763 4772 scope.go:117] "RemoveContainer" containerID="8931f0cc38dd8c453e687a0b65ac6a9c2d9a0265440b30f5480c6ac0483f9860" Jan 27 16:00:08 crc kubenswrapper[4772]: I0127 16:00:08.663262 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 16:00:08 crc kubenswrapper[4772]: E0127 16:00:08.663800 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:00:09 crc kubenswrapper[4772]: I0127 16:00:09.372021 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8lq7p"] Jan 27 16:00:09 crc kubenswrapper[4772]: E0127 16:00:09.372449 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b4afc52-82aa-4768-9cc8-5e9236fc4330" containerName="collect-profiles" Jan 27 16:00:09 crc kubenswrapper[4772]: I0127 16:00:09.372472 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b4afc52-82aa-4768-9cc8-5e9236fc4330" containerName="collect-profiles" Jan 27 16:00:09 crc kubenswrapper[4772]: I0127 16:00:09.372632 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b4afc52-82aa-4768-9cc8-5e9236fc4330" containerName="collect-profiles" Jan 27 16:00:09 crc kubenswrapper[4772]: I0127 16:00:09.373858 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8lq7p" Jan 27 16:00:09 crc kubenswrapper[4772]: I0127 16:00:09.381030 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8lq7p"] Jan 27 16:00:09 crc kubenswrapper[4772]: I0127 16:00:09.411198 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e55c646-e887-492e-b27d-39536b38b245-catalog-content\") pod \"certified-operators-8lq7p\" (UID: \"0e55c646-e887-492e-b27d-39536b38b245\") " pod="openshift-marketplace/certified-operators-8lq7p" Jan 27 16:00:09 crc kubenswrapper[4772]: I0127 16:00:09.411563 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e55c646-e887-492e-b27d-39536b38b245-utilities\") pod \"certified-operators-8lq7p\" (UID: \"0e55c646-e887-492e-b27d-39536b38b245\") " pod="openshift-marketplace/certified-operators-8lq7p" Jan 27 16:00:09 crc kubenswrapper[4772]: I0127 16:00:09.411584 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfh4p\" (UniqueName: \"kubernetes.io/projected/0e55c646-e887-492e-b27d-39536b38b245-kube-api-access-gfh4p\") pod \"certified-operators-8lq7p\" (UID: \"0e55c646-e887-492e-b27d-39536b38b245\") " pod="openshift-marketplace/certified-operators-8lq7p" Jan 27 16:00:09 crc kubenswrapper[4772]: I0127 16:00:09.512713 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e55c646-e887-492e-b27d-39536b38b245-catalog-content\") pod \"certified-operators-8lq7p\" (UID: \"0e55c646-e887-492e-b27d-39536b38b245\") " pod="openshift-marketplace/certified-operators-8lq7p" Jan 27 16:00:09 crc kubenswrapper[4772]: I0127 16:00:09.512844 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e55c646-e887-492e-b27d-39536b38b245-utilities\") pod \"certified-operators-8lq7p\" (UID: \"0e55c646-e887-492e-b27d-39536b38b245\") " pod="openshift-marketplace/certified-operators-8lq7p" Jan 27 16:00:09 crc kubenswrapper[4772]: I0127 16:00:09.512876 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfh4p\" (UniqueName: \"kubernetes.io/projected/0e55c646-e887-492e-b27d-39536b38b245-kube-api-access-gfh4p\") pod \"certified-operators-8lq7p\" (UID: \"0e55c646-e887-492e-b27d-39536b38b245\") " pod="openshift-marketplace/certified-operators-8lq7p" Jan 27 16:00:09 crc kubenswrapper[4772]: I0127 16:00:09.513430 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e55c646-e887-492e-b27d-39536b38b245-catalog-content\") pod \"certified-operators-8lq7p\" (UID: \"0e55c646-e887-492e-b27d-39536b38b245\") " pod="openshift-marketplace/certified-operators-8lq7p" Jan 27 16:00:09 crc kubenswrapper[4772]: I0127 16:00:09.513563 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e55c646-e887-492e-b27d-39536b38b245-utilities\") pod \"certified-operators-8lq7p\" (UID: \"0e55c646-e887-492e-b27d-39536b38b245\") " pod="openshift-marketplace/certified-operators-8lq7p" Jan 27 16:00:09 crc kubenswrapper[4772]: I0127 16:00:09.531831 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfh4p\" (UniqueName: \"kubernetes.io/projected/0e55c646-e887-492e-b27d-39536b38b245-kube-api-access-gfh4p\") pod \"certified-operators-8lq7p\" (UID: \"0e55c646-e887-492e-b27d-39536b38b245\") " pod="openshift-marketplace/certified-operators-8lq7p" Jan 27 16:00:09 crc kubenswrapper[4772]: I0127 16:00:09.691732 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8lq7p" Jan 27 16:00:10 crc kubenswrapper[4772]: I0127 16:00:10.164367 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8lq7p"] Jan 27 16:00:11 crc kubenswrapper[4772]: I0127 16:00:11.173313 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8d9vw"] Jan 27 16:00:11 crc kubenswrapper[4772]: I0127 16:00:11.177439 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8d9vw" Jan 27 16:00:11 crc kubenswrapper[4772]: I0127 16:00:11.177782 4772 generic.go:334] "Generic (PLEG): container finished" podID="0e55c646-e887-492e-b27d-39536b38b245" containerID="5f1d8fdc7aed07858ca3363c0144e81cdc92e3922938515a33edf543ec95389e" exitCode=0 Jan 27 16:00:11 crc kubenswrapper[4772]: I0127 16:00:11.177829 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8lq7p" event={"ID":"0e55c646-e887-492e-b27d-39536b38b245","Type":"ContainerDied","Data":"5f1d8fdc7aed07858ca3363c0144e81cdc92e3922938515a33edf543ec95389e"} Jan 27 16:00:11 crc kubenswrapper[4772]: I0127 16:00:11.177856 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8lq7p" event={"ID":"0e55c646-e887-492e-b27d-39536b38b245","Type":"ContainerStarted","Data":"215607eb1ef46147418a0a5c60a763e046da6f15ce0cdc380aa944d9788c12db"} Jan 27 16:00:11 crc kubenswrapper[4772]: I0127 16:00:11.182602 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 16:00:11 crc kubenswrapper[4772]: I0127 16:00:11.192826 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8d9vw"] Jan 27 16:00:11 crc kubenswrapper[4772]: I0127 16:00:11.340980 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpxvx\" (UniqueName: \"kubernetes.io/projected/a4dcc32f-d75b-420b-8dc1-c1d45ce1390d-kube-api-access-cpxvx\") pod \"community-operators-8d9vw\" (UID: \"a4dcc32f-d75b-420b-8dc1-c1d45ce1390d\") " pod="openshift-marketplace/community-operators-8d9vw" Jan 27 16:00:11 crc kubenswrapper[4772]: I0127 16:00:11.341063 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4dcc32f-d75b-420b-8dc1-c1d45ce1390d-utilities\") pod \"community-operators-8d9vw\" (UID: \"a4dcc32f-d75b-420b-8dc1-c1d45ce1390d\") " pod="openshift-marketplace/community-operators-8d9vw" Jan 27 16:00:11 crc kubenswrapper[4772]: I0127 16:00:11.341104 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4dcc32f-d75b-420b-8dc1-c1d45ce1390d-catalog-content\") pod \"community-operators-8d9vw\" (UID: \"a4dcc32f-d75b-420b-8dc1-c1d45ce1390d\") " pod="openshift-marketplace/community-operators-8d9vw" Jan 27 16:00:11 crc kubenswrapper[4772]: I0127 16:00:11.442778 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpxvx\" (UniqueName: \"kubernetes.io/projected/a4dcc32f-d75b-420b-8dc1-c1d45ce1390d-kube-api-access-cpxvx\") pod \"community-operators-8d9vw\" (UID: \"a4dcc32f-d75b-420b-8dc1-c1d45ce1390d\") " pod="openshift-marketplace/community-operators-8d9vw" Jan 27 16:00:11 crc kubenswrapper[4772]: I0127 16:00:11.442868 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4dcc32f-d75b-420b-8dc1-c1d45ce1390d-utilities\") pod \"community-operators-8d9vw\" (UID: \"a4dcc32f-d75b-420b-8dc1-c1d45ce1390d\") " pod="openshift-marketplace/community-operators-8d9vw" Jan 27 16:00:11 crc kubenswrapper[4772]: I0127 16:00:11.442904 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4dcc32f-d75b-420b-8dc1-c1d45ce1390d-catalog-content\") pod \"community-operators-8d9vw\" (UID: \"a4dcc32f-d75b-420b-8dc1-c1d45ce1390d\") " pod="openshift-marketplace/community-operators-8d9vw" Jan 27 16:00:11 crc kubenswrapper[4772]: I0127 16:00:11.443434 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4dcc32f-d75b-420b-8dc1-c1d45ce1390d-catalog-content\") pod \"community-operators-8d9vw\" (UID: \"a4dcc32f-d75b-420b-8dc1-c1d45ce1390d\") " pod="openshift-marketplace/community-operators-8d9vw" Jan 27 16:00:11 crc kubenswrapper[4772]: I0127 16:00:11.443752 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4dcc32f-d75b-420b-8dc1-c1d45ce1390d-utilities\") pod \"community-operators-8d9vw\" (UID: \"a4dcc32f-d75b-420b-8dc1-c1d45ce1390d\") " pod="openshift-marketplace/community-operators-8d9vw" Jan 27 16:00:11 crc kubenswrapper[4772]: I0127 16:00:11.468427 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpxvx\" (UniqueName: \"kubernetes.io/projected/a4dcc32f-d75b-420b-8dc1-c1d45ce1390d-kube-api-access-cpxvx\") pod \"community-operators-8d9vw\" (UID: \"a4dcc32f-d75b-420b-8dc1-c1d45ce1390d\") " pod="openshift-marketplace/community-operators-8d9vw" Jan 27 16:00:11 crc kubenswrapper[4772]: I0127 16:00:11.510577 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8d9vw" Jan 27 16:00:11 crc kubenswrapper[4772]: I0127 16:00:11.842034 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8d9vw"] Jan 27 16:00:12 crc kubenswrapper[4772]: I0127 16:00:12.174687 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-l26xs"] Jan 27 16:00:12 crc kubenswrapper[4772]: I0127 16:00:12.176819 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l26xs" Jan 27 16:00:12 crc kubenswrapper[4772]: I0127 16:00:12.183343 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l26xs"] Jan 27 16:00:12 crc kubenswrapper[4772]: I0127 16:00:12.187386 4772 generic.go:334] "Generic (PLEG): container finished" podID="a4dcc32f-d75b-420b-8dc1-c1d45ce1390d" containerID="eb2b24b6937f9d97b2ef4c64dd3a412a15206c847cf6c127667af5183957718b" exitCode=0 Jan 27 16:00:12 crc kubenswrapper[4772]: I0127 16:00:12.187940 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8d9vw" event={"ID":"a4dcc32f-d75b-420b-8dc1-c1d45ce1390d","Type":"ContainerDied","Data":"eb2b24b6937f9d97b2ef4c64dd3a412a15206c847cf6c127667af5183957718b"} Jan 27 16:00:12 crc kubenswrapper[4772]: I0127 16:00:12.188010 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8d9vw" event={"ID":"a4dcc32f-d75b-420b-8dc1-c1d45ce1390d","Type":"ContainerStarted","Data":"2fd0ace984bed3e0eb564ba4cf8034119c99f0f4a375915de39fff1909a1dfe1"} Jan 27 16:00:12 crc kubenswrapper[4772]: I0127 16:00:12.192044 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8lq7p" event={"ID":"0e55c646-e887-492e-b27d-39536b38b245","Type":"ContainerStarted","Data":"9ff88942bb2ffd8b58cb3374157dc07b383481f7c4b4e26e88426fd9c218bfac"} Jan 27 16:00:12 crc kubenswrapper[4772]: I0127 16:00:12.261814 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52988c66-42d0-4aa5-a46d-d966ac84eb6c-utilities\") pod \"redhat-marketplace-l26xs\" (UID: \"52988c66-42d0-4aa5-a46d-d966ac84eb6c\") " pod="openshift-marketplace/redhat-marketplace-l26xs" Jan 27 16:00:12 crc kubenswrapper[4772]: I0127 16:00:12.261927 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvjcj\" (UniqueName: \"kubernetes.io/projected/52988c66-42d0-4aa5-a46d-d966ac84eb6c-kube-api-access-nvjcj\") pod \"redhat-marketplace-l26xs\" (UID: \"52988c66-42d0-4aa5-a46d-d966ac84eb6c\") " pod="openshift-marketplace/redhat-marketplace-l26xs" Jan 27 16:00:12 crc kubenswrapper[4772]: I0127 16:00:12.262050 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52988c66-42d0-4aa5-a46d-d966ac84eb6c-catalog-content\") pod \"redhat-marketplace-l26xs\" (UID: \"52988c66-42d0-4aa5-a46d-d966ac84eb6c\") " pod="openshift-marketplace/redhat-marketplace-l26xs" Jan 27 16:00:12 crc kubenswrapper[4772]: I0127 16:00:12.362751 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52988c66-42d0-4aa5-a46d-d966ac84eb6c-catalog-content\") pod \"redhat-marketplace-l26xs\" (UID: \"52988c66-42d0-4aa5-a46d-d966ac84eb6c\") " pod="openshift-marketplace/redhat-marketplace-l26xs" Jan 27 16:00:12 crc kubenswrapper[4772]: I0127 16:00:12.362817 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52988c66-42d0-4aa5-a46d-d966ac84eb6c-utilities\") pod \"redhat-marketplace-l26xs\" (UID: \"52988c66-42d0-4aa5-a46d-d966ac84eb6c\") " pod="openshift-marketplace/redhat-marketplace-l26xs" Jan 27 16:00:12 crc kubenswrapper[4772]: I0127 16:00:12.362858 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvjcj\" (UniqueName: \"kubernetes.io/projected/52988c66-42d0-4aa5-a46d-d966ac84eb6c-kube-api-access-nvjcj\") pod \"redhat-marketplace-l26xs\" (UID: \"52988c66-42d0-4aa5-a46d-d966ac84eb6c\") " pod="openshift-marketplace/redhat-marketplace-l26xs" Jan 27 16:00:12 crc kubenswrapper[4772]: I0127 16:00:12.363606 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52988c66-42d0-4aa5-a46d-d966ac84eb6c-catalog-content\") pod \"redhat-marketplace-l26xs\" (UID: \"52988c66-42d0-4aa5-a46d-d966ac84eb6c\") " pod="openshift-marketplace/redhat-marketplace-l26xs" Jan 27 16:00:12 crc kubenswrapper[4772]: I0127 16:00:12.363838 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52988c66-42d0-4aa5-a46d-d966ac84eb6c-utilities\") pod \"redhat-marketplace-l26xs\" (UID: \"52988c66-42d0-4aa5-a46d-d966ac84eb6c\") " pod="openshift-marketplace/redhat-marketplace-l26xs" Jan 27 16:00:12 crc kubenswrapper[4772]: I0127 16:00:12.398887 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvjcj\" (UniqueName: \"kubernetes.io/projected/52988c66-42d0-4aa5-a46d-d966ac84eb6c-kube-api-access-nvjcj\") pod \"redhat-marketplace-l26xs\" (UID: \"52988c66-42d0-4aa5-a46d-d966ac84eb6c\") " pod="openshift-marketplace/redhat-marketplace-l26xs" Jan 27 16:00:12 crc kubenswrapper[4772]: I0127 16:00:12.496668 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l26xs" Jan 27 16:00:12 crc kubenswrapper[4772]: I0127 16:00:12.718944 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l26xs"] Jan 27 16:00:13 crc kubenswrapper[4772]: I0127 16:00:13.199261 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8d9vw" event={"ID":"a4dcc32f-d75b-420b-8dc1-c1d45ce1390d","Type":"ContainerStarted","Data":"c0b6e0a68c3dc10ef4d1ff53c5444a6a40a8fa4bff91ff29c1779550fd4bf956"} Jan 27 16:00:13 crc kubenswrapper[4772]: I0127 16:00:13.200464 4772 generic.go:334] "Generic (PLEG): container finished" podID="52988c66-42d0-4aa5-a46d-d966ac84eb6c" containerID="e7fc82eac5cd511389005799fa7a2100c759a473c11e1b04bce834caa53ea989" exitCode=0 Jan 27 16:00:13 crc kubenswrapper[4772]: I0127 16:00:13.200521 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l26xs" event={"ID":"52988c66-42d0-4aa5-a46d-d966ac84eb6c","Type":"ContainerDied","Data":"e7fc82eac5cd511389005799fa7a2100c759a473c11e1b04bce834caa53ea989"} Jan 27 16:00:13 crc kubenswrapper[4772]: I0127 16:00:13.200549 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l26xs" event={"ID":"52988c66-42d0-4aa5-a46d-d966ac84eb6c","Type":"ContainerStarted","Data":"1b80836c83d3067947b0be4ac5034b810cd9fa2107cf4663070bd74214922382"} Jan 27 16:00:13 crc kubenswrapper[4772]: I0127 16:00:13.201939 4772 generic.go:334] "Generic (PLEG): container finished" podID="0e55c646-e887-492e-b27d-39536b38b245" containerID="9ff88942bb2ffd8b58cb3374157dc07b383481f7c4b4e26e88426fd9c218bfac" exitCode=0 Jan 27 16:00:13 crc kubenswrapper[4772]: I0127 16:00:13.201967 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8lq7p" event={"ID":"0e55c646-e887-492e-b27d-39536b38b245","Type":"ContainerDied","Data":"9ff88942bb2ffd8b58cb3374157dc07b383481f7c4b4e26e88426fd9c218bfac"} Jan 27 16:00:14 crc kubenswrapper[4772]: I0127 16:00:14.209842 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l26xs" event={"ID":"52988c66-42d0-4aa5-a46d-d966ac84eb6c","Type":"ContainerStarted","Data":"efda54e04a3b9158207e5837fd7a6528bdd8fc43bf202745462e3bbb8c32bdb2"} Jan 27 16:00:14 crc kubenswrapper[4772]: I0127 16:00:14.212180 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8lq7p" event={"ID":"0e55c646-e887-492e-b27d-39536b38b245","Type":"ContainerStarted","Data":"0e58bb490648f3a839abadeb3266c1fb8ad0a72f82a439d1c989dbde19289df2"} Jan 27 16:00:14 crc kubenswrapper[4772]: I0127 16:00:14.214591 4772 generic.go:334] "Generic (PLEG): container finished" podID="a4dcc32f-d75b-420b-8dc1-c1d45ce1390d" containerID="c0b6e0a68c3dc10ef4d1ff53c5444a6a40a8fa4bff91ff29c1779550fd4bf956" exitCode=0 Jan 27 16:00:14 crc kubenswrapper[4772]: I0127 16:00:14.214630 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8d9vw" event={"ID":"a4dcc32f-d75b-420b-8dc1-c1d45ce1390d","Type":"ContainerDied","Data":"c0b6e0a68c3dc10ef4d1ff53c5444a6a40a8fa4bff91ff29c1779550fd4bf956"} Jan 27 16:00:14 crc kubenswrapper[4772]: I0127 16:00:14.252110 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8lq7p" podStartSLOduration=2.661111567 podStartE2EDuration="5.252089884s" podCreationTimestamp="2026-01-27 16:00:09 +0000 UTC" firstStartedPulling="2026-01-27 16:00:11.182282904 +0000 UTC m=+3197.162892002" lastFinishedPulling="2026-01-27 16:00:13.773261221 +0000 UTC m=+3199.753870319" observedRunningTime="2026-01-27 16:00:14.249905951 +0000 UTC m=+3200.230515049" watchObservedRunningTime="2026-01-27 16:00:14.252089884 +0000 UTC m=+3200.232698982" Jan 27 16:00:15 crc kubenswrapper[4772]: I0127 16:00:15.232594 4772 generic.go:334] "Generic (PLEG): container finished" podID="52988c66-42d0-4aa5-a46d-d966ac84eb6c" containerID="efda54e04a3b9158207e5837fd7a6528bdd8fc43bf202745462e3bbb8c32bdb2" exitCode=0 Jan 27 16:00:15 crc kubenswrapper[4772]: I0127 16:00:15.232662 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l26xs" event={"ID":"52988c66-42d0-4aa5-a46d-d966ac84eb6c","Type":"ContainerDied","Data":"efda54e04a3b9158207e5837fd7a6528bdd8fc43bf202745462e3bbb8c32bdb2"} Jan 27 16:00:15 crc kubenswrapper[4772]: I0127 16:00:15.238156 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8d9vw" event={"ID":"a4dcc32f-d75b-420b-8dc1-c1d45ce1390d","Type":"ContainerStarted","Data":"70571922f32975b959e9d6c31230572ab940e3dd080afcd55f53dd0e4c7d8039"} Jan 27 16:00:15 crc kubenswrapper[4772]: I0127 16:00:15.274022 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8d9vw" podStartSLOduration=1.844111055 podStartE2EDuration="4.274000154s" podCreationTimestamp="2026-01-27 16:00:11 +0000 UTC" firstStartedPulling="2026-01-27 16:00:12.190458679 +0000 UTC m=+3198.171067777" lastFinishedPulling="2026-01-27 16:00:14.620347778 +0000 UTC m=+3200.600956876" observedRunningTime="2026-01-27 16:00:15.272259634 +0000 UTC m=+3201.252868752" watchObservedRunningTime="2026-01-27 16:00:15.274000154 +0000 UTC m=+3201.254609252" Jan 27 16:00:16 crc kubenswrapper[4772]: I0127 16:00:16.245362 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l26xs" event={"ID":"52988c66-42d0-4aa5-a46d-d966ac84eb6c","Type":"ContainerStarted","Data":"dfe7a2bf50079c2e751ce5574fccab4a5353b71f051b83c79197ffe548a713a9"} Jan 27 16:00:16 crc kubenswrapper[4772]: I0127 16:00:16.264406 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-l26xs" podStartSLOduration=1.725218721 podStartE2EDuration="4.264387719s" podCreationTimestamp="2026-01-27 16:00:12 +0000 UTC" firstStartedPulling="2026-01-27 16:00:13.201487228 +0000 UTC m=+3199.182096326" lastFinishedPulling="2026-01-27 16:00:15.740656226 +0000 UTC m=+3201.721265324" observedRunningTime="2026-01-27 16:00:16.262750542 +0000 UTC m=+3202.243359640" watchObservedRunningTime="2026-01-27 16:00:16.264387719 +0000 UTC m=+3202.244996817" Jan 27 16:00:19 crc kubenswrapper[4772]: I0127 16:00:19.692520 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8lq7p" Jan 27 16:00:19 crc kubenswrapper[4772]: I0127 16:00:19.692908 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8lq7p" Jan 27 16:00:19 crc kubenswrapper[4772]: I0127 16:00:19.744948 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8lq7p" Jan 27 16:00:20 crc kubenswrapper[4772]: I0127 16:00:20.314905 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8lq7p" Jan 27 16:00:20 crc kubenswrapper[4772]: I0127 16:00:20.662583 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 16:00:21 crc kubenswrapper[4772]: I0127 16:00:21.511366 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8d9vw" Jan 27 16:00:21 crc kubenswrapper[4772]: I0127 16:00:21.511413 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8d9vw" Jan 27 16:00:21 crc kubenswrapper[4772]: I0127 16:00:21.556934 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8d9vw" Jan 27 16:00:22 crc kubenswrapper[4772]: I0127 16:00:22.173384 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8lq7p"] Jan 27 16:00:22 crc kubenswrapper[4772]: I0127 16:00:22.367056 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8lq7p" podUID="0e55c646-e887-492e-b27d-39536b38b245" containerName="registry-server" containerID="cri-o://0e58bb490648f3a839abadeb3266c1fb8ad0a72f82a439d1c989dbde19289df2" gracePeriod=2 Jan 27 16:00:22 crc kubenswrapper[4772]: I0127 16:00:22.367435 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"21ef3f853d795985962b240174bbec1611d9f0f58af15b07556fa41617b20592"} Jan 27 16:00:22 crc kubenswrapper[4772]: I0127 16:00:22.481515 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8d9vw" Jan 27 16:00:22 crc kubenswrapper[4772]: I0127 16:00:22.498455 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-l26xs" Jan 27 16:00:22 crc kubenswrapper[4772]: I0127 16:00:22.499241 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-l26xs" Jan 27 16:00:22 crc kubenswrapper[4772]: I0127 16:00:22.552876 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-l26xs" Jan 27 16:00:23 crc kubenswrapper[4772]: I0127 16:00:23.425466 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-l26xs" Jan 27 16:00:23 crc kubenswrapper[4772]: I0127 16:00:23.957078 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8d9vw"] Jan 27 16:00:24 crc kubenswrapper[4772]: I0127 16:00:24.386188 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8d9vw" podUID="a4dcc32f-d75b-420b-8dc1-c1d45ce1390d" containerName="registry-server" containerID="cri-o://70571922f32975b959e9d6c31230572ab940e3dd080afcd55f53dd0e4c7d8039" gracePeriod=2 Jan 27 16:00:25 crc kubenswrapper[4772]: I0127 16:00:25.395506 4772 generic.go:334] "Generic (PLEG): container finished" podID="0e55c646-e887-492e-b27d-39536b38b245" containerID="0e58bb490648f3a839abadeb3266c1fb8ad0a72f82a439d1c989dbde19289df2" exitCode=0 Jan 27 16:00:25 crc kubenswrapper[4772]: I0127 16:00:25.396514 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8lq7p" event={"ID":"0e55c646-e887-492e-b27d-39536b38b245","Type":"ContainerDied","Data":"0e58bb490648f3a839abadeb3266c1fb8ad0a72f82a439d1c989dbde19289df2"} Jan 27 16:00:25 crc kubenswrapper[4772]: I0127 16:00:25.860507 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8lq7p" Jan 27 16:00:25 crc kubenswrapper[4772]: I0127 16:00:25.987279 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfh4p\" (UniqueName: \"kubernetes.io/projected/0e55c646-e887-492e-b27d-39536b38b245-kube-api-access-gfh4p\") pod \"0e55c646-e887-492e-b27d-39536b38b245\" (UID: \"0e55c646-e887-492e-b27d-39536b38b245\") " Jan 27 16:00:25 crc kubenswrapper[4772]: I0127 16:00:25.987344 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e55c646-e887-492e-b27d-39536b38b245-utilities\") pod \"0e55c646-e887-492e-b27d-39536b38b245\" (UID: \"0e55c646-e887-492e-b27d-39536b38b245\") " Jan 27 16:00:25 crc kubenswrapper[4772]: I0127 16:00:25.987415 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e55c646-e887-492e-b27d-39536b38b245-catalog-content\") pod \"0e55c646-e887-492e-b27d-39536b38b245\" (UID: \"0e55c646-e887-492e-b27d-39536b38b245\") " Jan 27 16:00:25 crc kubenswrapper[4772]: I0127 16:00:25.988382 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e55c646-e887-492e-b27d-39536b38b245-utilities" (OuterVolumeSpecName: "utilities") pod "0e55c646-e887-492e-b27d-39536b38b245" (UID: "0e55c646-e887-492e-b27d-39536b38b245"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:00:25 crc kubenswrapper[4772]: I0127 16:00:25.996693 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e55c646-e887-492e-b27d-39536b38b245-kube-api-access-gfh4p" (OuterVolumeSpecName: "kube-api-access-gfh4p") pod "0e55c646-e887-492e-b27d-39536b38b245" (UID: "0e55c646-e887-492e-b27d-39536b38b245"). InnerVolumeSpecName "kube-api-access-gfh4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.036278 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e55c646-e887-492e-b27d-39536b38b245-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0e55c646-e887-492e-b27d-39536b38b245" (UID: "0e55c646-e887-492e-b27d-39536b38b245"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.088807 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfh4p\" (UniqueName: \"kubernetes.io/projected/0e55c646-e887-492e-b27d-39536b38b245-kube-api-access-gfh4p\") on node \"crc\" DevicePath \"\"" Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.088871 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e55c646-e887-492e-b27d-39536b38b245-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.088891 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e55c646-e887-492e-b27d-39536b38b245-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.356555 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l26xs"] Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.405735 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8lq7p" Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.405729 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8lq7p" event={"ID":"0e55c646-e887-492e-b27d-39536b38b245","Type":"ContainerDied","Data":"215607eb1ef46147418a0a5c60a763e046da6f15ce0cdc380aa944d9788c12db"} Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.405895 4772 scope.go:117] "RemoveContainer" containerID="0e58bb490648f3a839abadeb3266c1fb8ad0a72f82a439d1c989dbde19289df2" Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.414316 4772 generic.go:334] "Generic (PLEG): container finished" podID="a4dcc32f-d75b-420b-8dc1-c1d45ce1390d" containerID="70571922f32975b959e9d6c31230572ab940e3dd080afcd55f53dd0e4c7d8039" exitCode=0 Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.414437 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8d9vw" event={"ID":"a4dcc32f-d75b-420b-8dc1-c1d45ce1390d","Type":"ContainerDied","Data":"70571922f32975b959e9d6c31230572ab940e3dd080afcd55f53dd0e4c7d8039"} Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.414549 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-l26xs" podUID="52988c66-42d0-4aa5-a46d-d966ac84eb6c" containerName="registry-server" containerID="cri-o://dfe7a2bf50079c2e751ce5574fccab4a5353b71f051b83c79197ffe548a713a9" gracePeriod=2 Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.423922 4772 scope.go:117] "RemoveContainer" containerID="9ff88942bb2ffd8b58cb3374157dc07b383481f7c4b4e26e88426fd9c218bfac" Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.444345 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8lq7p"] Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.453080 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8lq7p"] Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.467687 4772 scope.go:117] "RemoveContainer" containerID="5f1d8fdc7aed07858ca3363c0144e81cdc92e3922938515a33edf543ec95389e" Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.651459 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8d9vw" Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.678631 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e55c646-e887-492e-b27d-39536b38b245" path="/var/lib/kubelet/pods/0e55c646-e887-492e-b27d-39536b38b245/volumes" Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.805955 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4dcc32f-d75b-420b-8dc1-c1d45ce1390d-utilities\") pod \"a4dcc32f-d75b-420b-8dc1-c1d45ce1390d\" (UID: \"a4dcc32f-d75b-420b-8dc1-c1d45ce1390d\") " Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.805996 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4dcc32f-d75b-420b-8dc1-c1d45ce1390d-catalog-content\") pod \"a4dcc32f-d75b-420b-8dc1-c1d45ce1390d\" (UID: \"a4dcc32f-d75b-420b-8dc1-c1d45ce1390d\") " Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.806096 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpxvx\" (UniqueName: \"kubernetes.io/projected/a4dcc32f-d75b-420b-8dc1-c1d45ce1390d-kube-api-access-cpxvx\") pod \"a4dcc32f-d75b-420b-8dc1-c1d45ce1390d\" (UID: \"a4dcc32f-d75b-420b-8dc1-c1d45ce1390d\") " Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.806886 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4dcc32f-d75b-420b-8dc1-c1d45ce1390d-utilities" (OuterVolumeSpecName: "utilities") pod "a4dcc32f-d75b-420b-8dc1-c1d45ce1390d" (UID: "a4dcc32f-d75b-420b-8dc1-c1d45ce1390d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.807557 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4dcc32f-d75b-420b-8dc1-c1d45ce1390d-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.809413 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4dcc32f-d75b-420b-8dc1-c1d45ce1390d-kube-api-access-cpxvx" (OuterVolumeSpecName: "kube-api-access-cpxvx") pod "a4dcc32f-d75b-420b-8dc1-c1d45ce1390d" (UID: "a4dcc32f-d75b-420b-8dc1-c1d45ce1390d"). InnerVolumeSpecName "kube-api-access-cpxvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.857775 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4dcc32f-d75b-420b-8dc1-c1d45ce1390d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a4dcc32f-d75b-420b-8dc1-c1d45ce1390d" (UID: "a4dcc32f-d75b-420b-8dc1-c1d45ce1390d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.909065 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpxvx\" (UniqueName: \"kubernetes.io/projected/a4dcc32f-d75b-420b-8dc1-c1d45ce1390d-kube-api-access-cpxvx\") on node \"crc\" DevicePath \"\"" Jan 27 16:00:26 crc kubenswrapper[4772]: I0127 16:00:26.909100 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4dcc32f-d75b-420b-8dc1-c1d45ce1390d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 16:00:27 crc kubenswrapper[4772]: I0127 16:00:27.447414 4772 generic.go:334] "Generic (PLEG): container finished" podID="52988c66-42d0-4aa5-a46d-d966ac84eb6c" containerID="dfe7a2bf50079c2e751ce5574fccab4a5353b71f051b83c79197ffe548a713a9" exitCode=0 Jan 27 16:00:27 crc kubenswrapper[4772]: I0127 16:00:27.447484 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l26xs" event={"ID":"52988c66-42d0-4aa5-a46d-d966ac84eb6c","Type":"ContainerDied","Data":"dfe7a2bf50079c2e751ce5574fccab4a5353b71f051b83c79197ffe548a713a9"} Jan 27 16:00:27 crc kubenswrapper[4772]: I0127 16:00:27.484876 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8d9vw" event={"ID":"a4dcc32f-d75b-420b-8dc1-c1d45ce1390d","Type":"ContainerDied","Data":"2fd0ace984bed3e0eb564ba4cf8034119c99f0f4a375915de39fff1909a1dfe1"} Jan 27 16:00:27 crc kubenswrapper[4772]: I0127 16:00:27.484927 4772 scope.go:117] "RemoveContainer" containerID="70571922f32975b959e9d6c31230572ab940e3dd080afcd55f53dd0e4c7d8039" Jan 27 16:00:27 crc kubenswrapper[4772]: I0127 16:00:27.485057 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8d9vw" Jan 27 16:00:27 crc kubenswrapper[4772]: I0127 16:00:27.535454 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8d9vw"] Jan 27 16:00:27 crc kubenswrapper[4772]: I0127 16:00:27.543881 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8d9vw"] Jan 27 16:00:27 crc kubenswrapper[4772]: I0127 16:00:27.544327 4772 scope.go:117] "RemoveContainer" containerID="c0b6e0a68c3dc10ef4d1ff53c5444a6a40a8fa4bff91ff29c1779550fd4bf956" Jan 27 16:00:27 crc kubenswrapper[4772]: I0127 16:00:27.583061 4772 scope.go:117] "RemoveContainer" containerID="eb2b24b6937f9d97b2ef4c64dd3a412a15206c847cf6c127667af5183957718b" Jan 27 16:00:27 crc kubenswrapper[4772]: I0127 16:00:27.648884 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l26xs" Jan 27 16:00:27 crc kubenswrapper[4772]: I0127 16:00:27.750970 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52988c66-42d0-4aa5-a46d-d966ac84eb6c-utilities\") pod \"52988c66-42d0-4aa5-a46d-d966ac84eb6c\" (UID: \"52988c66-42d0-4aa5-a46d-d966ac84eb6c\") " Jan 27 16:00:27 crc kubenswrapper[4772]: I0127 16:00:27.751049 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52988c66-42d0-4aa5-a46d-d966ac84eb6c-catalog-content\") pod \"52988c66-42d0-4aa5-a46d-d966ac84eb6c\" (UID: \"52988c66-42d0-4aa5-a46d-d966ac84eb6c\") " Jan 27 16:00:27 crc kubenswrapper[4772]: I0127 16:00:27.751085 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvjcj\" (UniqueName: \"kubernetes.io/projected/52988c66-42d0-4aa5-a46d-d966ac84eb6c-kube-api-access-nvjcj\") pod \"52988c66-42d0-4aa5-a46d-d966ac84eb6c\" (UID: \"52988c66-42d0-4aa5-a46d-d966ac84eb6c\") " Jan 27 16:00:27 crc kubenswrapper[4772]: I0127 16:00:27.752378 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52988c66-42d0-4aa5-a46d-d966ac84eb6c-utilities" (OuterVolumeSpecName: "utilities") pod "52988c66-42d0-4aa5-a46d-d966ac84eb6c" (UID: "52988c66-42d0-4aa5-a46d-d966ac84eb6c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:00:27 crc kubenswrapper[4772]: I0127 16:00:27.755034 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52988c66-42d0-4aa5-a46d-d966ac84eb6c-kube-api-access-nvjcj" (OuterVolumeSpecName: "kube-api-access-nvjcj") pod "52988c66-42d0-4aa5-a46d-d966ac84eb6c" (UID: "52988c66-42d0-4aa5-a46d-d966ac84eb6c"). InnerVolumeSpecName "kube-api-access-nvjcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:00:27 crc kubenswrapper[4772]: I0127 16:00:27.774300 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52988c66-42d0-4aa5-a46d-d966ac84eb6c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "52988c66-42d0-4aa5-a46d-d966ac84eb6c" (UID: "52988c66-42d0-4aa5-a46d-d966ac84eb6c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:00:27 crc kubenswrapper[4772]: I0127 16:00:27.852942 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52988c66-42d0-4aa5-a46d-d966ac84eb6c-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 16:00:27 crc kubenswrapper[4772]: I0127 16:00:27.852981 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvjcj\" (UniqueName: \"kubernetes.io/projected/52988c66-42d0-4aa5-a46d-d966ac84eb6c-kube-api-access-nvjcj\") on node \"crc\" DevicePath \"\"" Jan 27 16:00:27 crc kubenswrapper[4772]: I0127 16:00:27.852992 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52988c66-42d0-4aa5-a46d-d966ac84eb6c-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 16:00:28 crc kubenswrapper[4772]: I0127 16:00:28.494953 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l26xs" event={"ID":"52988c66-42d0-4aa5-a46d-d966ac84eb6c","Type":"ContainerDied","Data":"1b80836c83d3067947b0be4ac5034b810cd9fa2107cf4663070bd74214922382"} Jan 27 16:00:28 crc kubenswrapper[4772]: I0127 16:00:28.494965 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l26xs" Jan 27 16:00:28 crc kubenswrapper[4772]: I0127 16:00:28.495438 4772 scope.go:117] "RemoveContainer" containerID="dfe7a2bf50079c2e751ce5574fccab4a5353b71f051b83c79197ffe548a713a9" Jan 27 16:00:28 crc kubenswrapper[4772]: I0127 16:00:28.511681 4772 scope.go:117] "RemoveContainer" containerID="efda54e04a3b9158207e5837fd7a6528bdd8fc43bf202745462e3bbb8c32bdb2" Jan 27 16:00:28 crc kubenswrapper[4772]: I0127 16:00:28.527237 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l26xs"] Jan 27 16:00:28 crc kubenswrapper[4772]: I0127 16:00:28.533519 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-l26xs"] Jan 27 16:00:28 crc kubenswrapper[4772]: I0127 16:00:28.550493 4772 scope.go:117] "RemoveContainer" containerID="e7fc82eac5cd511389005799fa7a2100c759a473c11e1b04bce834caa53ea989" Jan 27 16:00:28 crc kubenswrapper[4772]: I0127 16:00:28.673672 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52988c66-42d0-4aa5-a46d-d966ac84eb6c" path="/var/lib/kubelet/pods/52988c66-42d0-4aa5-a46d-d966ac84eb6c/volumes" Jan 27 16:00:28 crc kubenswrapper[4772]: I0127 16:00:28.675668 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4dcc32f-d75b-420b-8dc1-c1d45ce1390d" path="/var/lib/kubelet/pods/a4dcc32f-d75b-420b-8dc1-c1d45ce1390d/volumes" Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.551050 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tz78w"] Jan 27 16:01:03 crc kubenswrapper[4772]: E0127 16:01:03.555688 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e55c646-e887-492e-b27d-39536b38b245" containerName="registry-server" Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.555723 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e55c646-e887-492e-b27d-39536b38b245" containerName="registry-server" Jan 27 16:01:03 crc kubenswrapper[4772]: E0127 16:01:03.555748 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52988c66-42d0-4aa5-a46d-d966ac84eb6c" containerName="extract-content" Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.555757 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="52988c66-42d0-4aa5-a46d-d966ac84eb6c" containerName="extract-content" Jan 27 16:01:03 crc kubenswrapper[4772]: E0127 16:01:03.555775 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52988c66-42d0-4aa5-a46d-d966ac84eb6c" containerName="registry-server" Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.555783 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="52988c66-42d0-4aa5-a46d-d966ac84eb6c" containerName="registry-server" Jan 27 16:01:03 crc kubenswrapper[4772]: E0127 16:01:03.555802 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52988c66-42d0-4aa5-a46d-d966ac84eb6c" containerName="extract-utilities" Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.555810 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="52988c66-42d0-4aa5-a46d-d966ac84eb6c" containerName="extract-utilities" Jan 27 16:01:03 crc kubenswrapper[4772]: E0127 16:01:03.555855 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4dcc32f-d75b-420b-8dc1-c1d45ce1390d" containerName="registry-server" Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.555863 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4dcc32f-d75b-420b-8dc1-c1d45ce1390d" containerName="registry-server" Jan 27 16:01:03 crc kubenswrapper[4772]: E0127 16:01:03.555878 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e55c646-e887-492e-b27d-39536b38b245" containerName="extract-content" Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.555886 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e55c646-e887-492e-b27d-39536b38b245" containerName="extract-content" Jan 27 16:01:03 crc kubenswrapper[4772]: E0127 16:01:03.555907 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4dcc32f-d75b-420b-8dc1-c1d45ce1390d" containerName="extract-content" Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.555915 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4dcc32f-d75b-420b-8dc1-c1d45ce1390d" containerName="extract-content" Jan 27 16:01:03 crc kubenswrapper[4772]: E0127 16:01:03.555928 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4dcc32f-d75b-420b-8dc1-c1d45ce1390d" containerName="extract-utilities" Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.555941 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4dcc32f-d75b-420b-8dc1-c1d45ce1390d" containerName="extract-utilities" Jan 27 16:01:03 crc kubenswrapper[4772]: E0127 16:01:03.555953 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e55c646-e887-492e-b27d-39536b38b245" containerName="extract-utilities" Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.555962 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e55c646-e887-492e-b27d-39536b38b245" containerName="extract-utilities" Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.556525 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4dcc32f-d75b-420b-8dc1-c1d45ce1390d" containerName="registry-server" Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.556563 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e55c646-e887-492e-b27d-39536b38b245" containerName="registry-server" Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.556590 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="52988c66-42d0-4aa5-a46d-d966ac84eb6c" containerName="registry-server" Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.559209 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tz78w" Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.573552 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tz78w"] Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.674079 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cd0f6c5-43f1-4120-841d-76e540249886-catalog-content\") pod \"redhat-operators-tz78w\" (UID: \"3cd0f6c5-43f1-4120-841d-76e540249886\") " pod="openshift-marketplace/redhat-operators-tz78w" Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.674436 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cd0f6c5-43f1-4120-841d-76e540249886-utilities\") pod \"redhat-operators-tz78w\" (UID: \"3cd0f6c5-43f1-4120-841d-76e540249886\") " pod="openshift-marketplace/redhat-operators-tz78w" Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.674546 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg8nn\" (UniqueName: \"kubernetes.io/projected/3cd0f6c5-43f1-4120-841d-76e540249886-kube-api-access-xg8nn\") pod \"redhat-operators-tz78w\" (UID: \"3cd0f6c5-43f1-4120-841d-76e540249886\") " pod="openshift-marketplace/redhat-operators-tz78w" Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.775865 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cd0f6c5-43f1-4120-841d-76e540249886-utilities\") pod \"redhat-operators-tz78w\" (UID: \"3cd0f6c5-43f1-4120-841d-76e540249886\") " pod="openshift-marketplace/redhat-operators-tz78w" Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.776212 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg8nn\" (UniqueName: \"kubernetes.io/projected/3cd0f6c5-43f1-4120-841d-76e540249886-kube-api-access-xg8nn\") pod \"redhat-operators-tz78w\" (UID: \"3cd0f6c5-43f1-4120-841d-76e540249886\") " pod="openshift-marketplace/redhat-operators-tz78w" Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.776426 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cd0f6c5-43f1-4120-841d-76e540249886-utilities\") pod \"redhat-operators-tz78w\" (UID: \"3cd0f6c5-43f1-4120-841d-76e540249886\") " pod="openshift-marketplace/redhat-operators-tz78w" Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.776553 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cd0f6c5-43f1-4120-841d-76e540249886-catalog-content\") pod \"redhat-operators-tz78w\" (UID: \"3cd0f6c5-43f1-4120-841d-76e540249886\") " pod="openshift-marketplace/redhat-operators-tz78w" Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.776821 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cd0f6c5-43f1-4120-841d-76e540249886-catalog-content\") pod \"redhat-operators-tz78w\" (UID: \"3cd0f6c5-43f1-4120-841d-76e540249886\") " pod="openshift-marketplace/redhat-operators-tz78w" Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.797014 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg8nn\" (UniqueName: \"kubernetes.io/projected/3cd0f6c5-43f1-4120-841d-76e540249886-kube-api-access-xg8nn\") pod \"redhat-operators-tz78w\" (UID: \"3cd0f6c5-43f1-4120-841d-76e540249886\") " pod="openshift-marketplace/redhat-operators-tz78w" Jan 27 16:01:03 crc kubenswrapper[4772]: I0127 16:01:03.893136 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tz78w" Jan 27 16:01:04 crc kubenswrapper[4772]: I0127 16:01:04.374470 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tz78w"] Jan 27 16:01:04 crc kubenswrapper[4772]: I0127 16:01:04.755364 4772 generic.go:334] "Generic (PLEG): container finished" podID="3cd0f6c5-43f1-4120-841d-76e540249886" containerID="5c076ec7e9f7ce861de2aeb2d2ce1407d2581066fe8432d268d48aa28c136d02" exitCode=0 Jan 27 16:01:04 crc kubenswrapper[4772]: I0127 16:01:04.755633 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tz78w" event={"ID":"3cd0f6c5-43f1-4120-841d-76e540249886","Type":"ContainerDied","Data":"5c076ec7e9f7ce861de2aeb2d2ce1407d2581066fe8432d268d48aa28c136d02"} Jan 27 16:01:04 crc kubenswrapper[4772]: I0127 16:01:04.755660 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tz78w" event={"ID":"3cd0f6c5-43f1-4120-841d-76e540249886","Type":"ContainerStarted","Data":"61dc4f44513fd78fc78365968bdd4024b29d704ed02e0432d82b9f4fbc0894cd"} Jan 27 16:01:06 crc kubenswrapper[4772]: I0127 16:01:06.770798 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tz78w" event={"ID":"3cd0f6c5-43f1-4120-841d-76e540249886","Type":"ContainerStarted","Data":"d626ad0bd0e93bf253da87fff3d7e5845272782b38c1fe8f821ce4c09122879f"} Jan 27 16:01:07 crc kubenswrapper[4772]: I0127 16:01:07.794314 4772 generic.go:334] "Generic (PLEG): container finished" podID="3cd0f6c5-43f1-4120-841d-76e540249886" containerID="d626ad0bd0e93bf253da87fff3d7e5845272782b38c1fe8f821ce4c09122879f" exitCode=0 Jan 27 16:01:07 crc kubenswrapper[4772]: I0127 16:01:07.794406 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tz78w" event={"ID":"3cd0f6c5-43f1-4120-841d-76e540249886","Type":"ContainerDied","Data":"d626ad0bd0e93bf253da87fff3d7e5845272782b38c1fe8f821ce4c09122879f"} Jan 27 16:01:08 crc kubenswrapper[4772]: I0127 16:01:08.803414 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tz78w" event={"ID":"3cd0f6c5-43f1-4120-841d-76e540249886","Type":"ContainerStarted","Data":"f6f9a4e35dfde3b6f98ddcc01331fdfec4af28b5b9e64b6ca58bee7f89fbaddc"} Jan 27 16:01:13 crc kubenswrapper[4772]: I0127 16:01:13.893941 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tz78w" Jan 27 16:01:13 crc kubenswrapper[4772]: I0127 16:01:13.894479 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tz78w" Jan 27 16:01:13 crc kubenswrapper[4772]: I0127 16:01:13.936273 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tz78w" Jan 27 16:01:13 crc kubenswrapper[4772]: I0127 16:01:13.962428 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tz78w" podStartSLOduration=7.533237794 podStartE2EDuration="10.962396342s" podCreationTimestamp="2026-01-27 16:01:03 +0000 UTC" firstStartedPulling="2026-01-27 16:01:04.757229925 +0000 UTC m=+3250.737839023" lastFinishedPulling="2026-01-27 16:01:08.186388453 +0000 UTC m=+3254.166997571" observedRunningTime="2026-01-27 16:01:08.83103225 +0000 UTC m=+3254.811641358" watchObservedRunningTime="2026-01-27 16:01:13.962396342 +0000 UTC m=+3259.943005440" Jan 27 16:01:14 crc kubenswrapper[4772]: I0127 16:01:14.891015 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tz78w" Jan 27 16:01:14 crc kubenswrapper[4772]: I0127 16:01:14.932318 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tz78w"] Jan 27 16:01:16 crc kubenswrapper[4772]: I0127 16:01:16.850589 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tz78w" podUID="3cd0f6c5-43f1-4120-841d-76e540249886" containerName="registry-server" containerID="cri-o://f6f9a4e35dfde3b6f98ddcc01331fdfec4af28b5b9e64b6ca58bee7f89fbaddc" gracePeriod=2 Jan 27 16:01:18 crc kubenswrapper[4772]: I0127 16:01:18.578232 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tz78w" Jan 27 16:01:18 crc kubenswrapper[4772]: I0127 16:01:18.692189 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cd0f6c5-43f1-4120-841d-76e540249886-utilities\") pod \"3cd0f6c5-43f1-4120-841d-76e540249886\" (UID: \"3cd0f6c5-43f1-4120-841d-76e540249886\") " Jan 27 16:01:18 crc kubenswrapper[4772]: I0127 16:01:18.692565 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cd0f6c5-43f1-4120-841d-76e540249886-catalog-content\") pod \"3cd0f6c5-43f1-4120-841d-76e540249886\" (UID: \"3cd0f6c5-43f1-4120-841d-76e540249886\") " Jan 27 16:01:18 crc kubenswrapper[4772]: I0127 16:01:18.692598 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xg8nn\" (UniqueName: \"kubernetes.io/projected/3cd0f6c5-43f1-4120-841d-76e540249886-kube-api-access-xg8nn\") pod \"3cd0f6c5-43f1-4120-841d-76e540249886\" (UID: \"3cd0f6c5-43f1-4120-841d-76e540249886\") " Jan 27 16:01:18 crc kubenswrapper[4772]: I0127 16:01:18.693294 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cd0f6c5-43f1-4120-841d-76e540249886-utilities" (OuterVolumeSpecName: "utilities") pod "3cd0f6c5-43f1-4120-841d-76e540249886" (UID: "3cd0f6c5-43f1-4120-841d-76e540249886"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:01:18 crc kubenswrapper[4772]: I0127 16:01:18.698480 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cd0f6c5-43f1-4120-841d-76e540249886-kube-api-access-xg8nn" (OuterVolumeSpecName: "kube-api-access-xg8nn") pod "3cd0f6c5-43f1-4120-841d-76e540249886" (UID: "3cd0f6c5-43f1-4120-841d-76e540249886"). InnerVolumeSpecName "kube-api-access-xg8nn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:01:18 crc kubenswrapper[4772]: I0127 16:01:18.794402 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xg8nn\" (UniqueName: \"kubernetes.io/projected/3cd0f6c5-43f1-4120-841d-76e540249886-kube-api-access-xg8nn\") on node \"crc\" DevicePath \"\"" Jan 27 16:01:18 crc kubenswrapper[4772]: I0127 16:01:18.794438 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cd0f6c5-43f1-4120-841d-76e540249886-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 16:01:18 crc kubenswrapper[4772]: I0127 16:01:18.823016 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cd0f6c5-43f1-4120-841d-76e540249886-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3cd0f6c5-43f1-4120-841d-76e540249886" (UID: "3cd0f6c5-43f1-4120-841d-76e540249886"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:01:18 crc kubenswrapper[4772]: I0127 16:01:18.865961 4772 generic.go:334] "Generic (PLEG): container finished" podID="3cd0f6c5-43f1-4120-841d-76e540249886" containerID="f6f9a4e35dfde3b6f98ddcc01331fdfec4af28b5b9e64b6ca58bee7f89fbaddc" exitCode=0 Jan 27 16:01:18 crc kubenswrapper[4772]: I0127 16:01:18.866011 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tz78w" event={"ID":"3cd0f6c5-43f1-4120-841d-76e540249886","Type":"ContainerDied","Data":"f6f9a4e35dfde3b6f98ddcc01331fdfec4af28b5b9e64b6ca58bee7f89fbaddc"} Jan 27 16:01:18 crc kubenswrapper[4772]: I0127 16:01:18.866042 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tz78w" event={"ID":"3cd0f6c5-43f1-4120-841d-76e540249886","Type":"ContainerDied","Data":"61dc4f44513fd78fc78365968bdd4024b29d704ed02e0432d82b9f4fbc0894cd"} Jan 27 16:01:18 crc kubenswrapper[4772]: I0127 16:01:18.866057 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tz78w" Jan 27 16:01:18 crc kubenswrapper[4772]: I0127 16:01:18.866066 4772 scope.go:117] "RemoveContainer" containerID="f6f9a4e35dfde3b6f98ddcc01331fdfec4af28b5b9e64b6ca58bee7f89fbaddc" Jan 27 16:01:18 crc kubenswrapper[4772]: I0127 16:01:18.895836 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cd0f6c5-43f1-4120-841d-76e540249886-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 16:01:18 crc kubenswrapper[4772]: I0127 16:01:18.900309 4772 scope.go:117] "RemoveContainer" containerID="d626ad0bd0e93bf253da87fff3d7e5845272782b38c1fe8f821ce4c09122879f" Jan 27 16:01:18 crc kubenswrapper[4772]: I0127 16:01:18.929544 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tz78w"] Jan 27 16:01:18 crc kubenswrapper[4772]: I0127 16:01:18.932581 4772 scope.go:117] "RemoveContainer" containerID="5c076ec7e9f7ce861de2aeb2d2ce1407d2581066fe8432d268d48aa28c136d02" Jan 27 16:01:18 crc kubenswrapper[4772]: I0127 16:01:18.946907 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tz78w"] Jan 27 16:01:18 crc kubenswrapper[4772]: I0127 16:01:18.986364 4772 scope.go:117] "RemoveContainer" containerID="f6f9a4e35dfde3b6f98ddcc01331fdfec4af28b5b9e64b6ca58bee7f89fbaddc" Jan 27 16:01:18 crc kubenswrapper[4772]: E0127 16:01:18.990098 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6f9a4e35dfde3b6f98ddcc01331fdfec4af28b5b9e64b6ca58bee7f89fbaddc\": container with ID starting with f6f9a4e35dfde3b6f98ddcc01331fdfec4af28b5b9e64b6ca58bee7f89fbaddc not found: ID does not exist" containerID="f6f9a4e35dfde3b6f98ddcc01331fdfec4af28b5b9e64b6ca58bee7f89fbaddc" Jan 27 16:01:18 crc kubenswrapper[4772]: I0127 16:01:18.990152 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6f9a4e35dfde3b6f98ddcc01331fdfec4af28b5b9e64b6ca58bee7f89fbaddc"} err="failed to get container status \"f6f9a4e35dfde3b6f98ddcc01331fdfec4af28b5b9e64b6ca58bee7f89fbaddc\": rpc error: code = NotFound desc = could not find container \"f6f9a4e35dfde3b6f98ddcc01331fdfec4af28b5b9e64b6ca58bee7f89fbaddc\": container with ID starting with f6f9a4e35dfde3b6f98ddcc01331fdfec4af28b5b9e64b6ca58bee7f89fbaddc not found: ID does not exist" Jan 27 16:01:18 crc kubenswrapper[4772]: I0127 16:01:18.990200 4772 scope.go:117] "RemoveContainer" containerID="d626ad0bd0e93bf253da87fff3d7e5845272782b38c1fe8f821ce4c09122879f" Jan 27 16:01:18 crc kubenswrapper[4772]: E0127 16:01:18.997214 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d626ad0bd0e93bf253da87fff3d7e5845272782b38c1fe8f821ce4c09122879f\": container with ID starting with d626ad0bd0e93bf253da87fff3d7e5845272782b38c1fe8f821ce4c09122879f not found: ID does not exist" containerID="d626ad0bd0e93bf253da87fff3d7e5845272782b38c1fe8f821ce4c09122879f" Jan 27 16:01:18 crc kubenswrapper[4772]: I0127 16:01:18.997266 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d626ad0bd0e93bf253da87fff3d7e5845272782b38c1fe8f821ce4c09122879f"} err="failed to get container status \"d626ad0bd0e93bf253da87fff3d7e5845272782b38c1fe8f821ce4c09122879f\": rpc error: code = NotFound desc = could not find container \"d626ad0bd0e93bf253da87fff3d7e5845272782b38c1fe8f821ce4c09122879f\": container with ID starting with d626ad0bd0e93bf253da87fff3d7e5845272782b38c1fe8f821ce4c09122879f not found: ID does not exist" Jan 27 16:01:18 crc kubenswrapper[4772]: I0127 16:01:18.997298 4772 scope.go:117] "RemoveContainer" containerID="5c076ec7e9f7ce861de2aeb2d2ce1407d2581066fe8432d268d48aa28c136d02" Jan 27 16:01:19 crc kubenswrapper[4772]: E0127 16:01:19.002447 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c076ec7e9f7ce861de2aeb2d2ce1407d2581066fe8432d268d48aa28c136d02\": container with ID starting with 5c076ec7e9f7ce861de2aeb2d2ce1407d2581066fe8432d268d48aa28c136d02 not found: ID does not exist" containerID="5c076ec7e9f7ce861de2aeb2d2ce1407d2581066fe8432d268d48aa28c136d02" Jan 27 16:01:19 crc kubenswrapper[4772]: I0127 16:01:19.002501 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c076ec7e9f7ce861de2aeb2d2ce1407d2581066fe8432d268d48aa28c136d02"} err="failed to get container status \"5c076ec7e9f7ce861de2aeb2d2ce1407d2581066fe8432d268d48aa28c136d02\": rpc error: code = NotFound desc = could not find container \"5c076ec7e9f7ce861de2aeb2d2ce1407d2581066fe8432d268d48aa28c136d02\": container with ID starting with 5c076ec7e9f7ce861de2aeb2d2ce1407d2581066fe8432d268d48aa28c136d02 not found: ID does not exist" Jan 27 16:01:20 crc kubenswrapper[4772]: I0127 16:01:20.672507 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cd0f6c5-43f1-4120-841d-76e540249886" path="/var/lib/kubelet/pods/3cd0f6c5-43f1-4120-841d-76e540249886/volumes" Jan 27 16:02:42 crc kubenswrapper[4772]: I0127 16:02:42.058695 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:02:42 crc kubenswrapper[4772]: I0127 16:02:42.059286 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:03:12 crc kubenswrapper[4772]: I0127 16:03:12.058256 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:03:12 crc kubenswrapper[4772]: I0127 16:03:12.058874 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:03:42 crc kubenswrapper[4772]: I0127 16:03:42.058556 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:03:42 crc kubenswrapper[4772]: I0127 16:03:42.059312 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:03:42 crc kubenswrapper[4772]: I0127 16:03:42.059397 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 16:03:42 crc kubenswrapper[4772]: I0127 16:03:42.060126 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"21ef3f853d795985962b240174bbec1611d9f0f58af15b07556fa41617b20592"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 16:03:42 crc kubenswrapper[4772]: I0127 16:03:42.060240 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://21ef3f853d795985962b240174bbec1611d9f0f58af15b07556fa41617b20592" gracePeriod=600 Jan 27 16:03:42 crc kubenswrapper[4772]: I0127 16:03:42.845843 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="21ef3f853d795985962b240174bbec1611d9f0f58af15b07556fa41617b20592" exitCode=0 Jan 27 16:03:42 crc kubenswrapper[4772]: I0127 16:03:42.845954 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"21ef3f853d795985962b240174bbec1611d9f0f58af15b07556fa41617b20592"} Jan 27 16:03:42 crc kubenswrapper[4772]: I0127 16:03:42.846321 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28"} Jan 27 16:03:42 crc kubenswrapper[4772]: I0127 16:03:42.846363 4772 scope.go:117] "RemoveContainer" containerID="bf80c85ff055e5b66481b1fb0c03a4a19bc2dadb96e8c295e0086beb0bb97a51" Jan 27 16:05:42 crc kubenswrapper[4772]: I0127 16:05:42.059162 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:05:42 crc kubenswrapper[4772]: I0127 16:05:42.060059 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:06:12 crc kubenswrapper[4772]: I0127 16:06:12.058524 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:06:12 crc kubenswrapper[4772]: I0127 16:06:12.059398 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:06:42 crc kubenswrapper[4772]: I0127 16:06:42.059401 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:06:42 crc kubenswrapper[4772]: I0127 16:06:42.060378 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:06:42 crc kubenswrapper[4772]: I0127 16:06:42.060432 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 16:06:42 crc kubenswrapper[4772]: I0127 16:06:42.061244 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 16:06:42 crc kubenswrapper[4772]: I0127 16:06:42.061333 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" gracePeriod=600 Jan 27 16:06:42 crc kubenswrapper[4772]: E0127 16:06:42.194895 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:06:42 crc kubenswrapper[4772]: I0127 16:06:42.513959 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" exitCode=0 Jan 27 16:06:42 crc kubenswrapper[4772]: I0127 16:06:42.514352 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28"} Jan 27 16:06:42 crc kubenswrapper[4772]: I0127 16:06:42.514522 4772 scope.go:117] "RemoveContainer" containerID="21ef3f853d795985962b240174bbec1611d9f0f58af15b07556fa41617b20592" Jan 27 16:06:42 crc kubenswrapper[4772]: I0127 16:06:42.515389 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:06:42 crc kubenswrapper[4772]: E0127 16:06:42.515936 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:06:55 crc kubenswrapper[4772]: I0127 16:06:55.662858 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:06:55 crc kubenswrapper[4772]: E0127 16:06:55.664669 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:07:10 crc kubenswrapper[4772]: I0127 16:07:10.663907 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:07:10 crc kubenswrapper[4772]: E0127 16:07:10.666051 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:07:21 crc kubenswrapper[4772]: I0127 16:07:21.663378 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:07:21 crc kubenswrapper[4772]: E0127 16:07:21.664153 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:07:33 crc kubenswrapper[4772]: I0127 16:07:33.663237 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:07:33 crc kubenswrapper[4772]: E0127 16:07:33.664123 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:07:45 crc kubenswrapper[4772]: I0127 16:07:45.662683 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:07:45 crc kubenswrapper[4772]: E0127 16:07:45.663503 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:08:00 crc kubenswrapper[4772]: I0127 16:08:00.664179 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:08:00 crc kubenswrapper[4772]: E0127 16:08:00.664892 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:08:11 crc kubenswrapper[4772]: I0127 16:08:11.663489 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:08:11 crc kubenswrapper[4772]: E0127 16:08:11.664469 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:08:22 crc kubenswrapper[4772]: I0127 16:08:22.662708 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:08:22 crc kubenswrapper[4772]: E0127 16:08:22.663400 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:08:33 crc kubenswrapper[4772]: I0127 16:08:33.663360 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:08:33 crc kubenswrapper[4772]: E0127 16:08:33.664434 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:08:45 crc kubenswrapper[4772]: I0127 16:08:45.663725 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:08:45 crc kubenswrapper[4772]: E0127 16:08:45.664663 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:08:56 crc kubenswrapper[4772]: I0127 16:08:56.665589 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:08:56 crc kubenswrapper[4772]: E0127 16:08:56.666917 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:09:10 crc kubenswrapper[4772]: I0127 16:09:10.663920 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:09:10 crc kubenswrapper[4772]: E0127 16:09:10.684547 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:09:25 crc kubenswrapper[4772]: I0127 16:09:25.663143 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:09:25 crc kubenswrapper[4772]: E0127 16:09:25.663938 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:09:37 crc kubenswrapper[4772]: I0127 16:09:37.662876 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:09:37 crc kubenswrapper[4772]: E0127 16:09:37.663694 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:09:51 crc kubenswrapper[4772]: I0127 16:09:51.663403 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:09:51 crc kubenswrapper[4772]: E0127 16:09:51.664290 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:10:02 crc kubenswrapper[4772]: I0127 16:10:02.663717 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:10:02 crc kubenswrapper[4772]: E0127 16:10:02.664594 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:10:14 crc kubenswrapper[4772]: I0127 16:10:14.096932 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-z6dng"] Jan 27 16:10:14 crc kubenswrapper[4772]: E0127 16:10:14.102143 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd0f6c5-43f1-4120-841d-76e540249886" containerName="registry-server" Jan 27 16:10:14 crc kubenswrapper[4772]: I0127 16:10:14.102210 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd0f6c5-43f1-4120-841d-76e540249886" containerName="registry-server" Jan 27 16:10:14 crc kubenswrapper[4772]: E0127 16:10:14.102244 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd0f6c5-43f1-4120-841d-76e540249886" containerName="extract-utilities" Jan 27 16:10:14 crc kubenswrapper[4772]: I0127 16:10:14.102256 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd0f6c5-43f1-4120-841d-76e540249886" containerName="extract-utilities" Jan 27 16:10:14 crc kubenswrapper[4772]: E0127 16:10:14.102274 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd0f6c5-43f1-4120-841d-76e540249886" containerName="extract-content" Jan 27 16:10:14 crc kubenswrapper[4772]: I0127 16:10:14.102287 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd0f6c5-43f1-4120-841d-76e540249886" containerName="extract-content" Jan 27 16:10:14 crc kubenswrapper[4772]: I0127 16:10:14.102540 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cd0f6c5-43f1-4120-841d-76e540249886" containerName="registry-server" Jan 27 16:10:14 crc kubenswrapper[4772]: I0127 16:10:14.104222 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z6dng" Jan 27 16:10:14 crc kubenswrapper[4772]: I0127 16:10:14.111779 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z6dng"] Jan 27 16:10:14 crc kubenswrapper[4772]: I0127 16:10:14.164909 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e-utilities\") pod \"certified-operators-z6dng\" (UID: \"c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e\") " pod="openshift-marketplace/certified-operators-z6dng" Jan 27 16:10:14 crc kubenswrapper[4772]: I0127 16:10:14.165025 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw26c\" (UniqueName: \"kubernetes.io/projected/c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e-kube-api-access-jw26c\") pod \"certified-operators-z6dng\" (UID: \"c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e\") " pod="openshift-marketplace/certified-operators-z6dng" Jan 27 16:10:14 crc kubenswrapper[4772]: I0127 16:10:14.165052 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e-catalog-content\") pod \"certified-operators-z6dng\" (UID: \"c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e\") " pod="openshift-marketplace/certified-operators-z6dng" Jan 27 16:10:14 crc kubenswrapper[4772]: I0127 16:10:14.266157 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e-utilities\") pod \"certified-operators-z6dng\" (UID: \"c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e\") " pod="openshift-marketplace/certified-operators-z6dng" Jan 27 16:10:14 crc kubenswrapper[4772]: I0127 16:10:14.266287 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw26c\" (UniqueName: \"kubernetes.io/projected/c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e-kube-api-access-jw26c\") pod \"certified-operators-z6dng\" (UID: \"c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e\") " pod="openshift-marketplace/certified-operators-z6dng" Jan 27 16:10:14 crc kubenswrapper[4772]: I0127 16:10:14.266309 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e-catalog-content\") pod \"certified-operators-z6dng\" (UID: \"c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e\") " pod="openshift-marketplace/certified-operators-z6dng" Jan 27 16:10:14 crc kubenswrapper[4772]: I0127 16:10:14.266774 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e-catalog-content\") pod \"certified-operators-z6dng\" (UID: \"c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e\") " pod="openshift-marketplace/certified-operators-z6dng" Jan 27 16:10:14 crc kubenswrapper[4772]: I0127 16:10:14.267214 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e-utilities\") pod \"certified-operators-z6dng\" (UID: \"c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e\") " pod="openshift-marketplace/certified-operators-z6dng" Jan 27 16:10:14 crc kubenswrapper[4772]: I0127 16:10:14.290469 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw26c\" (UniqueName: \"kubernetes.io/projected/c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e-kube-api-access-jw26c\") pod \"certified-operators-z6dng\" (UID: \"c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e\") " pod="openshift-marketplace/certified-operators-z6dng" Jan 27 16:10:14 crc kubenswrapper[4772]: I0127 16:10:14.435454 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z6dng" Jan 27 16:10:14 crc kubenswrapper[4772]: I0127 16:10:14.910775 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z6dng"] Jan 27 16:10:15 crc kubenswrapper[4772]: I0127 16:10:15.175196 4772 generic.go:334] "Generic (PLEG): container finished" podID="c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e" containerID="35168c6609a7407b4fc3f7bf925b654f6a0d8a97426e7a546461834bf8cb717f" exitCode=0 Jan 27 16:10:15 crc kubenswrapper[4772]: I0127 16:10:15.175266 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z6dng" event={"ID":"c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e","Type":"ContainerDied","Data":"35168c6609a7407b4fc3f7bf925b654f6a0d8a97426e7a546461834bf8cb717f"} Jan 27 16:10:15 crc kubenswrapper[4772]: I0127 16:10:15.175532 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z6dng" event={"ID":"c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e","Type":"ContainerStarted","Data":"558e1f6a78cf7d4a5320f1cf0bc3c39c426c541b60e924bae60ed80904aad6b9"} Jan 27 16:10:15 crc kubenswrapper[4772]: I0127 16:10:15.177859 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 16:10:17 crc kubenswrapper[4772]: I0127 16:10:17.200080 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z6dng" event={"ID":"c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e","Type":"ContainerStarted","Data":"6aa400c6da0aa1e2baf4abe44b967b9a0669127281173590e101b0edb111df62"} Jan 27 16:10:17 crc kubenswrapper[4772]: I0127 16:10:17.662524 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:10:17 crc kubenswrapper[4772]: E0127 16:10:17.662783 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:10:18 crc kubenswrapper[4772]: I0127 16:10:18.211946 4772 generic.go:334] "Generic (PLEG): container finished" podID="c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e" containerID="6aa400c6da0aa1e2baf4abe44b967b9a0669127281173590e101b0edb111df62" exitCode=0 Jan 27 16:10:18 crc kubenswrapper[4772]: I0127 16:10:18.212026 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z6dng" event={"ID":"c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e","Type":"ContainerDied","Data":"6aa400c6da0aa1e2baf4abe44b967b9a0669127281173590e101b0edb111df62"} Jan 27 16:10:19 crc kubenswrapper[4772]: I0127 16:10:19.219780 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z6dng" event={"ID":"c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e","Type":"ContainerStarted","Data":"e0d8b213023473fef3cf7c744bfb81b72647509211c96ef3e3c4b874720f6ece"} Jan 27 16:10:19 crc kubenswrapper[4772]: I0127 16:10:19.243403 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-z6dng" podStartSLOduration=1.810063081 podStartE2EDuration="5.243382351s" podCreationTimestamp="2026-01-27 16:10:14 +0000 UTC" firstStartedPulling="2026-01-27 16:10:15.177568788 +0000 UTC m=+3801.158177886" lastFinishedPulling="2026-01-27 16:10:18.610888058 +0000 UTC m=+3804.591497156" observedRunningTime="2026-01-27 16:10:19.235441942 +0000 UTC m=+3805.216051060" watchObservedRunningTime="2026-01-27 16:10:19.243382351 +0000 UTC m=+3805.223991459" Jan 27 16:10:24 crc kubenswrapper[4772]: I0127 16:10:24.436427 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-z6dng" Jan 27 16:10:24 crc kubenswrapper[4772]: I0127 16:10:24.437326 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-z6dng" Jan 27 16:10:24 crc kubenswrapper[4772]: I0127 16:10:24.517223 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-z6dng" Jan 27 16:10:25 crc kubenswrapper[4772]: I0127 16:10:25.311218 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-z6dng" Jan 27 16:10:25 crc kubenswrapper[4772]: I0127 16:10:25.358624 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z6dng"] Jan 27 16:10:27 crc kubenswrapper[4772]: I0127 16:10:27.275386 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-z6dng" podUID="c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e" containerName="registry-server" containerID="cri-o://e0d8b213023473fef3cf7c744bfb81b72647509211c96ef3e3c4b874720f6ece" gracePeriod=2 Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.244432 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z6dng" Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.285033 4772 generic.go:334] "Generic (PLEG): container finished" podID="c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e" containerID="e0d8b213023473fef3cf7c744bfb81b72647509211c96ef3e3c4b874720f6ece" exitCode=0 Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.285072 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z6dng" event={"ID":"c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e","Type":"ContainerDied","Data":"e0d8b213023473fef3cf7c744bfb81b72647509211c96ef3e3c4b874720f6ece"} Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.285098 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z6dng" event={"ID":"c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e","Type":"ContainerDied","Data":"558e1f6a78cf7d4a5320f1cf0bc3c39c426c541b60e924bae60ed80904aad6b9"} Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.285114 4772 scope.go:117] "RemoveContainer" containerID="e0d8b213023473fef3cf7c744bfb81b72647509211c96ef3e3c4b874720f6ece" Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.285124 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z6dng" Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.309700 4772 scope.go:117] "RemoveContainer" containerID="6aa400c6da0aa1e2baf4abe44b967b9a0669127281173590e101b0edb111df62" Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.324909 4772 scope.go:117] "RemoveContainer" containerID="35168c6609a7407b4fc3f7bf925b654f6a0d8a97426e7a546461834bf8cb717f" Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.354644 4772 scope.go:117] "RemoveContainer" containerID="e0d8b213023473fef3cf7c744bfb81b72647509211c96ef3e3c4b874720f6ece" Jan 27 16:10:28 crc kubenswrapper[4772]: E0127 16:10:28.354976 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0d8b213023473fef3cf7c744bfb81b72647509211c96ef3e3c4b874720f6ece\": container with ID starting with e0d8b213023473fef3cf7c744bfb81b72647509211c96ef3e3c4b874720f6ece not found: ID does not exist" containerID="e0d8b213023473fef3cf7c744bfb81b72647509211c96ef3e3c4b874720f6ece" Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.355016 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0d8b213023473fef3cf7c744bfb81b72647509211c96ef3e3c4b874720f6ece"} err="failed to get container status \"e0d8b213023473fef3cf7c744bfb81b72647509211c96ef3e3c4b874720f6ece\": rpc error: code = NotFound desc = could not find container \"e0d8b213023473fef3cf7c744bfb81b72647509211c96ef3e3c4b874720f6ece\": container with ID starting with e0d8b213023473fef3cf7c744bfb81b72647509211c96ef3e3c4b874720f6ece not found: ID does not exist" Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.355047 4772 scope.go:117] "RemoveContainer" containerID="6aa400c6da0aa1e2baf4abe44b967b9a0669127281173590e101b0edb111df62" Jan 27 16:10:28 crc kubenswrapper[4772]: E0127 16:10:28.355347 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6aa400c6da0aa1e2baf4abe44b967b9a0669127281173590e101b0edb111df62\": container with ID starting with 6aa400c6da0aa1e2baf4abe44b967b9a0669127281173590e101b0edb111df62 not found: ID does not exist" containerID="6aa400c6da0aa1e2baf4abe44b967b9a0669127281173590e101b0edb111df62" Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.355373 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6aa400c6da0aa1e2baf4abe44b967b9a0669127281173590e101b0edb111df62"} err="failed to get container status \"6aa400c6da0aa1e2baf4abe44b967b9a0669127281173590e101b0edb111df62\": rpc error: code = NotFound desc = could not find container \"6aa400c6da0aa1e2baf4abe44b967b9a0669127281173590e101b0edb111df62\": container with ID starting with 6aa400c6da0aa1e2baf4abe44b967b9a0669127281173590e101b0edb111df62 not found: ID does not exist" Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.355392 4772 scope.go:117] "RemoveContainer" containerID="35168c6609a7407b4fc3f7bf925b654f6a0d8a97426e7a546461834bf8cb717f" Jan 27 16:10:28 crc kubenswrapper[4772]: E0127 16:10:28.355657 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35168c6609a7407b4fc3f7bf925b654f6a0d8a97426e7a546461834bf8cb717f\": container with ID starting with 35168c6609a7407b4fc3f7bf925b654f6a0d8a97426e7a546461834bf8cb717f not found: ID does not exist" containerID="35168c6609a7407b4fc3f7bf925b654f6a0d8a97426e7a546461834bf8cb717f" Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.355688 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35168c6609a7407b4fc3f7bf925b654f6a0d8a97426e7a546461834bf8cb717f"} err="failed to get container status \"35168c6609a7407b4fc3f7bf925b654f6a0d8a97426e7a546461834bf8cb717f\": rpc error: code = NotFound desc = could not find container \"35168c6609a7407b4fc3f7bf925b654f6a0d8a97426e7a546461834bf8cb717f\": container with ID starting with 35168c6609a7407b4fc3f7bf925b654f6a0d8a97426e7a546461834bf8cb717f not found: ID does not exist" Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.400230 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e-catalog-content\") pod \"c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e\" (UID: \"c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e\") " Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.400306 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e-utilities\") pod \"c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e\" (UID: \"c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e\") " Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.400342 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jw26c\" (UniqueName: \"kubernetes.io/projected/c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e-kube-api-access-jw26c\") pod \"c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e\" (UID: \"c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e\") " Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.401211 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e-utilities" (OuterVolumeSpecName: "utilities") pod "c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e" (UID: "c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.406227 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e-kube-api-access-jw26c" (OuterVolumeSpecName: "kube-api-access-jw26c") pod "c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e" (UID: "c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e"). InnerVolumeSpecName "kube-api-access-jw26c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.456633 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e" (UID: "c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.502525 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.502583 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.502596 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jw26c\" (UniqueName: \"kubernetes.io/projected/c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e-kube-api-access-jw26c\") on node \"crc\" DevicePath \"\"" Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.627078 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z6dng"] Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.639549 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-z6dng"] Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.663409 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:10:28 crc kubenswrapper[4772]: E0127 16:10:28.663602 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:10:28 crc kubenswrapper[4772]: I0127 16:10:28.676709 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e" path="/var/lib/kubelet/pods/c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e/volumes" Jan 27 16:10:39 crc kubenswrapper[4772]: I0127 16:10:39.812964 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c4gj5"] Jan 27 16:10:39 crc kubenswrapper[4772]: E0127 16:10:39.815239 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e" containerName="extract-content" Jan 27 16:10:39 crc kubenswrapper[4772]: I0127 16:10:39.815375 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e" containerName="extract-content" Jan 27 16:10:39 crc kubenswrapper[4772]: E0127 16:10:39.815454 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e" containerName="extract-utilities" Jan 27 16:10:39 crc kubenswrapper[4772]: I0127 16:10:39.815536 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e" containerName="extract-utilities" Jan 27 16:10:39 crc kubenswrapper[4772]: E0127 16:10:39.815740 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e" containerName="registry-server" Jan 27 16:10:39 crc kubenswrapper[4772]: I0127 16:10:39.815823 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e" containerName="registry-server" Jan 27 16:10:39 crc kubenswrapper[4772]: I0127 16:10:39.816074 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9cc87d0-42e9-4591-8bd0-9d4614cfbb0e" containerName="registry-server" Jan 27 16:10:39 crc kubenswrapper[4772]: I0127 16:10:39.823354 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c4gj5" Jan 27 16:10:39 crc kubenswrapper[4772]: I0127 16:10:39.830305 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c4gj5"] Jan 27 16:10:39 crc kubenswrapper[4772]: I0127 16:10:39.959736 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2768246f-c1ba-4a6b-a591-3f2307bbb1ab-catalog-content\") pod \"redhat-marketplace-c4gj5\" (UID: \"2768246f-c1ba-4a6b-a591-3f2307bbb1ab\") " pod="openshift-marketplace/redhat-marketplace-c4gj5" Jan 27 16:10:39 crc kubenswrapper[4772]: I0127 16:10:39.959807 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jckq\" (UniqueName: \"kubernetes.io/projected/2768246f-c1ba-4a6b-a591-3f2307bbb1ab-kube-api-access-2jckq\") pod \"redhat-marketplace-c4gj5\" (UID: \"2768246f-c1ba-4a6b-a591-3f2307bbb1ab\") " pod="openshift-marketplace/redhat-marketplace-c4gj5" Jan 27 16:10:39 crc kubenswrapper[4772]: I0127 16:10:39.959853 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2768246f-c1ba-4a6b-a591-3f2307bbb1ab-utilities\") pod \"redhat-marketplace-c4gj5\" (UID: \"2768246f-c1ba-4a6b-a591-3f2307bbb1ab\") " pod="openshift-marketplace/redhat-marketplace-c4gj5" Jan 27 16:10:40 crc kubenswrapper[4772]: I0127 16:10:40.060682 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2768246f-c1ba-4a6b-a591-3f2307bbb1ab-catalog-content\") pod \"redhat-marketplace-c4gj5\" (UID: \"2768246f-c1ba-4a6b-a591-3f2307bbb1ab\") " pod="openshift-marketplace/redhat-marketplace-c4gj5" Jan 27 16:10:40 crc kubenswrapper[4772]: I0127 16:10:40.060737 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jckq\" (UniqueName: \"kubernetes.io/projected/2768246f-c1ba-4a6b-a591-3f2307bbb1ab-kube-api-access-2jckq\") pod \"redhat-marketplace-c4gj5\" (UID: \"2768246f-c1ba-4a6b-a591-3f2307bbb1ab\") " pod="openshift-marketplace/redhat-marketplace-c4gj5" Jan 27 16:10:40 crc kubenswrapper[4772]: I0127 16:10:40.060761 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2768246f-c1ba-4a6b-a591-3f2307bbb1ab-utilities\") pod \"redhat-marketplace-c4gj5\" (UID: \"2768246f-c1ba-4a6b-a591-3f2307bbb1ab\") " pod="openshift-marketplace/redhat-marketplace-c4gj5" Jan 27 16:10:40 crc kubenswrapper[4772]: I0127 16:10:40.061243 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2768246f-c1ba-4a6b-a591-3f2307bbb1ab-catalog-content\") pod \"redhat-marketplace-c4gj5\" (UID: \"2768246f-c1ba-4a6b-a591-3f2307bbb1ab\") " pod="openshift-marketplace/redhat-marketplace-c4gj5" Jan 27 16:10:40 crc kubenswrapper[4772]: I0127 16:10:40.061327 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2768246f-c1ba-4a6b-a591-3f2307bbb1ab-utilities\") pod \"redhat-marketplace-c4gj5\" (UID: \"2768246f-c1ba-4a6b-a591-3f2307bbb1ab\") " pod="openshift-marketplace/redhat-marketplace-c4gj5" Jan 27 16:10:40 crc kubenswrapper[4772]: I0127 16:10:40.079000 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jckq\" (UniqueName: \"kubernetes.io/projected/2768246f-c1ba-4a6b-a591-3f2307bbb1ab-kube-api-access-2jckq\") pod \"redhat-marketplace-c4gj5\" (UID: \"2768246f-c1ba-4a6b-a591-3f2307bbb1ab\") " pod="openshift-marketplace/redhat-marketplace-c4gj5" Jan 27 16:10:40 crc kubenswrapper[4772]: I0127 16:10:40.142153 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c4gj5" Jan 27 16:10:40 crc kubenswrapper[4772]: I0127 16:10:40.604203 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c4gj5"] Jan 27 16:10:40 crc kubenswrapper[4772]: I0127 16:10:40.664056 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:10:40 crc kubenswrapper[4772]: E0127 16:10:40.664325 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:10:41 crc kubenswrapper[4772]: I0127 16:10:41.379947 4772 generic.go:334] "Generic (PLEG): container finished" podID="2768246f-c1ba-4a6b-a591-3f2307bbb1ab" containerID="996c72dfdfa84fafb04e7cd89a50cece9a76678da4256459e90d7170571e1056" exitCode=0 Jan 27 16:10:41 crc kubenswrapper[4772]: I0127 16:10:41.380300 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c4gj5" event={"ID":"2768246f-c1ba-4a6b-a591-3f2307bbb1ab","Type":"ContainerDied","Data":"996c72dfdfa84fafb04e7cd89a50cece9a76678da4256459e90d7170571e1056"} Jan 27 16:10:41 crc kubenswrapper[4772]: I0127 16:10:41.380334 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c4gj5" event={"ID":"2768246f-c1ba-4a6b-a591-3f2307bbb1ab","Type":"ContainerStarted","Data":"492592aa862e27c3543cc1b061d716f233b3452f56f510b8b81b6ba4072a35ac"} Jan 27 16:10:42 crc kubenswrapper[4772]: I0127 16:10:42.388518 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c4gj5" event={"ID":"2768246f-c1ba-4a6b-a591-3f2307bbb1ab","Type":"ContainerStarted","Data":"2d5124e3976d77405e8d798c2f159be25a70440ff4f07c0ba4f655e8adb5a3fe"} Jan 27 16:10:43 crc kubenswrapper[4772]: I0127 16:10:43.397462 4772 generic.go:334] "Generic (PLEG): container finished" podID="2768246f-c1ba-4a6b-a591-3f2307bbb1ab" containerID="2d5124e3976d77405e8d798c2f159be25a70440ff4f07c0ba4f655e8adb5a3fe" exitCode=0 Jan 27 16:10:43 crc kubenswrapper[4772]: I0127 16:10:43.397522 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c4gj5" event={"ID":"2768246f-c1ba-4a6b-a591-3f2307bbb1ab","Type":"ContainerDied","Data":"2d5124e3976d77405e8d798c2f159be25a70440ff4f07c0ba4f655e8adb5a3fe"} Jan 27 16:10:44 crc kubenswrapper[4772]: I0127 16:10:44.407055 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c4gj5" event={"ID":"2768246f-c1ba-4a6b-a591-3f2307bbb1ab","Type":"ContainerStarted","Data":"1af866d659fa58e50a0610b6b8c8e12def86017e3a2f8f96a984121bf650c258"} Jan 27 16:10:44 crc kubenswrapper[4772]: I0127 16:10:44.441789 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c4gj5" podStartSLOduration=2.816111893 podStartE2EDuration="5.441754435s" podCreationTimestamp="2026-01-27 16:10:39 +0000 UTC" firstStartedPulling="2026-01-27 16:10:41.383180199 +0000 UTC m=+3827.363789297" lastFinishedPulling="2026-01-27 16:10:44.008822751 +0000 UTC m=+3829.989431839" observedRunningTime="2026-01-27 16:10:44.432018074 +0000 UTC m=+3830.412627192" watchObservedRunningTime="2026-01-27 16:10:44.441754435 +0000 UTC m=+3830.422363583" Jan 27 16:10:50 crc kubenswrapper[4772]: I0127 16:10:50.142375 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c4gj5" Jan 27 16:10:50 crc kubenswrapper[4772]: I0127 16:10:50.142890 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c4gj5" Jan 27 16:10:50 crc kubenswrapper[4772]: I0127 16:10:50.179965 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c4gj5" Jan 27 16:10:50 crc kubenswrapper[4772]: I0127 16:10:50.506490 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c4gj5" Jan 27 16:10:50 crc kubenswrapper[4772]: I0127 16:10:50.563064 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c4gj5"] Jan 27 16:10:52 crc kubenswrapper[4772]: I0127 16:10:52.459435 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c4gj5" podUID="2768246f-c1ba-4a6b-a591-3f2307bbb1ab" containerName="registry-server" containerID="cri-o://1af866d659fa58e50a0610b6b8c8e12def86017e3a2f8f96a984121bf650c258" gracePeriod=2 Jan 27 16:10:52 crc kubenswrapper[4772]: I0127 16:10:52.663303 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:10:52 crc kubenswrapper[4772]: E0127 16:10:52.663531 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:10:53 crc kubenswrapper[4772]: I0127 16:10:53.471836 4772 generic.go:334] "Generic (PLEG): container finished" podID="2768246f-c1ba-4a6b-a591-3f2307bbb1ab" containerID="1af866d659fa58e50a0610b6b8c8e12def86017e3a2f8f96a984121bf650c258" exitCode=0 Jan 27 16:10:53 crc kubenswrapper[4772]: I0127 16:10:53.471879 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c4gj5" event={"ID":"2768246f-c1ba-4a6b-a591-3f2307bbb1ab","Type":"ContainerDied","Data":"1af866d659fa58e50a0610b6b8c8e12def86017e3a2f8f96a984121bf650c258"} Jan 27 16:10:53 crc kubenswrapper[4772]: I0127 16:10:53.986808 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c4gj5" Jan 27 16:10:54 crc kubenswrapper[4772]: I0127 16:10:54.161149 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2768246f-c1ba-4a6b-a591-3f2307bbb1ab-catalog-content\") pod \"2768246f-c1ba-4a6b-a591-3f2307bbb1ab\" (UID: \"2768246f-c1ba-4a6b-a591-3f2307bbb1ab\") " Jan 27 16:10:54 crc kubenswrapper[4772]: I0127 16:10:54.161278 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jckq\" (UniqueName: \"kubernetes.io/projected/2768246f-c1ba-4a6b-a591-3f2307bbb1ab-kube-api-access-2jckq\") pod \"2768246f-c1ba-4a6b-a591-3f2307bbb1ab\" (UID: \"2768246f-c1ba-4a6b-a591-3f2307bbb1ab\") " Jan 27 16:10:54 crc kubenswrapper[4772]: I0127 16:10:54.161372 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2768246f-c1ba-4a6b-a591-3f2307bbb1ab-utilities\") pod \"2768246f-c1ba-4a6b-a591-3f2307bbb1ab\" (UID: \"2768246f-c1ba-4a6b-a591-3f2307bbb1ab\") " Jan 27 16:10:54 crc kubenswrapper[4772]: I0127 16:10:54.162245 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2768246f-c1ba-4a6b-a591-3f2307bbb1ab-utilities" (OuterVolumeSpecName: "utilities") pod "2768246f-c1ba-4a6b-a591-3f2307bbb1ab" (UID: "2768246f-c1ba-4a6b-a591-3f2307bbb1ab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:10:54 crc kubenswrapper[4772]: I0127 16:10:54.182790 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2768246f-c1ba-4a6b-a591-3f2307bbb1ab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2768246f-c1ba-4a6b-a591-3f2307bbb1ab" (UID: "2768246f-c1ba-4a6b-a591-3f2307bbb1ab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:10:54 crc kubenswrapper[4772]: I0127 16:10:54.263549 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2768246f-c1ba-4a6b-a591-3f2307bbb1ab-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 16:10:54 crc kubenswrapper[4772]: I0127 16:10:54.263587 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2768246f-c1ba-4a6b-a591-3f2307bbb1ab-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 16:10:54 crc kubenswrapper[4772]: I0127 16:10:54.481206 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c4gj5" event={"ID":"2768246f-c1ba-4a6b-a591-3f2307bbb1ab","Type":"ContainerDied","Data":"492592aa862e27c3543cc1b061d716f233b3452f56f510b8b81b6ba4072a35ac"} Jan 27 16:10:54 crc kubenswrapper[4772]: I0127 16:10:54.481279 4772 scope.go:117] "RemoveContainer" containerID="1af866d659fa58e50a0610b6b8c8e12def86017e3a2f8f96a984121bf650c258" Jan 27 16:10:54 crc kubenswrapper[4772]: I0127 16:10:54.481314 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c4gj5" Jan 27 16:10:54 crc kubenswrapper[4772]: I0127 16:10:54.485599 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2768246f-c1ba-4a6b-a591-3f2307bbb1ab-kube-api-access-2jckq" (OuterVolumeSpecName: "kube-api-access-2jckq") pod "2768246f-c1ba-4a6b-a591-3f2307bbb1ab" (UID: "2768246f-c1ba-4a6b-a591-3f2307bbb1ab"). InnerVolumeSpecName "kube-api-access-2jckq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:10:54 crc kubenswrapper[4772]: I0127 16:10:54.531737 4772 scope.go:117] "RemoveContainer" containerID="2d5124e3976d77405e8d798c2f159be25a70440ff4f07c0ba4f655e8adb5a3fe" Jan 27 16:10:54 crc kubenswrapper[4772]: I0127 16:10:54.547754 4772 scope.go:117] "RemoveContainer" containerID="996c72dfdfa84fafb04e7cd89a50cece9a76678da4256459e90d7170571e1056" Jan 27 16:10:54 crc kubenswrapper[4772]: I0127 16:10:54.567244 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jckq\" (UniqueName: \"kubernetes.io/projected/2768246f-c1ba-4a6b-a591-3f2307bbb1ab-kube-api-access-2jckq\") on node \"crc\" DevicePath \"\"" Jan 27 16:10:54 crc kubenswrapper[4772]: I0127 16:10:54.824423 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c4gj5"] Jan 27 16:10:54 crc kubenswrapper[4772]: I0127 16:10:54.825392 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c4gj5"] Jan 27 16:10:56 crc kubenswrapper[4772]: I0127 16:10:56.675070 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2768246f-c1ba-4a6b-a591-3f2307bbb1ab" path="/var/lib/kubelet/pods/2768246f-c1ba-4a6b-a591-3f2307bbb1ab/volumes" Jan 27 16:11:05 crc kubenswrapper[4772]: I0127 16:11:05.663131 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:11:05 crc kubenswrapper[4772]: E0127 16:11:05.664038 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:11:18 crc kubenswrapper[4772]: I0127 16:11:18.663062 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:11:18 crc kubenswrapper[4772]: E0127 16:11:18.664053 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:11:29 crc kubenswrapper[4772]: I0127 16:11:29.663511 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:11:29 crc kubenswrapper[4772]: E0127 16:11:29.664591 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:11:44 crc kubenswrapper[4772]: I0127 16:11:44.666924 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:11:45 crc kubenswrapper[4772]: I0127 16:11:45.855464 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"50cc7379b98d80b676c41cd73458def5401d0a1c59e714228729e6bc1cefe905"} Jan 27 16:14:12 crc kubenswrapper[4772]: I0127 16:14:12.058331 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:14:12 crc kubenswrapper[4772]: I0127 16:14:12.058898 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:14:42 crc kubenswrapper[4772]: I0127 16:14:42.058585 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:14:42 crc kubenswrapper[4772]: I0127 16:14:42.059109 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:15:00 crc kubenswrapper[4772]: I0127 16:15:00.188437 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492175-fg6x4"] Jan 27 16:15:00 crc kubenswrapper[4772]: E0127 16:15:00.189373 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2768246f-c1ba-4a6b-a591-3f2307bbb1ab" containerName="extract-content" Jan 27 16:15:00 crc kubenswrapper[4772]: I0127 16:15:00.189392 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="2768246f-c1ba-4a6b-a591-3f2307bbb1ab" containerName="extract-content" Jan 27 16:15:00 crc kubenswrapper[4772]: E0127 16:15:00.189420 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2768246f-c1ba-4a6b-a591-3f2307bbb1ab" containerName="extract-utilities" Jan 27 16:15:00 crc kubenswrapper[4772]: I0127 16:15:00.189430 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="2768246f-c1ba-4a6b-a591-3f2307bbb1ab" containerName="extract-utilities" Jan 27 16:15:00 crc kubenswrapper[4772]: E0127 16:15:00.189444 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2768246f-c1ba-4a6b-a591-3f2307bbb1ab" containerName="registry-server" Jan 27 16:15:00 crc kubenswrapper[4772]: I0127 16:15:00.189453 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="2768246f-c1ba-4a6b-a591-3f2307bbb1ab" containerName="registry-server" Jan 27 16:15:00 crc kubenswrapper[4772]: I0127 16:15:00.189639 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="2768246f-c1ba-4a6b-a591-3f2307bbb1ab" containerName="registry-server" Jan 27 16:15:00 crc kubenswrapper[4772]: I0127 16:15:00.190250 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492175-fg6x4" Jan 27 16:15:00 crc kubenswrapper[4772]: I0127 16:15:00.192308 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 27 16:15:00 crc kubenswrapper[4772]: I0127 16:15:00.193345 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 27 16:15:00 crc kubenswrapper[4772]: I0127 16:15:00.206771 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492175-fg6x4"] Jan 27 16:15:00 crc kubenswrapper[4772]: I0127 16:15:00.266691 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0be908ee-6173-4ee8-80c4-0738697898d2-config-volume\") pod \"collect-profiles-29492175-fg6x4\" (UID: \"0be908ee-6173-4ee8-80c4-0738697898d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492175-fg6x4" Jan 27 16:15:00 crc kubenswrapper[4772]: I0127 16:15:00.266753 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flppb\" (UniqueName: \"kubernetes.io/projected/0be908ee-6173-4ee8-80c4-0738697898d2-kube-api-access-flppb\") pod \"collect-profiles-29492175-fg6x4\" (UID: \"0be908ee-6173-4ee8-80c4-0738697898d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492175-fg6x4" Jan 27 16:15:00 crc kubenswrapper[4772]: I0127 16:15:00.266796 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0be908ee-6173-4ee8-80c4-0738697898d2-secret-volume\") pod \"collect-profiles-29492175-fg6x4\" (UID: \"0be908ee-6173-4ee8-80c4-0738697898d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492175-fg6x4" Jan 27 16:15:00 crc kubenswrapper[4772]: I0127 16:15:00.368070 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0be908ee-6173-4ee8-80c4-0738697898d2-config-volume\") pod \"collect-profiles-29492175-fg6x4\" (UID: \"0be908ee-6173-4ee8-80c4-0738697898d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492175-fg6x4" Jan 27 16:15:00 crc kubenswrapper[4772]: I0127 16:15:00.368184 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flppb\" (UniqueName: \"kubernetes.io/projected/0be908ee-6173-4ee8-80c4-0738697898d2-kube-api-access-flppb\") pod \"collect-profiles-29492175-fg6x4\" (UID: \"0be908ee-6173-4ee8-80c4-0738697898d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492175-fg6x4" Jan 27 16:15:00 crc kubenswrapper[4772]: I0127 16:15:00.368270 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0be908ee-6173-4ee8-80c4-0738697898d2-secret-volume\") pod \"collect-profiles-29492175-fg6x4\" (UID: \"0be908ee-6173-4ee8-80c4-0738697898d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492175-fg6x4" Jan 27 16:15:00 crc kubenswrapper[4772]: I0127 16:15:00.369085 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0be908ee-6173-4ee8-80c4-0738697898d2-config-volume\") pod \"collect-profiles-29492175-fg6x4\" (UID: \"0be908ee-6173-4ee8-80c4-0738697898d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492175-fg6x4" Jan 27 16:15:00 crc kubenswrapper[4772]: I0127 16:15:00.380028 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0be908ee-6173-4ee8-80c4-0738697898d2-secret-volume\") pod \"collect-profiles-29492175-fg6x4\" (UID: \"0be908ee-6173-4ee8-80c4-0738697898d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492175-fg6x4" Jan 27 16:15:00 crc kubenswrapper[4772]: I0127 16:15:00.384416 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flppb\" (UniqueName: \"kubernetes.io/projected/0be908ee-6173-4ee8-80c4-0738697898d2-kube-api-access-flppb\") pod \"collect-profiles-29492175-fg6x4\" (UID: \"0be908ee-6173-4ee8-80c4-0738697898d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492175-fg6x4" Jan 27 16:15:00 crc kubenswrapper[4772]: I0127 16:15:00.507597 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492175-fg6x4" Jan 27 16:15:00 crc kubenswrapper[4772]: I0127 16:15:00.748796 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492175-fg6x4"] Jan 27 16:15:00 crc kubenswrapper[4772]: W0127 16:15:00.751009 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0be908ee_6173_4ee8_80c4_0738697898d2.slice/crio-9bbba9000225be6d4721ad0c912ef6a2b8989be7041107ccb5241f8ca0ed8398 WatchSource:0}: Error finding container 9bbba9000225be6d4721ad0c912ef6a2b8989be7041107ccb5241f8ca0ed8398: Status 404 returned error can't find the container with id 9bbba9000225be6d4721ad0c912ef6a2b8989be7041107ccb5241f8ca0ed8398 Jan 27 16:15:01 crc kubenswrapper[4772]: I0127 16:15:01.297823 4772 generic.go:334] "Generic (PLEG): container finished" podID="0be908ee-6173-4ee8-80c4-0738697898d2" containerID="d013dea461e279e8b861558e82f04a509da66ccae91eabf32103d04803eb33bd" exitCode=0 Jan 27 16:15:01 crc kubenswrapper[4772]: I0127 16:15:01.297867 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492175-fg6x4" event={"ID":"0be908ee-6173-4ee8-80c4-0738697898d2","Type":"ContainerDied","Data":"d013dea461e279e8b861558e82f04a509da66ccae91eabf32103d04803eb33bd"} Jan 27 16:15:01 crc kubenswrapper[4772]: I0127 16:15:01.297901 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492175-fg6x4" event={"ID":"0be908ee-6173-4ee8-80c4-0738697898d2","Type":"ContainerStarted","Data":"9bbba9000225be6d4721ad0c912ef6a2b8989be7041107ccb5241f8ca0ed8398"} Jan 27 16:15:02 crc kubenswrapper[4772]: I0127 16:15:02.577490 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492175-fg6x4" Jan 27 16:15:02 crc kubenswrapper[4772]: I0127 16:15:02.698684 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0be908ee-6173-4ee8-80c4-0738697898d2-config-volume\") pod \"0be908ee-6173-4ee8-80c4-0738697898d2\" (UID: \"0be908ee-6173-4ee8-80c4-0738697898d2\") " Jan 27 16:15:02 crc kubenswrapper[4772]: I0127 16:15:02.698851 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flppb\" (UniqueName: \"kubernetes.io/projected/0be908ee-6173-4ee8-80c4-0738697898d2-kube-api-access-flppb\") pod \"0be908ee-6173-4ee8-80c4-0738697898d2\" (UID: \"0be908ee-6173-4ee8-80c4-0738697898d2\") " Jan 27 16:15:02 crc kubenswrapper[4772]: I0127 16:15:02.698891 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0be908ee-6173-4ee8-80c4-0738697898d2-secret-volume\") pod \"0be908ee-6173-4ee8-80c4-0738697898d2\" (UID: \"0be908ee-6173-4ee8-80c4-0738697898d2\") " Jan 27 16:15:02 crc kubenswrapper[4772]: I0127 16:15:02.699568 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0be908ee-6173-4ee8-80c4-0738697898d2-config-volume" (OuterVolumeSpecName: "config-volume") pod "0be908ee-6173-4ee8-80c4-0738697898d2" (UID: "0be908ee-6173-4ee8-80c4-0738697898d2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:15:02 crc kubenswrapper[4772]: I0127 16:15:02.703766 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0be908ee-6173-4ee8-80c4-0738697898d2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0be908ee-6173-4ee8-80c4-0738697898d2" (UID: "0be908ee-6173-4ee8-80c4-0738697898d2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:15:02 crc kubenswrapper[4772]: I0127 16:15:02.704011 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0be908ee-6173-4ee8-80c4-0738697898d2-kube-api-access-flppb" (OuterVolumeSpecName: "kube-api-access-flppb") pod "0be908ee-6173-4ee8-80c4-0738697898d2" (UID: "0be908ee-6173-4ee8-80c4-0738697898d2"). InnerVolumeSpecName "kube-api-access-flppb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:15:02 crc kubenswrapper[4772]: I0127 16:15:02.800325 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flppb\" (UniqueName: \"kubernetes.io/projected/0be908ee-6173-4ee8-80c4-0738697898d2-kube-api-access-flppb\") on node \"crc\" DevicePath \"\"" Jan 27 16:15:02 crc kubenswrapper[4772]: I0127 16:15:02.800354 4772 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0be908ee-6173-4ee8-80c4-0738697898d2-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 27 16:15:02 crc kubenswrapper[4772]: I0127 16:15:02.800364 4772 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0be908ee-6173-4ee8-80c4-0738697898d2-config-volume\") on node \"crc\" DevicePath \"\"" Jan 27 16:15:03 crc kubenswrapper[4772]: I0127 16:15:03.314621 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492175-fg6x4" event={"ID":"0be908ee-6173-4ee8-80c4-0738697898d2","Type":"ContainerDied","Data":"9bbba9000225be6d4721ad0c912ef6a2b8989be7041107ccb5241f8ca0ed8398"} Jan 27 16:15:03 crc kubenswrapper[4772]: I0127 16:15:03.314922 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9bbba9000225be6d4721ad0c912ef6a2b8989be7041107ccb5241f8ca0ed8398" Jan 27 16:15:03 crc kubenswrapper[4772]: I0127 16:15:03.314697 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492175-fg6x4" Jan 27 16:15:03 crc kubenswrapper[4772]: I0127 16:15:03.646178 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492130-7lgz6"] Jan 27 16:15:03 crc kubenswrapper[4772]: I0127 16:15:03.653045 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492130-7lgz6"] Jan 27 16:15:04 crc kubenswrapper[4772]: I0127 16:15:04.674225 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d103a19-1490-433a-abdb-3ebd279265f5" path="/var/lib/kubelet/pods/6d103a19-1490-433a-abdb-3ebd279265f5/volumes" Jan 27 16:15:06 crc kubenswrapper[4772]: I0127 16:15:06.014859 4772 scope.go:117] "RemoveContainer" containerID="58e0f9aeee1bc53c7d023bfdbaa2444440ab205390cfd9df2a1973966a2ae19f" Jan 27 16:15:12 crc kubenswrapper[4772]: I0127 16:15:12.058767 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:15:12 crc kubenswrapper[4772]: I0127 16:15:12.059417 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:15:12 crc kubenswrapper[4772]: I0127 16:15:12.059500 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 16:15:12 crc kubenswrapper[4772]: I0127 16:15:12.060155 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"50cc7379b98d80b676c41cd73458def5401d0a1c59e714228729e6bc1cefe905"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 16:15:12 crc kubenswrapper[4772]: I0127 16:15:12.060235 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://50cc7379b98d80b676c41cd73458def5401d0a1c59e714228729e6bc1cefe905" gracePeriod=600 Jan 27 16:15:12 crc kubenswrapper[4772]: I0127 16:15:12.371510 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="50cc7379b98d80b676c41cd73458def5401d0a1c59e714228729e6bc1cefe905" exitCode=0 Jan 27 16:15:12 crc kubenswrapper[4772]: I0127 16:15:12.371737 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"50cc7379b98d80b676c41cd73458def5401d0a1c59e714228729e6bc1cefe905"} Jan 27 16:15:12 crc kubenswrapper[4772]: I0127 16:15:12.371916 4772 scope.go:117] "RemoveContainer" containerID="619886d7924cc8d7020cb14dc925242ea7fcb59d2ec01dae8b4e97b29bb44e28" Jan 27 16:15:13 crc kubenswrapper[4772]: I0127 16:15:13.382641 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d"} Jan 27 16:15:57 crc kubenswrapper[4772]: I0127 16:15:57.000921 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-klr7d"] Jan 27 16:15:57 crc kubenswrapper[4772]: E0127 16:15:57.001834 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0be908ee-6173-4ee8-80c4-0738697898d2" containerName="collect-profiles" Jan 27 16:15:57 crc kubenswrapper[4772]: I0127 16:15:57.001849 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0be908ee-6173-4ee8-80c4-0738697898d2" containerName="collect-profiles" Jan 27 16:15:57 crc kubenswrapper[4772]: I0127 16:15:57.002027 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="0be908ee-6173-4ee8-80c4-0738697898d2" containerName="collect-profiles" Jan 27 16:15:57 crc kubenswrapper[4772]: I0127 16:15:57.003202 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-klr7d" Jan 27 16:15:57 crc kubenswrapper[4772]: I0127 16:15:57.013862 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-klr7d"] Jan 27 16:15:57 crc kubenswrapper[4772]: I0127 16:15:57.178102 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d53e726-f66e-4996-84fb-2f2547cadd29-utilities\") pod \"community-operators-klr7d\" (UID: \"0d53e726-f66e-4996-84fb-2f2547cadd29\") " pod="openshift-marketplace/community-operators-klr7d" Jan 27 16:15:57 crc kubenswrapper[4772]: I0127 16:15:57.178212 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d53e726-f66e-4996-84fb-2f2547cadd29-catalog-content\") pod \"community-operators-klr7d\" (UID: \"0d53e726-f66e-4996-84fb-2f2547cadd29\") " pod="openshift-marketplace/community-operators-klr7d" Jan 27 16:15:57 crc kubenswrapper[4772]: I0127 16:15:57.178245 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grbqf\" (UniqueName: \"kubernetes.io/projected/0d53e726-f66e-4996-84fb-2f2547cadd29-kube-api-access-grbqf\") pod \"community-operators-klr7d\" (UID: \"0d53e726-f66e-4996-84fb-2f2547cadd29\") " pod="openshift-marketplace/community-operators-klr7d" Jan 27 16:15:57 crc kubenswrapper[4772]: I0127 16:15:57.279508 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d53e726-f66e-4996-84fb-2f2547cadd29-catalog-content\") pod \"community-operators-klr7d\" (UID: \"0d53e726-f66e-4996-84fb-2f2547cadd29\") " pod="openshift-marketplace/community-operators-klr7d" Jan 27 16:15:57 crc kubenswrapper[4772]: I0127 16:15:57.279572 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grbqf\" (UniqueName: \"kubernetes.io/projected/0d53e726-f66e-4996-84fb-2f2547cadd29-kube-api-access-grbqf\") pod \"community-operators-klr7d\" (UID: \"0d53e726-f66e-4996-84fb-2f2547cadd29\") " pod="openshift-marketplace/community-operators-klr7d" Jan 27 16:15:57 crc kubenswrapper[4772]: I0127 16:15:57.279671 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d53e726-f66e-4996-84fb-2f2547cadd29-utilities\") pod \"community-operators-klr7d\" (UID: \"0d53e726-f66e-4996-84fb-2f2547cadd29\") " pod="openshift-marketplace/community-operators-klr7d" Jan 27 16:15:57 crc kubenswrapper[4772]: I0127 16:15:57.280117 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d53e726-f66e-4996-84fb-2f2547cadd29-catalog-content\") pod \"community-operators-klr7d\" (UID: \"0d53e726-f66e-4996-84fb-2f2547cadd29\") " pod="openshift-marketplace/community-operators-klr7d" Jan 27 16:15:57 crc kubenswrapper[4772]: I0127 16:15:57.280484 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d53e726-f66e-4996-84fb-2f2547cadd29-utilities\") pod \"community-operators-klr7d\" (UID: \"0d53e726-f66e-4996-84fb-2f2547cadd29\") " pod="openshift-marketplace/community-operators-klr7d" Jan 27 16:15:57 crc kubenswrapper[4772]: I0127 16:15:57.307076 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grbqf\" (UniqueName: \"kubernetes.io/projected/0d53e726-f66e-4996-84fb-2f2547cadd29-kube-api-access-grbqf\") pod \"community-operators-klr7d\" (UID: \"0d53e726-f66e-4996-84fb-2f2547cadd29\") " pod="openshift-marketplace/community-operators-klr7d" Jan 27 16:15:57 crc kubenswrapper[4772]: I0127 16:15:57.323907 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-klr7d" Jan 27 16:15:57 crc kubenswrapper[4772]: I0127 16:15:57.840766 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-klr7d"] Jan 27 16:15:58 crc kubenswrapper[4772]: I0127 16:15:58.723283 4772 generic.go:334] "Generic (PLEG): container finished" podID="0d53e726-f66e-4996-84fb-2f2547cadd29" containerID="dc989b02724c66a718647f7d51ad4282c746eb41e28bfd955a90b8dfc944f809" exitCode=0 Jan 27 16:15:58 crc kubenswrapper[4772]: I0127 16:15:58.723344 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klr7d" event={"ID":"0d53e726-f66e-4996-84fb-2f2547cadd29","Type":"ContainerDied","Data":"dc989b02724c66a718647f7d51ad4282c746eb41e28bfd955a90b8dfc944f809"} Jan 27 16:15:58 crc kubenswrapper[4772]: I0127 16:15:58.723696 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klr7d" event={"ID":"0d53e726-f66e-4996-84fb-2f2547cadd29","Type":"ContainerStarted","Data":"dc0a19ec7b4587b2e1794721655f1f30ca61ab62d049253d7b1c0a202d8024ba"} Jan 27 16:15:58 crc kubenswrapper[4772]: I0127 16:15:58.725134 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 16:15:59 crc kubenswrapper[4772]: I0127 16:15:59.733613 4772 generic.go:334] "Generic (PLEG): container finished" podID="0d53e726-f66e-4996-84fb-2f2547cadd29" containerID="30792775c2c579d6914fb2fed483b2e9a5316c264bffd3bdc2010be3ded311b9" exitCode=0 Jan 27 16:15:59 crc kubenswrapper[4772]: I0127 16:15:59.733670 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klr7d" event={"ID":"0d53e726-f66e-4996-84fb-2f2547cadd29","Type":"ContainerDied","Data":"30792775c2c579d6914fb2fed483b2e9a5316c264bffd3bdc2010be3ded311b9"} Jan 27 16:15:59 crc kubenswrapper[4772]: I0127 16:15:59.997278 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pqs7f"] Jan 27 16:15:59 crc kubenswrapper[4772]: I0127 16:15:59.998984 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pqs7f" Jan 27 16:16:00 crc kubenswrapper[4772]: I0127 16:16:00.012951 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pqs7f"] Jan 27 16:16:00 crc kubenswrapper[4772]: I0127 16:16:00.016700 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgnx5\" (UniqueName: \"kubernetes.io/projected/b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63-kube-api-access-tgnx5\") pod \"redhat-operators-pqs7f\" (UID: \"b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63\") " pod="openshift-marketplace/redhat-operators-pqs7f" Jan 27 16:16:00 crc kubenswrapper[4772]: I0127 16:16:00.016757 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63-catalog-content\") pod \"redhat-operators-pqs7f\" (UID: \"b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63\") " pod="openshift-marketplace/redhat-operators-pqs7f" Jan 27 16:16:00 crc kubenswrapper[4772]: I0127 16:16:00.016788 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63-utilities\") pod \"redhat-operators-pqs7f\" (UID: \"b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63\") " pod="openshift-marketplace/redhat-operators-pqs7f" Jan 27 16:16:00 crc kubenswrapper[4772]: I0127 16:16:00.117726 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgnx5\" (UniqueName: \"kubernetes.io/projected/b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63-kube-api-access-tgnx5\") pod \"redhat-operators-pqs7f\" (UID: \"b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63\") " pod="openshift-marketplace/redhat-operators-pqs7f" Jan 27 16:16:00 crc kubenswrapper[4772]: I0127 16:16:00.117791 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63-catalog-content\") pod \"redhat-operators-pqs7f\" (UID: \"b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63\") " pod="openshift-marketplace/redhat-operators-pqs7f" Jan 27 16:16:00 crc kubenswrapper[4772]: I0127 16:16:00.117821 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63-utilities\") pod \"redhat-operators-pqs7f\" (UID: \"b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63\") " pod="openshift-marketplace/redhat-operators-pqs7f" Jan 27 16:16:00 crc kubenswrapper[4772]: I0127 16:16:00.118336 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63-utilities\") pod \"redhat-operators-pqs7f\" (UID: \"b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63\") " pod="openshift-marketplace/redhat-operators-pqs7f" Jan 27 16:16:00 crc kubenswrapper[4772]: I0127 16:16:00.118527 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63-catalog-content\") pod \"redhat-operators-pqs7f\" (UID: \"b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63\") " pod="openshift-marketplace/redhat-operators-pqs7f" Jan 27 16:16:00 crc kubenswrapper[4772]: I0127 16:16:00.138302 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgnx5\" (UniqueName: \"kubernetes.io/projected/b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63-kube-api-access-tgnx5\") pod \"redhat-operators-pqs7f\" (UID: \"b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63\") " pod="openshift-marketplace/redhat-operators-pqs7f" Jan 27 16:16:00 crc kubenswrapper[4772]: I0127 16:16:00.327956 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pqs7f" Jan 27 16:16:00 crc kubenswrapper[4772]: I0127 16:16:00.742229 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klr7d" event={"ID":"0d53e726-f66e-4996-84fb-2f2547cadd29","Type":"ContainerStarted","Data":"28d4cd39be1138334db2b7783e62aa2ee0c778982c5eaf0a247d5529afdfc14d"} Jan 27 16:16:00 crc kubenswrapper[4772]: I0127 16:16:00.754092 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pqs7f"] Jan 27 16:16:00 crc kubenswrapper[4772]: W0127 16:16:00.765158 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4d3806b_ef7d_43c0_9c33_e08d0f8dcb63.slice/crio-a901b081ae362cf564826808ed535c7fccd3df2a57e808801afadcbe086a2b9b WatchSource:0}: Error finding container a901b081ae362cf564826808ed535c7fccd3df2a57e808801afadcbe086a2b9b: Status 404 returned error can't find the container with id a901b081ae362cf564826808ed535c7fccd3df2a57e808801afadcbe086a2b9b Jan 27 16:16:00 crc kubenswrapper[4772]: I0127 16:16:00.788957 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-klr7d" podStartSLOduration=3.370697252 podStartE2EDuration="4.788941905s" podCreationTimestamp="2026-01-27 16:15:56 +0000 UTC" firstStartedPulling="2026-01-27 16:15:58.724879859 +0000 UTC m=+4144.705488967" lastFinishedPulling="2026-01-27 16:16:00.143124522 +0000 UTC m=+4146.123733620" observedRunningTime="2026-01-27 16:16:00.786309829 +0000 UTC m=+4146.766918937" watchObservedRunningTime="2026-01-27 16:16:00.788941905 +0000 UTC m=+4146.769551003" Jan 27 16:16:01 crc kubenswrapper[4772]: I0127 16:16:01.750247 4772 generic.go:334] "Generic (PLEG): container finished" podID="b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63" containerID="6c518bdaf4840ef38edee20a0af898e5c9269c61f9c8cfe19cad3cd8770c5522" exitCode=0 Jan 27 16:16:01 crc kubenswrapper[4772]: I0127 16:16:01.751304 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqs7f" event={"ID":"b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63","Type":"ContainerDied","Data":"6c518bdaf4840ef38edee20a0af898e5c9269c61f9c8cfe19cad3cd8770c5522"} Jan 27 16:16:01 crc kubenswrapper[4772]: I0127 16:16:01.751365 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqs7f" event={"ID":"b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63","Type":"ContainerStarted","Data":"a901b081ae362cf564826808ed535c7fccd3df2a57e808801afadcbe086a2b9b"} Jan 27 16:16:02 crc kubenswrapper[4772]: I0127 16:16:02.760025 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqs7f" event={"ID":"b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63","Type":"ContainerStarted","Data":"f12cc8ff7554ccc6c80e45264e31a5ca39fa255c5b86b2f701314ef2b2c08a28"} Jan 27 16:16:03 crc kubenswrapper[4772]: I0127 16:16:03.771961 4772 generic.go:334] "Generic (PLEG): container finished" podID="b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63" containerID="f12cc8ff7554ccc6c80e45264e31a5ca39fa255c5b86b2f701314ef2b2c08a28" exitCode=0 Jan 27 16:16:03 crc kubenswrapper[4772]: I0127 16:16:03.772100 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqs7f" event={"ID":"b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63","Type":"ContainerDied","Data":"f12cc8ff7554ccc6c80e45264e31a5ca39fa255c5b86b2f701314ef2b2c08a28"} Jan 27 16:16:04 crc kubenswrapper[4772]: I0127 16:16:04.782836 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqs7f" event={"ID":"b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63","Type":"ContainerStarted","Data":"3ff98be81629d59868925742371033e6e9018417b497cb5940b98d61c9f755cf"} Jan 27 16:16:04 crc kubenswrapper[4772]: I0127 16:16:04.806599 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pqs7f" podStartSLOduration=3.232537149 podStartE2EDuration="5.806572403s" podCreationTimestamp="2026-01-27 16:15:59 +0000 UTC" firstStartedPulling="2026-01-27 16:16:01.75198679 +0000 UTC m=+4147.732595888" lastFinishedPulling="2026-01-27 16:16:04.326022044 +0000 UTC m=+4150.306631142" observedRunningTime="2026-01-27 16:16:04.800725635 +0000 UTC m=+4150.781334743" watchObservedRunningTime="2026-01-27 16:16:04.806572403 +0000 UTC m=+4150.787181501" Jan 27 16:16:07 crc kubenswrapper[4772]: I0127 16:16:07.324521 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-klr7d" Jan 27 16:16:07 crc kubenswrapper[4772]: I0127 16:16:07.324855 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-klr7d" Jan 27 16:16:07 crc kubenswrapper[4772]: I0127 16:16:07.365217 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-klr7d" Jan 27 16:16:07 crc kubenswrapper[4772]: I0127 16:16:07.840896 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-klr7d" Jan 27 16:16:08 crc kubenswrapper[4772]: I0127 16:16:08.793212 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-klr7d"] Jan 27 16:16:09 crc kubenswrapper[4772]: I0127 16:16:09.813463 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-klr7d" podUID="0d53e726-f66e-4996-84fb-2f2547cadd29" containerName="registry-server" containerID="cri-o://28d4cd39be1138334db2b7783e62aa2ee0c778982c5eaf0a247d5529afdfc14d" gracePeriod=2 Jan 27 16:16:10 crc kubenswrapper[4772]: I0127 16:16:10.329229 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pqs7f" Jan 27 16:16:10 crc kubenswrapper[4772]: I0127 16:16:10.329293 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pqs7f" Jan 27 16:16:10 crc kubenswrapper[4772]: I0127 16:16:10.371953 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pqs7f" Jan 27 16:16:10 crc kubenswrapper[4772]: I0127 16:16:10.836016 4772 generic.go:334] "Generic (PLEG): container finished" podID="0d53e726-f66e-4996-84fb-2f2547cadd29" containerID="28d4cd39be1138334db2b7783e62aa2ee0c778982c5eaf0a247d5529afdfc14d" exitCode=0 Jan 27 16:16:10 crc kubenswrapper[4772]: I0127 16:16:10.836856 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klr7d" event={"ID":"0d53e726-f66e-4996-84fb-2f2547cadd29","Type":"ContainerDied","Data":"28d4cd39be1138334db2b7783e62aa2ee0c778982c5eaf0a247d5529afdfc14d"} Jan 27 16:16:10 crc kubenswrapper[4772]: I0127 16:16:10.886442 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pqs7f" Jan 27 16:16:11 crc kubenswrapper[4772]: I0127 16:16:11.191967 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pqs7f"] Jan 27 16:16:11 crc kubenswrapper[4772]: I0127 16:16:11.437943 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-klr7d" Jan 27 16:16:11 crc kubenswrapper[4772]: I0127 16:16:11.480943 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grbqf\" (UniqueName: \"kubernetes.io/projected/0d53e726-f66e-4996-84fb-2f2547cadd29-kube-api-access-grbqf\") pod \"0d53e726-f66e-4996-84fb-2f2547cadd29\" (UID: \"0d53e726-f66e-4996-84fb-2f2547cadd29\") " Jan 27 16:16:11 crc kubenswrapper[4772]: I0127 16:16:11.481020 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d53e726-f66e-4996-84fb-2f2547cadd29-catalog-content\") pod \"0d53e726-f66e-4996-84fb-2f2547cadd29\" (UID: \"0d53e726-f66e-4996-84fb-2f2547cadd29\") " Jan 27 16:16:11 crc kubenswrapper[4772]: I0127 16:16:11.481206 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d53e726-f66e-4996-84fb-2f2547cadd29-utilities\") pod \"0d53e726-f66e-4996-84fb-2f2547cadd29\" (UID: \"0d53e726-f66e-4996-84fb-2f2547cadd29\") " Jan 27 16:16:11 crc kubenswrapper[4772]: I0127 16:16:11.482220 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d53e726-f66e-4996-84fb-2f2547cadd29-utilities" (OuterVolumeSpecName: "utilities") pod "0d53e726-f66e-4996-84fb-2f2547cadd29" (UID: "0d53e726-f66e-4996-84fb-2f2547cadd29"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:16:11 crc kubenswrapper[4772]: I0127 16:16:11.486134 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d53e726-f66e-4996-84fb-2f2547cadd29-kube-api-access-grbqf" (OuterVolumeSpecName: "kube-api-access-grbqf") pod "0d53e726-f66e-4996-84fb-2f2547cadd29" (UID: "0d53e726-f66e-4996-84fb-2f2547cadd29"). InnerVolumeSpecName "kube-api-access-grbqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:16:11 crc kubenswrapper[4772]: I0127 16:16:11.531677 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d53e726-f66e-4996-84fb-2f2547cadd29-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0d53e726-f66e-4996-84fb-2f2547cadd29" (UID: "0d53e726-f66e-4996-84fb-2f2547cadd29"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:16:11 crc kubenswrapper[4772]: I0127 16:16:11.583540 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grbqf\" (UniqueName: \"kubernetes.io/projected/0d53e726-f66e-4996-84fb-2f2547cadd29-kube-api-access-grbqf\") on node \"crc\" DevicePath \"\"" Jan 27 16:16:11 crc kubenswrapper[4772]: I0127 16:16:11.583590 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d53e726-f66e-4996-84fb-2f2547cadd29-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 16:16:11 crc kubenswrapper[4772]: I0127 16:16:11.583604 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d53e726-f66e-4996-84fb-2f2547cadd29-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 16:16:11 crc kubenswrapper[4772]: I0127 16:16:11.846305 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-klr7d" event={"ID":"0d53e726-f66e-4996-84fb-2f2547cadd29","Type":"ContainerDied","Data":"dc0a19ec7b4587b2e1794721655f1f30ca61ab62d049253d7b1c0a202d8024ba"} Jan 27 16:16:11 crc kubenswrapper[4772]: I0127 16:16:11.846388 4772 scope.go:117] "RemoveContainer" containerID="28d4cd39be1138334db2b7783e62aa2ee0c778982c5eaf0a247d5529afdfc14d" Jan 27 16:16:11 crc kubenswrapper[4772]: I0127 16:16:11.846609 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-klr7d" Jan 27 16:16:11 crc kubenswrapper[4772]: I0127 16:16:11.864328 4772 scope.go:117] "RemoveContainer" containerID="30792775c2c579d6914fb2fed483b2e9a5316c264bffd3bdc2010be3ded311b9" Jan 27 16:16:11 crc kubenswrapper[4772]: I0127 16:16:11.877632 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-klr7d"] Jan 27 16:16:11 crc kubenswrapper[4772]: I0127 16:16:11.888388 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-klr7d"] Jan 27 16:16:11 crc kubenswrapper[4772]: I0127 16:16:11.994755 4772 scope.go:117] "RemoveContainer" containerID="dc989b02724c66a718647f7d51ad4282c746eb41e28bfd955a90b8dfc944f809" Jan 27 16:16:12 crc kubenswrapper[4772]: I0127 16:16:12.673586 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d53e726-f66e-4996-84fb-2f2547cadd29" path="/var/lib/kubelet/pods/0d53e726-f66e-4996-84fb-2f2547cadd29/volumes" Jan 27 16:16:12 crc kubenswrapper[4772]: I0127 16:16:12.853459 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pqs7f" podUID="b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63" containerName="registry-server" containerID="cri-o://3ff98be81629d59868925742371033e6e9018417b497cb5940b98d61c9f755cf" gracePeriod=2 Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.336797 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pqs7f" Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.410376 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgnx5\" (UniqueName: \"kubernetes.io/projected/b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63-kube-api-access-tgnx5\") pod \"b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63\" (UID: \"b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63\") " Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.410433 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63-catalog-content\") pod \"b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63\" (UID: \"b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63\") " Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.410497 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63-utilities\") pod \"b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63\" (UID: \"b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63\") " Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.411758 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63-utilities" (OuterVolumeSpecName: "utilities") pod "b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63" (UID: "b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.414739 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63-kube-api-access-tgnx5" (OuterVolumeSpecName: "kube-api-access-tgnx5") pod "b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63" (UID: "b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63"). InnerVolumeSpecName "kube-api-access-tgnx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.511716 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgnx5\" (UniqueName: \"kubernetes.io/projected/b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63-kube-api-access-tgnx5\") on node \"crc\" DevicePath \"\"" Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.511753 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.797403 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63" (UID: "b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.815680 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.862067 4772 generic.go:334] "Generic (PLEG): container finished" podID="b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63" containerID="3ff98be81629d59868925742371033e6e9018417b497cb5940b98d61c9f755cf" exitCode=0 Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.862159 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqs7f" event={"ID":"b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63","Type":"ContainerDied","Data":"3ff98be81629d59868925742371033e6e9018417b497cb5940b98d61c9f755cf"} Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.862192 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pqs7f" Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.862243 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqs7f" event={"ID":"b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63","Type":"ContainerDied","Data":"a901b081ae362cf564826808ed535c7fccd3df2a57e808801afadcbe086a2b9b"} Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.862270 4772 scope.go:117] "RemoveContainer" containerID="3ff98be81629d59868925742371033e6e9018417b497cb5940b98d61c9f755cf" Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.882449 4772 scope.go:117] "RemoveContainer" containerID="f12cc8ff7554ccc6c80e45264e31a5ca39fa255c5b86b2f701314ef2b2c08a28" Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.903943 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pqs7f"] Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.909125 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pqs7f"] Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.910114 4772 scope.go:117] "RemoveContainer" containerID="6c518bdaf4840ef38edee20a0af898e5c9269c61f9c8cfe19cad3cd8770c5522" Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.935201 4772 scope.go:117] "RemoveContainer" containerID="3ff98be81629d59868925742371033e6e9018417b497cb5940b98d61c9f755cf" Jan 27 16:16:13 crc kubenswrapper[4772]: E0127 16:16:13.935757 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ff98be81629d59868925742371033e6e9018417b497cb5940b98d61c9f755cf\": container with ID starting with 3ff98be81629d59868925742371033e6e9018417b497cb5940b98d61c9f755cf not found: ID does not exist" containerID="3ff98be81629d59868925742371033e6e9018417b497cb5940b98d61c9f755cf" Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.935796 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ff98be81629d59868925742371033e6e9018417b497cb5940b98d61c9f755cf"} err="failed to get container status \"3ff98be81629d59868925742371033e6e9018417b497cb5940b98d61c9f755cf\": rpc error: code = NotFound desc = could not find container \"3ff98be81629d59868925742371033e6e9018417b497cb5940b98d61c9f755cf\": container with ID starting with 3ff98be81629d59868925742371033e6e9018417b497cb5940b98d61c9f755cf not found: ID does not exist" Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.935817 4772 scope.go:117] "RemoveContainer" containerID="f12cc8ff7554ccc6c80e45264e31a5ca39fa255c5b86b2f701314ef2b2c08a28" Jan 27 16:16:13 crc kubenswrapper[4772]: E0127 16:16:13.936155 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f12cc8ff7554ccc6c80e45264e31a5ca39fa255c5b86b2f701314ef2b2c08a28\": container with ID starting with f12cc8ff7554ccc6c80e45264e31a5ca39fa255c5b86b2f701314ef2b2c08a28 not found: ID does not exist" containerID="f12cc8ff7554ccc6c80e45264e31a5ca39fa255c5b86b2f701314ef2b2c08a28" Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.936235 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f12cc8ff7554ccc6c80e45264e31a5ca39fa255c5b86b2f701314ef2b2c08a28"} err="failed to get container status \"f12cc8ff7554ccc6c80e45264e31a5ca39fa255c5b86b2f701314ef2b2c08a28\": rpc error: code = NotFound desc = could not find container \"f12cc8ff7554ccc6c80e45264e31a5ca39fa255c5b86b2f701314ef2b2c08a28\": container with ID starting with f12cc8ff7554ccc6c80e45264e31a5ca39fa255c5b86b2f701314ef2b2c08a28 not found: ID does not exist" Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.936284 4772 scope.go:117] "RemoveContainer" containerID="6c518bdaf4840ef38edee20a0af898e5c9269c61f9c8cfe19cad3cd8770c5522" Jan 27 16:16:13 crc kubenswrapper[4772]: E0127 16:16:13.936548 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c518bdaf4840ef38edee20a0af898e5c9269c61f9c8cfe19cad3cd8770c5522\": container with ID starting with 6c518bdaf4840ef38edee20a0af898e5c9269c61f9c8cfe19cad3cd8770c5522 not found: ID does not exist" containerID="6c518bdaf4840ef38edee20a0af898e5c9269c61f9c8cfe19cad3cd8770c5522" Jan 27 16:16:13 crc kubenswrapper[4772]: I0127 16:16:13.936579 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c518bdaf4840ef38edee20a0af898e5c9269c61f9c8cfe19cad3cd8770c5522"} err="failed to get container status \"6c518bdaf4840ef38edee20a0af898e5c9269c61f9c8cfe19cad3cd8770c5522\": rpc error: code = NotFound desc = could not find container \"6c518bdaf4840ef38edee20a0af898e5c9269c61f9c8cfe19cad3cd8770c5522\": container with ID starting with 6c518bdaf4840ef38edee20a0af898e5c9269c61f9c8cfe19cad3cd8770c5522 not found: ID does not exist" Jan 27 16:16:14 crc kubenswrapper[4772]: I0127 16:16:14.675905 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63" path="/var/lib/kubelet/pods/b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63/volumes" Jan 27 16:17:12 crc kubenswrapper[4772]: I0127 16:17:12.058349 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:17:12 crc kubenswrapper[4772]: I0127 16:17:12.059154 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:17:42 crc kubenswrapper[4772]: I0127 16:17:42.059217 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:17:42 crc kubenswrapper[4772]: I0127 16:17:42.059784 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:18:12 crc kubenswrapper[4772]: I0127 16:18:12.058707 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:18:12 crc kubenswrapper[4772]: I0127 16:18:12.059301 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:18:12 crc kubenswrapper[4772]: I0127 16:18:12.059345 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 16:18:12 crc kubenswrapper[4772]: I0127 16:18:12.059899 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 16:18:12 crc kubenswrapper[4772]: I0127 16:18:12.059946 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" gracePeriod=600 Jan 27 16:18:12 crc kubenswrapper[4772]: E0127 16:18:12.188522 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:18:12 crc kubenswrapper[4772]: I0127 16:18:12.769899 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" exitCode=0 Jan 27 16:18:12 crc kubenswrapper[4772]: I0127 16:18:12.770003 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d"} Jan 27 16:18:12 crc kubenswrapper[4772]: I0127 16:18:12.770142 4772 scope.go:117] "RemoveContainer" containerID="50cc7379b98d80b676c41cd73458def5401d0a1c59e714228729e6bc1cefe905" Jan 27 16:18:12 crc kubenswrapper[4772]: I0127 16:18:12.771258 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:18:12 crc kubenswrapper[4772]: E0127 16:18:12.771666 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:18:27 crc kubenswrapper[4772]: I0127 16:18:27.663598 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:18:27 crc kubenswrapper[4772]: E0127 16:18:27.664508 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:18:42 crc kubenswrapper[4772]: I0127 16:18:42.663483 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:18:42 crc kubenswrapper[4772]: E0127 16:18:42.664540 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:18:53 crc kubenswrapper[4772]: I0127 16:18:53.667423 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:18:53 crc kubenswrapper[4772]: E0127 16:18:53.669531 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:19:06 crc kubenswrapper[4772]: I0127 16:19:06.664753 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:19:06 crc kubenswrapper[4772]: E0127 16:19:06.665842 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:19:17 crc kubenswrapper[4772]: I0127 16:19:17.663268 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:19:17 crc kubenswrapper[4772]: E0127 16:19:17.664006 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:19:28 crc kubenswrapper[4772]: I0127 16:19:28.663658 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:19:28 crc kubenswrapper[4772]: E0127 16:19:28.664463 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:19:41 crc kubenswrapper[4772]: I0127 16:19:41.663619 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:19:41 crc kubenswrapper[4772]: E0127 16:19:41.664447 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:19:56 crc kubenswrapper[4772]: I0127 16:19:56.663423 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:19:56 crc kubenswrapper[4772]: E0127 16:19:56.664228 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:20:07 crc kubenswrapper[4772]: I0127 16:20:07.663599 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:20:07 crc kubenswrapper[4772]: E0127 16:20:07.664419 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:20:20 crc kubenswrapper[4772]: I0127 16:20:20.663473 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:20:20 crc kubenswrapper[4772]: E0127 16:20:20.664417 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:20:32 crc kubenswrapper[4772]: I0127 16:20:32.662968 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:20:32 crc kubenswrapper[4772]: E0127 16:20:32.664131 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:20:43 crc kubenswrapper[4772]: I0127 16:20:43.051941 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rd4fm"] Jan 27 16:20:43 crc kubenswrapper[4772]: E0127 16:20:43.052830 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63" containerName="extract-utilities" Jan 27 16:20:43 crc kubenswrapper[4772]: I0127 16:20:43.052848 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63" containerName="extract-utilities" Jan 27 16:20:43 crc kubenswrapper[4772]: E0127 16:20:43.052865 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d53e726-f66e-4996-84fb-2f2547cadd29" containerName="extract-content" Jan 27 16:20:43 crc kubenswrapper[4772]: I0127 16:20:43.052872 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d53e726-f66e-4996-84fb-2f2547cadd29" containerName="extract-content" Jan 27 16:20:43 crc kubenswrapper[4772]: E0127 16:20:43.052891 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63" containerName="extract-content" Jan 27 16:20:43 crc kubenswrapper[4772]: I0127 16:20:43.052899 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63" containerName="extract-content" Jan 27 16:20:43 crc kubenswrapper[4772]: E0127 16:20:43.052909 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d53e726-f66e-4996-84fb-2f2547cadd29" containerName="registry-server" Jan 27 16:20:43 crc kubenswrapper[4772]: I0127 16:20:43.052918 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d53e726-f66e-4996-84fb-2f2547cadd29" containerName="registry-server" Jan 27 16:20:43 crc kubenswrapper[4772]: E0127 16:20:43.052928 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d53e726-f66e-4996-84fb-2f2547cadd29" containerName="extract-utilities" Jan 27 16:20:43 crc kubenswrapper[4772]: I0127 16:20:43.052934 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d53e726-f66e-4996-84fb-2f2547cadd29" containerName="extract-utilities" Jan 27 16:20:43 crc kubenswrapper[4772]: E0127 16:20:43.052950 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63" containerName="registry-server" Jan 27 16:20:43 crc kubenswrapper[4772]: I0127 16:20:43.052957 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63" containerName="registry-server" Jan 27 16:20:43 crc kubenswrapper[4772]: I0127 16:20:43.053123 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4d3806b-ef7d-43c0-9c33-e08d0f8dcb63" containerName="registry-server" Jan 27 16:20:43 crc kubenswrapper[4772]: I0127 16:20:43.053136 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d53e726-f66e-4996-84fb-2f2547cadd29" containerName="registry-server" Jan 27 16:20:43 crc kubenswrapper[4772]: I0127 16:20:43.054634 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rd4fm" Jan 27 16:20:43 crc kubenswrapper[4772]: I0127 16:20:43.066101 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rd4fm"] Jan 27 16:20:43 crc kubenswrapper[4772]: I0127 16:20:43.237233 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47938e85-9f22-40fc-a57a-6ba9649553eb-catalog-content\") pod \"redhat-marketplace-rd4fm\" (UID: \"47938e85-9f22-40fc-a57a-6ba9649553eb\") " pod="openshift-marketplace/redhat-marketplace-rd4fm" Jan 27 16:20:43 crc kubenswrapper[4772]: I0127 16:20:43.237462 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47938e85-9f22-40fc-a57a-6ba9649553eb-utilities\") pod \"redhat-marketplace-rd4fm\" (UID: \"47938e85-9f22-40fc-a57a-6ba9649553eb\") " pod="openshift-marketplace/redhat-marketplace-rd4fm" Jan 27 16:20:43 crc kubenswrapper[4772]: I0127 16:20:43.237648 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfjhc\" (UniqueName: \"kubernetes.io/projected/47938e85-9f22-40fc-a57a-6ba9649553eb-kube-api-access-sfjhc\") pod \"redhat-marketplace-rd4fm\" (UID: \"47938e85-9f22-40fc-a57a-6ba9649553eb\") " pod="openshift-marketplace/redhat-marketplace-rd4fm" Jan 27 16:20:43 crc kubenswrapper[4772]: I0127 16:20:43.338719 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfjhc\" (UniqueName: \"kubernetes.io/projected/47938e85-9f22-40fc-a57a-6ba9649553eb-kube-api-access-sfjhc\") pod \"redhat-marketplace-rd4fm\" (UID: \"47938e85-9f22-40fc-a57a-6ba9649553eb\") " pod="openshift-marketplace/redhat-marketplace-rd4fm" Jan 27 16:20:43 crc kubenswrapper[4772]: I0127 16:20:43.339093 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47938e85-9f22-40fc-a57a-6ba9649553eb-catalog-content\") pod \"redhat-marketplace-rd4fm\" (UID: \"47938e85-9f22-40fc-a57a-6ba9649553eb\") " pod="openshift-marketplace/redhat-marketplace-rd4fm" Jan 27 16:20:43 crc kubenswrapper[4772]: I0127 16:20:43.339308 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47938e85-9f22-40fc-a57a-6ba9649553eb-utilities\") pod \"redhat-marketplace-rd4fm\" (UID: \"47938e85-9f22-40fc-a57a-6ba9649553eb\") " pod="openshift-marketplace/redhat-marketplace-rd4fm" Jan 27 16:20:43 crc kubenswrapper[4772]: I0127 16:20:43.339626 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47938e85-9f22-40fc-a57a-6ba9649553eb-catalog-content\") pod \"redhat-marketplace-rd4fm\" (UID: \"47938e85-9f22-40fc-a57a-6ba9649553eb\") " pod="openshift-marketplace/redhat-marketplace-rd4fm" Jan 27 16:20:43 crc kubenswrapper[4772]: I0127 16:20:43.339770 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47938e85-9f22-40fc-a57a-6ba9649553eb-utilities\") pod \"redhat-marketplace-rd4fm\" (UID: \"47938e85-9f22-40fc-a57a-6ba9649553eb\") " pod="openshift-marketplace/redhat-marketplace-rd4fm" Jan 27 16:20:43 crc kubenswrapper[4772]: I0127 16:20:43.370281 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfjhc\" (UniqueName: \"kubernetes.io/projected/47938e85-9f22-40fc-a57a-6ba9649553eb-kube-api-access-sfjhc\") pod \"redhat-marketplace-rd4fm\" (UID: \"47938e85-9f22-40fc-a57a-6ba9649553eb\") " pod="openshift-marketplace/redhat-marketplace-rd4fm" Jan 27 16:20:43 crc kubenswrapper[4772]: I0127 16:20:43.405732 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rd4fm" Jan 27 16:20:43 crc kubenswrapper[4772]: I0127 16:20:43.843278 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rd4fm"] Jan 27 16:20:43 crc kubenswrapper[4772]: I0127 16:20:43.982544 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rd4fm" event={"ID":"47938e85-9f22-40fc-a57a-6ba9649553eb","Type":"ContainerStarted","Data":"53686951f0ef185c4075e7e81b31f486c3ba81d625f01a117518d6499db1b0d8"} Jan 27 16:20:44 crc kubenswrapper[4772]: E0127 16:20:44.214731 4772 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47938e85_9f22_40fc_a57a_6ba9649553eb.slice/crio-b20566d45bb50067ab86a9301c3513da2cd70c1f536b0bb394b306a49b9d08fc.scope\": RecentStats: unable to find data in memory cache]" Jan 27 16:20:44 crc kubenswrapper[4772]: I0127 16:20:44.996481 4772 generic.go:334] "Generic (PLEG): container finished" podID="47938e85-9f22-40fc-a57a-6ba9649553eb" containerID="b20566d45bb50067ab86a9301c3513da2cd70c1f536b0bb394b306a49b9d08fc" exitCode=0 Jan 27 16:20:44 crc kubenswrapper[4772]: I0127 16:20:44.996580 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rd4fm" event={"ID":"47938e85-9f22-40fc-a57a-6ba9649553eb","Type":"ContainerDied","Data":"b20566d45bb50067ab86a9301c3513da2cd70c1f536b0bb394b306a49b9d08fc"} Jan 27 16:20:45 crc kubenswrapper[4772]: I0127 16:20:45.663402 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:20:45 crc kubenswrapper[4772]: E0127 16:20:45.663692 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:20:46 crc kubenswrapper[4772]: I0127 16:20:46.005933 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rd4fm" event={"ID":"47938e85-9f22-40fc-a57a-6ba9649553eb","Type":"ContainerStarted","Data":"1f505067a0b82eec42c5b3819599d6f717108d18b5afebdd2d913d27113deb78"} Jan 27 16:20:47 crc kubenswrapper[4772]: I0127 16:20:47.013809 4772 generic.go:334] "Generic (PLEG): container finished" podID="47938e85-9f22-40fc-a57a-6ba9649553eb" containerID="1f505067a0b82eec42c5b3819599d6f717108d18b5afebdd2d913d27113deb78" exitCode=0 Jan 27 16:20:47 crc kubenswrapper[4772]: I0127 16:20:47.013859 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rd4fm" event={"ID":"47938e85-9f22-40fc-a57a-6ba9649553eb","Type":"ContainerDied","Data":"1f505067a0b82eec42c5b3819599d6f717108d18b5afebdd2d913d27113deb78"} Jan 27 16:20:48 crc kubenswrapper[4772]: I0127 16:20:48.024793 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rd4fm" event={"ID":"47938e85-9f22-40fc-a57a-6ba9649553eb","Type":"ContainerStarted","Data":"2049e20fe6b41b29f6a3b38b6b5f1356130ccd8fa55ed8d34829d3cca6d72ae0"} Jan 27 16:20:48 crc kubenswrapper[4772]: I0127 16:20:48.042153 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rd4fm" podStartSLOduration=2.589876625 podStartE2EDuration="5.042134067s" podCreationTimestamp="2026-01-27 16:20:43 +0000 UTC" firstStartedPulling="2026-01-27 16:20:45.000552811 +0000 UTC m=+4430.981161919" lastFinishedPulling="2026-01-27 16:20:47.452810263 +0000 UTC m=+4433.433419361" observedRunningTime="2026-01-27 16:20:48.039974575 +0000 UTC m=+4434.020583703" watchObservedRunningTime="2026-01-27 16:20:48.042134067 +0000 UTC m=+4434.022743185" Jan 27 16:20:53 crc kubenswrapper[4772]: I0127 16:20:53.406845 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rd4fm" Jan 27 16:20:53 crc kubenswrapper[4772]: I0127 16:20:53.407690 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rd4fm" Jan 27 16:20:53 crc kubenswrapper[4772]: I0127 16:20:53.464420 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rd4fm" Jan 27 16:20:54 crc kubenswrapper[4772]: I0127 16:20:54.186254 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rd4fm" Jan 27 16:20:54 crc kubenswrapper[4772]: I0127 16:20:54.252101 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rd4fm"] Jan 27 16:20:56 crc kubenswrapper[4772]: I0127 16:20:56.145217 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rd4fm" podUID="47938e85-9f22-40fc-a57a-6ba9649553eb" containerName="registry-server" containerID="cri-o://2049e20fe6b41b29f6a3b38b6b5f1356130ccd8fa55ed8d34829d3cca6d72ae0" gracePeriod=2 Jan 27 16:20:56 crc kubenswrapper[4772]: I0127 16:20:56.554673 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rd4fm" Jan 27 16:20:56 crc kubenswrapper[4772]: I0127 16:20:56.633021 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47938e85-9f22-40fc-a57a-6ba9649553eb-utilities\") pod \"47938e85-9f22-40fc-a57a-6ba9649553eb\" (UID: \"47938e85-9f22-40fc-a57a-6ba9649553eb\") " Jan 27 16:20:56 crc kubenswrapper[4772]: I0127 16:20:56.633070 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47938e85-9f22-40fc-a57a-6ba9649553eb-catalog-content\") pod \"47938e85-9f22-40fc-a57a-6ba9649553eb\" (UID: \"47938e85-9f22-40fc-a57a-6ba9649553eb\") " Jan 27 16:20:56 crc kubenswrapper[4772]: I0127 16:20:56.634282 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47938e85-9f22-40fc-a57a-6ba9649553eb-utilities" (OuterVolumeSpecName: "utilities") pod "47938e85-9f22-40fc-a57a-6ba9649553eb" (UID: "47938e85-9f22-40fc-a57a-6ba9649553eb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:20:56 crc kubenswrapper[4772]: I0127 16:20:56.662911 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:20:56 crc kubenswrapper[4772]: E0127 16:20:56.663314 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:20:56 crc kubenswrapper[4772]: I0127 16:20:56.665283 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47938e85-9f22-40fc-a57a-6ba9649553eb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "47938e85-9f22-40fc-a57a-6ba9649553eb" (UID: "47938e85-9f22-40fc-a57a-6ba9649553eb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:20:56 crc kubenswrapper[4772]: I0127 16:20:56.733853 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfjhc\" (UniqueName: \"kubernetes.io/projected/47938e85-9f22-40fc-a57a-6ba9649553eb-kube-api-access-sfjhc\") pod \"47938e85-9f22-40fc-a57a-6ba9649553eb\" (UID: \"47938e85-9f22-40fc-a57a-6ba9649553eb\") " Jan 27 16:20:56 crc kubenswrapper[4772]: I0127 16:20:56.734204 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47938e85-9f22-40fc-a57a-6ba9649553eb-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 16:20:56 crc kubenswrapper[4772]: I0127 16:20:56.734226 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47938e85-9f22-40fc-a57a-6ba9649553eb-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 16:20:56 crc kubenswrapper[4772]: I0127 16:20:56.740227 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47938e85-9f22-40fc-a57a-6ba9649553eb-kube-api-access-sfjhc" (OuterVolumeSpecName: "kube-api-access-sfjhc") pod "47938e85-9f22-40fc-a57a-6ba9649553eb" (UID: "47938e85-9f22-40fc-a57a-6ba9649553eb"). InnerVolumeSpecName "kube-api-access-sfjhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:20:56 crc kubenswrapper[4772]: I0127 16:20:56.835436 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfjhc\" (UniqueName: \"kubernetes.io/projected/47938e85-9f22-40fc-a57a-6ba9649553eb-kube-api-access-sfjhc\") on node \"crc\" DevicePath \"\"" Jan 27 16:20:57 crc kubenswrapper[4772]: I0127 16:20:57.153224 4772 generic.go:334] "Generic (PLEG): container finished" podID="47938e85-9f22-40fc-a57a-6ba9649553eb" containerID="2049e20fe6b41b29f6a3b38b6b5f1356130ccd8fa55ed8d34829d3cca6d72ae0" exitCode=0 Jan 27 16:20:57 crc kubenswrapper[4772]: I0127 16:20:57.153272 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rd4fm" event={"ID":"47938e85-9f22-40fc-a57a-6ba9649553eb","Type":"ContainerDied","Data":"2049e20fe6b41b29f6a3b38b6b5f1356130ccd8fa55ed8d34829d3cca6d72ae0"} Jan 27 16:20:57 crc kubenswrapper[4772]: I0127 16:20:57.153306 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rd4fm" event={"ID":"47938e85-9f22-40fc-a57a-6ba9649553eb","Type":"ContainerDied","Data":"53686951f0ef185c4075e7e81b31f486c3ba81d625f01a117518d6499db1b0d8"} Jan 27 16:20:57 crc kubenswrapper[4772]: I0127 16:20:57.153310 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rd4fm" Jan 27 16:20:57 crc kubenswrapper[4772]: I0127 16:20:57.153326 4772 scope.go:117] "RemoveContainer" containerID="2049e20fe6b41b29f6a3b38b6b5f1356130ccd8fa55ed8d34829d3cca6d72ae0" Jan 27 16:20:57 crc kubenswrapper[4772]: I0127 16:20:57.182062 4772 scope.go:117] "RemoveContainer" containerID="1f505067a0b82eec42c5b3819599d6f717108d18b5afebdd2d913d27113deb78" Jan 27 16:20:57 crc kubenswrapper[4772]: I0127 16:20:57.187401 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rd4fm"] Jan 27 16:20:57 crc kubenswrapper[4772]: I0127 16:20:57.192108 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rd4fm"] Jan 27 16:20:57 crc kubenswrapper[4772]: I0127 16:20:57.209717 4772 scope.go:117] "RemoveContainer" containerID="b20566d45bb50067ab86a9301c3513da2cd70c1f536b0bb394b306a49b9d08fc" Jan 27 16:20:57 crc kubenswrapper[4772]: I0127 16:20:57.237947 4772 scope.go:117] "RemoveContainer" containerID="2049e20fe6b41b29f6a3b38b6b5f1356130ccd8fa55ed8d34829d3cca6d72ae0" Jan 27 16:20:57 crc kubenswrapper[4772]: E0127 16:20:57.238474 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2049e20fe6b41b29f6a3b38b6b5f1356130ccd8fa55ed8d34829d3cca6d72ae0\": container with ID starting with 2049e20fe6b41b29f6a3b38b6b5f1356130ccd8fa55ed8d34829d3cca6d72ae0 not found: ID does not exist" containerID="2049e20fe6b41b29f6a3b38b6b5f1356130ccd8fa55ed8d34829d3cca6d72ae0" Jan 27 16:20:57 crc kubenswrapper[4772]: I0127 16:20:57.238512 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2049e20fe6b41b29f6a3b38b6b5f1356130ccd8fa55ed8d34829d3cca6d72ae0"} err="failed to get container status \"2049e20fe6b41b29f6a3b38b6b5f1356130ccd8fa55ed8d34829d3cca6d72ae0\": rpc error: code = NotFound desc = could not find container \"2049e20fe6b41b29f6a3b38b6b5f1356130ccd8fa55ed8d34829d3cca6d72ae0\": container with ID starting with 2049e20fe6b41b29f6a3b38b6b5f1356130ccd8fa55ed8d34829d3cca6d72ae0 not found: ID does not exist" Jan 27 16:20:57 crc kubenswrapper[4772]: I0127 16:20:57.238533 4772 scope.go:117] "RemoveContainer" containerID="1f505067a0b82eec42c5b3819599d6f717108d18b5afebdd2d913d27113deb78" Jan 27 16:20:57 crc kubenswrapper[4772]: E0127 16:20:57.238932 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f505067a0b82eec42c5b3819599d6f717108d18b5afebdd2d913d27113deb78\": container with ID starting with 1f505067a0b82eec42c5b3819599d6f717108d18b5afebdd2d913d27113deb78 not found: ID does not exist" containerID="1f505067a0b82eec42c5b3819599d6f717108d18b5afebdd2d913d27113deb78" Jan 27 16:20:57 crc kubenswrapper[4772]: I0127 16:20:57.238951 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f505067a0b82eec42c5b3819599d6f717108d18b5afebdd2d913d27113deb78"} err="failed to get container status \"1f505067a0b82eec42c5b3819599d6f717108d18b5afebdd2d913d27113deb78\": rpc error: code = NotFound desc = could not find container \"1f505067a0b82eec42c5b3819599d6f717108d18b5afebdd2d913d27113deb78\": container with ID starting with 1f505067a0b82eec42c5b3819599d6f717108d18b5afebdd2d913d27113deb78 not found: ID does not exist" Jan 27 16:20:57 crc kubenswrapper[4772]: I0127 16:20:57.238963 4772 scope.go:117] "RemoveContainer" containerID="b20566d45bb50067ab86a9301c3513da2cd70c1f536b0bb394b306a49b9d08fc" Jan 27 16:20:57 crc kubenswrapper[4772]: E0127 16:20:57.240763 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b20566d45bb50067ab86a9301c3513da2cd70c1f536b0bb394b306a49b9d08fc\": container with ID starting with b20566d45bb50067ab86a9301c3513da2cd70c1f536b0bb394b306a49b9d08fc not found: ID does not exist" containerID="b20566d45bb50067ab86a9301c3513da2cd70c1f536b0bb394b306a49b9d08fc" Jan 27 16:20:57 crc kubenswrapper[4772]: I0127 16:20:57.240929 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b20566d45bb50067ab86a9301c3513da2cd70c1f536b0bb394b306a49b9d08fc"} err="failed to get container status \"b20566d45bb50067ab86a9301c3513da2cd70c1f536b0bb394b306a49b9d08fc\": rpc error: code = NotFound desc = could not find container \"b20566d45bb50067ab86a9301c3513da2cd70c1f536b0bb394b306a49b9d08fc\": container with ID starting with b20566d45bb50067ab86a9301c3513da2cd70c1f536b0bb394b306a49b9d08fc not found: ID does not exist" Jan 27 16:20:58 crc kubenswrapper[4772]: I0127 16:20:58.678386 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47938e85-9f22-40fc-a57a-6ba9649553eb" path="/var/lib/kubelet/pods/47938e85-9f22-40fc-a57a-6ba9649553eb/volumes" Jan 27 16:21:00 crc kubenswrapper[4772]: I0127 16:21:00.977792 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8cfm5"] Jan 27 16:21:00 crc kubenswrapper[4772]: E0127 16:21:00.978156 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47938e85-9f22-40fc-a57a-6ba9649553eb" containerName="extract-utilities" Jan 27 16:21:00 crc kubenswrapper[4772]: I0127 16:21:00.978190 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="47938e85-9f22-40fc-a57a-6ba9649553eb" containerName="extract-utilities" Jan 27 16:21:00 crc kubenswrapper[4772]: E0127 16:21:00.978218 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47938e85-9f22-40fc-a57a-6ba9649553eb" containerName="registry-server" Jan 27 16:21:00 crc kubenswrapper[4772]: I0127 16:21:00.978227 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="47938e85-9f22-40fc-a57a-6ba9649553eb" containerName="registry-server" Jan 27 16:21:00 crc kubenswrapper[4772]: E0127 16:21:00.978244 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47938e85-9f22-40fc-a57a-6ba9649553eb" containerName="extract-content" Jan 27 16:21:00 crc kubenswrapper[4772]: I0127 16:21:00.978252 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="47938e85-9f22-40fc-a57a-6ba9649553eb" containerName="extract-content" Jan 27 16:21:00 crc kubenswrapper[4772]: I0127 16:21:00.978466 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="47938e85-9f22-40fc-a57a-6ba9649553eb" containerName="registry-server" Jan 27 16:21:00 crc kubenswrapper[4772]: I0127 16:21:00.979798 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8cfm5" Jan 27 16:21:00 crc kubenswrapper[4772]: I0127 16:21:00.991992 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8cfm5"] Jan 27 16:21:00 crc kubenswrapper[4772]: I0127 16:21:00.995199 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5854174b-59dc-43b7-aea1-321b0762e938-utilities\") pod \"certified-operators-8cfm5\" (UID: \"5854174b-59dc-43b7-aea1-321b0762e938\") " pod="openshift-marketplace/certified-operators-8cfm5" Jan 27 16:21:00 crc kubenswrapper[4772]: I0127 16:21:00.995402 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5854174b-59dc-43b7-aea1-321b0762e938-catalog-content\") pod \"certified-operators-8cfm5\" (UID: \"5854174b-59dc-43b7-aea1-321b0762e938\") " pod="openshift-marketplace/certified-operators-8cfm5" Jan 27 16:21:00 crc kubenswrapper[4772]: I0127 16:21:00.995610 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv9q7\" (UniqueName: \"kubernetes.io/projected/5854174b-59dc-43b7-aea1-321b0762e938-kube-api-access-tv9q7\") pod \"certified-operators-8cfm5\" (UID: \"5854174b-59dc-43b7-aea1-321b0762e938\") " pod="openshift-marketplace/certified-operators-8cfm5" Jan 27 16:21:01 crc kubenswrapper[4772]: I0127 16:21:01.097307 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5854174b-59dc-43b7-aea1-321b0762e938-utilities\") pod \"certified-operators-8cfm5\" (UID: \"5854174b-59dc-43b7-aea1-321b0762e938\") " pod="openshift-marketplace/certified-operators-8cfm5" Jan 27 16:21:01 crc kubenswrapper[4772]: I0127 16:21:01.097382 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5854174b-59dc-43b7-aea1-321b0762e938-catalog-content\") pod \"certified-operators-8cfm5\" (UID: \"5854174b-59dc-43b7-aea1-321b0762e938\") " pod="openshift-marketplace/certified-operators-8cfm5" Jan 27 16:21:01 crc kubenswrapper[4772]: I0127 16:21:01.097452 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tv9q7\" (UniqueName: \"kubernetes.io/projected/5854174b-59dc-43b7-aea1-321b0762e938-kube-api-access-tv9q7\") pod \"certified-operators-8cfm5\" (UID: \"5854174b-59dc-43b7-aea1-321b0762e938\") " pod="openshift-marketplace/certified-operators-8cfm5" Jan 27 16:21:01 crc kubenswrapper[4772]: I0127 16:21:01.098418 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5854174b-59dc-43b7-aea1-321b0762e938-utilities\") pod \"certified-operators-8cfm5\" (UID: \"5854174b-59dc-43b7-aea1-321b0762e938\") " pod="openshift-marketplace/certified-operators-8cfm5" Jan 27 16:21:01 crc kubenswrapper[4772]: I0127 16:21:01.098682 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5854174b-59dc-43b7-aea1-321b0762e938-catalog-content\") pod \"certified-operators-8cfm5\" (UID: \"5854174b-59dc-43b7-aea1-321b0762e938\") " pod="openshift-marketplace/certified-operators-8cfm5" Jan 27 16:21:01 crc kubenswrapper[4772]: I0127 16:21:01.117271 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv9q7\" (UniqueName: \"kubernetes.io/projected/5854174b-59dc-43b7-aea1-321b0762e938-kube-api-access-tv9q7\") pod \"certified-operators-8cfm5\" (UID: \"5854174b-59dc-43b7-aea1-321b0762e938\") " pod="openshift-marketplace/certified-operators-8cfm5" Jan 27 16:21:01 crc kubenswrapper[4772]: I0127 16:21:01.298092 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8cfm5" Jan 27 16:21:01 crc kubenswrapper[4772]: I0127 16:21:01.557621 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8cfm5"] Jan 27 16:21:02 crc kubenswrapper[4772]: I0127 16:21:02.191743 4772 generic.go:334] "Generic (PLEG): container finished" podID="5854174b-59dc-43b7-aea1-321b0762e938" containerID="a6ca15ff68296c29c6cc30ddf45249de19f9df9c7b117f0c3db8c30a8ef62894" exitCode=0 Jan 27 16:21:02 crc kubenswrapper[4772]: I0127 16:21:02.191853 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8cfm5" event={"ID":"5854174b-59dc-43b7-aea1-321b0762e938","Type":"ContainerDied","Data":"a6ca15ff68296c29c6cc30ddf45249de19f9df9c7b117f0c3db8c30a8ef62894"} Jan 27 16:21:02 crc kubenswrapper[4772]: I0127 16:21:02.192134 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8cfm5" event={"ID":"5854174b-59dc-43b7-aea1-321b0762e938","Type":"ContainerStarted","Data":"a25e0c08dab3d89878a502715c0666ee0ba53b97d33dc1e0735cbdd3a0adbe26"} Jan 27 16:21:02 crc kubenswrapper[4772]: I0127 16:21:02.193892 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 16:21:04 crc kubenswrapper[4772]: I0127 16:21:04.219029 4772 generic.go:334] "Generic (PLEG): container finished" podID="5854174b-59dc-43b7-aea1-321b0762e938" containerID="09b6440ec58b2621d6bef188feee41fb9bd95c35cb9a627aa339612e9aef4d84" exitCode=0 Jan 27 16:21:04 crc kubenswrapper[4772]: I0127 16:21:04.219098 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8cfm5" event={"ID":"5854174b-59dc-43b7-aea1-321b0762e938","Type":"ContainerDied","Data":"09b6440ec58b2621d6bef188feee41fb9bd95c35cb9a627aa339612e9aef4d84"} Jan 27 16:21:05 crc kubenswrapper[4772]: I0127 16:21:05.228900 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8cfm5" event={"ID":"5854174b-59dc-43b7-aea1-321b0762e938","Type":"ContainerStarted","Data":"64a13b2a934cc6a3e677cf6801cefaba161f5440fb5054836df1740f71d8d1af"} Jan 27 16:21:05 crc kubenswrapper[4772]: I0127 16:21:05.251978 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8cfm5" podStartSLOduration=2.48031297 podStartE2EDuration="5.251961878s" podCreationTimestamp="2026-01-27 16:21:00 +0000 UTC" firstStartedPulling="2026-01-27 16:21:02.193654958 +0000 UTC m=+4448.174264056" lastFinishedPulling="2026-01-27 16:21:04.965303826 +0000 UTC m=+4450.945912964" observedRunningTime="2026-01-27 16:21:05.249311372 +0000 UTC m=+4451.229920470" watchObservedRunningTime="2026-01-27 16:21:05.251961878 +0000 UTC m=+4451.232570976" Jan 27 16:21:07 crc kubenswrapper[4772]: I0127 16:21:07.662665 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:21:07 crc kubenswrapper[4772]: E0127 16:21:07.663198 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:21:11 crc kubenswrapper[4772]: I0127 16:21:11.298473 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8cfm5" Jan 27 16:21:11 crc kubenswrapper[4772]: I0127 16:21:11.300284 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8cfm5" Jan 27 16:21:11 crc kubenswrapper[4772]: I0127 16:21:11.346822 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8cfm5" Jan 27 16:21:12 crc kubenswrapper[4772]: I0127 16:21:12.315269 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8cfm5" Jan 27 16:21:12 crc kubenswrapper[4772]: I0127 16:21:12.764773 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8cfm5"] Jan 27 16:21:14 crc kubenswrapper[4772]: I0127 16:21:14.292340 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8cfm5" podUID="5854174b-59dc-43b7-aea1-321b0762e938" containerName="registry-server" containerID="cri-o://64a13b2a934cc6a3e677cf6801cefaba161f5440fb5054836df1740f71d8d1af" gracePeriod=2 Jan 27 16:21:14 crc kubenswrapper[4772]: I0127 16:21:14.779794 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8cfm5" Jan 27 16:21:14 crc kubenswrapper[4772]: I0127 16:21:14.909221 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5854174b-59dc-43b7-aea1-321b0762e938-catalog-content\") pod \"5854174b-59dc-43b7-aea1-321b0762e938\" (UID: \"5854174b-59dc-43b7-aea1-321b0762e938\") " Jan 27 16:21:14 crc kubenswrapper[4772]: I0127 16:21:14.909288 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tv9q7\" (UniqueName: \"kubernetes.io/projected/5854174b-59dc-43b7-aea1-321b0762e938-kube-api-access-tv9q7\") pod \"5854174b-59dc-43b7-aea1-321b0762e938\" (UID: \"5854174b-59dc-43b7-aea1-321b0762e938\") " Jan 27 16:21:14 crc kubenswrapper[4772]: I0127 16:21:14.909328 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5854174b-59dc-43b7-aea1-321b0762e938-utilities\") pod \"5854174b-59dc-43b7-aea1-321b0762e938\" (UID: \"5854174b-59dc-43b7-aea1-321b0762e938\") " Jan 27 16:21:14 crc kubenswrapper[4772]: I0127 16:21:14.910851 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5854174b-59dc-43b7-aea1-321b0762e938-utilities" (OuterVolumeSpecName: "utilities") pod "5854174b-59dc-43b7-aea1-321b0762e938" (UID: "5854174b-59dc-43b7-aea1-321b0762e938"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:21:14 crc kubenswrapper[4772]: I0127 16:21:14.915072 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5854174b-59dc-43b7-aea1-321b0762e938-kube-api-access-tv9q7" (OuterVolumeSpecName: "kube-api-access-tv9q7") pod "5854174b-59dc-43b7-aea1-321b0762e938" (UID: "5854174b-59dc-43b7-aea1-321b0762e938"). InnerVolumeSpecName "kube-api-access-tv9q7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:21:15 crc kubenswrapper[4772]: I0127 16:21:15.011624 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tv9q7\" (UniqueName: \"kubernetes.io/projected/5854174b-59dc-43b7-aea1-321b0762e938-kube-api-access-tv9q7\") on node \"crc\" DevicePath \"\"" Jan 27 16:21:15 crc kubenswrapper[4772]: I0127 16:21:15.011672 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5854174b-59dc-43b7-aea1-321b0762e938-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 16:21:15 crc kubenswrapper[4772]: I0127 16:21:15.113727 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5854174b-59dc-43b7-aea1-321b0762e938-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5854174b-59dc-43b7-aea1-321b0762e938" (UID: "5854174b-59dc-43b7-aea1-321b0762e938"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:21:15 crc kubenswrapper[4772]: I0127 16:21:15.214505 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5854174b-59dc-43b7-aea1-321b0762e938-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 16:21:15 crc kubenswrapper[4772]: I0127 16:21:15.300843 4772 generic.go:334] "Generic (PLEG): container finished" podID="5854174b-59dc-43b7-aea1-321b0762e938" containerID="64a13b2a934cc6a3e677cf6801cefaba161f5440fb5054836df1740f71d8d1af" exitCode=0 Jan 27 16:21:15 crc kubenswrapper[4772]: I0127 16:21:15.300898 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8cfm5" event={"ID":"5854174b-59dc-43b7-aea1-321b0762e938","Type":"ContainerDied","Data":"64a13b2a934cc6a3e677cf6801cefaba161f5440fb5054836df1740f71d8d1af"} Jan 27 16:21:15 crc kubenswrapper[4772]: I0127 16:21:15.300931 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8cfm5" Jan 27 16:21:15 crc kubenswrapper[4772]: I0127 16:21:15.300954 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8cfm5" event={"ID":"5854174b-59dc-43b7-aea1-321b0762e938","Type":"ContainerDied","Data":"a25e0c08dab3d89878a502715c0666ee0ba53b97d33dc1e0735cbdd3a0adbe26"} Jan 27 16:21:15 crc kubenswrapper[4772]: I0127 16:21:15.300971 4772 scope.go:117] "RemoveContainer" containerID="64a13b2a934cc6a3e677cf6801cefaba161f5440fb5054836df1740f71d8d1af" Jan 27 16:21:15 crc kubenswrapper[4772]: I0127 16:21:15.317738 4772 scope.go:117] "RemoveContainer" containerID="09b6440ec58b2621d6bef188feee41fb9bd95c35cb9a627aa339612e9aef4d84" Jan 27 16:21:15 crc kubenswrapper[4772]: I0127 16:21:15.345142 4772 scope.go:117] "RemoveContainer" containerID="a6ca15ff68296c29c6cc30ddf45249de19f9df9c7b117f0c3db8c30a8ef62894" Jan 27 16:21:15 crc kubenswrapper[4772]: I0127 16:21:15.346256 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8cfm5"] Jan 27 16:21:15 crc kubenswrapper[4772]: I0127 16:21:15.359109 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8cfm5"] Jan 27 16:21:15 crc kubenswrapper[4772]: I0127 16:21:15.365200 4772 scope.go:117] "RemoveContainer" containerID="64a13b2a934cc6a3e677cf6801cefaba161f5440fb5054836df1740f71d8d1af" Jan 27 16:21:15 crc kubenswrapper[4772]: E0127 16:21:15.365840 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64a13b2a934cc6a3e677cf6801cefaba161f5440fb5054836df1740f71d8d1af\": container with ID starting with 64a13b2a934cc6a3e677cf6801cefaba161f5440fb5054836df1740f71d8d1af not found: ID does not exist" containerID="64a13b2a934cc6a3e677cf6801cefaba161f5440fb5054836df1740f71d8d1af" Jan 27 16:21:15 crc kubenswrapper[4772]: I0127 16:21:15.365971 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64a13b2a934cc6a3e677cf6801cefaba161f5440fb5054836df1740f71d8d1af"} err="failed to get container status \"64a13b2a934cc6a3e677cf6801cefaba161f5440fb5054836df1740f71d8d1af\": rpc error: code = NotFound desc = could not find container \"64a13b2a934cc6a3e677cf6801cefaba161f5440fb5054836df1740f71d8d1af\": container with ID starting with 64a13b2a934cc6a3e677cf6801cefaba161f5440fb5054836df1740f71d8d1af not found: ID does not exist" Jan 27 16:21:15 crc kubenswrapper[4772]: I0127 16:21:15.366054 4772 scope.go:117] "RemoveContainer" containerID="09b6440ec58b2621d6bef188feee41fb9bd95c35cb9a627aa339612e9aef4d84" Jan 27 16:21:15 crc kubenswrapper[4772]: E0127 16:21:15.366484 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09b6440ec58b2621d6bef188feee41fb9bd95c35cb9a627aa339612e9aef4d84\": container with ID starting with 09b6440ec58b2621d6bef188feee41fb9bd95c35cb9a627aa339612e9aef4d84 not found: ID does not exist" containerID="09b6440ec58b2621d6bef188feee41fb9bd95c35cb9a627aa339612e9aef4d84" Jan 27 16:21:15 crc kubenswrapper[4772]: I0127 16:21:15.366530 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09b6440ec58b2621d6bef188feee41fb9bd95c35cb9a627aa339612e9aef4d84"} err="failed to get container status \"09b6440ec58b2621d6bef188feee41fb9bd95c35cb9a627aa339612e9aef4d84\": rpc error: code = NotFound desc = could not find container \"09b6440ec58b2621d6bef188feee41fb9bd95c35cb9a627aa339612e9aef4d84\": container with ID starting with 09b6440ec58b2621d6bef188feee41fb9bd95c35cb9a627aa339612e9aef4d84 not found: ID does not exist" Jan 27 16:21:15 crc kubenswrapper[4772]: I0127 16:21:15.366558 4772 scope.go:117] "RemoveContainer" containerID="a6ca15ff68296c29c6cc30ddf45249de19f9df9c7b117f0c3db8c30a8ef62894" Jan 27 16:21:15 crc kubenswrapper[4772]: E0127 16:21:15.366831 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6ca15ff68296c29c6cc30ddf45249de19f9df9c7b117f0c3db8c30a8ef62894\": container with ID starting with a6ca15ff68296c29c6cc30ddf45249de19f9df9c7b117f0c3db8c30a8ef62894 not found: ID does not exist" containerID="a6ca15ff68296c29c6cc30ddf45249de19f9df9c7b117f0c3db8c30a8ef62894" Jan 27 16:21:15 crc kubenswrapper[4772]: I0127 16:21:15.366921 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6ca15ff68296c29c6cc30ddf45249de19f9df9c7b117f0c3db8c30a8ef62894"} err="failed to get container status \"a6ca15ff68296c29c6cc30ddf45249de19f9df9c7b117f0c3db8c30a8ef62894\": rpc error: code = NotFound desc = could not find container \"a6ca15ff68296c29c6cc30ddf45249de19f9df9c7b117f0c3db8c30a8ef62894\": container with ID starting with a6ca15ff68296c29c6cc30ddf45249de19f9df9c7b117f0c3db8c30a8ef62894 not found: ID does not exist" Jan 27 16:21:16 crc kubenswrapper[4772]: I0127 16:21:16.675285 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5854174b-59dc-43b7-aea1-321b0762e938" path="/var/lib/kubelet/pods/5854174b-59dc-43b7-aea1-321b0762e938/volumes" Jan 27 16:21:20 crc kubenswrapper[4772]: I0127 16:21:20.663278 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:21:20 crc kubenswrapper[4772]: E0127 16:21:20.664208 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:21:34 crc kubenswrapper[4772]: I0127 16:21:34.670583 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:21:34 crc kubenswrapper[4772]: E0127 16:21:34.671314 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:21:48 crc kubenswrapper[4772]: I0127 16:21:48.663646 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:21:48 crc kubenswrapper[4772]: E0127 16:21:48.664500 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:22:02 crc kubenswrapper[4772]: I0127 16:22:02.663695 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:22:02 crc kubenswrapper[4772]: E0127 16:22:02.664639 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:22:13 crc kubenswrapper[4772]: I0127 16:22:13.663487 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:22:13 crc kubenswrapper[4772]: E0127 16:22:13.664394 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:22:27 crc kubenswrapper[4772]: I0127 16:22:27.663201 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:22:27 crc kubenswrapper[4772]: E0127 16:22:27.663982 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:22:38 crc kubenswrapper[4772]: I0127 16:22:38.197229 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-pmrs5"] Jan 27 16:22:38 crc kubenswrapper[4772]: I0127 16:22:38.204183 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-pmrs5"] Jan 27 16:22:38 crc kubenswrapper[4772]: I0127 16:22:38.314393 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-zb2ds"] Jan 27 16:22:38 crc kubenswrapper[4772]: E0127 16:22:38.318620 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5854174b-59dc-43b7-aea1-321b0762e938" containerName="extract-utilities" Jan 27 16:22:38 crc kubenswrapper[4772]: I0127 16:22:38.318880 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="5854174b-59dc-43b7-aea1-321b0762e938" containerName="extract-utilities" Jan 27 16:22:38 crc kubenswrapper[4772]: E0127 16:22:38.318960 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5854174b-59dc-43b7-aea1-321b0762e938" containerName="extract-content" Jan 27 16:22:38 crc kubenswrapper[4772]: I0127 16:22:38.319036 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="5854174b-59dc-43b7-aea1-321b0762e938" containerName="extract-content" Jan 27 16:22:38 crc kubenswrapper[4772]: E0127 16:22:38.319100 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5854174b-59dc-43b7-aea1-321b0762e938" containerName="registry-server" Jan 27 16:22:38 crc kubenswrapper[4772]: I0127 16:22:38.319157 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="5854174b-59dc-43b7-aea1-321b0762e938" containerName="registry-server" Jan 27 16:22:38 crc kubenswrapper[4772]: I0127 16:22:38.319379 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="5854174b-59dc-43b7-aea1-321b0762e938" containerName="registry-server" Jan 27 16:22:38 crc kubenswrapper[4772]: I0127 16:22:38.319934 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-zb2ds" Jan 27 16:22:38 crc kubenswrapper[4772]: I0127 16:22:38.322277 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Jan 27 16:22:38 crc kubenswrapper[4772]: I0127 16:22:38.322626 4772 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-r9nmx" Jan 27 16:22:38 crc kubenswrapper[4772]: I0127 16:22:38.322969 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Jan 27 16:22:38 crc kubenswrapper[4772]: I0127 16:22:38.323059 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Jan 27 16:22:38 crc kubenswrapper[4772]: I0127 16:22:38.336373 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-zb2ds"] Jan 27 16:22:38 crc kubenswrapper[4772]: I0127 16:22:38.397827 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s64tn\" (UniqueName: \"kubernetes.io/projected/edf192be-2985-42cc-94da-3e4523dffe67-kube-api-access-s64tn\") pod \"crc-storage-crc-zb2ds\" (UID: \"edf192be-2985-42cc-94da-3e4523dffe67\") " pod="crc-storage/crc-storage-crc-zb2ds" Jan 27 16:22:38 crc kubenswrapper[4772]: I0127 16:22:38.398133 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/edf192be-2985-42cc-94da-3e4523dffe67-crc-storage\") pod \"crc-storage-crc-zb2ds\" (UID: \"edf192be-2985-42cc-94da-3e4523dffe67\") " pod="crc-storage/crc-storage-crc-zb2ds" Jan 27 16:22:38 crc kubenswrapper[4772]: I0127 16:22:38.398304 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/edf192be-2985-42cc-94da-3e4523dffe67-node-mnt\") pod \"crc-storage-crc-zb2ds\" (UID: \"edf192be-2985-42cc-94da-3e4523dffe67\") " pod="crc-storage/crc-storage-crc-zb2ds" Jan 27 16:22:38 crc kubenswrapper[4772]: I0127 16:22:38.499594 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s64tn\" (UniqueName: \"kubernetes.io/projected/edf192be-2985-42cc-94da-3e4523dffe67-kube-api-access-s64tn\") pod \"crc-storage-crc-zb2ds\" (UID: \"edf192be-2985-42cc-94da-3e4523dffe67\") " pod="crc-storage/crc-storage-crc-zb2ds" Jan 27 16:22:38 crc kubenswrapper[4772]: I0127 16:22:38.500007 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/edf192be-2985-42cc-94da-3e4523dffe67-crc-storage\") pod \"crc-storage-crc-zb2ds\" (UID: \"edf192be-2985-42cc-94da-3e4523dffe67\") " pod="crc-storage/crc-storage-crc-zb2ds" Jan 27 16:22:38 crc kubenswrapper[4772]: I0127 16:22:38.500309 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/edf192be-2985-42cc-94da-3e4523dffe67-node-mnt\") pod \"crc-storage-crc-zb2ds\" (UID: \"edf192be-2985-42cc-94da-3e4523dffe67\") " pod="crc-storage/crc-storage-crc-zb2ds" Jan 27 16:22:38 crc kubenswrapper[4772]: I0127 16:22:38.500608 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/edf192be-2985-42cc-94da-3e4523dffe67-node-mnt\") pod \"crc-storage-crc-zb2ds\" (UID: \"edf192be-2985-42cc-94da-3e4523dffe67\") " pod="crc-storage/crc-storage-crc-zb2ds" Jan 27 16:22:38 crc kubenswrapper[4772]: I0127 16:22:38.501516 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/edf192be-2985-42cc-94da-3e4523dffe67-crc-storage\") pod \"crc-storage-crc-zb2ds\" (UID: \"edf192be-2985-42cc-94da-3e4523dffe67\") " pod="crc-storage/crc-storage-crc-zb2ds" Jan 27 16:22:38 crc kubenswrapper[4772]: I0127 16:22:38.523842 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s64tn\" (UniqueName: \"kubernetes.io/projected/edf192be-2985-42cc-94da-3e4523dffe67-kube-api-access-s64tn\") pod \"crc-storage-crc-zb2ds\" (UID: \"edf192be-2985-42cc-94da-3e4523dffe67\") " pod="crc-storage/crc-storage-crc-zb2ds" Jan 27 16:22:38 crc kubenswrapper[4772]: I0127 16:22:38.645011 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-zb2ds" Jan 27 16:22:38 crc kubenswrapper[4772]: I0127 16:22:38.673401 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32a5cafc-0519-4e90-9456-acb182176c41" path="/var/lib/kubelet/pods/32a5cafc-0519-4e90-9456-acb182176c41/volumes" Jan 27 16:22:39 crc kubenswrapper[4772]: I0127 16:22:39.064815 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-zb2ds"] Jan 27 16:22:39 crc kubenswrapper[4772]: I0127 16:22:39.936837 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-zb2ds" event={"ID":"edf192be-2985-42cc-94da-3e4523dffe67","Type":"ContainerStarted","Data":"a88311208e71190cb8b4c34f99be89fcb0a6b7eb71b5dd4a8ea37aa6dbbc506b"} Jan 27 16:22:40 crc kubenswrapper[4772]: I0127 16:22:40.663456 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:22:40 crc kubenswrapper[4772]: E0127 16:22:40.663900 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:22:40 crc kubenswrapper[4772]: I0127 16:22:40.945517 4772 generic.go:334] "Generic (PLEG): container finished" podID="edf192be-2985-42cc-94da-3e4523dffe67" containerID="a62dadf36906064bb1b0580332d53e82d1766f5edc230560463a0bad481701be" exitCode=0 Jan 27 16:22:40 crc kubenswrapper[4772]: I0127 16:22:40.945700 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-zb2ds" event={"ID":"edf192be-2985-42cc-94da-3e4523dffe67","Type":"ContainerDied","Data":"a62dadf36906064bb1b0580332d53e82d1766f5edc230560463a0bad481701be"} Jan 27 16:22:42 crc kubenswrapper[4772]: I0127 16:22:42.352260 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-zb2ds" Jan 27 16:22:42 crc kubenswrapper[4772]: I0127 16:22:42.459064 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s64tn\" (UniqueName: \"kubernetes.io/projected/edf192be-2985-42cc-94da-3e4523dffe67-kube-api-access-s64tn\") pod \"edf192be-2985-42cc-94da-3e4523dffe67\" (UID: \"edf192be-2985-42cc-94da-3e4523dffe67\") " Jan 27 16:22:42 crc kubenswrapper[4772]: I0127 16:22:42.459604 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/edf192be-2985-42cc-94da-3e4523dffe67-node-mnt\") pod \"edf192be-2985-42cc-94da-3e4523dffe67\" (UID: \"edf192be-2985-42cc-94da-3e4523dffe67\") " Jan 27 16:22:42 crc kubenswrapper[4772]: I0127 16:22:42.459686 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/edf192be-2985-42cc-94da-3e4523dffe67-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "edf192be-2985-42cc-94da-3e4523dffe67" (UID: "edf192be-2985-42cc-94da-3e4523dffe67"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 16:22:42 crc kubenswrapper[4772]: I0127 16:22:42.459747 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/edf192be-2985-42cc-94da-3e4523dffe67-crc-storage\") pod \"edf192be-2985-42cc-94da-3e4523dffe67\" (UID: \"edf192be-2985-42cc-94da-3e4523dffe67\") " Jan 27 16:22:42 crc kubenswrapper[4772]: I0127 16:22:42.460047 4772 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/edf192be-2985-42cc-94da-3e4523dffe67-node-mnt\") on node \"crc\" DevicePath \"\"" Jan 27 16:22:42 crc kubenswrapper[4772]: I0127 16:22:42.466325 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edf192be-2985-42cc-94da-3e4523dffe67-kube-api-access-s64tn" (OuterVolumeSpecName: "kube-api-access-s64tn") pod "edf192be-2985-42cc-94da-3e4523dffe67" (UID: "edf192be-2985-42cc-94da-3e4523dffe67"). InnerVolumeSpecName "kube-api-access-s64tn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:22:42 crc kubenswrapper[4772]: I0127 16:22:42.488690 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edf192be-2985-42cc-94da-3e4523dffe67-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "edf192be-2985-42cc-94da-3e4523dffe67" (UID: "edf192be-2985-42cc-94da-3e4523dffe67"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:22:42 crc kubenswrapper[4772]: I0127 16:22:42.561097 4772 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/edf192be-2985-42cc-94da-3e4523dffe67-crc-storage\") on node \"crc\" DevicePath \"\"" Jan 27 16:22:42 crc kubenswrapper[4772]: I0127 16:22:42.561136 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s64tn\" (UniqueName: \"kubernetes.io/projected/edf192be-2985-42cc-94da-3e4523dffe67-kube-api-access-s64tn\") on node \"crc\" DevicePath \"\"" Jan 27 16:22:42 crc kubenswrapper[4772]: I0127 16:22:42.963377 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-zb2ds" event={"ID":"edf192be-2985-42cc-94da-3e4523dffe67","Type":"ContainerDied","Data":"a88311208e71190cb8b4c34f99be89fcb0a6b7eb71b5dd4a8ea37aa6dbbc506b"} Jan 27 16:22:42 crc kubenswrapper[4772]: I0127 16:22:42.963420 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a88311208e71190cb8b4c34f99be89fcb0a6b7eb71b5dd4a8ea37aa6dbbc506b" Jan 27 16:22:42 crc kubenswrapper[4772]: I0127 16:22:42.963535 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-zb2ds" Jan 27 16:22:44 crc kubenswrapper[4772]: I0127 16:22:44.454958 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-zb2ds"] Jan 27 16:22:44 crc kubenswrapper[4772]: I0127 16:22:44.461442 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-zb2ds"] Jan 27 16:22:44 crc kubenswrapper[4772]: I0127 16:22:44.580903 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-bnw5d"] Jan 27 16:22:44 crc kubenswrapper[4772]: E0127 16:22:44.581283 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edf192be-2985-42cc-94da-3e4523dffe67" containerName="storage" Jan 27 16:22:44 crc kubenswrapper[4772]: I0127 16:22:44.581310 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="edf192be-2985-42cc-94da-3e4523dffe67" containerName="storage" Jan 27 16:22:44 crc kubenswrapper[4772]: I0127 16:22:44.581499 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="edf192be-2985-42cc-94da-3e4523dffe67" containerName="storage" Jan 27 16:22:44 crc kubenswrapper[4772]: I0127 16:22:44.582064 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-bnw5d" Jan 27 16:22:44 crc kubenswrapper[4772]: I0127 16:22:44.584224 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Jan 27 16:22:44 crc kubenswrapper[4772]: I0127 16:22:44.584284 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Jan 27 16:22:44 crc kubenswrapper[4772]: I0127 16:22:44.585799 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Jan 27 16:22:44 crc kubenswrapper[4772]: I0127 16:22:44.585881 4772 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-r9nmx" Jan 27 16:22:44 crc kubenswrapper[4772]: I0127 16:22:44.591668 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-bnw5d"] Jan 27 16:22:44 crc kubenswrapper[4772]: I0127 16:22:44.671856 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edf192be-2985-42cc-94da-3e4523dffe67" path="/var/lib/kubelet/pods/edf192be-2985-42cc-94da-3e4523dffe67/volumes" Jan 27 16:22:44 crc kubenswrapper[4772]: I0127 16:22:44.691525 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/764343c1-a6b8-4600-8cb4-a0aa386d4cd2-node-mnt\") pod \"crc-storage-crc-bnw5d\" (UID: \"764343c1-a6b8-4600-8cb4-a0aa386d4cd2\") " pod="crc-storage/crc-storage-crc-bnw5d" Jan 27 16:22:44 crc kubenswrapper[4772]: I0127 16:22:44.691727 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpt6h\" (UniqueName: \"kubernetes.io/projected/764343c1-a6b8-4600-8cb4-a0aa386d4cd2-kube-api-access-wpt6h\") pod \"crc-storage-crc-bnw5d\" (UID: \"764343c1-a6b8-4600-8cb4-a0aa386d4cd2\") " pod="crc-storage/crc-storage-crc-bnw5d" Jan 27 16:22:44 crc kubenswrapper[4772]: I0127 16:22:44.691790 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/764343c1-a6b8-4600-8cb4-a0aa386d4cd2-crc-storage\") pod \"crc-storage-crc-bnw5d\" (UID: \"764343c1-a6b8-4600-8cb4-a0aa386d4cd2\") " pod="crc-storage/crc-storage-crc-bnw5d" Jan 27 16:22:44 crc kubenswrapper[4772]: I0127 16:22:44.792771 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/764343c1-a6b8-4600-8cb4-a0aa386d4cd2-node-mnt\") pod \"crc-storage-crc-bnw5d\" (UID: \"764343c1-a6b8-4600-8cb4-a0aa386d4cd2\") " pod="crc-storage/crc-storage-crc-bnw5d" Jan 27 16:22:44 crc kubenswrapper[4772]: I0127 16:22:44.792908 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpt6h\" (UniqueName: \"kubernetes.io/projected/764343c1-a6b8-4600-8cb4-a0aa386d4cd2-kube-api-access-wpt6h\") pod \"crc-storage-crc-bnw5d\" (UID: \"764343c1-a6b8-4600-8cb4-a0aa386d4cd2\") " pod="crc-storage/crc-storage-crc-bnw5d" Jan 27 16:22:44 crc kubenswrapper[4772]: I0127 16:22:44.792940 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/764343c1-a6b8-4600-8cb4-a0aa386d4cd2-crc-storage\") pod \"crc-storage-crc-bnw5d\" (UID: \"764343c1-a6b8-4600-8cb4-a0aa386d4cd2\") " pod="crc-storage/crc-storage-crc-bnw5d" Jan 27 16:22:44 crc kubenswrapper[4772]: I0127 16:22:44.793115 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/764343c1-a6b8-4600-8cb4-a0aa386d4cd2-node-mnt\") pod \"crc-storage-crc-bnw5d\" (UID: \"764343c1-a6b8-4600-8cb4-a0aa386d4cd2\") " pod="crc-storage/crc-storage-crc-bnw5d" Jan 27 16:22:44 crc kubenswrapper[4772]: I0127 16:22:44.793987 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/764343c1-a6b8-4600-8cb4-a0aa386d4cd2-crc-storage\") pod \"crc-storage-crc-bnw5d\" (UID: \"764343c1-a6b8-4600-8cb4-a0aa386d4cd2\") " pod="crc-storage/crc-storage-crc-bnw5d" Jan 27 16:22:44 crc kubenswrapper[4772]: I0127 16:22:44.813130 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpt6h\" (UniqueName: \"kubernetes.io/projected/764343c1-a6b8-4600-8cb4-a0aa386d4cd2-kube-api-access-wpt6h\") pod \"crc-storage-crc-bnw5d\" (UID: \"764343c1-a6b8-4600-8cb4-a0aa386d4cd2\") " pod="crc-storage/crc-storage-crc-bnw5d" Jan 27 16:22:44 crc kubenswrapper[4772]: I0127 16:22:44.900186 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-bnw5d" Jan 27 16:22:45 crc kubenswrapper[4772]: I0127 16:22:45.512346 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-bnw5d"] Jan 27 16:22:45 crc kubenswrapper[4772]: I0127 16:22:45.987631 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-bnw5d" event={"ID":"764343c1-a6b8-4600-8cb4-a0aa386d4cd2","Type":"ContainerStarted","Data":"5806739c0c411d54a997b2b6fdd6d93080c6d25dfe4b6a90ab2738dfa909d575"} Jan 27 16:22:48 crc kubenswrapper[4772]: I0127 16:22:48.001570 4772 generic.go:334] "Generic (PLEG): container finished" podID="764343c1-a6b8-4600-8cb4-a0aa386d4cd2" containerID="657bc401ac778c0e03331638c4f0d729dea942292c16996b872934a4ca92ef14" exitCode=0 Jan 27 16:22:48 crc kubenswrapper[4772]: I0127 16:22:48.001656 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-bnw5d" event={"ID":"764343c1-a6b8-4600-8cb4-a0aa386d4cd2","Type":"ContainerDied","Data":"657bc401ac778c0e03331638c4f0d729dea942292c16996b872934a4ca92ef14"} Jan 27 16:22:49 crc kubenswrapper[4772]: I0127 16:22:49.360155 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-bnw5d" Jan 27 16:22:49 crc kubenswrapper[4772]: I0127 16:22:49.502209 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/764343c1-a6b8-4600-8cb4-a0aa386d4cd2-crc-storage\") pod \"764343c1-a6b8-4600-8cb4-a0aa386d4cd2\" (UID: \"764343c1-a6b8-4600-8cb4-a0aa386d4cd2\") " Jan 27 16:22:49 crc kubenswrapper[4772]: I0127 16:22:49.502272 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/764343c1-a6b8-4600-8cb4-a0aa386d4cd2-node-mnt\") pod \"764343c1-a6b8-4600-8cb4-a0aa386d4cd2\" (UID: \"764343c1-a6b8-4600-8cb4-a0aa386d4cd2\") " Jan 27 16:22:49 crc kubenswrapper[4772]: I0127 16:22:49.502307 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpt6h\" (UniqueName: \"kubernetes.io/projected/764343c1-a6b8-4600-8cb4-a0aa386d4cd2-kube-api-access-wpt6h\") pod \"764343c1-a6b8-4600-8cb4-a0aa386d4cd2\" (UID: \"764343c1-a6b8-4600-8cb4-a0aa386d4cd2\") " Jan 27 16:22:49 crc kubenswrapper[4772]: I0127 16:22:49.502408 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/764343c1-a6b8-4600-8cb4-a0aa386d4cd2-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "764343c1-a6b8-4600-8cb4-a0aa386d4cd2" (UID: "764343c1-a6b8-4600-8cb4-a0aa386d4cd2"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 16:22:49 crc kubenswrapper[4772]: I0127 16:22:49.502573 4772 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/764343c1-a6b8-4600-8cb4-a0aa386d4cd2-node-mnt\") on node \"crc\" DevicePath \"\"" Jan 27 16:22:49 crc kubenswrapper[4772]: I0127 16:22:49.511456 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/764343c1-a6b8-4600-8cb4-a0aa386d4cd2-kube-api-access-wpt6h" (OuterVolumeSpecName: "kube-api-access-wpt6h") pod "764343c1-a6b8-4600-8cb4-a0aa386d4cd2" (UID: "764343c1-a6b8-4600-8cb4-a0aa386d4cd2"). InnerVolumeSpecName "kube-api-access-wpt6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:22:49 crc kubenswrapper[4772]: I0127 16:22:49.522717 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/764343c1-a6b8-4600-8cb4-a0aa386d4cd2-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "764343c1-a6b8-4600-8cb4-a0aa386d4cd2" (UID: "764343c1-a6b8-4600-8cb4-a0aa386d4cd2"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:22:49 crc kubenswrapper[4772]: I0127 16:22:49.606283 4772 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/764343c1-a6b8-4600-8cb4-a0aa386d4cd2-crc-storage\") on node \"crc\" DevicePath \"\"" Jan 27 16:22:49 crc kubenswrapper[4772]: I0127 16:22:49.606325 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpt6h\" (UniqueName: \"kubernetes.io/projected/764343c1-a6b8-4600-8cb4-a0aa386d4cd2-kube-api-access-wpt6h\") on node \"crc\" DevicePath \"\"" Jan 27 16:22:50 crc kubenswrapper[4772]: I0127 16:22:50.015222 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-bnw5d" event={"ID":"764343c1-a6b8-4600-8cb4-a0aa386d4cd2","Type":"ContainerDied","Data":"5806739c0c411d54a997b2b6fdd6d93080c6d25dfe4b6a90ab2738dfa909d575"} Jan 27 16:22:50 crc kubenswrapper[4772]: I0127 16:22:50.015269 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5806739c0c411d54a997b2b6fdd6d93080c6d25dfe4b6a90ab2738dfa909d575" Jan 27 16:22:50 crc kubenswrapper[4772]: I0127 16:22:50.015394 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-bnw5d" Jan 27 16:22:52 crc kubenswrapper[4772]: I0127 16:22:52.663314 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:22:52 crc kubenswrapper[4772]: E0127 16:22:52.664317 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:23:06 crc kubenswrapper[4772]: I0127 16:23:06.192801 4772 scope.go:117] "RemoveContainer" containerID="f4e8f8b6e9c9139e4588eb373fc8616c60521a1a5d0cbfb79f4f8c9d4dc676b9" Jan 27 16:23:07 crc kubenswrapper[4772]: I0127 16:23:07.662965 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:23:07 crc kubenswrapper[4772]: E0127 16:23:07.663483 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:23:18 crc kubenswrapper[4772]: I0127 16:23:18.663406 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:23:19 crc kubenswrapper[4772]: I0127 16:23:19.243759 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"f3a08a71f69d769f4a6a29d6cef13873c9dceaa6515bc086fcafc82c5f73a041"} Jan 27 16:25:42 crc kubenswrapper[4772]: I0127 16:25:42.058529 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:25:42 crc kubenswrapper[4772]: I0127 16:25:42.059100 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.249918 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-rvc76"] Jan 27 16:25:57 crc kubenswrapper[4772]: E0127 16:25:57.254680 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="764343c1-a6b8-4600-8cb4-a0aa386d4cd2" containerName="storage" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.254702 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="764343c1-a6b8-4600-8cb4-a0aa386d4cd2" containerName="storage" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.254868 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="764343c1-a6b8-4600-8cb4-a0aa386d4cd2" containerName="storage" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.255781 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-rvc76" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.262427 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.262530 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.262696 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.266485 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-bzcld" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.266751 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.273785 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-rvc76"] Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.384820 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-566ms\" (UniqueName: \"kubernetes.io/projected/67aa19ec-be98-4d61-b758-0fd0f7f77f42-kube-api-access-566ms\") pod \"dnsmasq-dns-5d7b5456f5-rvc76\" (UID: \"67aa19ec-be98-4d61-b758-0fd0f7f77f42\") " pod="openstack/dnsmasq-dns-5d7b5456f5-rvc76" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.385060 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67aa19ec-be98-4d61-b758-0fd0f7f77f42-config\") pod \"dnsmasq-dns-5d7b5456f5-rvc76\" (UID: \"67aa19ec-be98-4d61-b758-0fd0f7f77f42\") " pod="openstack/dnsmasq-dns-5d7b5456f5-rvc76" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.385124 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67aa19ec-be98-4d61-b758-0fd0f7f77f42-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-rvc76\" (UID: \"67aa19ec-be98-4d61-b758-0fd0f7f77f42\") " pod="openstack/dnsmasq-dns-5d7b5456f5-rvc76" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.486552 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67aa19ec-be98-4d61-b758-0fd0f7f77f42-config\") pod \"dnsmasq-dns-5d7b5456f5-rvc76\" (UID: \"67aa19ec-be98-4d61-b758-0fd0f7f77f42\") " pod="openstack/dnsmasq-dns-5d7b5456f5-rvc76" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.486635 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67aa19ec-be98-4d61-b758-0fd0f7f77f42-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-rvc76\" (UID: \"67aa19ec-be98-4d61-b758-0fd0f7f77f42\") " pod="openstack/dnsmasq-dns-5d7b5456f5-rvc76" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.486708 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-566ms\" (UniqueName: \"kubernetes.io/projected/67aa19ec-be98-4d61-b758-0fd0f7f77f42-kube-api-access-566ms\") pod \"dnsmasq-dns-5d7b5456f5-rvc76\" (UID: \"67aa19ec-be98-4d61-b758-0fd0f7f77f42\") " pod="openstack/dnsmasq-dns-5d7b5456f5-rvc76" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.487475 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67aa19ec-be98-4d61-b758-0fd0f7f77f42-config\") pod \"dnsmasq-dns-5d7b5456f5-rvc76\" (UID: \"67aa19ec-be98-4d61-b758-0fd0f7f77f42\") " pod="openstack/dnsmasq-dns-5d7b5456f5-rvc76" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.487514 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67aa19ec-be98-4d61-b758-0fd0f7f77f42-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-rvc76\" (UID: \"67aa19ec-be98-4d61-b758-0fd0f7f77f42\") " pod="openstack/dnsmasq-dns-5d7b5456f5-rvc76" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.520935 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-566ms\" (UniqueName: \"kubernetes.io/projected/67aa19ec-be98-4d61-b758-0fd0f7f77f42-kube-api-access-566ms\") pod \"dnsmasq-dns-5d7b5456f5-rvc76\" (UID: \"67aa19ec-be98-4d61-b758-0fd0f7f77f42\") " pod="openstack/dnsmasq-dns-5d7b5456f5-rvc76" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.577453 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-rvc76" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.613629 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-mmzht"] Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.617329 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-mmzht" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.629876 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-mmzht"] Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.790051 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqg6h\" (UniqueName: \"kubernetes.io/projected/bb8b7780-142e-4fd6-967f-a42e112a0b2e-kube-api-access-cqg6h\") pod \"dnsmasq-dns-98ddfc8f-mmzht\" (UID: \"bb8b7780-142e-4fd6-967f-a42e112a0b2e\") " pod="openstack/dnsmasq-dns-98ddfc8f-mmzht" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.790134 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb8b7780-142e-4fd6-967f-a42e112a0b2e-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-mmzht\" (UID: \"bb8b7780-142e-4fd6-967f-a42e112a0b2e\") " pod="openstack/dnsmasq-dns-98ddfc8f-mmzht" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.790227 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb8b7780-142e-4fd6-967f-a42e112a0b2e-config\") pod \"dnsmasq-dns-98ddfc8f-mmzht\" (UID: \"bb8b7780-142e-4fd6-967f-a42e112a0b2e\") " pod="openstack/dnsmasq-dns-98ddfc8f-mmzht" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.891634 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqg6h\" (UniqueName: \"kubernetes.io/projected/bb8b7780-142e-4fd6-967f-a42e112a0b2e-kube-api-access-cqg6h\") pod \"dnsmasq-dns-98ddfc8f-mmzht\" (UID: \"bb8b7780-142e-4fd6-967f-a42e112a0b2e\") " pod="openstack/dnsmasq-dns-98ddfc8f-mmzht" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.891697 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb8b7780-142e-4fd6-967f-a42e112a0b2e-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-mmzht\" (UID: \"bb8b7780-142e-4fd6-967f-a42e112a0b2e\") " pod="openstack/dnsmasq-dns-98ddfc8f-mmzht" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.891723 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb8b7780-142e-4fd6-967f-a42e112a0b2e-config\") pod \"dnsmasq-dns-98ddfc8f-mmzht\" (UID: \"bb8b7780-142e-4fd6-967f-a42e112a0b2e\") " pod="openstack/dnsmasq-dns-98ddfc8f-mmzht" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.892690 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb8b7780-142e-4fd6-967f-a42e112a0b2e-config\") pod \"dnsmasq-dns-98ddfc8f-mmzht\" (UID: \"bb8b7780-142e-4fd6-967f-a42e112a0b2e\") " pod="openstack/dnsmasq-dns-98ddfc8f-mmzht" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.892955 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb8b7780-142e-4fd6-967f-a42e112a0b2e-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-mmzht\" (UID: \"bb8b7780-142e-4fd6-967f-a42e112a0b2e\") " pod="openstack/dnsmasq-dns-98ddfc8f-mmzht" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.908625 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqg6h\" (UniqueName: \"kubernetes.io/projected/bb8b7780-142e-4fd6-967f-a42e112a0b2e-kube-api-access-cqg6h\") pod \"dnsmasq-dns-98ddfc8f-mmzht\" (UID: \"bb8b7780-142e-4fd6-967f-a42e112a0b2e\") " pod="openstack/dnsmasq-dns-98ddfc8f-mmzht" Jan 27 16:25:57 crc kubenswrapper[4772]: I0127 16:25:57.964031 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-mmzht" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.099989 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-rvc76"] Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.411078 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.412692 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.417659 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.417785 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-kn7v7" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.417912 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.418023 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.418057 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.426374 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-mmzht"] Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.440310 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.474148 4772 generic.go:334] "Generic (PLEG): container finished" podID="67aa19ec-be98-4d61-b758-0fd0f7f77f42" containerID="ee06b3482fdb47bc70c8781ce53754e851c0140b4c8f1a7fccfeea4aa97ef68b" exitCode=0 Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.474237 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-rvc76" event={"ID":"67aa19ec-be98-4d61-b758-0fd0f7f77f42","Type":"ContainerDied","Data":"ee06b3482fdb47bc70c8781ce53754e851c0140b4c8f1a7fccfeea4aa97ef68b"} Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.474266 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-rvc76" event={"ID":"67aa19ec-be98-4d61-b758-0fd0f7f77f42","Type":"ContainerStarted","Data":"855fc44f7201e7a6bbe04d32cb4e1bdbca9adcf771a78c25383154b1e38bbac4"} Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.477311 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-mmzht" event={"ID":"bb8b7780-142e-4fd6-967f-a42e112a0b2e","Type":"ContainerStarted","Data":"78ea743dbcac291e88b556c9f353cf72b6797716babc1be9562796c00fc48cc9"} Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.501533 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f38de303-3271-4d8a-b114-4fca1e36c6a3-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.501582 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f38de303-3271-4d8a-b114-4fca1e36c6a3-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.501618 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f38de303-3271-4d8a-b114-4fca1e36c6a3-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.501638 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hj9fs\" (UniqueName: \"kubernetes.io/projected/f38de303-3271-4d8a-b114-4fca1e36c6a3-kube-api-access-hj9fs\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.501660 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f4d1f856-0902-41db-b052-d29ccd2349fb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f4d1f856-0902-41db-b052-d29ccd2349fb\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.501721 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f38de303-3271-4d8a-b114-4fca1e36c6a3-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.501754 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f38de303-3271-4d8a-b114-4fca1e36c6a3-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.501969 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f38de303-3271-4d8a-b114-4fca1e36c6a3-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.502000 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f38de303-3271-4d8a-b114-4fca1e36c6a3-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.605936 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f38de303-3271-4d8a-b114-4fca1e36c6a3-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.606009 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f38de303-3271-4d8a-b114-4fca1e36c6a3-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.606040 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hj9fs\" (UniqueName: \"kubernetes.io/projected/f38de303-3271-4d8a-b114-4fca1e36c6a3-kube-api-access-hj9fs\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.606066 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f38de303-3271-4d8a-b114-4fca1e36c6a3-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.606101 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f4d1f856-0902-41db-b052-d29ccd2349fb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f4d1f856-0902-41db-b052-d29ccd2349fb\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.606178 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f38de303-3271-4d8a-b114-4fca1e36c6a3-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.606231 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f38de303-3271-4d8a-b114-4fca1e36c6a3-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.606267 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f38de303-3271-4d8a-b114-4fca1e36c6a3-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.606309 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f38de303-3271-4d8a-b114-4fca1e36c6a3-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.607302 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f38de303-3271-4d8a-b114-4fca1e36c6a3-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.607312 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f38de303-3271-4d8a-b114-4fca1e36c6a3-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.607590 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f38de303-3271-4d8a-b114-4fca1e36c6a3-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.607758 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f38de303-3271-4d8a-b114-4fca1e36c6a3-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.612021 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f38de303-3271-4d8a-b114-4fca1e36c6a3-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.612498 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f38de303-3271-4d8a-b114-4fca1e36c6a3-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.614358 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f38de303-3271-4d8a-b114-4fca1e36c6a3-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.614620 4772 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.614677 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f4d1f856-0902-41db-b052-d29ccd2349fb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f4d1f856-0902-41db-b052-d29ccd2349fb\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/36a183de2c7abb5d8abee5f0c83592d4960d3f7dbd03e4d4afd32924fe238d72/globalmount\"" pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.634662 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hj9fs\" (UniqueName: \"kubernetes.io/projected/f38de303-3271-4d8a-b114-4fca1e36c6a3-kube-api-access-hj9fs\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.659768 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f4d1f856-0902-41db-b052-d29ccd2349fb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f4d1f856-0902-41db-b052-d29ccd2349fb\") pod \"rabbitmq-server-0\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.729215 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: E0127 16:25:58.753476 4772 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Jan 27 16:25:58 crc kubenswrapper[4772]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/67aa19ec-be98-4d61-b758-0fd0f7f77f42/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Jan 27 16:25:58 crc kubenswrapper[4772]: > podSandboxID="855fc44f7201e7a6bbe04d32cb4e1bdbca9adcf771a78c25383154b1e38bbac4" Jan 27 16:25:58 crc kubenswrapper[4772]: E0127 16:25:58.753830 4772 kuberuntime_manager.go:1274] "Unhandled Error" err=< Jan 27 16:25:58 crc kubenswrapper[4772]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n8chc6h5bh56fh546hb7hc8h67h5bchffh577h697h5b5h5bdh59bhf6hf4h558hb5h578h595h5cchfbh644h59ch7fh654h547h587h5cbh5d5h8fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-566ms,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5d7b5456f5-rvc76_openstack(67aa19ec-be98-4d61-b758-0fd0f7f77f42): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/67aa19ec-be98-4d61-b758-0fd0f7f77f42/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Jan 27 16:25:58 crc kubenswrapper[4772]: > logger="UnhandledError" Jan 27 16:25:58 crc kubenswrapper[4772]: E0127 16:25:58.755251 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/67aa19ec-be98-4d61-b758-0fd0f7f77f42/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-5d7b5456f5-rvc76" podUID="67aa19ec-be98-4d61-b758-0fd0f7f77f42" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.788906 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.790558 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.794743 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.794743 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.794851 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.794954 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-554gz" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.795332 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.809625 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.910269 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.910836 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.910881 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.910928 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qc4j\" (UniqueName: \"kubernetes.io/projected/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-kube-api-access-8qc4j\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.910952 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.911051 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.911091 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.911123 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:58 crc kubenswrapper[4772]: I0127 16:25:58.911151 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.012138 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.012215 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.012262 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qc4j\" (UniqueName: \"kubernetes.io/projected/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-kube-api-access-8qc4j\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.012286 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.012390 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.013528 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.013614 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.013960 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.014098 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.014125 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.014159 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.014441 4772 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.014471 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/74f6c4054e7c826304958ef416594be6d4b6260f90a6b43d068948e9c0dc0fa0/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.014573 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.014897 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.020214 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.022219 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.022801 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.033362 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qc4j\" (UniqueName: \"kubernetes.io/projected/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-kube-api-access-8qc4j\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.045882 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899\") pod \"rabbitmq-cell1-server-0\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.192279 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.198542 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.484167 4772 generic.go:334] "Generic (PLEG): container finished" podID="bb8b7780-142e-4fd6-967f-a42e112a0b2e" containerID="16ae7072b9a91932c05ae95c0de53fbb1486c6e51d07f50b89c0848be4b668b5" exitCode=0 Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.484294 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-mmzht" event={"ID":"bb8b7780-142e-4fd6-967f-a42e112a0b2e","Type":"ContainerDied","Data":"16ae7072b9a91932c05ae95c0de53fbb1486c6e51d07f50b89c0848be4b668b5"} Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.487028 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f38de303-3271-4d8a-b114-4fca1e36c6a3","Type":"ContainerStarted","Data":"10715889097bdfd0b2f4c8a7bc95c59af267c1bbc009f6e01130ee1ccb028c38"} Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.624584 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.734228 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.736014 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.742244 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.742968 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.743100 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.743270 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.743563 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-c4cld" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.745085 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.835778 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eefd7ff4-5222-45cf-aaad-20ebfd50a2ff-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") " pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.835833 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/eefd7ff4-5222-45cf-aaad-20ebfd50a2ff-config-data-generated\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") " pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.835966 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcqcw\" (UniqueName: \"kubernetes.io/projected/eefd7ff4-5222-45cf-aaad-20ebfd50a2ff-kube-api-access-xcqcw\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") " pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.836055 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/eefd7ff4-5222-45cf-aaad-20ebfd50a2ff-config-data-default\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") " pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.836082 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eefd7ff4-5222-45cf-aaad-20ebfd50a2ff-operator-scripts\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") " pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.836234 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/eefd7ff4-5222-45cf-aaad-20ebfd50a2ff-kolla-config\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") " pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.836273 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/eefd7ff4-5222-45cf-aaad-20ebfd50a2ff-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") " pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.836336 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-40b3b8a3-9743-4a00-b998-26333b0bfb3e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-40b3b8a3-9743-4a00-b998-26333b0bfb3e\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") " pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.937071 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/eefd7ff4-5222-45cf-aaad-20ebfd50a2ff-config-data-default\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") " pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.937121 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eefd7ff4-5222-45cf-aaad-20ebfd50a2ff-operator-scripts\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") " pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.937163 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/eefd7ff4-5222-45cf-aaad-20ebfd50a2ff-kolla-config\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") " pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.937190 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/eefd7ff4-5222-45cf-aaad-20ebfd50a2ff-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") " pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.937246 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-40b3b8a3-9743-4a00-b998-26333b0bfb3e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-40b3b8a3-9743-4a00-b998-26333b0bfb3e\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") " pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.937272 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eefd7ff4-5222-45cf-aaad-20ebfd50a2ff-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") " pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.937289 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/eefd7ff4-5222-45cf-aaad-20ebfd50a2ff-config-data-generated\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") " pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.937317 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcqcw\" (UniqueName: \"kubernetes.io/projected/eefd7ff4-5222-45cf-aaad-20ebfd50a2ff-kube-api-access-xcqcw\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") " pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.938382 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/eefd7ff4-5222-45cf-aaad-20ebfd50a2ff-config-data-default\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") " pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.939659 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eefd7ff4-5222-45cf-aaad-20ebfd50a2ff-operator-scripts\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") " pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.941301 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/eefd7ff4-5222-45cf-aaad-20ebfd50a2ff-kolla-config\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") " pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.943471 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/eefd7ff4-5222-45cf-aaad-20ebfd50a2ff-config-data-generated\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") " pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.945529 4772 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.945581 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-40b3b8a3-9743-4a00-b998-26333b0bfb3e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-40b3b8a3-9743-4a00-b998-26333b0bfb3e\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5764f99e1512039e35052d84fb68ae5f0622866614c46458b82d06b647e9a2a6/globalmount\"" pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.945668 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eefd7ff4-5222-45cf-aaad-20ebfd50a2ff-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") " pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.946985 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/eefd7ff4-5222-45cf-aaad-20ebfd50a2ff-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") " pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.955813 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcqcw\" (UniqueName: \"kubernetes.io/projected/eefd7ff4-5222-45cf-aaad-20ebfd50a2ff-kube-api-access-xcqcw\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") " pod="openstack/openstack-galera-0" Jan 27 16:25:59 crc kubenswrapper[4772]: I0127 16:25:59.979132 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-40b3b8a3-9743-4a00-b998-26333b0bfb3e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-40b3b8a3-9743-4a00-b998-26333b0bfb3e\") pod \"openstack-galera-0\" (UID: \"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff\") " pod="openstack/openstack-galera-0" Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.074127 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.075188 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.078869 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.079157 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-gv4lb" Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.085072 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.088493 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.141892 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/36e53353-e817-4d3d-878e-2b34f7c9192f-kolla-config\") pod \"memcached-0\" (UID: \"36e53353-e817-4d3d-878e-2b34f7c9192f\") " pod="openstack/memcached-0" Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.142007 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8pmp\" (UniqueName: \"kubernetes.io/projected/36e53353-e817-4d3d-878e-2b34f7c9192f-kube-api-access-z8pmp\") pod \"memcached-0\" (UID: \"36e53353-e817-4d3d-878e-2b34f7c9192f\") " pod="openstack/memcached-0" Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.142056 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/36e53353-e817-4d3d-878e-2b34f7c9192f-config-data\") pod \"memcached-0\" (UID: \"36e53353-e817-4d3d-878e-2b34f7c9192f\") " pod="openstack/memcached-0" Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.243196 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/36e53353-e817-4d3d-878e-2b34f7c9192f-kolla-config\") pod \"memcached-0\" (UID: \"36e53353-e817-4d3d-878e-2b34f7c9192f\") " pod="openstack/memcached-0" Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.243615 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8pmp\" (UniqueName: \"kubernetes.io/projected/36e53353-e817-4d3d-878e-2b34f7c9192f-kube-api-access-z8pmp\") pod \"memcached-0\" (UID: \"36e53353-e817-4d3d-878e-2b34f7c9192f\") " pod="openstack/memcached-0" Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.243670 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/36e53353-e817-4d3d-878e-2b34f7c9192f-config-data\") pod \"memcached-0\" (UID: \"36e53353-e817-4d3d-878e-2b34f7c9192f\") " pod="openstack/memcached-0" Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.245091 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/36e53353-e817-4d3d-878e-2b34f7c9192f-kolla-config\") pod \"memcached-0\" (UID: \"36e53353-e817-4d3d-878e-2b34f7c9192f\") " pod="openstack/memcached-0" Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.245112 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/36e53353-e817-4d3d-878e-2b34f7c9192f-config-data\") pod \"memcached-0\" (UID: \"36e53353-e817-4d3d-878e-2b34f7c9192f\") " pod="openstack/memcached-0" Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.261145 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8pmp\" (UniqueName: \"kubernetes.io/projected/36e53353-e817-4d3d-878e-2b34f7c9192f-kube-api-access-z8pmp\") pod \"memcached-0\" (UID: \"36e53353-e817-4d3d-878e-2b34f7c9192f\") " pod="openstack/memcached-0" Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.392144 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.508953 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-mmzht" event={"ID":"bb8b7780-142e-4fd6-967f-a42e112a0b2e","Type":"ContainerStarted","Data":"9c160ab266039b95d6f27552251a6a39ff7c08fcb4346eec1f5f3d5a708d9f94"} Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.509013 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-98ddfc8f-mmzht" Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.517170 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b","Type":"ContainerStarted","Data":"b96f230c3d38a86522fef8eccf06bafd1c3858ca3dbe42112ea60613f2b942f3"} Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.519317 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-rvc76" event={"ID":"67aa19ec-be98-4d61-b758-0fd0f7f77f42","Type":"ContainerStarted","Data":"5350f301f4b930945ac0ca2e13712ac15b2d17386ce0426b5517080245adcfb5"} Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.520527 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d7b5456f5-rvc76" Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.528768 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-98ddfc8f-mmzht" podStartSLOduration=3.528745176 podStartE2EDuration="3.528745176s" podCreationTimestamp="2026-01-27 16:25:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:26:00.526572213 +0000 UTC m=+4746.507181321" watchObservedRunningTime="2026-01-27 16:26:00.528745176 +0000 UTC m=+4746.509354274" Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.552687 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d7b5456f5-rvc76" podStartSLOduration=3.552662405 podStartE2EDuration="3.552662405s" podCreationTimestamp="2026-01-27 16:25:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:26:00.547875397 +0000 UTC m=+4746.528484525" watchObservedRunningTime="2026-01-27 16:26:00.552662405 +0000 UTC m=+4746.533271513" Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.627231 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Jan 27 16:26:00 crc kubenswrapper[4772]: W0127 16:26:00.699475 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeefd7ff4_5222_45cf_aaad_20ebfd50a2ff.slice/crio-9bd01a4fe738455348a843ce7b9fb4f1111f8cd4f653833eee18272b680d13f1 WatchSource:0}: Error finding container 9bd01a4fe738455348a843ce7b9fb4f1111f8cd4f653833eee18272b680d13f1: Status 404 returned error can't find the container with id 9bd01a4fe738455348a843ce7b9fb4f1111f8cd4f653833eee18272b680d13f1 Jan 27 16:26:00 crc kubenswrapper[4772]: I0127 16:26:00.822979 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Jan 27 16:26:00 crc kubenswrapper[4772]: W0127 16:26:00.827776 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36e53353_e817_4d3d_878e_2b34f7c9192f.slice/crio-4022283e879f8ad28b8e8063f012de6623cca74757ef7b4e4af54829c24d00b0 WatchSource:0}: Error finding container 4022283e879f8ad28b8e8063f012de6623cca74757ef7b4e4af54829c24d00b0: Status 404 returned error can't find the container with id 4022283e879f8ad28b8e8063f012de6623cca74757ef7b4e4af54829c24d00b0 Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.236313 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.237838 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.242420 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.242651 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.242668 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-gbdqz" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.242810 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.251965 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.258161 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-eb9d0e9c-9416-463e-af03-585faac9acf4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-eb9d0e9c-9416-463e-af03-585faac9acf4\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") " pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.258217 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/be7c27a4-64d2-4581-8271-5aaf74103b04-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") " pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.258249 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/be7c27a4-64d2-4581-8271-5aaf74103b04-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") " pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.258273 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/be7c27a4-64d2-4581-8271-5aaf74103b04-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") " pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.258291 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be7c27a4-64d2-4581-8271-5aaf74103b04-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") " pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.258331 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvjj5\" (UniqueName: \"kubernetes.io/projected/be7c27a4-64d2-4581-8271-5aaf74103b04-kube-api-access-mvjj5\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") " pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.258823 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/be7c27a4-64d2-4581-8271-5aaf74103b04-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") " pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.258880 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be7c27a4-64d2-4581-8271-5aaf74103b04-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") " pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.360497 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be7c27a4-64d2-4581-8271-5aaf74103b04-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") " pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.360549 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvjj5\" (UniqueName: \"kubernetes.io/projected/be7c27a4-64d2-4581-8271-5aaf74103b04-kube-api-access-mvjj5\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") " pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.360612 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/be7c27a4-64d2-4581-8271-5aaf74103b04-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") " pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.360641 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be7c27a4-64d2-4581-8271-5aaf74103b04-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") " pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.360667 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-eb9d0e9c-9416-463e-af03-585faac9acf4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-eb9d0e9c-9416-463e-af03-585faac9acf4\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") " pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.360685 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/be7c27a4-64d2-4581-8271-5aaf74103b04-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") " pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.360708 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/be7c27a4-64d2-4581-8271-5aaf74103b04-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") " pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.360727 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/be7c27a4-64d2-4581-8271-5aaf74103b04-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") " pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.361141 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/be7c27a4-64d2-4581-8271-5aaf74103b04-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") " pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.363209 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/be7c27a4-64d2-4581-8271-5aaf74103b04-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") " pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.363558 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/be7c27a4-64d2-4581-8271-5aaf74103b04-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") " pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.364160 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be7c27a4-64d2-4581-8271-5aaf74103b04-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") " pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.364949 4772 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.365023 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-eb9d0e9c-9416-463e-af03-585faac9acf4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-eb9d0e9c-9416-463e-af03-585faac9acf4\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/3b9ee123aaeb872cc0de9d19f51d95a4c0a289a7f718b9f548cfb57c8ed00262/globalmount\"" pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.367049 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be7c27a4-64d2-4581-8271-5aaf74103b04-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") " pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.369233 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/be7c27a4-64d2-4581-8271-5aaf74103b04-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") " pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.381569 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvjj5\" (UniqueName: \"kubernetes.io/projected/be7c27a4-64d2-4581-8271-5aaf74103b04-kube-api-access-mvjj5\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") " pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.386637 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-eb9d0e9c-9416-463e-af03-585faac9acf4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-eb9d0e9c-9416-463e-af03-585faac9acf4\") pod \"openstack-cell1-galera-0\" (UID: \"be7c27a4-64d2-4581-8271-5aaf74103b04\") " pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.536413 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff","Type":"ContainerStarted","Data":"319d2bf04906b7ea2601932e6eda46f58f2c951a1eef20989d0df442b8498b3f"} Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.536466 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff","Type":"ContainerStarted","Data":"9bd01a4fe738455348a843ce7b9fb4f1111f8cd4f653833eee18272b680d13f1"} Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.539971 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f38de303-3271-4d8a-b114-4fca1e36c6a3","Type":"ContainerStarted","Data":"69ba34cf36586fe5c71196b7621cd2b9ba30a359ba359f88ecfede8d86e1b355"} Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.541596 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b","Type":"ContainerStarted","Data":"480ea6652ba12d615e1f6b8fb144377c490339e8e4bed427eade476083ce2cbb"} Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.543600 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"36e53353-e817-4d3d-878e-2b34f7c9192f","Type":"ContainerStarted","Data":"2aab1cfcdac8f489550d996cadf179ce2496b80131aeb9ca866dfb4aa5cfbdfe"} Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.543634 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"36e53353-e817-4d3d-878e-2b34f7c9192f","Type":"ContainerStarted","Data":"4022283e879f8ad28b8e8063f012de6623cca74757ef7b4e4af54829c24d00b0"} Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.544099 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.552639 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:01 crc kubenswrapper[4772]: I0127 16:26:01.584436 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=1.584418339 podStartE2EDuration="1.584418339s" podCreationTimestamp="2026-01-27 16:26:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:26:01.57646983 +0000 UTC m=+4747.557078938" watchObservedRunningTime="2026-01-27 16:26:01.584418339 +0000 UTC m=+4747.565027437" Jan 27 16:26:02 crc kubenswrapper[4772]: I0127 16:26:02.007357 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 27 16:26:02 crc kubenswrapper[4772]: I0127 16:26:02.551247 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"be7c27a4-64d2-4581-8271-5aaf74103b04","Type":"ContainerStarted","Data":"5edb0a193224020eb6e999d1c89efe76ea6edde4800d8fbee94b0bbbba0c753e"} Jan 27 16:26:02 crc kubenswrapper[4772]: I0127 16:26:02.551733 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"be7c27a4-64d2-4581-8271-5aaf74103b04","Type":"ContainerStarted","Data":"22a6b00ac3e7621a53a0417bda2f532421c802801737fb94d2efef84383cc17a"} Jan 27 16:26:04 crc kubenswrapper[4772]: I0127 16:26:04.568768 4772 generic.go:334] "Generic (PLEG): container finished" podID="eefd7ff4-5222-45cf-aaad-20ebfd50a2ff" containerID="319d2bf04906b7ea2601932e6eda46f58f2c951a1eef20989d0df442b8498b3f" exitCode=0 Jan 27 16:26:04 crc kubenswrapper[4772]: I0127 16:26:04.569116 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff","Type":"ContainerDied","Data":"319d2bf04906b7ea2601932e6eda46f58f2c951a1eef20989d0df442b8498b3f"} Jan 27 16:26:05 crc kubenswrapper[4772]: I0127 16:26:05.393923 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Jan 27 16:26:05 crc kubenswrapper[4772]: I0127 16:26:05.576633 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"eefd7ff4-5222-45cf-aaad-20ebfd50a2ff","Type":"ContainerStarted","Data":"31c0f86da05308b57c6fff62a573d0982988a4b2b9c9952ee68b4f90a379009a"} Jan 27 16:26:05 crc kubenswrapper[4772]: I0127 16:26:05.596329 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=7.596305481 podStartE2EDuration="7.596305481s" podCreationTimestamp="2026-01-27 16:25:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:26:05.596198278 +0000 UTC m=+4751.576807376" watchObservedRunningTime="2026-01-27 16:26:05.596305481 +0000 UTC m=+4751.576914579" Jan 27 16:26:06 crc kubenswrapper[4772]: I0127 16:26:06.226960 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rcww9"] Jan 27 16:26:06 crc kubenswrapper[4772]: I0127 16:26:06.228524 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rcww9" Jan 27 16:26:06 crc kubenswrapper[4772]: I0127 16:26:06.245611 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rcww9"] Jan 27 16:26:06 crc kubenswrapper[4772]: I0127 16:26:06.332239 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5cpj\" (UniqueName: \"kubernetes.io/projected/0f798a7f-3f79-4bad-81ce-927deac7748c-kube-api-access-w5cpj\") pod \"redhat-operators-rcww9\" (UID: \"0f798a7f-3f79-4bad-81ce-927deac7748c\") " pod="openshift-marketplace/redhat-operators-rcww9" Jan 27 16:26:06 crc kubenswrapper[4772]: I0127 16:26:06.332299 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f798a7f-3f79-4bad-81ce-927deac7748c-utilities\") pod \"redhat-operators-rcww9\" (UID: \"0f798a7f-3f79-4bad-81ce-927deac7748c\") " pod="openshift-marketplace/redhat-operators-rcww9" Jan 27 16:26:06 crc kubenswrapper[4772]: I0127 16:26:06.332363 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f798a7f-3f79-4bad-81ce-927deac7748c-catalog-content\") pod \"redhat-operators-rcww9\" (UID: \"0f798a7f-3f79-4bad-81ce-927deac7748c\") " pod="openshift-marketplace/redhat-operators-rcww9" Jan 27 16:26:06 crc kubenswrapper[4772]: I0127 16:26:06.434217 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f798a7f-3f79-4bad-81ce-927deac7748c-utilities\") pod \"redhat-operators-rcww9\" (UID: \"0f798a7f-3f79-4bad-81ce-927deac7748c\") " pod="openshift-marketplace/redhat-operators-rcww9" Jan 27 16:26:06 crc kubenswrapper[4772]: I0127 16:26:06.434647 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f798a7f-3f79-4bad-81ce-927deac7748c-catalog-content\") pod \"redhat-operators-rcww9\" (UID: \"0f798a7f-3f79-4bad-81ce-927deac7748c\") " pod="openshift-marketplace/redhat-operators-rcww9" Jan 27 16:26:06 crc kubenswrapper[4772]: I0127 16:26:06.434751 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5cpj\" (UniqueName: \"kubernetes.io/projected/0f798a7f-3f79-4bad-81ce-927deac7748c-kube-api-access-w5cpj\") pod \"redhat-operators-rcww9\" (UID: \"0f798a7f-3f79-4bad-81ce-927deac7748c\") " pod="openshift-marketplace/redhat-operators-rcww9" Jan 27 16:26:06 crc kubenswrapper[4772]: I0127 16:26:06.434911 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f798a7f-3f79-4bad-81ce-927deac7748c-utilities\") pod \"redhat-operators-rcww9\" (UID: \"0f798a7f-3f79-4bad-81ce-927deac7748c\") " pod="openshift-marketplace/redhat-operators-rcww9" Jan 27 16:26:06 crc kubenswrapper[4772]: I0127 16:26:06.435028 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f798a7f-3f79-4bad-81ce-927deac7748c-catalog-content\") pod \"redhat-operators-rcww9\" (UID: \"0f798a7f-3f79-4bad-81ce-927deac7748c\") " pod="openshift-marketplace/redhat-operators-rcww9" Jan 27 16:26:06 crc kubenswrapper[4772]: I0127 16:26:06.458175 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5cpj\" (UniqueName: \"kubernetes.io/projected/0f798a7f-3f79-4bad-81ce-927deac7748c-kube-api-access-w5cpj\") pod \"redhat-operators-rcww9\" (UID: \"0f798a7f-3f79-4bad-81ce-927deac7748c\") " pod="openshift-marketplace/redhat-operators-rcww9" Jan 27 16:26:06 crc kubenswrapper[4772]: I0127 16:26:06.548201 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rcww9" Jan 27 16:26:06 crc kubenswrapper[4772]: I0127 16:26:06.587413 4772 generic.go:334] "Generic (PLEG): container finished" podID="be7c27a4-64d2-4581-8271-5aaf74103b04" containerID="5edb0a193224020eb6e999d1c89efe76ea6edde4800d8fbee94b0bbbba0c753e" exitCode=0 Jan 27 16:26:06 crc kubenswrapper[4772]: I0127 16:26:06.587466 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"be7c27a4-64d2-4581-8271-5aaf74103b04","Type":"ContainerDied","Data":"5edb0a193224020eb6e999d1c89efe76ea6edde4800d8fbee94b0bbbba0c753e"} Jan 27 16:26:07 crc kubenswrapper[4772]: I0127 16:26:07.011386 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rcww9"] Jan 27 16:26:07 crc kubenswrapper[4772]: W0127 16:26:07.014415 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f798a7f_3f79_4bad_81ce_927deac7748c.slice/crio-968e17dd30db85c3b57c00c0b8a56c178398486e19424f28d377e321bb687c89 WatchSource:0}: Error finding container 968e17dd30db85c3b57c00c0b8a56c178398486e19424f28d377e321bb687c89: Status 404 returned error can't find the container with id 968e17dd30db85c3b57c00c0b8a56c178398486e19424f28d377e321bb687c89 Jan 27 16:26:07 crc kubenswrapper[4772]: I0127 16:26:07.579499 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d7b5456f5-rvc76" Jan 27 16:26:07 crc kubenswrapper[4772]: I0127 16:26:07.597295 4772 generic.go:334] "Generic (PLEG): container finished" podID="0f798a7f-3f79-4bad-81ce-927deac7748c" containerID="1e79f6f181330775c970fd32e9724d3a5685272b702ae330691c90a1a8494c58" exitCode=0 Jan 27 16:26:07 crc kubenswrapper[4772]: I0127 16:26:07.597396 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcww9" event={"ID":"0f798a7f-3f79-4bad-81ce-927deac7748c","Type":"ContainerDied","Data":"1e79f6f181330775c970fd32e9724d3a5685272b702ae330691c90a1a8494c58"} Jan 27 16:26:07 crc kubenswrapper[4772]: I0127 16:26:07.597487 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcww9" event={"ID":"0f798a7f-3f79-4bad-81ce-927deac7748c","Type":"ContainerStarted","Data":"968e17dd30db85c3b57c00c0b8a56c178398486e19424f28d377e321bb687c89"} Jan 27 16:26:07 crc kubenswrapper[4772]: I0127 16:26:07.599532 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 16:26:07 crc kubenswrapper[4772]: I0127 16:26:07.608730 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"be7c27a4-64d2-4581-8271-5aaf74103b04","Type":"ContainerStarted","Data":"e7e64e0f199444d267c2d7492662163ae8edb4f6a71875f0a5607f2ec36a95ba"} Jan 27 16:26:07 crc kubenswrapper[4772]: I0127 16:26:07.676712 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=7.6766865079999995 podStartE2EDuration="7.676686508s" podCreationTimestamp="2026-01-27 16:26:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:26:07.670141059 +0000 UTC m=+4753.650750167" watchObservedRunningTime="2026-01-27 16:26:07.676686508 +0000 UTC m=+4753.657295616" Jan 27 16:26:07 crc kubenswrapper[4772]: I0127 16:26:07.966359 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-98ddfc8f-mmzht" Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.021377 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-rvc76"] Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.021611 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d7b5456f5-rvc76" podUID="67aa19ec-be98-4d61-b758-0fd0f7f77f42" containerName="dnsmasq-dns" containerID="cri-o://5350f301f4b930945ac0ca2e13712ac15b2d17386ce0426b5517080245adcfb5" gracePeriod=10 Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.466658 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-rvc76" Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.569715 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-566ms\" (UniqueName: \"kubernetes.io/projected/67aa19ec-be98-4d61-b758-0fd0f7f77f42-kube-api-access-566ms\") pod \"67aa19ec-be98-4d61-b758-0fd0f7f77f42\" (UID: \"67aa19ec-be98-4d61-b758-0fd0f7f77f42\") " Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.569859 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67aa19ec-be98-4d61-b758-0fd0f7f77f42-config\") pod \"67aa19ec-be98-4d61-b758-0fd0f7f77f42\" (UID: \"67aa19ec-be98-4d61-b758-0fd0f7f77f42\") " Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.569916 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67aa19ec-be98-4d61-b758-0fd0f7f77f42-dns-svc\") pod \"67aa19ec-be98-4d61-b758-0fd0f7f77f42\" (UID: \"67aa19ec-be98-4d61-b758-0fd0f7f77f42\") " Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.587207 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67aa19ec-be98-4d61-b758-0fd0f7f77f42-kube-api-access-566ms" (OuterVolumeSpecName: "kube-api-access-566ms") pod "67aa19ec-be98-4d61-b758-0fd0f7f77f42" (UID: "67aa19ec-be98-4d61-b758-0fd0f7f77f42"). InnerVolumeSpecName "kube-api-access-566ms". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.604228 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67aa19ec-be98-4d61-b758-0fd0f7f77f42-config" (OuterVolumeSpecName: "config") pod "67aa19ec-be98-4d61-b758-0fd0f7f77f42" (UID: "67aa19ec-be98-4d61-b758-0fd0f7f77f42"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.608245 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67aa19ec-be98-4d61-b758-0fd0f7f77f42-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "67aa19ec-be98-4d61-b758-0fd0f7f77f42" (UID: "67aa19ec-be98-4d61-b758-0fd0f7f77f42"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.618334 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcww9" event={"ID":"0f798a7f-3f79-4bad-81ce-927deac7748c","Type":"ContainerStarted","Data":"7fc27289a60fa452c350532a44e942b0d6e7cd80328c4d362e21d2a68688c24e"} Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.620532 4772 generic.go:334] "Generic (PLEG): container finished" podID="67aa19ec-be98-4d61-b758-0fd0f7f77f42" containerID="5350f301f4b930945ac0ca2e13712ac15b2d17386ce0426b5517080245adcfb5" exitCode=0 Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.620591 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-rvc76" event={"ID":"67aa19ec-be98-4d61-b758-0fd0f7f77f42","Type":"ContainerDied","Data":"5350f301f4b930945ac0ca2e13712ac15b2d17386ce0426b5517080245adcfb5"} Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.620624 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-rvc76" event={"ID":"67aa19ec-be98-4d61-b758-0fd0f7f77f42","Type":"ContainerDied","Data":"855fc44f7201e7a6bbe04d32cb4e1bdbca9adcf771a78c25383154b1e38bbac4"} Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.620646 4772 scope.go:117] "RemoveContainer" containerID="5350f301f4b930945ac0ca2e13712ac15b2d17386ce0426b5517080245adcfb5" Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.620802 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-rvc76" Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.675352 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67aa19ec-be98-4d61-b758-0fd0f7f77f42-config\") on node \"crc\" DevicePath \"\"" Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.675386 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67aa19ec-be98-4d61-b758-0fd0f7f77f42-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.675399 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-566ms\" (UniqueName: \"kubernetes.io/projected/67aa19ec-be98-4d61-b758-0fd0f7f77f42-kube-api-access-566ms\") on node \"crc\" DevicePath \"\"" Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.680361 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-rvc76"] Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.681114 4772 scope.go:117] "RemoveContainer" containerID="ee06b3482fdb47bc70c8781ce53754e851c0140b4c8f1a7fccfeea4aa97ef68b" Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.683047 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-rvc76"] Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.697706 4772 scope.go:117] "RemoveContainer" containerID="5350f301f4b930945ac0ca2e13712ac15b2d17386ce0426b5517080245adcfb5" Jan 27 16:26:08 crc kubenswrapper[4772]: E0127 16:26:08.698151 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5350f301f4b930945ac0ca2e13712ac15b2d17386ce0426b5517080245adcfb5\": container with ID starting with 5350f301f4b930945ac0ca2e13712ac15b2d17386ce0426b5517080245adcfb5 not found: ID does not exist" containerID="5350f301f4b930945ac0ca2e13712ac15b2d17386ce0426b5517080245adcfb5" Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.698212 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5350f301f4b930945ac0ca2e13712ac15b2d17386ce0426b5517080245adcfb5"} err="failed to get container status \"5350f301f4b930945ac0ca2e13712ac15b2d17386ce0426b5517080245adcfb5\": rpc error: code = NotFound desc = could not find container \"5350f301f4b930945ac0ca2e13712ac15b2d17386ce0426b5517080245adcfb5\": container with ID starting with 5350f301f4b930945ac0ca2e13712ac15b2d17386ce0426b5517080245adcfb5 not found: ID does not exist" Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.698273 4772 scope.go:117] "RemoveContainer" containerID="ee06b3482fdb47bc70c8781ce53754e851c0140b4c8f1a7fccfeea4aa97ef68b" Jan 27 16:26:08 crc kubenswrapper[4772]: E0127 16:26:08.698640 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee06b3482fdb47bc70c8781ce53754e851c0140b4c8f1a7fccfeea4aa97ef68b\": container with ID starting with ee06b3482fdb47bc70c8781ce53754e851c0140b4c8f1a7fccfeea4aa97ef68b not found: ID does not exist" containerID="ee06b3482fdb47bc70c8781ce53754e851c0140b4c8f1a7fccfeea4aa97ef68b" Jan 27 16:26:08 crc kubenswrapper[4772]: I0127 16:26:08.698670 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee06b3482fdb47bc70c8781ce53754e851c0140b4c8f1a7fccfeea4aa97ef68b"} err="failed to get container status \"ee06b3482fdb47bc70c8781ce53754e851c0140b4c8f1a7fccfeea4aa97ef68b\": rpc error: code = NotFound desc = could not find container \"ee06b3482fdb47bc70c8781ce53754e851c0140b4c8f1a7fccfeea4aa97ef68b\": container with ID starting with ee06b3482fdb47bc70c8781ce53754e851c0140b4c8f1a7fccfeea4aa97ef68b not found: ID does not exist" Jan 27 16:26:09 crc kubenswrapper[4772]: I0127 16:26:09.629594 4772 generic.go:334] "Generic (PLEG): container finished" podID="0f798a7f-3f79-4bad-81ce-927deac7748c" containerID="7fc27289a60fa452c350532a44e942b0d6e7cd80328c4d362e21d2a68688c24e" exitCode=0 Jan 27 16:26:09 crc kubenswrapper[4772]: I0127 16:26:09.629668 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcww9" event={"ID":"0f798a7f-3f79-4bad-81ce-927deac7748c","Type":"ContainerDied","Data":"7fc27289a60fa452c350532a44e942b0d6e7cd80328c4d362e21d2a68688c24e"} Jan 27 16:26:10 crc kubenswrapper[4772]: I0127 16:26:10.085864 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Jan 27 16:26:10 crc kubenswrapper[4772]: I0127 16:26:10.086318 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Jan 27 16:26:10 crc kubenswrapper[4772]: I0127 16:26:10.641838 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcww9" event={"ID":"0f798a7f-3f79-4bad-81ce-927deac7748c","Type":"ContainerStarted","Data":"c53284a092cb30eb41fcbd2150f2e0435fbbe105e96d6d7b176f53ce99e3588d"} Jan 27 16:26:10 crc kubenswrapper[4772]: I0127 16:26:10.666603 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rcww9" podStartSLOduration=2.010051006 podStartE2EDuration="4.666383151s" podCreationTimestamp="2026-01-27 16:26:06 +0000 UTC" firstStartedPulling="2026-01-27 16:26:07.599263447 +0000 UTC m=+4753.579872535" lastFinishedPulling="2026-01-27 16:26:10.255595582 +0000 UTC m=+4756.236204680" observedRunningTime="2026-01-27 16:26:10.65942213 +0000 UTC m=+4756.640031228" watchObservedRunningTime="2026-01-27 16:26:10.666383151 +0000 UTC m=+4756.646992249" Jan 27 16:26:10 crc kubenswrapper[4772]: I0127 16:26:10.673218 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67aa19ec-be98-4d61-b758-0fd0f7f77f42" path="/var/lib/kubelet/pods/67aa19ec-be98-4d61-b758-0fd0f7f77f42/volumes" Jan 27 16:26:11 crc kubenswrapper[4772]: I0127 16:26:11.554097 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:11 crc kubenswrapper[4772]: I0127 16:26:11.554154 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:12 crc kubenswrapper[4772]: I0127 16:26:12.020494 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:12 crc kubenswrapper[4772]: I0127 16:26:12.059053 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:26:12 crc kubenswrapper[4772]: I0127 16:26:12.059129 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:26:12 crc kubenswrapper[4772]: I0127 16:26:12.089565 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Jan 27 16:26:12 crc kubenswrapper[4772]: I0127 16:26:12.369068 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Jan 27 16:26:12 crc kubenswrapper[4772]: I0127 16:26:12.563437 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Jan 27 16:26:16 crc kubenswrapper[4772]: I0127 16:26:16.549437 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rcww9" Jan 27 16:26:16 crc kubenswrapper[4772]: I0127 16:26:16.549800 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rcww9" Jan 27 16:26:16 crc kubenswrapper[4772]: I0127 16:26:16.599446 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rcww9" Jan 27 16:26:16 crc kubenswrapper[4772]: I0127 16:26:16.724333 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rcww9" Jan 27 16:26:16 crc kubenswrapper[4772]: I0127 16:26:16.833443 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rcww9"] Jan 27 16:26:18 crc kubenswrapper[4772]: I0127 16:26:18.692719 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-4pgr2"] Jan 27 16:26:18 crc kubenswrapper[4772]: E0127 16:26:18.693509 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67aa19ec-be98-4d61-b758-0fd0f7f77f42" containerName="dnsmasq-dns" Jan 27 16:26:18 crc kubenswrapper[4772]: I0127 16:26:18.693532 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="67aa19ec-be98-4d61-b758-0fd0f7f77f42" containerName="dnsmasq-dns" Jan 27 16:26:18 crc kubenswrapper[4772]: E0127 16:26:18.693562 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67aa19ec-be98-4d61-b758-0fd0f7f77f42" containerName="init" Jan 27 16:26:18 crc kubenswrapper[4772]: I0127 16:26:18.693574 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="67aa19ec-be98-4d61-b758-0fd0f7f77f42" containerName="init" Jan 27 16:26:18 crc kubenswrapper[4772]: I0127 16:26:18.694059 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="67aa19ec-be98-4d61-b758-0fd0f7f77f42" containerName="dnsmasq-dns" Jan 27 16:26:18 crc kubenswrapper[4772]: I0127 16:26:18.695082 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-4pgr2" Jan 27 16:26:18 crc kubenswrapper[4772]: I0127 16:26:18.695609 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rcww9" podUID="0f798a7f-3f79-4bad-81ce-927deac7748c" containerName="registry-server" containerID="cri-o://c53284a092cb30eb41fcbd2150f2e0435fbbe105e96d6d7b176f53ce99e3588d" gracePeriod=2 Jan 27 16:26:18 crc kubenswrapper[4772]: I0127 16:26:18.702785 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Jan 27 16:26:18 crc kubenswrapper[4772]: I0127 16:26:18.713597 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-4pgr2"] Jan 27 16:26:18 crc kubenswrapper[4772]: I0127 16:26:18.822531 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eeb30f86-cf93-47e2-8dfa-0bec7d656b74-operator-scripts\") pod \"root-account-create-update-4pgr2\" (UID: \"eeb30f86-cf93-47e2-8dfa-0bec7d656b74\") " pod="openstack/root-account-create-update-4pgr2" Jan 27 16:26:18 crc kubenswrapper[4772]: I0127 16:26:18.822629 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2t2c\" (UniqueName: \"kubernetes.io/projected/eeb30f86-cf93-47e2-8dfa-0bec7d656b74-kube-api-access-z2t2c\") pod \"root-account-create-update-4pgr2\" (UID: \"eeb30f86-cf93-47e2-8dfa-0bec7d656b74\") " pod="openstack/root-account-create-update-4pgr2" Jan 27 16:26:18 crc kubenswrapper[4772]: I0127 16:26:18.923994 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eeb30f86-cf93-47e2-8dfa-0bec7d656b74-operator-scripts\") pod \"root-account-create-update-4pgr2\" (UID: \"eeb30f86-cf93-47e2-8dfa-0bec7d656b74\") " pod="openstack/root-account-create-update-4pgr2" Jan 27 16:26:18 crc kubenswrapper[4772]: I0127 16:26:18.924073 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2t2c\" (UniqueName: \"kubernetes.io/projected/eeb30f86-cf93-47e2-8dfa-0bec7d656b74-kube-api-access-z2t2c\") pod \"root-account-create-update-4pgr2\" (UID: \"eeb30f86-cf93-47e2-8dfa-0bec7d656b74\") " pod="openstack/root-account-create-update-4pgr2" Jan 27 16:26:18 crc kubenswrapper[4772]: I0127 16:26:18.924905 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eeb30f86-cf93-47e2-8dfa-0bec7d656b74-operator-scripts\") pod \"root-account-create-update-4pgr2\" (UID: \"eeb30f86-cf93-47e2-8dfa-0bec7d656b74\") " pod="openstack/root-account-create-update-4pgr2" Jan 27 16:26:18 crc kubenswrapper[4772]: I0127 16:26:18.952918 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2t2c\" (UniqueName: \"kubernetes.io/projected/eeb30f86-cf93-47e2-8dfa-0bec7d656b74-kube-api-access-z2t2c\") pod \"root-account-create-update-4pgr2\" (UID: \"eeb30f86-cf93-47e2-8dfa-0bec7d656b74\") " pod="openstack/root-account-create-update-4pgr2" Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.025546 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-4pgr2" Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.118958 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rcww9" Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.228638 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f798a7f-3f79-4bad-81ce-927deac7748c-utilities\") pod \"0f798a7f-3f79-4bad-81ce-927deac7748c\" (UID: \"0f798a7f-3f79-4bad-81ce-927deac7748c\") " Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.228834 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5cpj\" (UniqueName: \"kubernetes.io/projected/0f798a7f-3f79-4bad-81ce-927deac7748c-kube-api-access-w5cpj\") pod \"0f798a7f-3f79-4bad-81ce-927deac7748c\" (UID: \"0f798a7f-3f79-4bad-81ce-927deac7748c\") " Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.228901 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f798a7f-3f79-4bad-81ce-927deac7748c-catalog-content\") pod \"0f798a7f-3f79-4bad-81ce-927deac7748c\" (UID: \"0f798a7f-3f79-4bad-81ce-927deac7748c\") " Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.229771 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f798a7f-3f79-4bad-81ce-927deac7748c-utilities" (OuterVolumeSpecName: "utilities") pod "0f798a7f-3f79-4bad-81ce-927deac7748c" (UID: "0f798a7f-3f79-4bad-81ce-927deac7748c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.236115 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f798a7f-3f79-4bad-81ce-927deac7748c-kube-api-access-w5cpj" (OuterVolumeSpecName: "kube-api-access-w5cpj") pod "0f798a7f-3f79-4bad-81ce-927deac7748c" (UID: "0f798a7f-3f79-4bad-81ce-927deac7748c"). InnerVolumeSpecName "kube-api-access-w5cpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.330368 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5cpj\" (UniqueName: \"kubernetes.io/projected/0f798a7f-3f79-4bad-81ce-927deac7748c-kube-api-access-w5cpj\") on node \"crc\" DevicePath \"\"" Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.330400 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f798a7f-3f79-4bad-81ce-927deac7748c-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.484074 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-4pgr2"] Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.704091 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-4pgr2" event={"ID":"eeb30f86-cf93-47e2-8dfa-0bec7d656b74","Type":"ContainerStarted","Data":"dbf03a502ec31eeb1c32c76f40edb6250c7c04cd55512165e350a627fdf8e1b7"} Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.704554 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-4pgr2" event={"ID":"eeb30f86-cf93-47e2-8dfa-0bec7d656b74","Type":"ContainerStarted","Data":"2ddf6ca655defe81bb7463fb72db8f2faf5a112847354fe4ff6d98a7bbcf6b41"} Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.705902 4772 generic.go:334] "Generic (PLEG): container finished" podID="0f798a7f-3f79-4bad-81ce-927deac7748c" containerID="c53284a092cb30eb41fcbd2150f2e0435fbbe105e96d6d7b176f53ce99e3588d" exitCode=0 Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.705956 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcww9" event={"ID":"0f798a7f-3f79-4bad-81ce-927deac7748c","Type":"ContainerDied","Data":"c53284a092cb30eb41fcbd2150f2e0435fbbe105e96d6d7b176f53ce99e3588d"} Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.705992 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcww9" event={"ID":"0f798a7f-3f79-4bad-81ce-927deac7748c","Type":"ContainerDied","Data":"968e17dd30db85c3b57c00c0b8a56c178398486e19424f28d377e321bb687c89"} Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.706013 4772 scope.go:117] "RemoveContainer" containerID="c53284a092cb30eb41fcbd2150f2e0435fbbe105e96d6d7b176f53ce99e3588d" Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.706054 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rcww9" Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.725141 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-4pgr2" podStartSLOduration=1.725124503 podStartE2EDuration="1.725124503s" podCreationTimestamp="2026-01-27 16:26:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:26:19.720464549 +0000 UTC m=+4765.701073677" watchObservedRunningTime="2026-01-27 16:26:19.725124503 +0000 UTC m=+4765.705733601" Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.728746 4772 scope.go:117] "RemoveContainer" containerID="7fc27289a60fa452c350532a44e942b0d6e7cd80328c4d362e21d2a68688c24e" Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.745550 4772 scope.go:117] "RemoveContainer" containerID="1e79f6f181330775c970fd32e9724d3a5685272b702ae330691c90a1a8494c58" Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.761545 4772 scope.go:117] "RemoveContainer" containerID="c53284a092cb30eb41fcbd2150f2e0435fbbe105e96d6d7b176f53ce99e3588d" Jan 27 16:26:19 crc kubenswrapper[4772]: E0127 16:26:19.761945 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c53284a092cb30eb41fcbd2150f2e0435fbbe105e96d6d7b176f53ce99e3588d\": container with ID starting with c53284a092cb30eb41fcbd2150f2e0435fbbe105e96d6d7b176f53ce99e3588d not found: ID does not exist" containerID="c53284a092cb30eb41fcbd2150f2e0435fbbe105e96d6d7b176f53ce99e3588d" Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.761989 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c53284a092cb30eb41fcbd2150f2e0435fbbe105e96d6d7b176f53ce99e3588d"} err="failed to get container status \"c53284a092cb30eb41fcbd2150f2e0435fbbe105e96d6d7b176f53ce99e3588d\": rpc error: code = NotFound desc = could not find container \"c53284a092cb30eb41fcbd2150f2e0435fbbe105e96d6d7b176f53ce99e3588d\": container with ID starting with c53284a092cb30eb41fcbd2150f2e0435fbbe105e96d6d7b176f53ce99e3588d not found: ID does not exist" Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.762017 4772 scope.go:117] "RemoveContainer" containerID="7fc27289a60fa452c350532a44e942b0d6e7cd80328c4d362e21d2a68688c24e" Jan 27 16:26:19 crc kubenswrapper[4772]: E0127 16:26:19.762332 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fc27289a60fa452c350532a44e942b0d6e7cd80328c4d362e21d2a68688c24e\": container with ID starting with 7fc27289a60fa452c350532a44e942b0d6e7cd80328c4d362e21d2a68688c24e not found: ID does not exist" containerID="7fc27289a60fa452c350532a44e942b0d6e7cd80328c4d362e21d2a68688c24e" Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.762353 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fc27289a60fa452c350532a44e942b0d6e7cd80328c4d362e21d2a68688c24e"} err="failed to get container status \"7fc27289a60fa452c350532a44e942b0d6e7cd80328c4d362e21d2a68688c24e\": rpc error: code = NotFound desc = could not find container \"7fc27289a60fa452c350532a44e942b0d6e7cd80328c4d362e21d2a68688c24e\": container with ID starting with 7fc27289a60fa452c350532a44e942b0d6e7cd80328c4d362e21d2a68688c24e not found: ID does not exist" Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.762366 4772 scope.go:117] "RemoveContainer" containerID="1e79f6f181330775c970fd32e9724d3a5685272b702ae330691c90a1a8494c58" Jan 27 16:26:19 crc kubenswrapper[4772]: E0127 16:26:19.762644 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e79f6f181330775c970fd32e9724d3a5685272b702ae330691c90a1a8494c58\": container with ID starting with 1e79f6f181330775c970fd32e9724d3a5685272b702ae330691c90a1a8494c58 not found: ID does not exist" containerID="1e79f6f181330775c970fd32e9724d3a5685272b702ae330691c90a1a8494c58" Jan 27 16:26:19 crc kubenswrapper[4772]: I0127 16:26:19.762682 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e79f6f181330775c970fd32e9724d3a5685272b702ae330691c90a1a8494c58"} err="failed to get container status \"1e79f6f181330775c970fd32e9724d3a5685272b702ae330691c90a1a8494c58\": rpc error: code = NotFound desc = could not find container \"1e79f6f181330775c970fd32e9724d3a5685272b702ae330691c90a1a8494c58\": container with ID starting with 1e79f6f181330775c970fd32e9724d3a5685272b702ae330691c90a1a8494c58 not found: ID does not exist" Jan 27 16:26:20 crc kubenswrapper[4772]: I0127 16:26:20.794062 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f798a7f-3f79-4bad-81ce-927deac7748c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0f798a7f-3f79-4bad-81ce-927deac7748c" (UID: "0f798a7f-3f79-4bad-81ce-927deac7748c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:26:20 crc kubenswrapper[4772]: I0127 16:26:20.851987 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f798a7f-3f79-4bad-81ce-927deac7748c-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 16:26:20 crc kubenswrapper[4772]: I0127 16:26:20.942775 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rcww9"] Jan 27 16:26:20 crc kubenswrapper[4772]: I0127 16:26:20.949339 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rcww9"] Jan 27 16:26:21 crc kubenswrapper[4772]: I0127 16:26:21.724572 4772 generic.go:334] "Generic (PLEG): container finished" podID="eeb30f86-cf93-47e2-8dfa-0bec7d656b74" containerID="dbf03a502ec31eeb1c32c76f40edb6250c7c04cd55512165e350a627fdf8e1b7" exitCode=0 Jan 27 16:26:21 crc kubenswrapper[4772]: I0127 16:26:21.724689 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-4pgr2" event={"ID":"eeb30f86-cf93-47e2-8dfa-0bec7d656b74","Type":"ContainerDied","Data":"dbf03a502ec31eeb1c32c76f40edb6250c7c04cd55512165e350a627fdf8e1b7"} Jan 27 16:26:22 crc kubenswrapper[4772]: I0127 16:26:22.679547 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f798a7f-3f79-4bad-81ce-927deac7748c" path="/var/lib/kubelet/pods/0f798a7f-3f79-4bad-81ce-927deac7748c/volumes" Jan 27 16:26:23 crc kubenswrapper[4772]: I0127 16:26:23.157462 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-4pgr2" Jan 27 16:26:23 crc kubenswrapper[4772]: I0127 16:26:23.290664 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eeb30f86-cf93-47e2-8dfa-0bec7d656b74-operator-scripts\") pod \"eeb30f86-cf93-47e2-8dfa-0bec7d656b74\" (UID: \"eeb30f86-cf93-47e2-8dfa-0bec7d656b74\") " Jan 27 16:26:23 crc kubenswrapper[4772]: I0127 16:26:23.290708 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2t2c\" (UniqueName: \"kubernetes.io/projected/eeb30f86-cf93-47e2-8dfa-0bec7d656b74-kube-api-access-z2t2c\") pod \"eeb30f86-cf93-47e2-8dfa-0bec7d656b74\" (UID: \"eeb30f86-cf93-47e2-8dfa-0bec7d656b74\") " Jan 27 16:26:23 crc kubenswrapper[4772]: I0127 16:26:23.291768 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eeb30f86-cf93-47e2-8dfa-0bec7d656b74-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "eeb30f86-cf93-47e2-8dfa-0bec7d656b74" (UID: "eeb30f86-cf93-47e2-8dfa-0bec7d656b74"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:26:23 crc kubenswrapper[4772]: I0127 16:26:23.297572 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eeb30f86-cf93-47e2-8dfa-0bec7d656b74-kube-api-access-z2t2c" (OuterVolumeSpecName: "kube-api-access-z2t2c") pod "eeb30f86-cf93-47e2-8dfa-0bec7d656b74" (UID: "eeb30f86-cf93-47e2-8dfa-0bec7d656b74"). InnerVolumeSpecName "kube-api-access-z2t2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:26:23 crc kubenswrapper[4772]: I0127 16:26:23.392963 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eeb30f86-cf93-47e2-8dfa-0bec7d656b74-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:26:23 crc kubenswrapper[4772]: I0127 16:26:23.393060 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2t2c\" (UniqueName: \"kubernetes.io/projected/eeb30f86-cf93-47e2-8dfa-0bec7d656b74-kube-api-access-z2t2c\") on node \"crc\" DevicePath \"\"" Jan 27 16:26:23 crc kubenswrapper[4772]: I0127 16:26:23.759691 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-4pgr2" event={"ID":"eeb30f86-cf93-47e2-8dfa-0bec7d656b74","Type":"ContainerDied","Data":"2ddf6ca655defe81bb7463fb72db8f2faf5a112847354fe4ff6d98a7bbcf6b41"} Jan 27 16:26:23 crc kubenswrapper[4772]: I0127 16:26:23.759728 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ddf6ca655defe81bb7463fb72db8f2faf5a112847354fe4ff6d98a7bbcf6b41" Jan 27 16:26:23 crc kubenswrapper[4772]: I0127 16:26:23.759751 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-4pgr2" Jan 27 16:26:25 crc kubenswrapper[4772]: I0127 16:26:25.236228 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-4pgr2"] Jan 27 16:26:25 crc kubenswrapper[4772]: I0127 16:26:25.241998 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-4pgr2"] Jan 27 16:26:26 crc kubenswrapper[4772]: I0127 16:26:26.675265 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eeb30f86-cf93-47e2-8dfa-0bec7d656b74" path="/var/lib/kubelet/pods/eeb30f86-cf93-47e2-8dfa-0bec7d656b74/volumes" Jan 27 16:26:28 crc kubenswrapper[4772]: I0127 16:26:28.727875 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-269mm"] Jan 27 16:26:28 crc kubenswrapper[4772]: E0127 16:26:28.728591 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eeb30f86-cf93-47e2-8dfa-0bec7d656b74" containerName="mariadb-account-create-update" Jan 27 16:26:28 crc kubenswrapper[4772]: I0127 16:26:28.728608 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="eeb30f86-cf93-47e2-8dfa-0bec7d656b74" containerName="mariadb-account-create-update" Jan 27 16:26:28 crc kubenswrapper[4772]: E0127 16:26:28.728627 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f798a7f-3f79-4bad-81ce-927deac7748c" containerName="extract-content" Jan 27 16:26:28 crc kubenswrapper[4772]: I0127 16:26:28.728635 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f798a7f-3f79-4bad-81ce-927deac7748c" containerName="extract-content" Jan 27 16:26:28 crc kubenswrapper[4772]: E0127 16:26:28.728653 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f798a7f-3f79-4bad-81ce-927deac7748c" containerName="registry-server" Jan 27 16:26:28 crc kubenswrapper[4772]: I0127 16:26:28.728661 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f798a7f-3f79-4bad-81ce-927deac7748c" containerName="registry-server" Jan 27 16:26:28 crc kubenswrapper[4772]: E0127 16:26:28.728673 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f798a7f-3f79-4bad-81ce-927deac7748c" containerName="extract-utilities" Jan 27 16:26:28 crc kubenswrapper[4772]: I0127 16:26:28.728681 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f798a7f-3f79-4bad-81ce-927deac7748c" containerName="extract-utilities" Jan 27 16:26:28 crc kubenswrapper[4772]: I0127 16:26:28.728851 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f798a7f-3f79-4bad-81ce-927deac7748c" containerName="registry-server" Jan 27 16:26:28 crc kubenswrapper[4772]: I0127 16:26:28.728907 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="eeb30f86-cf93-47e2-8dfa-0bec7d656b74" containerName="mariadb-account-create-update" Jan 27 16:26:28 crc kubenswrapper[4772]: I0127 16:26:28.729511 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-269mm" Jan 27 16:26:28 crc kubenswrapper[4772]: I0127 16:26:28.731852 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Jan 27 16:26:28 crc kubenswrapper[4772]: I0127 16:26:28.734361 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-269mm"] Jan 27 16:26:28 crc kubenswrapper[4772]: I0127 16:26:28.874634 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de29a5f9-d1c3-413b-8e86-77d1e9f10602-operator-scripts\") pod \"root-account-create-update-269mm\" (UID: \"de29a5f9-d1c3-413b-8e86-77d1e9f10602\") " pod="openstack/root-account-create-update-269mm" Jan 27 16:26:28 crc kubenswrapper[4772]: I0127 16:26:28.874736 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtcgq\" (UniqueName: \"kubernetes.io/projected/de29a5f9-d1c3-413b-8e86-77d1e9f10602-kube-api-access-qtcgq\") pod \"root-account-create-update-269mm\" (UID: \"de29a5f9-d1c3-413b-8e86-77d1e9f10602\") " pod="openstack/root-account-create-update-269mm" Jan 27 16:26:28 crc kubenswrapper[4772]: I0127 16:26:28.975858 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de29a5f9-d1c3-413b-8e86-77d1e9f10602-operator-scripts\") pod \"root-account-create-update-269mm\" (UID: \"de29a5f9-d1c3-413b-8e86-77d1e9f10602\") " pod="openstack/root-account-create-update-269mm" Jan 27 16:26:28 crc kubenswrapper[4772]: I0127 16:26:28.976016 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtcgq\" (UniqueName: \"kubernetes.io/projected/de29a5f9-d1c3-413b-8e86-77d1e9f10602-kube-api-access-qtcgq\") pod \"root-account-create-update-269mm\" (UID: \"de29a5f9-d1c3-413b-8e86-77d1e9f10602\") " pod="openstack/root-account-create-update-269mm" Jan 27 16:26:28 crc kubenswrapper[4772]: I0127 16:26:28.976836 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de29a5f9-d1c3-413b-8e86-77d1e9f10602-operator-scripts\") pod \"root-account-create-update-269mm\" (UID: \"de29a5f9-d1c3-413b-8e86-77d1e9f10602\") " pod="openstack/root-account-create-update-269mm" Jan 27 16:26:28 crc kubenswrapper[4772]: I0127 16:26:28.994090 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtcgq\" (UniqueName: \"kubernetes.io/projected/de29a5f9-d1c3-413b-8e86-77d1e9f10602-kube-api-access-qtcgq\") pod \"root-account-create-update-269mm\" (UID: \"de29a5f9-d1c3-413b-8e86-77d1e9f10602\") " pod="openstack/root-account-create-update-269mm" Jan 27 16:26:29 crc kubenswrapper[4772]: I0127 16:26:29.057462 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-269mm" Jan 27 16:26:29 crc kubenswrapper[4772]: I0127 16:26:29.452678 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-269mm"] Jan 27 16:26:29 crc kubenswrapper[4772]: W0127 16:26:29.460338 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde29a5f9_d1c3_413b_8e86_77d1e9f10602.slice/crio-f42c60f151ca72489735671502b041bd6301ac421d9bf54bebf0fa9452852292 WatchSource:0}: Error finding container f42c60f151ca72489735671502b041bd6301ac421d9bf54bebf0fa9452852292: Status 404 returned error can't find the container with id f42c60f151ca72489735671502b041bd6301ac421d9bf54bebf0fa9452852292 Jan 27 16:26:29 crc kubenswrapper[4772]: I0127 16:26:29.808281 4772 generic.go:334] "Generic (PLEG): container finished" podID="de29a5f9-d1c3-413b-8e86-77d1e9f10602" containerID="0bdb0516e5ad0fcd11824f097428db46c1768aa20c8111ee97d7a876d3f00649" exitCode=0 Jan 27 16:26:29 crc kubenswrapper[4772]: I0127 16:26:29.808318 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-269mm" event={"ID":"de29a5f9-d1c3-413b-8e86-77d1e9f10602","Type":"ContainerDied","Data":"0bdb0516e5ad0fcd11824f097428db46c1768aa20c8111ee97d7a876d3f00649"} Jan 27 16:26:29 crc kubenswrapper[4772]: I0127 16:26:29.808342 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-269mm" event={"ID":"de29a5f9-d1c3-413b-8e86-77d1e9f10602","Type":"ContainerStarted","Data":"f42c60f151ca72489735671502b041bd6301ac421d9bf54bebf0fa9452852292"} Jan 27 16:26:31 crc kubenswrapper[4772]: I0127 16:26:31.154379 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-269mm" Jan 27 16:26:31 crc kubenswrapper[4772]: I0127 16:26:31.209891 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de29a5f9-d1c3-413b-8e86-77d1e9f10602-operator-scripts\") pod \"de29a5f9-d1c3-413b-8e86-77d1e9f10602\" (UID: \"de29a5f9-d1c3-413b-8e86-77d1e9f10602\") " Jan 27 16:26:31 crc kubenswrapper[4772]: I0127 16:26:31.210003 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtcgq\" (UniqueName: \"kubernetes.io/projected/de29a5f9-d1c3-413b-8e86-77d1e9f10602-kube-api-access-qtcgq\") pod \"de29a5f9-d1c3-413b-8e86-77d1e9f10602\" (UID: \"de29a5f9-d1c3-413b-8e86-77d1e9f10602\") " Jan 27 16:26:31 crc kubenswrapper[4772]: I0127 16:26:31.210981 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de29a5f9-d1c3-413b-8e86-77d1e9f10602-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "de29a5f9-d1c3-413b-8e86-77d1e9f10602" (UID: "de29a5f9-d1c3-413b-8e86-77d1e9f10602"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:26:31 crc kubenswrapper[4772]: I0127 16:26:31.217371 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de29a5f9-d1c3-413b-8e86-77d1e9f10602-kube-api-access-qtcgq" (OuterVolumeSpecName: "kube-api-access-qtcgq") pod "de29a5f9-d1c3-413b-8e86-77d1e9f10602" (UID: "de29a5f9-d1c3-413b-8e86-77d1e9f10602"). InnerVolumeSpecName "kube-api-access-qtcgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:26:31 crc kubenswrapper[4772]: I0127 16:26:31.312298 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtcgq\" (UniqueName: \"kubernetes.io/projected/de29a5f9-d1c3-413b-8e86-77d1e9f10602-kube-api-access-qtcgq\") on node \"crc\" DevicePath \"\"" Jan 27 16:26:31 crc kubenswrapper[4772]: I0127 16:26:31.312335 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de29a5f9-d1c3-413b-8e86-77d1e9f10602-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:26:31 crc kubenswrapper[4772]: I0127 16:26:31.823811 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-269mm" event={"ID":"de29a5f9-d1c3-413b-8e86-77d1e9f10602","Type":"ContainerDied","Data":"f42c60f151ca72489735671502b041bd6301ac421d9bf54bebf0fa9452852292"} Jan 27 16:26:31 crc kubenswrapper[4772]: I0127 16:26:31.823854 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f42c60f151ca72489735671502b041bd6301ac421d9bf54bebf0fa9452852292" Jan 27 16:26:31 crc kubenswrapper[4772]: I0127 16:26:31.823890 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-269mm" Jan 27 16:26:32 crc kubenswrapper[4772]: I0127 16:26:32.832340 4772 generic.go:334] "Generic (PLEG): container finished" podID="f38de303-3271-4d8a-b114-4fca1e36c6a3" containerID="69ba34cf36586fe5c71196b7621cd2b9ba30a359ba359f88ecfede8d86e1b355" exitCode=0 Jan 27 16:26:32 crc kubenswrapper[4772]: I0127 16:26:32.832390 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f38de303-3271-4d8a-b114-4fca1e36c6a3","Type":"ContainerDied","Data":"69ba34cf36586fe5c71196b7621cd2b9ba30a359ba359f88ecfede8d86e1b355"} Jan 27 16:26:33 crc kubenswrapper[4772]: I0127 16:26:33.840011 4772 generic.go:334] "Generic (PLEG): container finished" podID="18fc2383-1b4e-43c5-b6cb-8aa40600cf7b" containerID="480ea6652ba12d615e1f6b8fb144377c490339e8e4bed427eade476083ce2cbb" exitCode=0 Jan 27 16:26:33 crc kubenswrapper[4772]: I0127 16:26:33.840095 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b","Type":"ContainerDied","Data":"480ea6652ba12d615e1f6b8fb144377c490339e8e4bed427eade476083ce2cbb"} Jan 27 16:26:33 crc kubenswrapper[4772]: I0127 16:26:33.843883 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f38de303-3271-4d8a-b114-4fca1e36c6a3","Type":"ContainerStarted","Data":"9b0305c5b89b648ee767009d14d68e95e3436ecb734d1416e97e7782000adad3"} Jan 27 16:26:33 crc kubenswrapper[4772]: I0127 16:26:33.844231 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Jan 27 16:26:33 crc kubenswrapper[4772]: I0127 16:26:33.893491 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.893469513 podStartE2EDuration="36.893469513s" podCreationTimestamp="2026-01-27 16:25:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:26:33.88641775 +0000 UTC m=+4779.867026858" watchObservedRunningTime="2026-01-27 16:26:33.893469513 +0000 UTC m=+4779.874078621" Jan 27 16:26:34 crc kubenswrapper[4772]: I0127 16:26:34.852067 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b","Type":"ContainerStarted","Data":"5308f6aebce7e906d821045f50b393cfaba5255e05a4b74e645d3d680c330291"} Jan 27 16:26:34 crc kubenswrapper[4772]: I0127 16:26:34.853094 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:26:34 crc kubenswrapper[4772]: I0127 16:26:34.882665 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.882644051 podStartE2EDuration="37.882644051s" podCreationTimestamp="2026-01-27 16:25:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:26:34.871049407 +0000 UTC m=+4780.851658515" watchObservedRunningTime="2026-01-27 16:26:34.882644051 +0000 UTC m=+4780.863253159" Jan 27 16:26:35 crc kubenswrapper[4772]: I0127 16:26:35.253111 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-269mm"] Jan 27 16:26:35 crc kubenswrapper[4772]: I0127 16:26:35.268039 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-269mm"] Jan 27 16:26:36 crc kubenswrapper[4772]: I0127 16:26:36.674416 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de29a5f9-d1c3-413b-8e86-77d1e9f10602" path="/var/lib/kubelet/pods/de29a5f9-d1c3-413b-8e86-77d1e9f10602/volumes" Jan 27 16:26:40 crc kubenswrapper[4772]: I0127 16:26:40.276221 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-bxrqf"] Jan 27 16:26:40 crc kubenswrapper[4772]: E0127 16:26:40.276883 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de29a5f9-d1c3-413b-8e86-77d1e9f10602" containerName="mariadb-account-create-update" Jan 27 16:26:40 crc kubenswrapper[4772]: I0127 16:26:40.276898 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="de29a5f9-d1c3-413b-8e86-77d1e9f10602" containerName="mariadb-account-create-update" Jan 27 16:26:40 crc kubenswrapper[4772]: I0127 16:26:40.277116 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="de29a5f9-d1c3-413b-8e86-77d1e9f10602" containerName="mariadb-account-create-update" Jan 27 16:26:40 crc kubenswrapper[4772]: I0127 16:26:40.277748 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-bxrqf" Jan 27 16:26:40 crc kubenswrapper[4772]: I0127 16:26:40.281005 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Jan 27 16:26:40 crc kubenswrapper[4772]: I0127 16:26:40.284288 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-bxrqf"] Jan 27 16:26:40 crc kubenswrapper[4772]: I0127 16:26:40.366819 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cb558533-27c2-4249-9beb-e01d5b918c58-operator-scripts\") pod \"root-account-create-update-bxrqf\" (UID: \"cb558533-27c2-4249-9beb-e01d5b918c58\") " pod="openstack/root-account-create-update-bxrqf" Jan 27 16:26:40 crc kubenswrapper[4772]: I0127 16:26:40.366892 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk5s5\" (UniqueName: \"kubernetes.io/projected/cb558533-27c2-4249-9beb-e01d5b918c58-kube-api-access-jk5s5\") pod \"root-account-create-update-bxrqf\" (UID: \"cb558533-27c2-4249-9beb-e01d5b918c58\") " pod="openstack/root-account-create-update-bxrqf" Jan 27 16:26:40 crc kubenswrapper[4772]: I0127 16:26:40.468107 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cb558533-27c2-4249-9beb-e01d5b918c58-operator-scripts\") pod \"root-account-create-update-bxrqf\" (UID: \"cb558533-27c2-4249-9beb-e01d5b918c58\") " pod="openstack/root-account-create-update-bxrqf" Jan 27 16:26:40 crc kubenswrapper[4772]: I0127 16:26:40.468197 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk5s5\" (UniqueName: \"kubernetes.io/projected/cb558533-27c2-4249-9beb-e01d5b918c58-kube-api-access-jk5s5\") pod \"root-account-create-update-bxrqf\" (UID: \"cb558533-27c2-4249-9beb-e01d5b918c58\") " pod="openstack/root-account-create-update-bxrqf" Jan 27 16:26:40 crc kubenswrapper[4772]: I0127 16:26:40.469387 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cb558533-27c2-4249-9beb-e01d5b918c58-operator-scripts\") pod \"root-account-create-update-bxrqf\" (UID: \"cb558533-27c2-4249-9beb-e01d5b918c58\") " pod="openstack/root-account-create-update-bxrqf" Jan 27 16:26:40 crc kubenswrapper[4772]: I0127 16:26:40.490237 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk5s5\" (UniqueName: \"kubernetes.io/projected/cb558533-27c2-4249-9beb-e01d5b918c58-kube-api-access-jk5s5\") pod \"root-account-create-update-bxrqf\" (UID: \"cb558533-27c2-4249-9beb-e01d5b918c58\") " pod="openstack/root-account-create-update-bxrqf" Jan 27 16:26:40 crc kubenswrapper[4772]: I0127 16:26:40.595391 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-bxrqf" Jan 27 16:26:40 crc kubenswrapper[4772]: I0127 16:26:40.836534 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-bxrqf"] Jan 27 16:26:40 crc kubenswrapper[4772]: I0127 16:26:40.899323 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-bxrqf" event={"ID":"cb558533-27c2-4249-9beb-e01d5b918c58","Type":"ContainerStarted","Data":"e9040036f87513970452b16f9fd192af689be570efd1157cc6b81ecb51d01f63"} Jan 27 16:26:41 crc kubenswrapper[4772]: I0127 16:26:41.910100 4772 generic.go:334] "Generic (PLEG): container finished" podID="cb558533-27c2-4249-9beb-e01d5b918c58" containerID="65777ccce3cb931b879ebb264390f1a957ffbebf2e9446690c38c79d1e3ddb7c" exitCode=0 Jan 27 16:26:41 crc kubenswrapper[4772]: I0127 16:26:41.910459 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-bxrqf" event={"ID":"cb558533-27c2-4249-9beb-e01d5b918c58","Type":"ContainerDied","Data":"65777ccce3cb931b879ebb264390f1a957ffbebf2e9446690c38c79d1e3ddb7c"} Jan 27 16:26:42 crc kubenswrapper[4772]: I0127 16:26:42.058031 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:26:42 crc kubenswrapper[4772]: I0127 16:26:42.058092 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:26:42 crc kubenswrapper[4772]: I0127 16:26:42.058141 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 16:26:42 crc kubenswrapper[4772]: I0127 16:26:42.058814 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f3a08a71f69d769f4a6a29d6cef13873c9dceaa6515bc086fcafc82c5f73a041"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 16:26:42 crc kubenswrapper[4772]: I0127 16:26:42.058873 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://f3a08a71f69d769f4a6a29d6cef13873c9dceaa6515bc086fcafc82c5f73a041" gracePeriod=600 Jan 27 16:26:42 crc kubenswrapper[4772]: I0127 16:26:42.921184 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="f3a08a71f69d769f4a6a29d6cef13873c9dceaa6515bc086fcafc82c5f73a041" exitCode=0 Jan 27 16:26:42 crc kubenswrapper[4772]: I0127 16:26:42.921282 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"f3a08a71f69d769f4a6a29d6cef13873c9dceaa6515bc086fcafc82c5f73a041"} Jan 27 16:26:42 crc kubenswrapper[4772]: I0127 16:26:42.921517 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b"} Jan 27 16:26:42 crc kubenswrapper[4772]: I0127 16:26:42.921548 4772 scope.go:117] "RemoveContainer" containerID="8604202f7fe20b38dd6ccc7e97fcf384e30e6ff4cf589a28a42b70c4dab8470d" Jan 27 16:26:43 crc kubenswrapper[4772]: I0127 16:26:43.170338 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-bxrqf" Jan 27 16:26:43 crc kubenswrapper[4772]: I0127 16:26:43.221237 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jk5s5\" (UniqueName: \"kubernetes.io/projected/cb558533-27c2-4249-9beb-e01d5b918c58-kube-api-access-jk5s5\") pod \"cb558533-27c2-4249-9beb-e01d5b918c58\" (UID: \"cb558533-27c2-4249-9beb-e01d5b918c58\") " Jan 27 16:26:43 crc kubenswrapper[4772]: I0127 16:26:43.221406 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cb558533-27c2-4249-9beb-e01d5b918c58-operator-scripts\") pod \"cb558533-27c2-4249-9beb-e01d5b918c58\" (UID: \"cb558533-27c2-4249-9beb-e01d5b918c58\") " Jan 27 16:26:43 crc kubenswrapper[4772]: I0127 16:26:43.222136 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb558533-27c2-4249-9beb-e01d5b918c58-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cb558533-27c2-4249-9beb-e01d5b918c58" (UID: "cb558533-27c2-4249-9beb-e01d5b918c58"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:26:43 crc kubenswrapper[4772]: I0127 16:26:43.227370 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb558533-27c2-4249-9beb-e01d5b918c58-kube-api-access-jk5s5" (OuterVolumeSpecName: "kube-api-access-jk5s5") pod "cb558533-27c2-4249-9beb-e01d5b918c58" (UID: "cb558533-27c2-4249-9beb-e01d5b918c58"). InnerVolumeSpecName "kube-api-access-jk5s5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:26:43 crc kubenswrapper[4772]: I0127 16:26:43.322877 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jk5s5\" (UniqueName: \"kubernetes.io/projected/cb558533-27c2-4249-9beb-e01d5b918c58-kube-api-access-jk5s5\") on node \"crc\" DevicePath \"\"" Jan 27 16:26:43 crc kubenswrapper[4772]: I0127 16:26:43.322913 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cb558533-27c2-4249-9beb-e01d5b918c58-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:26:43 crc kubenswrapper[4772]: I0127 16:26:43.929249 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-bxrqf" event={"ID":"cb558533-27c2-4249-9beb-e01d5b918c58","Type":"ContainerDied","Data":"e9040036f87513970452b16f9fd192af689be570efd1157cc6b81ecb51d01f63"} Jan 27 16:26:43 crc kubenswrapper[4772]: I0127 16:26:43.929268 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-bxrqf" Jan 27 16:26:43 crc kubenswrapper[4772]: I0127 16:26:43.929290 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9040036f87513970452b16f9fd192af689be570efd1157cc6b81ecb51d01f63" Jan 27 16:26:48 crc kubenswrapper[4772]: I0127 16:26:48.732358 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Jan 27 16:26:49 crc kubenswrapper[4772]: I0127 16:26:49.195526 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:26:54 crc kubenswrapper[4772]: I0127 16:26:54.789041 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-ndkll"] Jan 27 16:26:54 crc kubenswrapper[4772]: E0127 16:26:54.789830 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb558533-27c2-4249-9beb-e01d5b918c58" containerName="mariadb-account-create-update" Jan 27 16:26:54 crc kubenswrapper[4772]: I0127 16:26:54.789841 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb558533-27c2-4249-9beb-e01d5b918c58" containerName="mariadb-account-create-update" Jan 27 16:26:54 crc kubenswrapper[4772]: I0127 16:26:54.789983 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb558533-27c2-4249-9beb-e01d5b918c58" containerName="mariadb-account-create-update" Jan 27 16:26:54 crc kubenswrapper[4772]: I0127 16:26:54.790786 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-ndkll" Jan 27 16:26:54 crc kubenswrapper[4772]: I0127 16:26:54.798063 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-ndkll"] Jan 27 16:26:54 crc kubenswrapper[4772]: I0127 16:26:54.889775 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m72l\" (UniqueName: \"kubernetes.io/projected/51a5db5c-8de5-441d-a8e9-7c07acc7df31-kube-api-access-7m72l\") pod \"dnsmasq-dns-5b7946d7b9-ndkll\" (UID: \"51a5db5c-8de5-441d-a8e9-7c07acc7df31\") " pod="openstack/dnsmasq-dns-5b7946d7b9-ndkll" Jan 27 16:26:54 crc kubenswrapper[4772]: I0127 16:26:54.889818 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a5db5c-8de5-441d-a8e9-7c07acc7df31-config\") pod \"dnsmasq-dns-5b7946d7b9-ndkll\" (UID: \"51a5db5c-8de5-441d-a8e9-7c07acc7df31\") " pod="openstack/dnsmasq-dns-5b7946d7b9-ndkll" Jan 27 16:26:54 crc kubenswrapper[4772]: I0127 16:26:54.889851 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a5db5c-8de5-441d-a8e9-7c07acc7df31-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-ndkll\" (UID: \"51a5db5c-8de5-441d-a8e9-7c07acc7df31\") " pod="openstack/dnsmasq-dns-5b7946d7b9-ndkll" Jan 27 16:26:54 crc kubenswrapper[4772]: I0127 16:26:54.991362 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m72l\" (UniqueName: \"kubernetes.io/projected/51a5db5c-8de5-441d-a8e9-7c07acc7df31-kube-api-access-7m72l\") pod \"dnsmasq-dns-5b7946d7b9-ndkll\" (UID: \"51a5db5c-8de5-441d-a8e9-7c07acc7df31\") " pod="openstack/dnsmasq-dns-5b7946d7b9-ndkll" Jan 27 16:26:54 crc kubenswrapper[4772]: I0127 16:26:54.991405 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a5db5c-8de5-441d-a8e9-7c07acc7df31-config\") pod \"dnsmasq-dns-5b7946d7b9-ndkll\" (UID: \"51a5db5c-8de5-441d-a8e9-7c07acc7df31\") " pod="openstack/dnsmasq-dns-5b7946d7b9-ndkll" Jan 27 16:26:54 crc kubenswrapper[4772]: I0127 16:26:54.991438 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a5db5c-8de5-441d-a8e9-7c07acc7df31-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-ndkll\" (UID: \"51a5db5c-8de5-441d-a8e9-7c07acc7df31\") " pod="openstack/dnsmasq-dns-5b7946d7b9-ndkll" Jan 27 16:26:54 crc kubenswrapper[4772]: I0127 16:26:54.992318 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a5db5c-8de5-441d-a8e9-7c07acc7df31-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-ndkll\" (UID: \"51a5db5c-8de5-441d-a8e9-7c07acc7df31\") " pod="openstack/dnsmasq-dns-5b7946d7b9-ndkll" Jan 27 16:26:54 crc kubenswrapper[4772]: I0127 16:26:54.992369 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a5db5c-8de5-441d-a8e9-7c07acc7df31-config\") pod \"dnsmasq-dns-5b7946d7b9-ndkll\" (UID: \"51a5db5c-8de5-441d-a8e9-7c07acc7df31\") " pod="openstack/dnsmasq-dns-5b7946d7b9-ndkll" Jan 27 16:26:55 crc kubenswrapper[4772]: I0127 16:26:55.013651 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m72l\" (UniqueName: \"kubernetes.io/projected/51a5db5c-8de5-441d-a8e9-7c07acc7df31-kube-api-access-7m72l\") pod \"dnsmasq-dns-5b7946d7b9-ndkll\" (UID: \"51a5db5c-8de5-441d-a8e9-7c07acc7df31\") " pod="openstack/dnsmasq-dns-5b7946d7b9-ndkll" Jan 27 16:26:55 crc kubenswrapper[4772]: I0127 16:26:55.109085 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-ndkll" Jan 27 16:26:55 crc kubenswrapper[4772]: I0127 16:26:55.509300 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 27 16:26:55 crc kubenswrapper[4772]: I0127 16:26:55.555383 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-ndkll"] Jan 27 16:26:55 crc kubenswrapper[4772]: W0127 16:26:55.556839 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51a5db5c_8de5_441d_a8e9_7c07acc7df31.slice/crio-cd506cf5289047aba524d3325375f9767877e5e9af047a505f68883c25d7ad72 WatchSource:0}: Error finding container cd506cf5289047aba524d3325375f9767877e5e9af047a505f68883c25d7ad72: Status 404 returned error can't find the container with id cd506cf5289047aba524d3325375f9767877e5e9af047a505f68883c25d7ad72 Jan 27 16:26:56 crc kubenswrapper[4772]: I0127 16:26:56.034871 4772 generic.go:334] "Generic (PLEG): container finished" podID="51a5db5c-8de5-441d-a8e9-7c07acc7df31" containerID="2b48ddd2c1b4ddcb3e7d2673318ce52a0b89618a27b83de04520bd8160030f43" exitCode=0 Jan 27 16:26:56 crc kubenswrapper[4772]: I0127 16:26:56.034989 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-ndkll" event={"ID":"51a5db5c-8de5-441d-a8e9-7c07acc7df31","Type":"ContainerDied","Data":"2b48ddd2c1b4ddcb3e7d2673318ce52a0b89618a27b83de04520bd8160030f43"} Jan 27 16:26:56 crc kubenswrapper[4772]: I0127 16:26:56.035157 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-ndkll" event={"ID":"51a5db5c-8de5-441d-a8e9-7c07acc7df31","Type":"ContainerStarted","Data":"cd506cf5289047aba524d3325375f9767877e5e9af047a505f68883c25d7ad72"} Jan 27 16:26:56 crc kubenswrapper[4772]: I0127 16:26:56.122583 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 27 16:26:57 crc kubenswrapper[4772]: I0127 16:26:57.044656 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-ndkll" event={"ID":"51a5db5c-8de5-441d-a8e9-7c07acc7df31","Type":"ContainerStarted","Data":"a355488297e660433fbaef5e701f58797760e94b791d017a807af73346bc3756"} Jan 27 16:26:57 crc kubenswrapper[4772]: I0127 16:26:57.045992 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b7946d7b9-ndkll" Jan 27 16:26:57 crc kubenswrapper[4772]: I0127 16:26:57.069378 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b7946d7b9-ndkll" podStartSLOduration=3.069359651 podStartE2EDuration="3.069359651s" podCreationTimestamp="2026-01-27 16:26:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:26:57.069098184 +0000 UTC m=+4803.049707312" watchObservedRunningTime="2026-01-27 16:26:57.069359651 +0000 UTC m=+4803.049968749" Jan 27 16:26:57 crc kubenswrapper[4772]: I0127 16:26:57.501250 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="f38de303-3271-4d8a-b114-4fca1e36c6a3" containerName="rabbitmq" containerID="cri-o://9b0305c5b89b648ee767009d14d68e95e3436ecb734d1416e97e7782000adad3" gracePeriod=604799 Jan 27 16:26:58 crc kubenswrapper[4772]: I0127 16:26:58.001246 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="18fc2383-1b4e-43c5-b6cb-8aa40600cf7b" containerName="rabbitmq" containerID="cri-o://5308f6aebce7e906d821045f50b393cfaba5255e05a4b74e645d3d680c330291" gracePeriod=604799 Jan 27 16:26:58 crc kubenswrapper[4772]: I0127 16:26:58.730794 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="f38de303-3271-4d8a-b114-4fca1e36c6a3" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.235:5672: connect: connection refused" Jan 27 16:26:59 crc kubenswrapper[4772]: I0127 16:26:59.193464 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="18fc2383-1b4e-43c5-b6cb-8aa40600cf7b" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.236:5672: connect: connection refused" Jan 27 16:26:59 crc kubenswrapper[4772]: I0127 16:26:59.715104 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jjgpk"] Jan 27 16:26:59 crc kubenswrapper[4772]: I0127 16:26:59.720785 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jjgpk" Jan 27 16:26:59 crc kubenswrapper[4772]: I0127 16:26:59.739923 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jjgpk"] Jan 27 16:26:59 crc kubenswrapper[4772]: I0127 16:26:59.763310 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2025168-9cc0-417c-8d3b-1d336447a3ff-catalog-content\") pod \"community-operators-jjgpk\" (UID: \"b2025168-9cc0-417c-8d3b-1d336447a3ff\") " pod="openshift-marketplace/community-operators-jjgpk" Jan 27 16:26:59 crc kubenswrapper[4772]: I0127 16:26:59.763407 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2025168-9cc0-417c-8d3b-1d336447a3ff-utilities\") pod \"community-operators-jjgpk\" (UID: \"b2025168-9cc0-417c-8d3b-1d336447a3ff\") " pod="openshift-marketplace/community-operators-jjgpk" Jan 27 16:26:59 crc kubenswrapper[4772]: I0127 16:26:59.763524 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpgqq\" (UniqueName: \"kubernetes.io/projected/b2025168-9cc0-417c-8d3b-1d336447a3ff-kube-api-access-gpgqq\") pod \"community-operators-jjgpk\" (UID: \"b2025168-9cc0-417c-8d3b-1d336447a3ff\") " pod="openshift-marketplace/community-operators-jjgpk" Jan 27 16:26:59 crc kubenswrapper[4772]: I0127 16:26:59.864448 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpgqq\" (UniqueName: \"kubernetes.io/projected/b2025168-9cc0-417c-8d3b-1d336447a3ff-kube-api-access-gpgqq\") pod \"community-operators-jjgpk\" (UID: \"b2025168-9cc0-417c-8d3b-1d336447a3ff\") " pod="openshift-marketplace/community-operators-jjgpk" Jan 27 16:26:59 crc kubenswrapper[4772]: I0127 16:26:59.864519 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2025168-9cc0-417c-8d3b-1d336447a3ff-catalog-content\") pod \"community-operators-jjgpk\" (UID: \"b2025168-9cc0-417c-8d3b-1d336447a3ff\") " pod="openshift-marketplace/community-operators-jjgpk" Jan 27 16:26:59 crc kubenswrapper[4772]: I0127 16:26:59.864547 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2025168-9cc0-417c-8d3b-1d336447a3ff-utilities\") pod \"community-operators-jjgpk\" (UID: \"b2025168-9cc0-417c-8d3b-1d336447a3ff\") " pod="openshift-marketplace/community-operators-jjgpk" Jan 27 16:26:59 crc kubenswrapper[4772]: I0127 16:26:59.864983 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2025168-9cc0-417c-8d3b-1d336447a3ff-utilities\") pod \"community-operators-jjgpk\" (UID: \"b2025168-9cc0-417c-8d3b-1d336447a3ff\") " pod="openshift-marketplace/community-operators-jjgpk" Jan 27 16:26:59 crc kubenswrapper[4772]: I0127 16:26:59.865280 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2025168-9cc0-417c-8d3b-1d336447a3ff-catalog-content\") pod \"community-operators-jjgpk\" (UID: \"b2025168-9cc0-417c-8d3b-1d336447a3ff\") " pod="openshift-marketplace/community-operators-jjgpk" Jan 27 16:26:59 crc kubenswrapper[4772]: I0127 16:26:59.886099 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpgqq\" (UniqueName: \"kubernetes.io/projected/b2025168-9cc0-417c-8d3b-1d336447a3ff-kube-api-access-gpgqq\") pod \"community-operators-jjgpk\" (UID: \"b2025168-9cc0-417c-8d3b-1d336447a3ff\") " pod="openshift-marketplace/community-operators-jjgpk" Jan 27 16:27:00 crc kubenswrapper[4772]: I0127 16:27:00.045153 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jjgpk" Jan 27 16:27:00 crc kubenswrapper[4772]: I0127 16:27:00.542813 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jjgpk"] Jan 27 16:27:01 crc kubenswrapper[4772]: I0127 16:27:01.083864 4772 generic.go:334] "Generic (PLEG): container finished" podID="b2025168-9cc0-417c-8d3b-1d336447a3ff" containerID="9a45c145ee6112b35c7f85262dd11f2bab997eedc4836c666662e77f41944aa0" exitCode=0 Jan 27 16:27:01 crc kubenswrapper[4772]: I0127 16:27:01.083916 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jjgpk" event={"ID":"b2025168-9cc0-417c-8d3b-1d336447a3ff","Type":"ContainerDied","Data":"9a45c145ee6112b35c7f85262dd11f2bab997eedc4836c666662e77f41944aa0"} Jan 27 16:27:01 crc kubenswrapper[4772]: I0127 16:27:01.083956 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jjgpk" event={"ID":"b2025168-9cc0-417c-8d3b-1d336447a3ff","Type":"ContainerStarted","Data":"aaf303dc4f995cb6bff56fd0049dc32fd5acf74cce10552ba6fbab1b994a80c7"} Jan 27 16:27:02 crc kubenswrapper[4772]: I0127 16:27:02.093546 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jjgpk" event={"ID":"b2025168-9cc0-417c-8d3b-1d336447a3ff","Type":"ContainerStarted","Data":"020319dacd513bbd957b8817f14f5c128e22f4c1f10468a190b399e9005621c7"} Jan 27 16:27:03 crc kubenswrapper[4772]: I0127 16:27:03.100214 4772 generic.go:334] "Generic (PLEG): container finished" podID="b2025168-9cc0-417c-8d3b-1d336447a3ff" containerID="020319dacd513bbd957b8817f14f5c128e22f4c1f10468a190b399e9005621c7" exitCode=0 Jan 27 16:27:03 crc kubenswrapper[4772]: I0127 16:27:03.100267 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jjgpk" event={"ID":"b2025168-9cc0-417c-8d3b-1d336447a3ff","Type":"ContainerDied","Data":"020319dacd513bbd957b8817f14f5c128e22f4c1f10468a190b399e9005621c7"} Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.047796 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.113184 4772 generic.go:334] "Generic (PLEG): container finished" podID="f38de303-3271-4d8a-b114-4fca1e36c6a3" containerID="9b0305c5b89b648ee767009d14d68e95e3436ecb734d1416e97e7782000adad3" exitCode=0 Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.113270 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f38de303-3271-4d8a-b114-4fca1e36c6a3","Type":"ContainerDied","Data":"9b0305c5b89b648ee767009d14d68e95e3436ecb734d1416e97e7782000adad3"} Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.113313 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f38de303-3271-4d8a-b114-4fca1e36c6a3","Type":"ContainerDied","Data":"10715889097bdfd0b2f4c8a7bc95c59af267c1bbc009f6e01130ee1ccb028c38"} Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.113311 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.113371 4772 scope.go:117] "RemoveContainer" containerID="9b0305c5b89b648ee767009d14d68e95e3436ecb734d1416e97e7782000adad3" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.118292 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jjgpk" event={"ID":"b2025168-9cc0-417c-8d3b-1d336447a3ff","Type":"ContainerStarted","Data":"647af5829c88ac17630d560d584a1467c6061b43b60545e27fdaa4193667e8cf"} Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.126546 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f38de303-3271-4d8a-b114-4fca1e36c6a3-plugins-conf\") pod \"f38de303-3271-4d8a-b114-4fca1e36c6a3\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.126612 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f38de303-3271-4d8a-b114-4fca1e36c6a3-rabbitmq-plugins\") pod \"f38de303-3271-4d8a-b114-4fca1e36c6a3\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.126802 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f4d1f856-0902-41db-b052-d29ccd2349fb\") pod \"f38de303-3271-4d8a-b114-4fca1e36c6a3\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.126924 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f38de303-3271-4d8a-b114-4fca1e36c6a3-erlang-cookie-secret\") pod \"f38de303-3271-4d8a-b114-4fca1e36c6a3\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.126959 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f38de303-3271-4d8a-b114-4fca1e36c6a3-rabbitmq-erlang-cookie\") pod \"f38de303-3271-4d8a-b114-4fca1e36c6a3\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.126989 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hj9fs\" (UniqueName: \"kubernetes.io/projected/f38de303-3271-4d8a-b114-4fca1e36c6a3-kube-api-access-hj9fs\") pod \"f38de303-3271-4d8a-b114-4fca1e36c6a3\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.127032 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f38de303-3271-4d8a-b114-4fca1e36c6a3-server-conf\") pod \"f38de303-3271-4d8a-b114-4fca1e36c6a3\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.127086 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f38de303-3271-4d8a-b114-4fca1e36c6a3-pod-info\") pod \"f38de303-3271-4d8a-b114-4fca1e36c6a3\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.127108 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f38de303-3271-4d8a-b114-4fca1e36c6a3-rabbitmq-confd\") pod \"f38de303-3271-4d8a-b114-4fca1e36c6a3\" (UID: \"f38de303-3271-4d8a-b114-4fca1e36c6a3\") " Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.128208 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f38de303-3271-4d8a-b114-4fca1e36c6a3-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "f38de303-3271-4d8a-b114-4fca1e36c6a3" (UID: "f38de303-3271-4d8a-b114-4fca1e36c6a3"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.128666 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f38de303-3271-4d8a-b114-4fca1e36c6a3-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "f38de303-3271-4d8a-b114-4fca1e36c6a3" (UID: "f38de303-3271-4d8a-b114-4fca1e36c6a3"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.129339 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f38de303-3271-4d8a-b114-4fca1e36c6a3-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "f38de303-3271-4d8a-b114-4fca1e36c6a3" (UID: "f38de303-3271-4d8a-b114-4fca1e36c6a3"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.134814 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f38de303-3271-4d8a-b114-4fca1e36c6a3-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "f38de303-3271-4d8a-b114-4fca1e36c6a3" (UID: "f38de303-3271-4d8a-b114-4fca1e36c6a3"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.134890 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f38de303-3271-4d8a-b114-4fca1e36c6a3-kube-api-access-hj9fs" (OuterVolumeSpecName: "kube-api-access-hj9fs") pod "f38de303-3271-4d8a-b114-4fca1e36c6a3" (UID: "f38de303-3271-4d8a-b114-4fca1e36c6a3"). InnerVolumeSpecName "kube-api-access-hj9fs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.144225 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f4d1f856-0902-41db-b052-d29ccd2349fb" (OuterVolumeSpecName: "persistence") pod "f38de303-3271-4d8a-b114-4fca1e36c6a3" (UID: "f38de303-3271-4d8a-b114-4fca1e36c6a3"). InnerVolumeSpecName "pvc-f4d1f856-0902-41db-b052-d29ccd2349fb". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.149169 4772 scope.go:117] "RemoveContainer" containerID="69ba34cf36586fe5c71196b7621cd2b9ba30a359ba359f88ecfede8d86e1b355" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.149206 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/f38de303-3271-4d8a-b114-4fca1e36c6a3-pod-info" (OuterVolumeSpecName: "pod-info") pod "f38de303-3271-4d8a-b114-4fca1e36c6a3" (UID: "f38de303-3271-4d8a-b114-4fca1e36c6a3"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.152819 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jjgpk" podStartSLOduration=2.749817112 podStartE2EDuration="5.152803326s" podCreationTimestamp="2026-01-27 16:26:59 +0000 UTC" firstStartedPulling="2026-01-27 16:27:01.087527085 +0000 UTC m=+4807.068136183" lastFinishedPulling="2026-01-27 16:27:03.490513289 +0000 UTC m=+4809.471122397" observedRunningTime="2026-01-27 16:27:04.144440225 +0000 UTC m=+4810.125049333" watchObservedRunningTime="2026-01-27 16:27:04.152803326 +0000 UTC m=+4810.133412424" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.172822 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f38de303-3271-4d8a-b114-4fca1e36c6a3-server-conf" (OuterVolumeSpecName: "server-conf") pod "f38de303-3271-4d8a-b114-4fca1e36c6a3" (UID: "f38de303-3271-4d8a-b114-4fca1e36c6a3"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.228443 4772 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f38de303-3271-4d8a-b114-4fca1e36c6a3-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.228476 4772 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f38de303-3271-4d8a-b114-4fca1e36c6a3-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.228488 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hj9fs\" (UniqueName: \"kubernetes.io/projected/f38de303-3271-4d8a-b114-4fca1e36c6a3-kube-api-access-hj9fs\") on node \"crc\" DevicePath \"\"" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.228497 4772 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f38de303-3271-4d8a-b114-4fca1e36c6a3-server-conf\") on node \"crc\" DevicePath \"\"" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.228505 4772 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f38de303-3271-4d8a-b114-4fca1e36c6a3-pod-info\") on node \"crc\" DevicePath \"\"" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.228514 4772 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f38de303-3271-4d8a-b114-4fca1e36c6a3-plugins-conf\") on node \"crc\" DevicePath \"\"" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.228522 4772 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f38de303-3271-4d8a-b114-4fca1e36c6a3-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.228551 4772 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-f4d1f856-0902-41db-b052-d29ccd2349fb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f4d1f856-0902-41db-b052-d29ccd2349fb\") on node \"crc\" " Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.231429 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f38de303-3271-4d8a-b114-4fca1e36c6a3-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "f38de303-3271-4d8a-b114-4fca1e36c6a3" (UID: "f38de303-3271-4d8a-b114-4fca1e36c6a3"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.247803 4772 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.247963 4772 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-f4d1f856-0902-41db-b052-d29ccd2349fb" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f4d1f856-0902-41db-b052-d29ccd2349fb") on node "crc" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.286106 4772 scope.go:117] "RemoveContainer" containerID="9b0305c5b89b648ee767009d14d68e95e3436ecb734d1416e97e7782000adad3" Jan 27 16:27:04 crc kubenswrapper[4772]: E0127 16:27:04.286518 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b0305c5b89b648ee767009d14d68e95e3436ecb734d1416e97e7782000adad3\": container with ID starting with 9b0305c5b89b648ee767009d14d68e95e3436ecb734d1416e97e7782000adad3 not found: ID does not exist" containerID="9b0305c5b89b648ee767009d14d68e95e3436ecb734d1416e97e7782000adad3" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.286572 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b0305c5b89b648ee767009d14d68e95e3436ecb734d1416e97e7782000adad3"} err="failed to get container status \"9b0305c5b89b648ee767009d14d68e95e3436ecb734d1416e97e7782000adad3\": rpc error: code = NotFound desc = could not find container \"9b0305c5b89b648ee767009d14d68e95e3436ecb734d1416e97e7782000adad3\": container with ID starting with 9b0305c5b89b648ee767009d14d68e95e3436ecb734d1416e97e7782000adad3 not found: ID does not exist" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.286653 4772 scope.go:117] "RemoveContainer" containerID="69ba34cf36586fe5c71196b7621cd2b9ba30a359ba359f88ecfede8d86e1b355" Jan 27 16:27:04 crc kubenswrapper[4772]: E0127 16:27:04.286986 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69ba34cf36586fe5c71196b7621cd2b9ba30a359ba359f88ecfede8d86e1b355\": container with ID starting with 69ba34cf36586fe5c71196b7621cd2b9ba30a359ba359f88ecfede8d86e1b355 not found: ID does not exist" containerID="69ba34cf36586fe5c71196b7621cd2b9ba30a359ba359f88ecfede8d86e1b355" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.287036 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69ba34cf36586fe5c71196b7621cd2b9ba30a359ba359f88ecfede8d86e1b355"} err="failed to get container status \"69ba34cf36586fe5c71196b7621cd2b9ba30a359ba359f88ecfede8d86e1b355\": rpc error: code = NotFound desc = could not find container \"69ba34cf36586fe5c71196b7621cd2b9ba30a359ba359f88ecfede8d86e1b355\": container with ID starting with 69ba34cf36586fe5c71196b7621cd2b9ba30a359ba359f88ecfede8d86e1b355 not found: ID does not exist" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.330559 4772 reconciler_common.go:293] "Volume detached for volume \"pvc-f4d1f856-0902-41db-b052-d29ccd2349fb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f4d1f856-0902-41db-b052-d29ccd2349fb\") on node \"crc\" DevicePath \"\"" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.330610 4772 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f38de303-3271-4d8a-b114-4fca1e36c6a3-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.448856 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.454735 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.477737 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Jan 27 16:27:04 crc kubenswrapper[4772]: E0127 16:27:04.478104 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f38de303-3271-4d8a-b114-4fca1e36c6a3" containerName="setup-container" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.478126 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f38de303-3271-4d8a-b114-4fca1e36c6a3" containerName="setup-container" Jan 27 16:27:04 crc kubenswrapper[4772]: E0127 16:27:04.478150 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f38de303-3271-4d8a-b114-4fca1e36c6a3" containerName="rabbitmq" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.478158 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f38de303-3271-4d8a-b114-4fca1e36c6a3" containerName="rabbitmq" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.478373 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="f38de303-3271-4d8a-b114-4fca1e36c6a3" containerName="rabbitmq" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.479268 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.481429 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.481685 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.481875 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.482018 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-kn7v7" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.484879 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.499877 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.534931 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f3b5f224-602e-454a-b35e-2e55160016b5-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.535265 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f3b5f224-602e-454a-b35e-2e55160016b5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.535379 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f3b5f224-602e-454a-b35e-2e55160016b5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.537234 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f3b5f224-602e-454a-b35e-2e55160016b5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.537439 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f3b5f224-602e-454a-b35e-2e55160016b5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.537715 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f3b5f224-602e-454a-b35e-2e55160016b5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.537945 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f3b5f224-602e-454a-b35e-2e55160016b5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.538050 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f4d1f856-0902-41db-b052-d29ccd2349fb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f4d1f856-0902-41db-b052-d29ccd2349fb\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.538160 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4qzb\" (UniqueName: \"kubernetes.io/projected/f3b5f224-602e-454a-b35e-2e55160016b5-kube-api-access-h4qzb\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.579639 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.639514 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-server-conf\") pod \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.639585 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-rabbitmq-plugins\") pod \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.639639 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qc4j\" (UniqueName: \"kubernetes.io/projected/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-kube-api-access-8qc4j\") pod \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.639659 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-plugins-conf\") pod \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.639794 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899\") pod \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.639831 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-erlang-cookie-secret\") pod \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.639853 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-rabbitmq-confd\") pod \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.639896 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-pod-info\") pod \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.639927 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-rabbitmq-erlang-cookie\") pod \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\" (UID: \"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b\") " Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.640115 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f3b5f224-602e-454a-b35e-2e55160016b5-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.640141 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f3b5f224-602e-454a-b35e-2e55160016b5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.640161 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f3b5f224-602e-454a-b35e-2e55160016b5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.640870 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "18fc2383-1b4e-43c5-b6cb-8aa40600cf7b" (UID: "18fc2383-1b4e-43c5-b6cb-8aa40600cf7b"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.640940 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "18fc2383-1b4e-43c5-b6cb-8aa40600cf7b" (UID: "18fc2383-1b4e-43c5-b6cb-8aa40600cf7b"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.640985 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f3b5f224-602e-454a-b35e-2e55160016b5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.641012 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "18fc2383-1b4e-43c5-b6cb-8aa40600cf7b" (UID: "18fc2383-1b4e-43c5-b6cb-8aa40600cf7b"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.641267 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f3b5f224-602e-454a-b35e-2e55160016b5-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.641337 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f3b5f224-602e-454a-b35e-2e55160016b5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.641433 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f3b5f224-602e-454a-b35e-2e55160016b5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.641469 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f3b5f224-602e-454a-b35e-2e55160016b5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.641500 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f3b5f224-602e-454a-b35e-2e55160016b5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.641518 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f4d1f856-0902-41db-b052-d29ccd2349fb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f4d1f856-0902-41db-b052-d29ccd2349fb\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.641543 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4qzb\" (UniqueName: \"kubernetes.io/projected/f3b5f224-602e-454a-b35e-2e55160016b5-kube-api-access-h4qzb\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.641599 4772 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.641609 4772 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-plugins-conf\") on node \"crc\" DevicePath \"\"" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.641618 4772 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.642150 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f3b5f224-602e-454a-b35e-2e55160016b5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.645437 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-pod-info" (OuterVolumeSpecName: "pod-info") pod "18fc2383-1b4e-43c5-b6cb-8aa40600cf7b" (UID: "18fc2383-1b4e-43c5-b6cb-8aa40600cf7b"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.646028 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f3b5f224-602e-454a-b35e-2e55160016b5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.647720 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-kube-api-access-8qc4j" (OuterVolumeSpecName: "kube-api-access-8qc4j") pod "18fc2383-1b4e-43c5-b6cb-8aa40600cf7b" (UID: "18fc2383-1b4e-43c5-b6cb-8aa40600cf7b"). InnerVolumeSpecName "kube-api-access-8qc4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.648583 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f3b5f224-602e-454a-b35e-2e55160016b5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.648678 4772 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.648703 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "18fc2383-1b4e-43c5-b6cb-8aa40600cf7b" (UID: "18fc2383-1b4e-43c5-b6cb-8aa40600cf7b"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.648712 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f4d1f856-0902-41db-b052-d29ccd2349fb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f4d1f856-0902-41db-b052-d29ccd2349fb\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/36a183de2c7abb5d8abee5f0c83592d4960d3f7dbd03e4d4afd32924fe238d72/globalmount\"" pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.650787 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f3b5f224-602e-454a-b35e-2e55160016b5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.657651 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f3b5f224-602e-454a-b35e-2e55160016b5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.662861 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-server-conf" (OuterVolumeSpecName: "server-conf") pod "18fc2383-1b4e-43c5-b6cb-8aa40600cf7b" (UID: "18fc2383-1b4e-43c5-b6cb-8aa40600cf7b"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.664784 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899" (OuterVolumeSpecName: "persistence") pod "18fc2383-1b4e-43c5-b6cb-8aa40600cf7b" (UID: "18fc2383-1b4e-43c5-b6cb-8aa40600cf7b"). InnerVolumeSpecName "pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.670909 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4qzb\" (UniqueName: \"kubernetes.io/projected/f3b5f224-602e-454a-b35e-2e55160016b5-kube-api-access-h4qzb\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.693128 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f38de303-3271-4d8a-b114-4fca1e36c6a3" path="/var/lib/kubelet/pods/f38de303-3271-4d8a-b114-4fca1e36c6a3/volumes" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.708457 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f4d1f856-0902-41db-b052-d29ccd2349fb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f4d1f856-0902-41db-b052-d29ccd2349fb\") pod \"rabbitmq-server-0\" (UID: \"f3b5f224-602e-454a-b35e-2e55160016b5\") " pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.743135 4772 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-server-conf\") on node \"crc\" DevicePath \"\"" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.743458 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qc4j\" (UniqueName: \"kubernetes.io/projected/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-kube-api-access-8qc4j\") on node \"crc\" DevicePath \"\"" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.743608 4772 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899\") on node \"crc\" " Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.743708 4772 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.743809 4772 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-pod-info\") on node \"crc\" DevicePath \"\"" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.747445 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "18fc2383-1b4e-43c5-b6cb-8aa40600cf7b" (UID: "18fc2383-1b4e-43c5-b6cb-8aa40600cf7b"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.761746 4772 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.761936 4772 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899") on node "crc" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.806261 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.845684 4772 reconciler_common.go:293] "Volume detached for volume \"pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899\") on node \"crc\" DevicePath \"\"" Jan 27 16:27:04 crc kubenswrapper[4772]: I0127 16:27:04.845718 4772 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.111077 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b7946d7b9-ndkll" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.132604 4772 generic.go:334] "Generic (PLEG): container finished" podID="18fc2383-1b4e-43c5-b6cb-8aa40600cf7b" containerID="5308f6aebce7e906d821045f50b393cfaba5255e05a4b74e645d3d680c330291" exitCode=0 Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.133522 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.136379 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b","Type":"ContainerDied","Data":"5308f6aebce7e906d821045f50b393cfaba5255e05a4b74e645d3d680c330291"} Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.136457 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"18fc2383-1b4e-43c5-b6cb-8aa40600cf7b","Type":"ContainerDied","Data":"b96f230c3d38a86522fef8eccf06bafd1c3858ca3dbe42112ea60613f2b942f3"} Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.136496 4772 scope.go:117] "RemoveContainer" containerID="5308f6aebce7e906d821045f50b393cfaba5255e05a4b74e645d3d680c330291" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.169284 4772 scope.go:117] "RemoveContainer" containerID="480ea6652ba12d615e1f6b8fb144377c490339e8e4bed427eade476083ce2cbb" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.188913 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-mmzht"] Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.189186 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-98ddfc8f-mmzht" podUID="bb8b7780-142e-4fd6-967f-a42e112a0b2e" containerName="dnsmasq-dns" containerID="cri-o://9c160ab266039b95d6f27552251a6a39ff7c08fcb4346eec1f5f3d5a708d9f94" gracePeriod=10 Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.200229 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.207967 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.212251 4772 scope.go:117] "RemoveContainer" containerID="5308f6aebce7e906d821045f50b393cfaba5255e05a4b74e645d3d680c330291" Jan 27 16:27:05 crc kubenswrapper[4772]: E0127 16:27:05.212836 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5308f6aebce7e906d821045f50b393cfaba5255e05a4b74e645d3d680c330291\": container with ID starting with 5308f6aebce7e906d821045f50b393cfaba5255e05a4b74e645d3d680c330291 not found: ID does not exist" containerID="5308f6aebce7e906d821045f50b393cfaba5255e05a4b74e645d3d680c330291" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.212903 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5308f6aebce7e906d821045f50b393cfaba5255e05a4b74e645d3d680c330291"} err="failed to get container status \"5308f6aebce7e906d821045f50b393cfaba5255e05a4b74e645d3d680c330291\": rpc error: code = NotFound desc = could not find container \"5308f6aebce7e906d821045f50b393cfaba5255e05a4b74e645d3d680c330291\": container with ID starting with 5308f6aebce7e906d821045f50b393cfaba5255e05a4b74e645d3d680c330291 not found: ID does not exist" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.212946 4772 scope.go:117] "RemoveContainer" containerID="480ea6652ba12d615e1f6b8fb144377c490339e8e4bed427eade476083ce2cbb" Jan 27 16:27:05 crc kubenswrapper[4772]: E0127 16:27:05.213378 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"480ea6652ba12d615e1f6b8fb144377c490339e8e4bed427eade476083ce2cbb\": container with ID starting with 480ea6652ba12d615e1f6b8fb144377c490339e8e4bed427eade476083ce2cbb not found: ID does not exist" containerID="480ea6652ba12d615e1f6b8fb144377c490339e8e4bed427eade476083ce2cbb" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.213412 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"480ea6652ba12d615e1f6b8fb144377c490339e8e4bed427eade476083ce2cbb"} err="failed to get container status \"480ea6652ba12d615e1f6b8fb144377c490339e8e4bed427eade476083ce2cbb\": rpc error: code = NotFound desc = could not find container \"480ea6652ba12d615e1f6b8fb144377c490339e8e4bed427eade476083ce2cbb\": container with ID starting with 480ea6652ba12d615e1f6b8fb144377c490339e8e4bed427eade476083ce2cbb not found: ID does not exist" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.238454 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 27 16:27:05 crc kubenswrapper[4772]: E0127 16:27:05.238800 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18fc2383-1b4e-43c5-b6cb-8aa40600cf7b" containerName="rabbitmq" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.238820 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="18fc2383-1b4e-43c5-b6cb-8aa40600cf7b" containerName="rabbitmq" Jan 27 16:27:05 crc kubenswrapper[4772]: E0127 16:27:05.238840 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18fc2383-1b4e-43c5-b6cb-8aa40600cf7b" containerName="setup-container" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.238848 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="18fc2383-1b4e-43c5-b6cb-8aa40600cf7b" containerName="setup-container" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.239034 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="18fc2383-1b4e-43c5-b6cb-8aa40600cf7b" containerName="rabbitmq" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.243146 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.247830 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.249674 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.249741 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.250016 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-554gz" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.250046 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.255598 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 27 16:27:05 crc kubenswrapper[4772]: W0127 16:27:05.266772 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3b5f224_602e_454a_b35e_2e55160016b5.slice/crio-735e34770b8fa85917d67cdf87dff85c464a64682387bc5b15840a99a4ae68bd WatchSource:0}: Error finding container 735e34770b8fa85917d67cdf87dff85c464a64682387bc5b15840a99a4ae68bd: Status 404 returned error can't find the container with id 735e34770b8fa85917d67cdf87dff85c464a64682387bc5b15840a99a4ae68bd Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.268398 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.354230 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.354296 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.354319 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.354341 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.354383 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7dwn\" (UniqueName: \"kubernetes.io/projected/1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88-kube-api-access-d7dwn\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.354433 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.354472 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.354533 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.354565 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.459206 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.459284 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.459341 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.459375 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.459422 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.459479 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.459506 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.459531 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.459572 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7dwn\" (UniqueName: \"kubernetes.io/projected/1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88-kube-api-access-d7dwn\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.459963 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.460723 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.461531 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.461601 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.464322 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.464347 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.465030 4772 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.465069 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/74f6c4054e7c826304958ef416594be6d4b6260f90a6b43d068948e9c0dc0fa0/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.465228 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.475765 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7dwn\" (UniqueName: \"kubernetes.io/projected/1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88-kube-api-access-d7dwn\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.512400 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b481002b-0e4c-443b-9281-7c1ac6b1e899\") pod \"rabbitmq-cell1-server-0\" (UID: \"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88\") " pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.594443 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.722260 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-mmzht" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.865644 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb8b7780-142e-4fd6-967f-a42e112a0b2e-config\") pod \"bb8b7780-142e-4fd6-967f-a42e112a0b2e\" (UID: \"bb8b7780-142e-4fd6-967f-a42e112a0b2e\") " Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.865709 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb8b7780-142e-4fd6-967f-a42e112a0b2e-dns-svc\") pod \"bb8b7780-142e-4fd6-967f-a42e112a0b2e\" (UID: \"bb8b7780-142e-4fd6-967f-a42e112a0b2e\") " Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.865748 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqg6h\" (UniqueName: \"kubernetes.io/projected/bb8b7780-142e-4fd6-967f-a42e112a0b2e-kube-api-access-cqg6h\") pod \"bb8b7780-142e-4fd6-967f-a42e112a0b2e\" (UID: \"bb8b7780-142e-4fd6-967f-a42e112a0b2e\") " Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.874387 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb8b7780-142e-4fd6-967f-a42e112a0b2e-kube-api-access-cqg6h" (OuterVolumeSpecName: "kube-api-access-cqg6h") pod "bb8b7780-142e-4fd6-967f-a42e112a0b2e" (UID: "bb8b7780-142e-4fd6-967f-a42e112a0b2e"). InnerVolumeSpecName "kube-api-access-cqg6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.901413 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb8b7780-142e-4fd6-967f-a42e112a0b2e-config" (OuterVolumeSpecName: "config") pod "bb8b7780-142e-4fd6-967f-a42e112a0b2e" (UID: "bb8b7780-142e-4fd6-967f-a42e112a0b2e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.910527 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb8b7780-142e-4fd6-967f-a42e112a0b2e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bb8b7780-142e-4fd6-967f-a42e112a0b2e" (UID: "bb8b7780-142e-4fd6-967f-a42e112a0b2e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.967773 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb8b7780-142e-4fd6-967f-a42e112a0b2e-config\") on node \"crc\" DevicePath \"\"" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.968060 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb8b7780-142e-4fd6-967f-a42e112a0b2e-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 16:27:05 crc kubenswrapper[4772]: I0127 16:27:05.968145 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqg6h\" (UniqueName: \"kubernetes.io/projected/bb8b7780-142e-4fd6-967f-a42e112a0b2e-kube-api-access-cqg6h\") on node \"crc\" DevicePath \"\"" Jan 27 16:27:06 crc kubenswrapper[4772]: I0127 16:27:06.071497 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 27 16:27:06 crc kubenswrapper[4772]: W0127 16:27:06.086815 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c2904f0_5ba8_4bb4_9952_ca1ee06a4d88.slice/crio-e0c5cacd55009b64e1a8abc6f9b74ae8e9045b3dca625c9f4afd5c099f239b3b WatchSource:0}: Error finding container e0c5cacd55009b64e1a8abc6f9b74ae8e9045b3dca625c9f4afd5c099f239b3b: Status 404 returned error can't find the container with id e0c5cacd55009b64e1a8abc6f9b74ae8e9045b3dca625c9f4afd5c099f239b3b Jan 27 16:27:06 crc kubenswrapper[4772]: I0127 16:27:06.140200 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f3b5f224-602e-454a-b35e-2e55160016b5","Type":"ContainerStarted","Data":"735e34770b8fa85917d67cdf87dff85c464a64682387bc5b15840a99a4ae68bd"} Jan 27 16:27:06 crc kubenswrapper[4772]: I0127 16:27:06.141125 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88","Type":"ContainerStarted","Data":"e0c5cacd55009b64e1a8abc6f9b74ae8e9045b3dca625c9f4afd5c099f239b3b"} Jan 27 16:27:06 crc kubenswrapper[4772]: I0127 16:27:06.142623 4772 generic.go:334] "Generic (PLEG): container finished" podID="bb8b7780-142e-4fd6-967f-a42e112a0b2e" containerID="9c160ab266039b95d6f27552251a6a39ff7c08fcb4346eec1f5f3d5a708d9f94" exitCode=0 Jan 27 16:27:06 crc kubenswrapper[4772]: I0127 16:27:06.142647 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-mmzht" event={"ID":"bb8b7780-142e-4fd6-967f-a42e112a0b2e","Type":"ContainerDied","Data":"9c160ab266039b95d6f27552251a6a39ff7c08fcb4346eec1f5f3d5a708d9f94"} Jan 27 16:27:06 crc kubenswrapper[4772]: I0127 16:27:06.142676 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-mmzht" Jan 27 16:27:06 crc kubenswrapper[4772]: I0127 16:27:06.142688 4772 scope.go:117] "RemoveContainer" containerID="9c160ab266039b95d6f27552251a6a39ff7c08fcb4346eec1f5f3d5a708d9f94" Jan 27 16:27:06 crc kubenswrapper[4772]: I0127 16:27:06.142675 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-mmzht" event={"ID":"bb8b7780-142e-4fd6-967f-a42e112a0b2e","Type":"ContainerDied","Data":"78ea743dbcac291e88b556c9f353cf72b6797716babc1be9562796c00fc48cc9"} Jan 27 16:27:06 crc kubenswrapper[4772]: I0127 16:27:06.177339 4772 scope.go:117] "RemoveContainer" containerID="16ae7072b9a91932c05ae95c0de53fbb1486c6e51d07f50b89c0848be4b668b5" Jan 27 16:27:06 crc kubenswrapper[4772]: I0127 16:27:06.192044 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-mmzht"] Jan 27 16:27:06 crc kubenswrapper[4772]: I0127 16:27:06.205651 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-mmzht"] Jan 27 16:27:06 crc kubenswrapper[4772]: I0127 16:27:06.320244 4772 scope.go:117] "RemoveContainer" containerID="9c160ab266039b95d6f27552251a6a39ff7c08fcb4346eec1f5f3d5a708d9f94" Jan 27 16:27:06 crc kubenswrapper[4772]: E0127 16:27:06.320640 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c160ab266039b95d6f27552251a6a39ff7c08fcb4346eec1f5f3d5a708d9f94\": container with ID starting with 9c160ab266039b95d6f27552251a6a39ff7c08fcb4346eec1f5f3d5a708d9f94 not found: ID does not exist" containerID="9c160ab266039b95d6f27552251a6a39ff7c08fcb4346eec1f5f3d5a708d9f94" Jan 27 16:27:06 crc kubenswrapper[4772]: I0127 16:27:06.320677 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c160ab266039b95d6f27552251a6a39ff7c08fcb4346eec1f5f3d5a708d9f94"} err="failed to get container status \"9c160ab266039b95d6f27552251a6a39ff7c08fcb4346eec1f5f3d5a708d9f94\": rpc error: code = NotFound desc = could not find container \"9c160ab266039b95d6f27552251a6a39ff7c08fcb4346eec1f5f3d5a708d9f94\": container with ID starting with 9c160ab266039b95d6f27552251a6a39ff7c08fcb4346eec1f5f3d5a708d9f94 not found: ID does not exist" Jan 27 16:27:06 crc kubenswrapper[4772]: I0127 16:27:06.320700 4772 scope.go:117] "RemoveContainer" containerID="16ae7072b9a91932c05ae95c0de53fbb1486c6e51d07f50b89c0848be4b668b5" Jan 27 16:27:06 crc kubenswrapper[4772]: E0127 16:27:06.321052 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16ae7072b9a91932c05ae95c0de53fbb1486c6e51d07f50b89c0848be4b668b5\": container with ID starting with 16ae7072b9a91932c05ae95c0de53fbb1486c6e51d07f50b89c0848be4b668b5 not found: ID does not exist" containerID="16ae7072b9a91932c05ae95c0de53fbb1486c6e51d07f50b89c0848be4b668b5" Jan 27 16:27:06 crc kubenswrapper[4772]: I0127 16:27:06.321077 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16ae7072b9a91932c05ae95c0de53fbb1486c6e51d07f50b89c0848be4b668b5"} err="failed to get container status \"16ae7072b9a91932c05ae95c0de53fbb1486c6e51d07f50b89c0848be4b668b5\": rpc error: code = NotFound desc = could not find container \"16ae7072b9a91932c05ae95c0de53fbb1486c6e51d07f50b89c0848be4b668b5\": container with ID starting with 16ae7072b9a91932c05ae95c0de53fbb1486c6e51d07f50b89c0848be4b668b5 not found: ID does not exist" Jan 27 16:27:06 crc kubenswrapper[4772]: I0127 16:27:06.672831 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18fc2383-1b4e-43c5-b6cb-8aa40600cf7b" path="/var/lib/kubelet/pods/18fc2383-1b4e-43c5-b6cb-8aa40600cf7b/volumes" Jan 27 16:27:06 crc kubenswrapper[4772]: I0127 16:27:06.673479 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb8b7780-142e-4fd6-967f-a42e112a0b2e" path="/var/lib/kubelet/pods/bb8b7780-142e-4fd6-967f-a42e112a0b2e/volumes" Jan 27 16:27:07 crc kubenswrapper[4772]: I0127 16:27:07.152426 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f3b5f224-602e-454a-b35e-2e55160016b5","Type":"ContainerStarted","Data":"6d2656c9e69a02c468dda617caf117c0c2d45cfe2e8706e5cd1d82cc528442ff"} Jan 27 16:27:08 crc kubenswrapper[4772]: I0127 16:27:08.162214 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88","Type":"ContainerStarted","Data":"0980d84a86e629e38b050c4e0e81922329b893eae78161f17488b28b814908bb"} Jan 27 16:27:10 crc kubenswrapper[4772]: I0127 16:27:10.045317 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jjgpk" Jan 27 16:27:10 crc kubenswrapper[4772]: I0127 16:27:10.045534 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jjgpk" Jan 27 16:27:10 crc kubenswrapper[4772]: I0127 16:27:10.096538 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jjgpk" Jan 27 16:27:10 crc kubenswrapper[4772]: I0127 16:27:10.217086 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jjgpk" Jan 27 16:27:10 crc kubenswrapper[4772]: I0127 16:27:10.329146 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jjgpk"] Jan 27 16:27:12 crc kubenswrapper[4772]: I0127 16:27:12.192630 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jjgpk" podUID="b2025168-9cc0-417c-8d3b-1d336447a3ff" containerName="registry-server" containerID="cri-o://647af5829c88ac17630d560d584a1467c6061b43b60545e27fdaa4193667e8cf" gracePeriod=2 Jan 27 16:27:12 crc kubenswrapper[4772]: I0127 16:27:12.570582 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jjgpk" Jan 27 16:27:12 crc kubenswrapper[4772]: I0127 16:27:12.692900 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpgqq\" (UniqueName: \"kubernetes.io/projected/b2025168-9cc0-417c-8d3b-1d336447a3ff-kube-api-access-gpgqq\") pod \"b2025168-9cc0-417c-8d3b-1d336447a3ff\" (UID: \"b2025168-9cc0-417c-8d3b-1d336447a3ff\") " Jan 27 16:27:12 crc kubenswrapper[4772]: I0127 16:27:12.692999 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2025168-9cc0-417c-8d3b-1d336447a3ff-utilities\") pod \"b2025168-9cc0-417c-8d3b-1d336447a3ff\" (UID: \"b2025168-9cc0-417c-8d3b-1d336447a3ff\") " Jan 27 16:27:12 crc kubenswrapper[4772]: I0127 16:27:12.693096 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2025168-9cc0-417c-8d3b-1d336447a3ff-catalog-content\") pod \"b2025168-9cc0-417c-8d3b-1d336447a3ff\" (UID: \"b2025168-9cc0-417c-8d3b-1d336447a3ff\") " Jan 27 16:27:12 crc kubenswrapper[4772]: I0127 16:27:12.693875 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2025168-9cc0-417c-8d3b-1d336447a3ff-utilities" (OuterVolumeSpecName: "utilities") pod "b2025168-9cc0-417c-8d3b-1d336447a3ff" (UID: "b2025168-9cc0-417c-8d3b-1d336447a3ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:27:12 crc kubenswrapper[4772]: I0127 16:27:12.697992 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2025168-9cc0-417c-8d3b-1d336447a3ff-kube-api-access-gpgqq" (OuterVolumeSpecName: "kube-api-access-gpgqq") pod "b2025168-9cc0-417c-8d3b-1d336447a3ff" (UID: "b2025168-9cc0-417c-8d3b-1d336447a3ff"). InnerVolumeSpecName "kube-api-access-gpgqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:27:12 crc kubenswrapper[4772]: I0127 16:27:12.795789 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpgqq\" (UniqueName: \"kubernetes.io/projected/b2025168-9cc0-417c-8d3b-1d336447a3ff-kube-api-access-gpgqq\") on node \"crc\" DevicePath \"\"" Jan 27 16:27:12 crc kubenswrapper[4772]: I0127 16:27:12.795834 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2025168-9cc0-417c-8d3b-1d336447a3ff-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 16:27:13 crc kubenswrapper[4772]: I0127 16:27:13.201081 4772 generic.go:334] "Generic (PLEG): container finished" podID="b2025168-9cc0-417c-8d3b-1d336447a3ff" containerID="647af5829c88ac17630d560d584a1467c6061b43b60545e27fdaa4193667e8cf" exitCode=0 Jan 27 16:27:13 crc kubenswrapper[4772]: I0127 16:27:13.201131 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jjgpk" event={"ID":"b2025168-9cc0-417c-8d3b-1d336447a3ff","Type":"ContainerDied","Data":"647af5829c88ac17630d560d584a1467c6061b43b60545e27fdaa4193667e8cf"} Jan 27 16:27:13 crc kubenswrapper[4772]: I0127 16:27:13.201188 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jjgpk" Jan 27 16:27:13 crc kubenswrapper[4772]: I0127 16:27:13.202473 4772 scope.go:117] "RemoveContainer" containerID="647af5829c88ac17630d560d584a1467c6061b43b60545e27fdaa4193667e8cf" Jan 27 16:27:13 crc kubenswrapper[4772]: I0127 16:27:13.202364 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jjgpk" event={"ID":"b2025168-9cc0-417c-8d3b-1d336447a3ff","Type":"ContainerDied","Data":"aaf303dc4f995cb6bff56fd0049dc32fd5acf74cce10552ba6fbab1b994a80c7"} Jan 27 16:27:13 crc kubenswrapper[4772]: I0127 16:27:13.222146 4772 scope.go:117] "RemoveContainer" containerID="020319dacd513bbd957b8817f14f5c128e22f4c1f10468a190b399e9005621c7" Jan 27 16:27:13 crc kubenswrapper[4772]: I0127 16:27:13.248116 4772 scope.go:117] "RemoveContainer" containerID="9a45c145ee6112b35c7f85262dd11f2bab997eedc4836c666662e77f41944aa0" Jan 27 16:27:13 crc kubenswrapper[4772]: I0127 16:27:13.290462 4772 scope.go:117] "RemoveContainer" containerID="647af5829c88ac17630d560d584a1467c6061b43b60545e27fdaa4193667e8cf" Jan 27 16:27:13 crc kubenswrapper[4772]: E0127 16:27:13.290863 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"647af5829c88ac17630d560d584a1467c6061b43b60545e27fdaa4193667e8cf\": container with ID starting with 647af5829c88ac17630d560d584a1467c6061b43b60545e27fdaa4193667e8cf not found: ID does not exist" containerID="647af5829c88ac17630d560d584a1467c6061b43b60545e27fdaa4193667e8cf" Jan 27 16:27:13 crc kubenswrapper[4772]: I0127 16:27:13.290898 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"647af5829c88ac17630d560d584a1467c6061b43b60545e27fdaa4193667e8cf"} err="failed to get container status \"647af5829c88ac17630d560d584a1467c6061b43b60545e27fdaa4193667e8cf\": rpc error: code = NotFound desc = could not find container \"647af5829c88ac17630d560d584a1467c6061b43b60545e27fdaa4193667e8cf\": container with ID starting with 647af5829c88ac17630d560d584a1467c6061b43b60545e27fdaa4193667e8cf not found: ID does not exist" Jan 27 16:27:13 crc kubenswrapper[4772]: I0127 16:27:13.290917 4772 scope.go:117] "RemoveContainer" containerID="020319dacd513bbd957b8817f14f5c128e22f4c1f10468a190b399e9005621c7" Jan 27 16:27:13 crc kubenswrapper[4772]: E0127 16:27:13.291302 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"020319dacd513bbd957b8817f14f5c128e22f4c1f10468a190b399e9005621c7\": container with ID starting with 020319dacd513bbd957b8817f14f5c128e22f4c1f10468a190b399e9005621c7 not found: ID does not exist" containerID="020319dacd513bbd957b8817f14f5c128e22f4c1f10468a190b399e9005621c7" Jan 27 16:27:13 crc kubenswrapper[4772]: I0127 16:27:13.291362 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"020319dacd513bbd957b8817f14f5c128e22f4c1f10468a190b399e9005621c7"} err="failed to get container status \"020319dacd513bbd957b8817f14f5c128e22f4c1f10468a190b399e9005621c7\": rpc error: code = NotFound desc = could not find container \"020319dacd513bbd957b8817f14f5c128e22f4c1f10468a190b399e9005621c7\": container with ID starting with 020319dacd513bbd957b8817f14f5c128e22f4c1f10468a190b399e9005621c7 not found: ID does not exist" Jan 27 16:27:13 crc kubenswrapper[4772]: I0127 16:27:13.291396 4772 scope.go:117] "RemoveContainer" containerID="9a45c145ee6112b35c7f85262dd11f2bab997eedc4836c666662e77f41944aa0" Jan 27 16:27:13 crc kubenswrapper[4772]: E0127 16:27:13.291725 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a45c145ee6112b35c7f85262dd11f2bab997eedc4836c666662e77f41944aa0\": container with ID starting with 9a45c145ee6112b35c7f85262dd11f2bab997eedc4836c666662e77f41944aa0 not found: ID does not exist" containerID="9a45c145ee6112b35c7f85262dd11f2bab997eedc4836c666662e77f41944aa0" Jan 27 16:27:13 crc kubenswrapper[4772]: I0127 16:27:13.291757 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a45c145ee6112b35c7f85262dd11f2bab997eedc4836c666662e77f41944aa0"} err="failed to get container status \"9a45c145ee6112b35c7f85262dd11f2bab997eedc4836c666662e77f41944aa0\": rpc error: code = NotFound desc = could not find container \"9a45c145ee6112b35c7f85262dd11f2bab997eedc4836c666662e77f41944aa0\": container with ID starting with 9a45c145ee6112b35c7f85262dd11f2bab997eedc4836c666662e77f41944aa0 not found: ID does not exist" Jan 27 16:27:13 crc kubenswrapper[4772]: I0127 16:27:13.559656 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2025168-9cc0-417c-8d3b-1d336447a3ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2025168-9cc0-417c-8d3b-1d336447a3ff" (UID: "b2025168-9cc0-417c-8d3b-1d336447a3ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:27:13 crc kubenswrapper[4772]: I0127 16:27:13.610525 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2025168-9cc0-417c-8d3b-1d336447a3ff-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 16:27:13 crc kubenswrapper[4772]: I0127 16:27:13.843861 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jjgpk"] Jan 27 16:27:13 crc kubenswrapper[4772]: I0127 16:27:13.851948 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jjgpk"] Jan 27 16:27:14 crc kubenswrapper[4772]: I0127 16:27:14.674308 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2025168-9cc0-417c-8d3b-1d336447a3ff" path="/var/lib/kubelet/pods/b2025168-9cc0-417c-8d3b-1d336447a3ff/volumes" Jan 27 16:27:39 crc kubenswrapper[4772]: I0127 16:27:39.405936 4772 generic.go:334] "Generic (PLEG): container finished" podID="f3b5f224-602e-454a-b35e-2e55160016b5" containerID="6d2656c9e69a02c468dda617caf117c0c2d45cfe2e8706e5cd1d82cc528442ff" exitCode=0 Jan 27 16:27:39 crc kubenswrapper[4772]: I0127 16:27:39.406031 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f3b5f224-602e-454a-b35e-2e55160016b5","Type":"ContainerDied","Data":"6d2656c9e69a02c468dda617caf117c0c2d45cfe2e8706e5cd1d82cc528442ff"} Jan 27 16:27:39 crc kubenswrapper[4772]: I0127 16:27:39.409005 4772 generic.go:334] "Generic (PLEG): container finished" podID="1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88" containerID="0980d84a86e629e38b050c4e0e81922329b893eae78161f17488b28b814908bb" exitCode=0 Jan 27 16:27:39 crc kubenswrapper[4772]: I0127 16:27:39.409049 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88","Type":"ContainerDied","Data":"0980d84a86e629e38b050c4e0e81922329b893eae78161f17488b28b814908bb"} Jan 27 16:27:40 crc kubenswrapper[4772]: I0127 16:27:40.418006 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f3b5f224-602e-454a-b35e-2e55160016b5","Type":"ContainerStarted","Data":"687181ce3ac97e8e7c2b77ea2674ed00ed45a1ee0b30edbe6112a0965279b62a"} Jan 27 16:27:40 crc kubenswrapper[4772]: I0127 16:27:40.418591 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Jan 27 16:27:40 crc kubenswrapper[4772]: I0127 16:27:40.420284 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88","Type":"ContainerStarted","Data":"ecc8aa55c8175131ece32925bb28b99bd308ecf7369edc88c3c039d474fb4cc3"} Jan 27 16:27:40 crc kubenswrapper[4772]: I0127 16:27:40.420512 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:27:40 crc kubenswrapper[4772]: I0127 16:27:40.448010 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.447989759 podStartE2EDuration="36.447989759s" podCreationTimestamp="2026-01-27 16:27:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:27:40.440655808 +0000 UTC m=+4846.421264926" watchObservedRunningTime="2026-01-27 16:27:40.447989759 +0000 UTC m=+4846.428598857" Jan 27 16:27:40 crc kubenswrapper[4772]: I0127 16:27:40.466575 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=35.466556794 podStartE2EDuration="35.466556794s" podCreationTimestamp="2026-01-27 16:27:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:27:40.460825959 +0000 UTC m=+4846.441435057" watchObservedRunningTime="2026-01-27 16:27:40.466556794 +0000 UTC m=+4846.447165882" Jan 27 16:27:54 crc kubenswrapper[4772]: I0127 16:27:54.810074 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Jan 27 16:27:55 crc kubenswrapper[4772]: I0127 16:27:55.598473 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Jan 27 16:28:06 crc kubenswrapper[4772]: I0127 16:28:06.523263 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Jan 27 16:28:06 crc kubenswrapper[4772]: E0127 16:28:06.524207 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2025168-9cc0-417c-8d3b-1d336447a3ff" containerName="registry-server" Jan 27 16:28:06 crc kubenswrapper[4772]: I0127 16:28:06.524225 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2025168-9cc0-417c-8d3b-1d336447a3ff" containerName="registry-server" Jan 27 16:28:06 crc kubenswrapper[4772]: E0127 16:28:06.524242 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2025168-9cc0-417c-8d3b-1d336447a3ff" containerName="extract-content" Jan 27 16:28:06 crc kubenswrapper[4772]: I0127 16:28:06.524253 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2025168-9cc0-417c-8d3b-1d336447a3ff" containerName="extract-content" Jan 27 16:28:06 crc kubenswrapper[4772]: E0127 16:28:06.524277 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb8b7780-142e-4fd6-967f-a42e112a0b2e" containerName="init" Jan 27 16:28:06 crc kubenswrapper[4772]: I0127 16:28:06.524286 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb8b7780-142e-4fd6-967f-a42e112a0b2e" containerName="init" Jan 27 16:28:06 crc kubenswrapper[4772]: E0127 16:28:06.524301 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2025168-9cc0-417c-8d3b-1d336447a3ff" containerName="extract-utilities" Jan 27 16:28:06 crc kubenswrapper[4772]: I0127 16:28:06.524311 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2025168-9cc0-417c-8d3b-1d336447a3ff" containerName="extract-utilities" Jan 27 16:28:06 crc kubenswrapper[4772]: E0127 16:28:06.524340 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb8b7780-142e-4fd6-967f-a42e112a0b2e" containerName="dnsmasq-dns" Jan 27 16:28:06 crc kubenswrapper[4772]: I0127 16:28:06.524348 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb8b7780-142e-4fd6-967f-a42e112a0b2e" containerName="dnsmasq-dns" Jan 27 16:28:06 crc kubenswrapper[4772]: I0127 16:28:06.524535 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb8b7780-142e-4fd6-967f-a42e112a0b2e" containerName="dnsmasq-dns" Jan 27 16:28:06 crc kubenswrapper[4772]: I0127 16:28:06.524566 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2025168-9cc0-417c-8d3b-1d336447a3ff" containerName="registry-server" Jan 27 16:28:06 crc kubenswrapper[4772]: I0127 16:28:06.525480 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Jan 27 16:28:06 crc kubenswrapper[4772]: I0127 16:28:06.533350 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-jd6dc" Jan 27 16:28:06 crc kubenswrapper[4772]: I0127 16:28:06.536318 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Jan 27 16:28:06 crc kubenswrapper[4772]: I0127 16:28:06.623394 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpdcs\" (UniqueName: \"kubernetes.io/projected/96c99cc4-7be7-49a2-bbc4-a16c3698e291-kube-api-access-hpdcs\") pod \"mariadb-client\" (UID: \"96c99cc4-7be7-49a2-bbc4-a16c3698e291\") " pod="openstack/mariadb-client" Jan 27 16:28:06 crc kubenswrapper[4772]: I0127 16:28:06.724868 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpdcs\" (UniqueName: \"kubernetes.io/projected/96c99cc4-7be7-49a2-bbc4-a16c3698e291-kube-api-access-hpdcs\") pod \"mariadb-client\" (UID: \"96c99cc4-7be7-49a2-bbc4-a16c3698e291\") " pod="openstack/mariadb-client" Jan 27 16:28:06 crc kubenswrapper[4772]: I0127 16:28:06.746719 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpdcs\" (UniqueName: \"kubernetes.io/projected/96c99cc4-7be7-49a2-bbc4-a16c3698e291-kube-api-access-hpdcs\") pod \"mariadb-client\" (UID: \"96c99cc4-7be7-49a2-bbc4-a16c3698e291\") " pod="openstack/mariadb-client" Jan 27 16:28:06 crc kubenswrapper[4772]: I0127 16:28:06.848583 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Jan 27 16:28:07 crc kubenswrapper[4772]: I0127 16:28:07.403370 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Jan 27 16:28:07 crc kubenswrapper[4772]: I0127 16:28:07.617681 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"96c99cc4-7be7-49a2-bbc4-a16c3698e291","Type":"ContainerStarted","Data":"e2528fc25164a9d3257d0c9f90fb442cd95239856af489f43f185023ba775a4f"} Jan 27 16:28:07 crc kubenswrapper[4772]: I0127 16:28:07.617724 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"96c99cc4-7be7-49a2-bbc4-a16c3698e291","Type":"ContainerStarted","Data":"9361cb30d4628d107a493b7c099f845f78b344e58ebf24b88567206da733e5ba"} Jan 27 16:28:07 crc kubenswrapper[4772]: I0127 16:28:07.630775 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client" podStartSLOduration=1.630755239 podStartE2EDuration="1.630755239s" podCreationTimestamp="2026-01-27 16:28:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:28:07.628401601 +0000 UTC m=+4873.609010699" watchObservedRunningTime="2026-01-27 16:28:07.630755239 +0000 UTC m=+4873.611364337" Jan 27 16:28:22 crc kubenswrapper[4772]: I0127 16:28:22.095448 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Jan 27 16:28:22 crc kubenswrapper[4772]: I0127 16:28:22.096383 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-client" podUID="96c99cc4-7be7-49a2-bbc4-a16c3698e291" containerName="mariadb-client" containerID="cri-o://e2528fc25164a9d3257d0c9f90fb442cd95239856af489f43f185023ba775a4f" gracePeriod=30 Jan 27 16:28:22 crc kubenswrapper[4772]: I0127 16:28:22.652921 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Jan 27 16:28:22 crc kubenswrapper[4772]: I0127 16:28:22.739692 4772 generic.go:334] "Generic (PLEG): container finished" podID="96c99cc4-7be7-49a2-bbc4-a16c3698e291" containerID="e2528fc25164a9d3257d0c9f90fb442cd95239856af489f43f185023ba775a4f" exitCode=143 Jan 27 16:28:22 crc kubenswrapper[4772]: I0127 16:28:22.739733 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"96c99cc4-7be7-49a2-bbc4-a16c3698e291","Type":"ContainerDied","Data":"e2528fc25164a9d3257d0c9f90fb442cd95239856af489f43f185023ba775a4f"} Jan 27 16:28:22 crc kubenswrapper[4772]: I0127 16:28:22.739738 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Jan 27 16:28:22 crc kubenswrapper[4772]: I0127 16:28:22.739757 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"96c99cc4-7be7-49a2-bbc4-a16c3698e291","Type":"ContainerDied","Data":"9361cb30d4628d107a493b7c099f845f78b344e58ebf24b88567206da733e5ba"} Jan 27 16:28:22 crc kubenswrapper[4772]: I0127 16:28:22.739771 4772 scope.go:117] "RemoveContainer" containerID="e2528fc25164a9d3257d0c9f90fb442cd95239856af489f43f185023ba775a4f" Jan 27 16:28:22 crc kubenswrapper[4772]: I0127 16:28:22.762844 4772 scope.go:117] "RemoveContainer" containerID="e2528fc25164a9d3257d0c9f90fb442cd95239856af489f43f185023ba775a4f" Jan 27 16:28:22 crc kubenswrapper[4772]: E0127 16:28:22.763287 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2528fc25164a9d3257d0c9f90fb442cd95239856af489f43f185023ba775a4f\": container with ID starting with e2528fc25164a9d3257d0c9f90fb442cd95239856af489f43f185023ba775a4f not found: ID does not exist" containerID="e2528fc25164a9d3257d0c9f90fb442cd95239856af489f43f185023ba775a4f" Jan 27 16:28:22 crc kubenswrapper[4772]: I0127 16:28:22.763327 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2528fc25164a9d3257d0c9f90fb442cd95239856af489f43f185023ba775a4f"} err="failed to get container status \"e2528fc25164a9d3257d0c9f90fb442cd95239856af489f43f185023ba775a4f\": rpc error: code = NotFound desc = could not find container \"e2528fc25164a9d3257d0c9f90fb442cd95239856af489f43f185023ba775a4f\": container with ID starting with e2528fc25164a9d3257d0c9f90fb442cd95239856af489f43f185023ba775a4f not found: ID does not exist" Jan 27 16:28:22 crc kubenswrapper[4772]: I0127 16:28:22.778102 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpdcs\" (UniqueName: \"kubernetes.io/projected/96c99cc4-7be7-49a2-bbc4-a16c3698e291-kube-api-access-hpdcs\") pod \"96c99cc4-7be7-49a2-bbc4-a16c3698e291\" (UID: \"96c99cc4-7be7-49a2-bbc4-a16c3698e291\") " Jan 27 16:28:22 crc kubenswrapper[4772]: I0127 16:28:22.783428 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96c99cc4-7be7-49a2-bbc4-a16c3698e291-kube-api-access-hpdcs" (OuterVolumeSpecName: "kube-api-access-hpdcs") pod "96c99cc4-7be7-49a2-bbc4-a16c3698e291" (UID: "96c99cc4-7be7-49a2-bbc4-a16c3698e291"). InnerVolumeSpecName "kube-api-access-hpdcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:28:22 crc kubenswrapper[4772]: I0127 16:28:22.879720 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpdcs\" (UniqueName: \"kubernetes.io/projected/96c99cc4-7be7-49a2-bbc4-a16c3698e291-kube-api-access-hpdcs\") on node \"crc\" DevicePath \"\"" Jan 27 16:28:23 crc kubenswrapper[4772]: I0127 16:28:23.084354 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Jan 27 16:28:23 crc kubenswrapper[4772]: I0127 16:28:23.091146 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Jan 27 16:28:24 crc kubenswrapper[4772]: I0127 16:28:24.672942 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96c99cc4-7be7-49a2-bbc4-a16c3698e291" path="/var/lib/kubelet/pods/96c99cc4-7be7-49a2-bbc4-a16c3698e291/volumes" Jan 27 16:28:42 crc kubenswrapper[4772]: I0127 16:28:42.059031 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:28:42 crc kubenswrapper[4772]: I0127 16:28:42.059660 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:29:06 crc kubenswrapper[4772]: I0127 16:29:06.412068 4772 scope.go:117] "RemoveContainer" containerID="a62dadf36906064bb1b0580332d53e82d1766f5edc230560463a0bad481701be" Jan 27 16:29:12 crc kubenswrapper[4772]: I0127 16:29:12.059043 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:29:12 crc kubenswrapper[4772]: I0127 16:29:12.059596 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:29:42 crc kubenswrapper[4772]: I0127 16:29:42.058744 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:29:42 crc kubenswrapper[4772]: I0127 16:29:42.059517 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:29:42 crc kubenswrapper[4772]: I0127 16:29:42.059598 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 16:29:42 crc kubenswrapper[4772]: I0127 16:29:42.060638 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 16:29:42 crc kubenswrapper[4772]: I0127 16:29:42.060765 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" gracePeriod=600 Jan 27 16:29:42 crc kubenswrapper[4772]: E0127 16:29:42.183259 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:29:42 crc kubenswrapper[4772]: I0127 16:29:42.349318 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" exitCode=0 Jan 27 16:29:42 crc kubenswrapper[4772]: I0127 16:29:42.349386 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b"} Jan 27 16:29:42 crc kubenswrapper[4772]: I0127 16:29:42.349463 4772 scope.go:117] "RemoveContainer" containerID="f3a08a71f69d769f4a6a29d6cef13873c9dceaa6515bc086fcafc82c5f73a041" Jan 27 16:29:42 crc kubenswrapper[4772]: I0127 16:29:42.350083 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:29:42 crc kubenswrapper[4772]: E0127 16:29:42.350331 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:29:55 crc kubenswrapper[4772]: I0127 16:29:55.663254 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:29:55 crc kubenswrapper[4772]: E0127 16:29:55.663990 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:30:00 crc kubenswrapper[4772]: I0127 16:30:00.144406 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492190-nkwl9"] Jan 27 16:30:00 crc kubenswrapper[4772]: E0127 16:30:00.145466 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96c99cc4-7be7-49a2-bbc4-a16c3698e291" containerName="mariadb-client" Jan 27 16:30:00 crc kubenswrapper[4772]: I0127 16:30:00.145484 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="96c99cc4-7be7-49a2-bbc4-a16c3698e291" containerName="mariadb-client" Jan 27 16:30:00 crc kubenswrapper[4772]: I0127 16:30:00.145711 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="96c99cc4-7be7-49a2-bbc4-a16c3698e291" containerName="mariadb-client" Jan 27 16:30:00 crc kubenswrapper[4772]: I0127 16:30:00.146409 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492190-nkwl9" Jan 27 16:30:00 crc kubenswrapper[4772]: I0127 16:30:00.151979 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 27 16:30:00 crc kubenswrapper[4772]: I0127 16:30:00.152222 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 27 16:30:00 crc kubenswrapper[4772]: I0127 16:30:00.153439 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492190-nkwl9"] Jan 27 16:30:00 crc kubenswrapper[4772]: I0127 16:30:00.235224 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aad11684-a5b7-4df1-9d18-5179c6113f66-secret-volume\") pod \"collect-profiles-29492190-nkwl9\" (UID: \"aad11684-a5b7-4df1-9d18-5179c6113f66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492190-nkwl9" Jan 27 16:30:00 crc kubenswrapper[4772]: I0127 16:30:00.235303 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cq4h\" (UniqueName: \"kubernetes.io/projected/aad11684-a5b7-4df1-9d18-5179c6113f66-kube-api-access-8cq4h\") pod \"collect-profiles-29492190-nkwl9\" (UID: \"aad11684-a5b7-4df1-9d18-5179c6113f66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492190-nkwl9" Jan 27 16:30:00 crc kubenswrapper[4772]: I0127 16:30:00.235411 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aad11684-a5b7-4df1-9d18-5179c6113f66-config-volume\") pod \"collect-profiles-29492190-nkwl9\" (UID: \"aad11684-a5b7-4df1-9d18-5179c6113f66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492190-nkwl9" Jan 27 16:30:00 crc kubenswrapper[4772]: I0127 16:30:00.336965 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aad11684-a5b7-4df1-9d18-5179c6113f66-secret-volume\") pod \"collect-profiles-29492190-nkwl9\" (UID: \"aad11684-a5b7-4df1-9d18-5179c6113f66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492190-nkwl9" Jan 27 16:30:00 crc kubenswrapper[4772]: I0127 16:30:00.337012 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cq4h\" (UniqueName: \"kubernetes.io/projected/aad11684-a5b7-4df1-9d18-5179c6113f66-kube-api-access-8cq4h\") pod \"collect-profiles-29492190-nkwl9\" (UID: \"aad11684-a5b7-4df1-9d18-5179c6113f66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492190-nkwl9" Jan 27 16:30:00 crc kubenswrapper[4772]: I0127 16:30:00.337056 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aad11684-a5b7-4df1-9d18-5179c6113f66-config-volume\") pod \"collect-profiles-29492190-nkwl9\" (UID: \"aad11684-a5b7-4df1-9d18-5179c6113f66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492190-nkwl9" Jan 27 16:30:00 crc kubenswrapper[4772]: I0127 16:30:00.337938 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aad11684-a5b7-4df1-9d18-5179c6113f66-config-volume\") pod \"collect-profiles-29492190-nkwl9\" (UID: \"aad11684-a5b7-4df1-9d18-5179c6113f66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492190-nkwl9" Jan 27 16:30:00 crc kubenswrapper[4772]: I0127 16:30:00.343293 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aad11684-a5b7-4df1-9d18-5179c6113f66-secret-volume\") pod \"collect-profiles-29492190-nkwl9\" (UID: \"aad11684-a5b7-4df1-9d18-5179c6113f66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492190-nkwl9" Jan 27 16:30:00 crc kubenswrapper[4772]: I0127 16:30:00.358382 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cq4h\" (UniqueName: \"kubernetes.io/projected/aad11684-a5b7-4df1-9d18-5179c6113f66-kube-api-access-8cq4h\") pod \"collect-profiles-29492190-nkwl9\" (UID: \"aad11684-a5b7-4df1-9d18-5179c6113f66\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492190-nkwl9" Jan 27 16:30:00 crc kubenswrapper[4772]: I0127 16:30:00.470198 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492190-nkwl9" Jan 27 16:30:00 crc kubenswrapper[4772]: I0127 16:30:00.884077 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492190-nkwl9"] Jan 27 16:30:01 crc kubenswrapper[4772]: I0127 16:30:01.504966 4772 generic.go:334] "Generic (PLEG): container finished" podID="aad11684-a5b7-4df1-9d18-5179c6113f66" containerID="114edebee04cbeb82762a8f7e28bf44b5665934fa2746ec43b3a0a20d9084515" exitCode=0 Jan 27 16:30:01 crc kubenswrapper[4772]: I0127 16:30:01.505013 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492190-nkwl9" event={"ID":"aad11684-a5b7-4df1-9d18-5179c6113f66","Type":"ContainerDied","Data":"114edebee04cbeb82762a8f7e28bf44b5665934fa2746ec43b3a0a20d9084515"} Jan 27 16:30:01 crc kubenswrapper[4772]: I0127 16:30:01.505039 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492190-nkwl9" event={"ID":"aad11684-a5b7-4df1-9d18-5179c6113f66","Type":"ContainerStarted","Data":"fb69b2bdd58763a29639c10b7607422660e39cd3e742d1e9cb84e4ab63d757cb"} Jan 27 16:30:02 crc kubenswrapper[4772]: I0127 16:30:02.778672 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492190-nkwl9" Jan 27 16:30:02 crc kubenswrapper[4772]: I0127 16:30:02.878995 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aad11684-a5b7-4df1-9d18-5179c6113f66-config-volume\") pod \"aad11684-a5b7-4df1-9d18-5179c6113f66\" (UID: \"aad11684-a5b7-4df1-9d18-5179c6113f66\") " Jan 27 16:30:02 crc kubenswrapper[4772]: I0127 16:30:02.879188 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cq4h\" (UniqueName: \"kubernetes.io/projected/aad11684-a5b7-4df1-9d18-5179c6113f66-kube-api-access-8cq4h\") pod \"aad11684-a5b7-4df1-9d18-5179c6113f66\" (UID: \"aad11684-a5b7-4df1-9d18-5179c6113f66\") " Jan 27 16:30:02 crc kubenswrapper[4772]: I0127 16:30:02.879315 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aad11684-a5b7-4df1-9d18-5179c6113f66-secret-volume\") pod \"aad11684-a5b7-4df1-9d18-5179c6113f66\" (UID: \"aad11684-a5b7-4df1-9d18-5179c6113f66\") " Jan 27 16:30:02 crc kubenswrapper[4772]: I0127 16:30:02.879928 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aad11684-a5b7-4df1-9d18-5179c6113f66-config-volume" (OuterVolumeSpecName: "config-volume") pod "aad11684-a5b7-4df1-9d18-5179c6113f66" (UID: "aad11684-a5b7-4df1-9d18-5179c6113f66"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:30:02 crc kubenswrapper[4772]: I0127 16:30:02.885922 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aad11684-a5b7-4df1-9d18-5179c6113f66-kube-api-access-8cq4h" (OuterVolumeSpecName: "kube-api-access-8cq4h") pod "aad11684-a5b7-4df1-9d18-5179c6113f66" (UID: "aad11684-a5b7-4df1-9d18-5179c6113f66"). InnerVolumeSpecName "kube-api-access-8cq4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:30:02 crc kubenswrapper[4772]: I0127 16:30:02.886301 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad11684-a5b7-4df1-9d18-5179c6113f66-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "aad11684-a5b7-4df1-9d18-5179c6113f66" (UID: "aad11684-a5b7-4df1-9d18-5179c6113f66"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:30:02 crc kubenswrapper[4772]: I0127 16:30:02.980531 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cq4h\" (UniqueName: \"kubernetes.io/projected/aad11684-a5b7-4df1-9d18-5179c6113f66-kube-api-access-8cq4h\") on node \"crc\" DevicePath \"\"" Jan 27 16:30:02 crc kubenswrapper[4772]: I0127 16:30:02.980786 4772 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aad11684-a5b7-4df1-9d18-5179c6113f66-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 27 16:30:02 crc kubenswrapper[4772]: I0127 16:30:02.980796 4772 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aad11684-a5b7-4df1-9d18-5179c6113f66-config-volume\") on node \"crc\" DevicePath \"\"" Jan 27 16:30:03 crc kubenswrapper[4772]: I0127 16:30:03.523036 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492190-nkwl9" event={"ID":"aad11684-a5b7-4df1-9d18-5179c6113f66","Type":"ContainerDied","Data":"fb69b2bdd58763a29639c10b7607422660e39cd3e742d1e9cb84e4ab63d757cb"} Jan 27 16:30:03 crc kubenswrapper[4772]: I0127 16:30:03.523082 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb69b2bdd58763a29639c10b7607422660e39cd3e742d1e9cb84e4ab63d757cb" Jan 27 16:30:03 crc kubenswrapper[4772]: I0127 16:30:03.523094 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492190-nkwl9" Jan 27 16:30:03 crc kubenswrapper[4772]: I0127 16:30:03.871853 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492145-st6dd"] Jan 27 16:30:03 crc kubenswrapper[4772]: I0127 16:30:03.876874 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492145-st6dd"] Jan 27 16:30:04 crc kubenswrapper[4772]: I0127 16:30:04.671488 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df82c0c4-9652-407e-b63d-17e2ccdb38aa" path="/var/lib/kubelet/pods/df82c0c4-9652-407e-b63d-17e2ccdb38aa/volumes" Jan 27 16:30:06 crc kubenswrapper[4772]: I0127 16:30:06.483484 4772 scope.go:117] "RemoveContainer" containerID="72ea0a33955c0509b888997e5b6ca0dc68de786a608fe5aae9035bbbf84ae773" Jan 27 16:30:06 crc kubenswrapper[4772]: I0127 16:30:06.664277 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:30:06 crc kubenswrapper[4772]: E0127 16:30:06.665023 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:30:18 crc kubenswrapper[4772]: I0127 16:30:18.664663 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:30:18 crc kubenswrapper[4772]: E0127 16:30:18.667681 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:30:29 crc kubenswrapper[4772]: I0127 16:30:29.663706 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:30:29 crc kubenswrapper[4772]: E0127 16:30:29.665805 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:30:41 crc kubenswrapper[4772]: I0127 16:30:41.664080 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:30:41 crc kubenswrapper[4772]: E0127 16:30:41.665415 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:30:44 crc kubenswrapper[4772]: I0127 16:30:44.996458 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qcsqq"] Jan 27 16:30:44 crc kubenswrapper[4772]: E0127 16:30:44.996988 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aad11684-a5b7-4df1-9d18-5179c6113f66" containerName="collect-profiles" Jan 27 16:30:44 crc kubenswrapper[4772]: I0127 16:30:44.996999 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="aad11684-a5b7-4df1-9d18-5179c6113f66" containerName="collect-profiles" Jan 27 16:30:44 crc kubenswrapper[4772]: I0127 16:30:44.997134 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="aad11684-a5b7-4df1-9d18-5179c6113f66" containerName="collect-profiles" Jan 27 16:30:44 crc kubenswrapper[4772]: I0127 16:30:44.998364 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qcsqq" Jan 27 16:30:45 crc kubenswrapper[4772]: I0127 16:30:45.012725 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qcsqq"] Jan 27 16:30:45 crc kubenswrapper[4772]: I0127 16:30:45.070442 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1546001c-59c1-4641-b1a6-cfd263698406-utilities\") pod \"redhat-marketplace-qcsqq\" (UID: \"1546001c-59c1-4641-b1a6-cfd263698406\") " pod="openshift-marketplace/redhat-marketplace-qcsqq" Jan 27 16:30:45 crc kubenswrapper[4772]: I0127 16:30:45.070507 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5r6z\" (UniqueName: \"kubernetes.io/projected/1546001c-59c1-4641-b1a6-cfd263698406-kube-api-access-n5r6z\") pod \"redhat-marketplace-qcsqq\" (UID: \"1546001c-59c1-4641-b1a6-cfd263698406\") " pod="openshift-marketplace/redhat-marketplace-qcsqq" Jan 27 16:30:45 crc kubenswrapper[4772]: I0127 16:30:45.070588 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1546001c-59c1-4641-b1a6-cfd263698406-catalog-content\") pod \"redhat-marketplace-qcsqq\" (UID: \"1546001c-59c1-4641-b1a6-cfd263698406\") " pod="openshift-marketplace/redhat-marketplace-qcsqq" Jan 27 16:30:45 crc kubenswrapper[4772]: I0127 16:30:45.172074 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1546001c-59c1-4641-b1a6-cfd263698406-utilities\") pod \"redhat-marketplace-qcsqq\" (UID: \"1546001c-59c1-4641-b1a6-cfd263698406\") " pod="openshift-marketplace/redhat-marketplace-qcsqq" Jan 27 16:30:45 crc kubenswrapper[4772]: I0127 16:30:45.172146 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5r6z\" (UniqueName: \"kubernetes.io/projected/1546001c-59c1-4641-b1a6-cfd263698406-kube-api-access-n5r6z\") pod \"redhat-marketplace-qcsqq\" (UID: \"1546001c-59c1-4641-b1a6-cfd263698406\") " pod="openshift-marketplace/redhat-marketplace-qcsqq" Jan 27 16:30:45 crc kubenswrapper[4772]: I0127 16:30:45.172224 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1546001c-59c1-4641-b1a6-cfd263698406-catalog-content\") pod \"redhat-marketplace-qcsqq\" (UID: \"1546001c-59c1-4641-b1a6-cfd263698406\") " pod="openshift-marketplace/redhat-marketplace-qcsqq" Jan 27 16:30:45 crc kubenswrapper[4772]: I0127 16:30:45.172788 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1546001c-59c1-4641-b1a6-cfd263698406-catalog-content\") pod \"redhat-marketplace-qcsqq\" (UID: \"1546001c-59c1-4641-b1a6-cfd263698406\") " pod="openshift-marketplace/redhat-marketplace-qcsqq" Jan 27 16:30:45 crc kubenswrapper[4772]: I0127 16:30:45.172784 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1546001c-59c1-4641-b1a6-cfd263698406-utilities\") pod \"redhat-marketplace-qcsqq\" (UID: \"1546001c-59c1-4641-b1a6-cfd263698406\") " pod="openshift-marketplace/redhat-marketplace-qcsqq" Jan 27 16:30:45 crc kubenswrapper[4772]: I0127 16:30:45.191767 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5r6z\" (UniqueName: \"kubernetes.io/projected/1546001c-59c1-4641-b1a6-cfd263698406-kube-api-access-n5r6z\") pod \"redhat-marketplace-qcsqq\" (UID: \"1546001c-59c1-4641-b1a6-cfd263698406\") " pod="openshift-marketplace/redhat-marketplace-qcsqq" Jan 27 16:30:45 crc kubenswrapper[4772]: I0127 16:30:45.373361 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qcsqq" Jan 27 16:30:45 crc kubenswrapper[4772]: I0127 16:30:45.849051 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qcsqq"] Jan 27 16:30:45 crc kubenswrapper[4772]: I0127 16:30:45.889353 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcsqq" event={"ID":"1546001c-59c1-4641-b1a6-cfd263698406","Type":"ContainerStarted","Data":"5264e359274425b89a615b65a92f4223cfc875211991d693035591b51c1a420b"} Jan 27 16:30:46 crc kubenswrapper[4772]: I0127 16:30:46.897386 4772 generic.go:334] "Generic (PLEG): container finished" podID="1546001c-59c1-4641-b1a6-cfd263698406" containerID="52ff2eaac3c0328895089314015b1a4fde9c687e4fe951903045624423c75984" exitCode=0 Jan 27 16:30:46 crc kubenswrapper[4772]: I0127 16:30:46.897719 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcsqq" event={"ID":"1546001c-59c1-4641-b1a6-cfd263698406","Type":"ContainerDied","Data":"52ff2eaac3c0328895089314015b1a4fde9c687e4fe951903045624423c75984"} Jan 27 16:30:48 crc kubenswrapper[4772]: I0127 16:30:47.912248 4772 generic.go:334] "Generic (PLEG): container finished" podID="1546001c-59c1-4641-b1a6-cfd263698406" containerID="453a76da621aa580189ea4add3101e0ade10bc3693bb7f7aaeabf9887f214ee1" exitCode=0 Jan 27 16:30:48 crc kubenswrapper[4772]: I0127 16:30:47.912371 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcsqq" event={"ID":"1546001c-59c1-4641-b1a6-cfd263698406","Type":"ContainerDied","Data":"453a76da621aa580189ea4add3101e0ade10bc3693bb7f7aaeabf9887f214ee1"} Jan 27 16:30:48 crc kubenswrapper[4772]: I0127 16:30:48.920441 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcsqq" event={"ID":"1546001c-59c1-4641-b1a6-cfd263698406","Type":"ContainerStarted","Data":"6c1eb65e8c06dfb0b2a366344e6cf4c14004706415f43851d993ecc534ead2cf"} Jan 27 16:30:48 crc kubenswrapper[4772]: I0127 16:30:48.936502 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qcsqq" podStartSLOduration=3.404606659 podStartE2EDuration="4.936482557s" podCreationTimestamp="2026-01-27 16:30:44 +0000 UTC" firstStartedPulling="2026-01-27 16:30:46.898997168 +0000 UTC m=+5032.879606266" lastFinishedPulling="2026-01-27 16:30:48.430873066 +0000 UTC m=+5034.411482164" observedRunningTime="2026-01-27 16:30:48.933737148 +0000 UTC m=+5034.914346236" watchObservedRunningTime="2026-01-27 16:30:48.936482557 +0000 UTC m=+5034.917091655" Jan 27 16:30:52 crc kubenswrapper[4772]: I0127 16:30:52.663410 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:30:52 crc kubenswrapper[4772]: E0127 16:30:52.663973 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:30:55 crc kubenswrapper[4772]: I0127 16:30:55.374180 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qcsqq" Jan 27 16:30:55 crc kubenswrapper[4772]: I0127 16:30:55.374490 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qcsqq" Jan 27 16:30:55 crc kubenswrapper[4772]: I0127 16:30:55.417398 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qcsqq" Jan 27 16:30:56 crc kubenswrapper[4772]: I0127 16:30:56.014519 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qcsqq" Jan 27 16:30:56 crc kubenswrapper[4772]: I0127 16:30:56.062428 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qcsqq"] Jan 27 16:30:57 crc kubenswrapper[4772]: I0127 16:30:57.989117 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qcsqq" podUID="1546001c-59c1-4641-b1a6-cfd263698406" containerName="registry-server" containerID="cri-o://6c1eb65e8c06dfb0b2a366344e6cf4c14004706415f43851d993ecc534ead2cf" gracePeriod=2 Jan 27 16:30:58 crc kubenswrapper[4772]: I0127 16:30:58.903995 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qcsqq" Jan 27 16:30:58 crc kubenswrapper[4772]: I0127 16:30:58.999408 4772 generic.go:334] "Generic (PLEG): container finished" podID="1546001c-59c1-4641-b1a6-cfd263698406" containerID="6c1eb65e8c06dfb0b2a366344e6cf4c14004706415f43851d993ecc534ead2cf" exitCode=0 Jan 27 16:30:58 crc kubenswrapper[4772]: I0127 16:30:58.999461 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcsqq" event={"ID":"1546001c-59c1-4641-b1a6-cfd263698406","Type":"ContainerDied","Data":"6c1eb65e8c06dfb0b2a366344e6cf4c14004706415f43851d993ecc534ead2cf"} Jan 27 16:30:58 crc kubenswrapper[4772]: I0127 16:30:58.999493 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qcsqq" event={"ID":"1546001c-59c1-4641-b1a6-cfd263698406","Type":"ContainerDied","Data":"5264e359274425b89a615b65a92f4223cfc875211991d693035591b51c1a420b"} Jan 27 16:30:59 crc kubenswrapper[4772]: I0127 16:30:58.999493 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qcsqq" Jan 27 16:30:59 crc kubenswrapper[4772]: I0127 16:30:58.999576 4772 scope.go:117] "RemoveContainer" containerID="6c1eb65e8c06dfb0b2a366344e6cf4c14004706415f43851d993ecc534ead2cf" Jan 27 16:30:59 crc kubenswrapper[4772]: I0127 16:30:59.021103 4772 scope.go:117] "RemoveContainer" containerID="453a76da621aa580189ea4add3101e0ade10bc3693bb7f7aaeabf9887f214ee1" Jan 27 16:30:59 crc kubenswrapper[4772]: I0127 16:30:59.041158 4772 scope.go:117] "RemoveContainer" containerID="52ff2eaac3c0328895089314015b1a4fde9c687e4fe951903045624423c75984" Jan 27 16:30:59 crc kubenswrapper[4772]: I0127 16:30:59.074403 4772 scope.go:117] "RemoveContainer" containerID="6c1eb65e8c06dfb0b2a366344e6cf4c14004706415f43851d993ecc534ead2cf" Jan 27 16:30:59 crc kubenswrapper[4772]: E0127 16:30:59.074808 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c1eb65e8c06dfb0b2a366344e6cf4c14004706415f43851d993ecc534ead2cf\": container with ID starting with 6c1eb65e8c06dfb0b2a366344e6cf4c14004706415f43851d993ecc534ead2cf not found: ID does not exist" containerID="6c1eb65e8c06dfb0b2a366344e6cf4c14004706415f43851d993ecc534ead2cf" Jan 27 16:30:59 crc kubenswrapper[4772]: I0127 16:30:59.074857 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c1eb65e8c06dfb0b2a366344e6cf4c14004706415f43851d993ecc534ead2cf"} err="failed to get container status \"6c1eb65e8c06dfb0b2a366344e6cf4c14004706415f43851d993ecc534ead2cf\": rpc error: code = NotFound desc = could not find container \"6c1eb65e8c06dfb0b2a366344e6cf4c14004706415f43851d993ecc534ead2cf\": container with ID starting with 6c1eb65e8c06dfb0b2a366344e6cf4c14004706415f43851d993ecc534ead2cf not found: ID does not exist" Jan 27 16:30:59 crc kubenswrapper[4772]: I0127 16:30:59.074883 4772 scope.go:117] "RemoveContainer" containerID="453a76da621aa580189ea4add3101e0ade10bc3693bb7f7aaeabf9887f214ee1" Jan 27 16:30:59 crc kubenswrapper[4772]: E0127 16:30:59.075112 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"453a76da621aa580189ea4add3101e0ade10bc3693bb7f7aaeabf9887f214ee1\": container with ID starting with 453a76da621aa580189ea4add3101e0ade10bc3693bb7f7aaeabf9887f214ee1 not found: ID does not exist" containerID="453a76da621aa580189ea4add3101e0ade10bc3693bb7f7aaeabf9887f214ee1" Jan 27 16:30:59 crc kubenswrapper[4772]: I0127 16:30:59.075136 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"453a76da621aa580189ea4add3101e0ade10bc3693bb7f7aaeabf9887f214ee1"} err="failed to get container status \"453a76da621aa580189ea4add3101e0ade10bc3693bb7f7aaeabf9887f214ee1\": rpc error: code = NotFound desc = could not find container \"453a76da621aa580189ea4add3101e0ade10bc3693bb7f7aaeabf9887f214ee1\": container with ID starting with 453a76da621aa580189ea4add3101e0ade10bc3693bb7f7aaeabf9887f214ee1 not found: ID does not exist" Jan 27 16:30:59 crc kubenswrapper[4772]: I0127 16:30:59.075152 4772 scope.go:117] "RemoveContainer" containerID="52ff2eaac3c0328895089314015b1a4fde9c687e4fe951903045624423c75984" Jan 27 16:30:59 crc kubenswrapper[4772]: E0127 16:30:59.075420 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52ff2eaac3c0328895089314015b1a4fde9c687e4fe951903045624423c75984\": container with ID starting with 52ff2eaac3c0328895089314015b1a4fde9c687e4fe951903045624423c75984 not found: ID does not exist" containerID="52ff2eaac3c0328895089314015b1a4fde9c687e4fe951903045624423c75984" Jan 27 16:30:59 crc kubenswrapper[4772]: I0127 16:30:59.075446 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52ff2eaac3c0328895089314015b1a4fde9c687e4fe951903045624423c75984"} err="failed to get container status \"52ff2eaac3c0328895089314015b1a4fde9c687e4fe951903045624423c75984\": rpc error: code = NotFound desc = could not find container \"52ff2eaac3c0328895089314015b1a4fde9c687e4fe951903045624423c75984\": container with ID starting with 52ff2eaac3c0328895089314015b1a4fde9c687e4fe951903045624423c75984 not found: ID does not exist" Jan 27 16:30:59 crc kubenswrapper[4772]: I0127 16:30:59.089680 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1546001c-59c1-4641-b1a6-cfd263698406-utilities\") pod \"1546001c-59c1-4641-b1a6-cfd263698406\" (UID: \"1546001c-59c1-4641-b1a6-cfd263698406\") " Jan 27 16:30:59 crc kubenswrapper[4772]: I0127 16:30:59.089780 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1546001c-59c1-4641-b1a6-cfd263698406-catalog-content\") pod \"1546001c-59c1-4641-b1a6-cfd263698406\" (UID: \"1546001c-59c1-4641-b1a6-cfd263698406\") " Jan 27 16:30:59 crc kubenswrapper[4772]: I0127 16:30:59.089875 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5r6z\" (UniqueName: \"kubernetes.io/projected/1546001c-59c1-4641-b1a6-cfd263698406-kube-api-access-n5r6z\") pod \"1546001c-59c1-4641-b1a6-cfd263698406\" (UID: \"1546001c-59c1-4641-b1a6-cfd263698406\") " Jan 27 16:30:59 crc kubenswrapper[4772]: I0127 16:30:59.090935 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1546001c-59c1-4641-b1a6-cfd263698406-utilities" (OuterVolumeSpecName: "utilities") pod "1546001c-59c1-4641-b1a6-cfd263698406" (UID: "1546001c-59c1-4641-b1a6-cfd263698406"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:30:59 crc kubenswrapper[4772]: I0127 16:30:59.097109 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1546001c-59c1-4641-b1a6-cfd263698406-kube-api-access-n5r6z" (OuterVolumeSpecName: "kube-api-access-n5r6z") pod "1546001c-59c1-4641-b1a6-cfd263698406" (UID: "1546001c-59c1-4641-b1a6-cfd263698406"). InnerVolumeSpecName "kube-api-access-n5r6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:30:59 crc kubenswrapper[4772]: I0127 16:30:59.112780 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1546001c-59c1-4641-b1a6-cfd263698406-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1546001c-59c1-4641-b1a6-cfd263698406" (UID: "1546001c-59c1-4641-b1a6-cfd263698406"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:30:59 crc kubenswrapper[4772]: I0127 16:30:59.192287 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1546001c-59c1-4641-b1a6-cfd263698406-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 16:30:59 crc kubenswrapper[4772]: I0127 16:30:59.192834 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1546001c-59c1-4641-b1a6-cfd263698406-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 16:30:59 crc kubenswrapper[4772]: I0127 16:30:59.192862 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5r6z\" (UniqueName: \"kubernetes.io/projected/1546001c-59c1-4641-b1a6-cfd263698406-kube-api-access-n5r6z\") on node \"crc\" DevicePath \"\"" Jan 27 16:30:59 crc kubenswrapper[4772]: I0127 16:30:59.330676 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qcsqq"] Jan 27 16:30:59 crc kubenswrapper[4772]: I0127 16:30:59.336340 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qcsqq"] Jan 27 16:31:00 crc kubenswrapper[4772]: I0127 16:31:00.674144 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1546001c-59c1-4641-b1a6-cfd263698406" path="/var/lib/kubelet/pods/1546001c-59c1-4641-b1a6-cfd263698406/volumes" Jan 27 16:31:05 crc kubenswrapper[4772]: I0127 16:31:05.663756 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:31:05 crc kubenswrapper[4772]: E0127 16:31:05.665042 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:31:17 crc kubenswrapper[4772]: I0127 16:31:17.663512 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:31:17 crc kubenswrapper[4772]: E0127 16:31:17.664352 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:31:31 crc kubenswrapper[4772]: I0127 16:31:31.662999 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:31:31 crc kubenswrapper[4772]: E0127 16:31:31.663868 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:31:46 crc kubenswrapper[4772]: I0127 16:31:46.663084 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:31:46 crc kubenswrapper[4772]: E0127 16:31:46.663816 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:31:47 crc kubenswrapper[4772]: I0127 16:31:47.435821 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4v5t2"] Jan 27 16:31:47 crc kubenswrapper[4772]: E0127 16:31:47.436289 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1546001c-59c1-4641-b1a6-cfd263698406" containerName="extract-utilities" Jan 27 16:31:47 crc kubenswrapper[4772]: I0127 16:31:47.436309 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="1546001c-59c1-4641-b1a6-cfd263698406" containerName="extract-utilities" Jan 27 16:31:47 crc kubenswrapper[4772]: E0127 16:31:47.436330 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1546001c-59c1-4641-b1a6-cfd263698406" containerName="registry-server" Jan 27 16:31:47 crc kubenswrapper[4772]: I0127 16:31:47.436338 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="1546001c-59c1-4641-b1a6-cfd263698406" containerName="registry-server" Jan 27 16:31:47 crc kubenswrapper[4772]: E0127 16:31:47.436359 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1546001c-59c1-4641-b1a6-cfd263698406" containerName="extract-content" Jan 27 16:31:47 crc kubenswrapper[4772]: I0127 16:31:47.436368 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="1546001c-59c1-4641-b1a6-cfd263698406" containerName="extract-content" Jan 27 16:31:47 crc kubenswrapper[4772]: I0127 16:31:47.436542 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="1546001c-59c1-4641-b1a6-cfd263698406" containerName="registry-server" Jan 27 16:31:47 crc kubenswrapper[4772]: I0127 16:31:47.437812 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4v5t2" Jan 27 16:31:47 crc kubenswrapper[4772]: I0127 16:31:47.449096 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4v5t2"] Jan 27 16:31:47 crc kubenswrapper[4772]: I0127 16:31:47.626065 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c92585ec-743a-472c-b4dd-c2626dea5440-catalog-content\") pod \"certified-operators-4v5t2\" (UID: \"c92585ec-743a-472c-b4dd-c2626dea5440\") " pod="openshift-marketplace/certified-operators-4v5t2" Jan 27 16:31:47 crc kubenswrapper[4772]: I0127 16:31:47.626147 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c92585ec-743a-472c-b4dd-c2626dea5440-utilities\") pod \"certified-operators-4v5t2\" (UID: \"c92585ec-743a-472c-b4dd-c2626dea5440\") " pod="openshift-marketplace/certified-operators-4v5t2" Jan 27 16:31:47 crc kubenswrapper[4772]: I0127 16:31:47.626431 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvdgz\" (UniqueName: \"kubernetes.io/projected/c92585ec-743a-472c-b4dd-c2626dea5440-kube-api-access-gvdgz\") pod \"certified-operators-4v5t2\" (UID: \"c92585ec-743a-472c-b4dd-c2626dea5440\") " pod="openshift-marketplace/certified-operators-4v5t2" Jan 27 16:31:47 crc kubenswrapper[4772]: I0127 16:31:47.727885 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c92585ec-743a-472c-b4dd-c2626dea5440-utilities\") pod \"certified-operators-4v5t2\" (UID: \"c92585ec-743a-472c-b4dd-c2626dea5440\") " pod="openshift-marketplace/certified-operators-4v5t2" Jan 27 16:31:47 crc kubenswrapper[4772]: I0127 16:31:47.727983 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvdgz\" (UniqueName: \"kubernetes.io/projected/c92585ec-743a-472c-b4dd-c2626dea5440-kube-api-access-gvdgz\") pod \"certified-operators-4v5t2\" (UID: \"c92585ec-743a-472c-b4dd-c2626dea5440\") " pod="openshift-marketplace/certified-operators-4v5t2" Jan 27 16:31:47 crc kubenswrapper[4772]: I0127 16:31:47.728015 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c92585ec-743a-472c-b4dd-c2626dea5440-catalog-content\") pod \"certified-operators-4v5t2\" (UID: \"c92585ec-743a-472c-b4dd-c2626dea5440\") " pod="openshift-marketplace/certified-operators-4v5t2" Jan 27 16:31:47 crc kubenswrapper[4772]: I0127 16:31:47.728842 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c92585ec-743a-472c-b4dd-c2626dea5440-catalog-content\") pod \"certified-operators-4v5t2\" (UID: \"c92585ec-743a-472c-b4dd-c2626dea5440\") " pod="openshift-marketplace/certified-operators-4v5t2" Jan 27 16:31:47 crc kubenswrapper[4772]: I0127 16:31:47.728965 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c92585ec-743a-472c-b4dd-c2626dea5440-utilities\") pod \"certified-operators-4v5t2\" (UID: \"c92585ec-743a-472c-b4dd-c2626dea5440\") " pod="openshift-marketplace/certified-operators-4v5t2" Jan 27 16:31:47 crc kubenswrapper[4772]: I0127 16:31:47.749649 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvdgz\" (UniqueName: \"kubernetes.io/projected/c92585ec-743a-472c-b4dd-c2626dea5440-kube-api-access-gvdgz\") pod \"certified-operators-4v5t2\" (UID: \"c92585ec-743a-472c-b4dd-c2626dea5440\") " pod="openshift-marketplace/certified-operators-4v5t2" Jan 27 16:31:47 crc kubenswrapper[4772]: I0127 16:31:47.755545 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4v5t2" Jan 27 16:31:48 crc kubenswrapper[4772]: I0127 16:31:48.205870 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4v5t2"] Jan 27 16:31:48 crc kubenswrapper[4772]: I0127 16:31:48.400859 4772 generic.go:334] "Generic (PLEG): container finished" podID="c92585ec-743a-472c-b4dd-c2626dea5440" containerID="2397a0459258571ef674ff88dc7d488cbbcfec7938b44eaae891914f57c1e0fb" exitCode=0 Jan 27 16:31:48 crc kubenswrapper[4772]: I0127 16:31:48.400903 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4v5t2" event={"ID":"c92585ec-743a-472c-b4dd-c2626dea5440","Type":"ContainerDied","Data":"2397a0459258571ef674ff88dc7d488cbbcfec7938b44eaae891914f57c1e0fb"} Jan 27 16:31:48 crc kubenswrapper[4772]: I0127 16:31:48.400954 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4v5t2" event={"ID":"c92585ec-743a-472c-b4dd-c2626dea5440","Type":"ContainerStarted","Data":"96225b9958d3fda24c7fec9dc8c8c03d59127b7c25a0ead4b4398213ed1f54aa"} Jan 27 16:31:48 crc kubenswrapper[4772]: I0127 16:31:48.402476 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 16:31:50 crc kubenswrapper[4772]: I0127 16:31:50.417809 4772 generic.go:334] "Generic (PLEG): container finished" podID="c92585ec-743a-472c-b4dd-c2626dea5440" containerID="b61d7ffda814b78a2d4e085143511eac06adda78e765120d5bba47717b008345" exitCode=0 Jan 27 16:31:50 crc kubenswrapper[4772]: I0127 16:31:50.417864 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4v5t2" event={"ID":"c92585ec-743a-472c-b4dd-c2626dea5440","Type":"ContainerDied","Data":"b61d7ffda814b78a2d4e085143511eac06adda78e765120d5bba47717b008345"} Jan 27 16:31:51 crc kubenswrapper[4772]: I0127 16:31:51.427364 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4v5t2" event={"ID":"c92585ec-743a-472c-b4dd-c2626dea5440","Type":"ContainerStarted","Data":"349c99af13ca6ed29ac408784b47f3d087af6dd7e6d509533e354880f0af0c54"} Jan 27 16:31:51 crc kubenswrapper[4772]: I0127 16:31:51.454575 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4v5t2" podStartSLOduration=2.045539772 podStartE2EDuration="4.45455218s" podCreationTimestamp="2026-01-27 16:31:47 +0000 UTC" firstStartedPulling="2026-01-27 16:31:48.402204391 +0000 UTC m=+5094.382813489" lastFinishedPulling="2026-01-27 16:31:50.811216799 +0000 UTC m=+5096.791825897" observedRunningTime="2026-01-27 16:31:51.445032145 +0000 UTC m=+5097.425641243" watchObservedRunningTime="2026-01-27 16:31:51.45455218 +0000 UTC m=+5097.435161278" Jan 27 16:31:57 crc kubenswrapper[4772]: I0127 16:31:57.756643 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4v5t2" Jan 27 16:31:57 crc kubenswrapper[4772]: I0127 16:31:57.758133 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4v5t2" Jan 27 16:31:57 crc kubenswrapper[4772]: I0127 16:31:57.800443 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4v5t2" Jan 27 16:31:58 crc kubenswrapper[4772]: I0127 16:31:58.522427 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4v5t2" Jan 27 16:31:58 crc kubenswrapper[4772]: I0127 16:31:58.573979 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4v5t2"] Jan 27 16:32:00 crc kubenswrapper[4772]: I0127 16:32:00.492585 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4v5t2" podUID="c92585ec-743a-472c-b4dd-c2626dea5440" containerName="registry-server" containerID="cri-o://349c99af13ca6ed29ac408784b47f3d087af6dd7e6d509533e354880f0af0c54" gracePeriod=2 Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.489870 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4v5t2" Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.508626 4772 generic.go:334] "Generic (PLEG): container finished" podID="c92585ec-743a-472c-b4dd-c2626dea5440" containerID="349c99af13ca6ed29ac408784b47f3d087af6dd7e6d509533e354880f0af0c54" exitCode=0 Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.508679 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4v5t2" event={"ID":"c92585ec-743a-472c-b4dd-c2626dea5440","Type":"ContainerDied","Data":"349c99af13ca6ed29ac408784b47f3d087af6dd7e6d509533e354880f0af0c54"} Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.508696 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4v5t2" Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.508712 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4v5t2" event={"ID":"c92585ec-743a-472c-b4dd-c2626dea5440","Type":"ContainerDied","Data":"96225b9958d3fda24c7fec9dc8c8c03d59127b7c25a0ead4b4398213ed1f54aa"} Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.508738 4772 scope.go:117] "RemoveContainer" containerID="349c99af13ca6ed29ac408784b47f3d087af6dd7e6d509533e354880f0af0c54" Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.527410 4772 scope.go:117] "RemoveContainer" containerID="b61d7ffda814b78a2d4e085143511eac06adda78e765120d5bba47717b008345" Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.545996 4772 scope.go:117] "RemoveContainer" containerID="2397a0459258571ef674ff88dc7d488cbbcfec7938b44eaae891914f57c1e0fb" Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.578543 4772 scope.go:117] "RemoveContainer" containerID="349c99af13ca6ed29ac408784b47f3d087af6dd7e6d509533e354880f0af0c54" Jan 27 16:32:01 crc kubenswrapper[4772]: E0127 16:32:01.578940 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"349c99af13ca6ed29ac408784b47f3d087af6dd7e6d509533e354880f0af0c54\": container with ID starting with 349c99af13ca6ed29ac408784b47f3d087af6dd7e6d509533e354880f0af0c54 not found: ID does not exist" containerID="349c99af13ca6ed29ac408784b47f3d087af6dd7e6d509533e354880f0af0c54" Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.578973 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"349c99af13ca6ed29ac408784b47f3d087af6dd7e6d509533e354880f0af0c54"} err="failed to get container status \"349c99af13ca6ed29ac408784b47f3d087af6dd7e6d509533e354880f0af0c54\": rpc error: code = NotFound desc = could not find container \"349c99af13ca6ed29ac408784b47f3d087af6dd7e6d509533e354880f0af0c54\": container with ID starting with 349c99af13ca6ed29ac408784b47f3d087af6dd7e6d509533e354880f0af0c54 not found: ID does not exist" Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.578994 4772 scope.go:117] "RemoveContainer" containerID="b61d7ffda814b78a2d4e085143511eac06adda78e765120d5bba47717b008345" Jan 27 16:32:01 crc kubenswrapper[4772]: E0127 16:32:01.579389 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b61d7ffda814b78a2d4e085143511eac06adda78e765120d5bba47717b008345\": container with ID starting with b61d7ffda814b78a2d4e085143511eac06adda78e765120d5bba47717b008345 not found: ID does not exist" containerID="b61d7ffda814b78a2d4e085143511eac06adda78e765120d5bba47717b008345" Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.579445 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b61d7ffda814b78a2d4e085143511eac06adda78e765120d5bba47717b008345"} err="failed to get container status \"b61d7ffda814b78a2d4e085143511eac06adda78e765120d5bba47717b008345\": rpc error: code = NotFound desc = could not find container \"b61d7ffda814b78a2d4e085143511eac06adda78e765120d5bba47717b008345\": container with ID starting with b61d7ffda814b78a2d4e085143511eac06adda78e765120d5bba47717b008345 not found: ID does not exist" Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.579490 4772 scope.go:117] "RemoveContainer" containerID="2397a0459258571ef674ff88dc7d488cbbcfec7938b44eaae891914f57c1e0fb" Jan 27 16:32:01 crc kubenswrapper[4772]: E0127 16:32:01.579824 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2397a0459258571ef674ff88dc7d488cbbcfec7938b44eaae891914f57c1e0fb\": container with ID starting with 2397a0459258571ef674ff88dc7d488cbbcfec7938b44eaae891914f57c1e0fb not found: ID does not exist" containerID="2397a0459258571ef674ff88dc7d488cbbcfec7938b44eaae891914f57c1e0fb" Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.579853 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2397a0459258571ef674ff88dc7d488cbbcfec7938b44eaae891914f57c1e0fb"} err="failed to get container status \"2397a0459258571ef674ff88dc7d488cbbcfec7938b44eaae891914f57c1e0fb\": rpc error: code = NotFound desc = could not find container \"2397a0459258571ef674ff88dc7d488cbbcfec7938b44eaae891914f57c1e0fb\": container with ID starting with 2397a0459258571ef674ff88dc7d488cbbcfec7938b44eaae891914f57c1e0fb not found: ID does not exist" Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.661747 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c92585ec-743a-472c-b4dd-c2626dea5440-utilities\") pod \"c92585ec-743a-472c-b4dd-c2626dea5440\" (UID: \"c92585ec-743a-472c-b4dd-c2626dea5440\") " Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.662375 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:32:01 crc kubenswrapper[4772]: E0127 16:32:01.662651 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.662959 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c92585ec-743a-472c-b4dd-c2626dea5440-utilities" (OuterVolumeSpecName: "utilities") pod "c92585ec-743a-472c-b4dd-c2626dea5440" (UID: "c92585ec-743a-472c-b4dd-c2626dea5440"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.663019 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c92585ec-743a-472c-b4dd-c2626dea5440-catalog-content\") pod \"c92585ec-743a-472c-b4dd-c2626dea5440\" (UID: \"c92585ec-743a-472c-b4dd-c2626dea5440\") " Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.666452 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvdgz\" (UniqueName: \"kubernetes.io/projected/c92585ec-743a-472c-b4dd-c2626dea5440-kube-api-access-gvdgz\") pod \"c92585ec-743a-472c-b4dd-c2626dea5440\" (UID: \"c92585ec-743a-472c-b4dd-c2626dea5440\") " Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.667101 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c92585ec-743a-472c-b4dd-c2626dea5440-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.673984 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c92585ec-743a-472c-b4dd-c2626dea5440-kube-api-access-gvdgz" (OuterVolumeSpecName: "kube-api-access-gvdgz") pod "c92585ec-743a-472c-b4dd-c2626dea5440" (UID: "c92585ec-743a-472c-b4dd-c2626dea5440"). InnerVolumeSpecName "kube-api-access-gvdgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.708955 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c92585ec-743a-472c-b4dd-c2626dea5440-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c92585ec-743a-472c-b4dd-c2626dea5440" (UID: "c92585ec-743a-472c-b4dd-c2626dea5440"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.768401 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c92585ec-743a-472c-b4dd-c2626dea5440-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.768434 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvdgz\" (UniqueName: \"kubernetes.io/projected/c92585ec-743a-472c-b4dd-c2626dea5440-kube-api-access-gvdgz\") on node \"crc\" DevicePath \"\"" Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.835624 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4v5t2"] Jan 27 16:32:01 crc kubenswrapper[4772]: I0127 16:32:01.842197 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4v5t2"] Jan 27 16:32:02 crc kubenswrapper[4772]: I0127 16:32:02.670762 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c92585ec-743a-472c-b4dd-c2626dea5440" path="/var/lib/kubelet/pods/c92585ec-743a-472c-b4dd-c2626dea5440/volumes" Jan 27 16:32:14 crc kubenswrapper[4772]: I0127 16:32:14.666761 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:32:14 crc kubenswrapper[4772]: E0127 16:32:14.667454 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:32:23 crc kubenswrapper[4772]: I0127 16:32:23.093196 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Jan 27 16:32:23 crc kubenswrapper[4772]: E0127 16:32:23.094234 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c92585ec-743a-472c-b4dd-c2626dea5440" containerName="registry-server" Jan 27 16:32:23 crc kubenswrapper[4772]: I0127 16:32:23.094248 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="c92585ec-743a-472c-b4dd-c2626dea5440" containerName="registry-server" Jan 27 16:32:23 crc kubenswrapper[4772]: E0127 16:32:23.094258 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c92585ec-743a-472c-b4dd-c2626dea5440" containerName="extract-content" Jan 27 16:32:23 crc kubenswrapper[4772]: I0127 16:32:23.094265 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="c92585ec-743a-472c-b4dd-c2626dea5440" containerName="extract-content" Jan 27 16:32:23 crc kubenswrapper[4772]: E0127 16:32:23.094280 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c92585ec-743a-472c-b4dd-c2626dea5440" containerName="extract-utilities" Jan 27 16:32:23 crc kubenswrapper[4772]: I0127 16:32:23.094286 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="c92585ec-743a-472c-b4dd-c2626dea5440" containerName="extract-utilities" Jan 27 16:32:23 crc kubenswrapper[4772]: I0127 16:32:23.094451 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="c92585ec-743a-472c-b4dd-c2626dea5440" containerName="registry-server" Jan 27 16:32:23 crc kubenswrapper[4772]: I0127 16:32:23.094966 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Jan 27 16:32:23 crc kubenswrapper[4772]: I0127 16:32:23.097565 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-jd6dc" Jan 27 16:32:23 crc kubenswrapper[4772]: I0127 16:32:23.100377 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Jan 27 16:32:23 crc kubenswrapper[4772]: I0127 16:32:23.191153 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f6fda35a-22da-4119-a41a-9a4f5c51027a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f6fda35a-22da-4119-a41a-9a4f5c51027a\") pod \"mariadb-copy-data\" (UID: \"7db35434-01e2-470d-bb27-8e30189936b3\") " pod="openstack/mariadb-copy-data" Jan 27 16:32:23 crc kubenswrapper[4772]: I0127 16:32:23.191285 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km8d9\" (UniqueName: \"kubernetes.io/projected/7db35434-01e2-470d-bb27-8e30189936b3-kube-api-access-km8d9\") pod \"mariadb-copy-data\" (UID: \"7db35434-01e2-470d-bb27-8e30189936b3\") " pod="openstack/mariadb-copy-data" Jan 27 16:32:23 crc kubenswrapper[4772]: I0127 16:32:23.292931 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f6fda35a-22da-4119-a41a-9a4f5c51027a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f6fda35a-22da-4119-a41a-9a4f5c51027a\") pod \"mariadb-copy-data\" (UID: \"7db35434-01e2-470d-bb27-8e30189936b3\") " pod="openstack/mariadb-copy-data" Jan 27 16:32:23 crc kubenswrapper[4772]: I0127 16:32:23.293027 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km8d9\" (UniqueName: \"kubernetes.io/projected/7db35434-01e2-470d-bb27-8e30189936b3-kube-api-access-km8d9\") pod \"mariadb-copy-data\" (UID: \"7db35434-01e2-470d-bb27-8e30189936b3\") " pod="openstack/mariadb-copy-data" Jan 27 16:32:23 crc kubenswrapper[4772]: I0127 16:32:23.296040 4772 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 27 16:32:23 crc kubenswrapper[4772]: I0127 16:32:23.296092 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f6fda35a-22da-4119-a41a-9a4f5c51027a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f6fda35a-22da-4119-a41a-9a4f5c51027a\") pod \"mariadb-copy-data\" (UID: \"7db35434-01e2-470d-bb27-8e30189936b3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/791abbe3041f2db8530b425c87a9deb2b029505cd287d16d9fe14f995d5e5eb5/globalmount\"" pod="openstack/mariadb-copy-data" Jan 27 16:32:23 crc kubenswrapper[4772]: I0127 16:32:23.317970 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km8d9\" (UniqueName: \"kubernetes.io/projected/7db35434-01e2-470d-bb27-8e30189936b3-kube-api-access-km8d9\") pod \"mariadb-copy-data\" (UID: \"7db35434-01e2-470d-bb27-8e30189936b3\") " pod="openstack/mariadb-copy-data" Jan 27 16:32:23 crc kubenswrapper[4772]: I0127 16:32:23.331720 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f6fda35a-22da-4119-a41a-9a4f5c51027a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f6fda35a-22da-4119-a41a-9a4f5c51027a\") pod \"mariadb-copy-data\" (UID: \"7db35434-01e2-470d-bb27-8e30189936b3\") " pod="openstack/mariadb-copy-data" Jan 27 16:32:23 crc kubenswrapper[4772]: I0127 16:32:23.417932 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Jan 27 16:32:23 crc kubenswrapper[4772]: I0127 16:32:23.936063 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Jan 27 16:32:24 crc kubenswrapper[4772]: I0127 16:32:24.680625 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"7db35434-01e2-470d-bb27-8e30189936b3","Type":"ContainerStarted","Data":"1d7029faad9dc2bef8435c8c3a74852f826537085f88f1da26eca62aad4328e8"} Jan 27 16:32:24 crc kubenswrapper[4772]: I0127 16:32:24.680988 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"7db35434-01e2-470d-bb27-8e30189936b3","Type":"ContainerStarted","Data":"db1e9e0db949f173719176328889059cc417d2987fd47f63a4b1be6e62827be5"} Jan 27 16:32:24 crc kubenswrapper[4772]: I0127 16:32:24.699649 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=2.699630659 podStartE2EDuration="2.699630659s" podCreationTimestamp="2026-01-27 16:32:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:32:24.698496767 +0000 UTC m=+5130.679105885" watchObservedRunningTime="2026-01-27 16:32:24.699630659 +0000 UTC m=+5130.680239757" Jan 27 16:32:27 crc kubenswrapper[4772]: I0127 16:32:27.357113 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Jan 27 16:32:27 crc kubenswrapper[4772]: I0127 16:32:27.358689 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Jan 27 16:32:27 crc kubenswrapper[4772]: I0127 16:32:27.368999 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Jan 27 16:32:27 crc kubenswrapper[4772]: I0127 16:32:27.454502 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fpxp\" (UniqueName: \"kubernetes.io/projected/4adf565d-c266-4781-9107-fd05a16d3a53-kube-api-access-5fpxp\") pod \"mariadb-client\" (UID: \"4adf565d-c266-4781-9107-fd05a16d3a53\") " pod="openstack/mariadb-client" Jan 27 16:32:27 crc kubenswrapper[4772]: I0127 16:32:27.555688 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fpxp\" (UniqueName: \"kubernetes.io/projected/4adf565d-c266-4781-9107-fd05a16d3a53-kube-api-access-5fpxp\") pod \"mariadb-client\" (UID: \"4adf565d-c266-4781-9107-fd05a16d3a53\") " pod="openstack/mariadb-client" Jan 27 16:32:27 crc kubenswrapper[4772]: I0127 16:32:27.576258 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fpxp\" (UniqueName: \"kubernetes.io/projected/4adf565d-c266-4781-9107-fd05a16d3a53-kube-api-access-5fpxp\") pod \"mariadb-client\" (UID: \"4adf565d-c266-4781-9107-fd05a16d3a53\") " pod="openstack/mariadb-client" Jan 27 16:32:27 crc kubenswrapper[4772]: I0127 16:32:27.692248 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Jan 27 16:32:28 crc kubenswrapper[4772]: I0127 16:32:28.106321 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Jan 27 16:32:28 crc kubenswrapper[4772]: I0127 16:32:28.708074 4772 generic.go:334] "Generic (PLEG): container finished" podID="4adf565d-c266-4781-9107-fd05a16d3a53" containerID="f861f93267509f86f0bf58147bbab5f93435b373f74dada3332239c3cced3199" exitCode=0 Jan 27 16:32:28 crc kubenswrapper[4772]: I0127 16:32:28.708196 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"4adf565d-c266-4781-9107-fd05a16d3a53","Type":"ContainerDied","Data":"f861f93267509f86f0bf58147bbab5f93435b373f74dada3332239c3cced3199"} Jan 27 16:32:28 crc kubenswrapper[4772]: I0127 16:32:28.708413 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"4adf565d-c266-4781-9107-fd05a16d3a53","Type":"ContainerStarted","Data":"0f87006e2d0c3ada96d5c2183b202a9d37b5b94d9a888b23022a549612fbd148"} Jan 27 16:32:29 crc kubenswrapper[4772]: I0127 16:32:29.663290 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:32:29 crc kubenswrapper[4772]: E0127 16:32:29.663583 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:32:30 crc kubenswrapper[4772]: I0127 16:32:30.028377 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Jan 27 16:32:30 crc kubenswrapper[4772]: I0127 16:32:30.050699 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_4adf565d-c266-4781-9107-fd05a16d3a53/mariadb-client/0.log" Jan 27 16:32:30 crc kubenswrapper[4772]: I0127 16:32:30.074111 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Jan 27 16:32:30 crc kubenswrapper[4772]: I0127 16:32:30.080340 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Jan 27 16:32:30 crc kubenswrapper[4772]: I0127 16:32:30.190707 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Jan 27 16:32:30 crc kubenswrapper[4772]: E0127 16:32:30.191417 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4adf565d-c266-4781-9107-fd05a16d3a53" containerName="mariadb-client" Jan 27 16:32:30 crc kubenswrapper[4772]: I0127 16:32:30.191440 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4adf565d-c266-4781-9107-fd05a16d3a53" containerName="mariadb-client" Jan 27 16:32:30 crc kubenswrapper[4772]: I0127 16:32:30.191649 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="4adf565d-c266-4781-9107-fd05a16d3a53" containerName="mariadb-client" Jan 27 16:32:30 crc kubenswrapper[4772]: I0127 16:32:30.192514 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Jan 27 16:32:30 crc kubenswrapper[4772]: I0127 16:32:30.197931 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Jan 27 16:32:30 crc kubenswrapper[4772]: I0127 16:32:30.198483 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fpxp\" (UniqueName: \"kubernetes.io/projected/4adf565d-c266-4781-9107-fd05a16d3a53-kube-api-access-5fpxp\") pod \"4adf565d-c266-4781-9107-fd05a16d3a53\" (UID: \"4adf565d-c266-4781-9107-fd05a16d3a53\") " Jan 27 16:32:30 crc kubenswrapper[4772]: I0127 16:32:30.207528 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4adf565d-c266-4781-9107-fd05a16d3a53-kube-api-access-5fpxp" (OuterVolumeSpecName: "kube-api-access-5fpxp") pod "4adf565d-c266-4781-9107-fd05a16d3a53" (UID: "4adf565d-c266-4781-9107-fd05a16d3a53"). InnerVolumeSpecName "kube-api-access-5fpxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:32:30 crc kubenswrapper[4772]: I0127 16:32:30.300438 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pv8r\" (UniqueName: \"kubernetes.io/projected/6752954c-12b9-4b60-94bb-2f4676de7e6c-kube-api-access-4pv8r\") pod \"mariadb-client\" (UID: \"6752954c-12b9-4b60-94bb-2f4676de7e6c\") " pod="openstack/mariadb-client" Jan 27 16:32:30 crc kubenswrapper[4772]: I0127 16:32:30.300600 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fpxp\" (UniqueName: \"kubernetes.io/projected/4adf565d-c266-4781-9107-fd05a16d3a53-kube-api-access-5fpxp\") on node \"crc\" DevicePath \"\"" Jan 27 16:32:30 crc kubenswrapper[4772]: I0127 16:32:30.402544 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pv8r\" (UniqueName: \"kubernetes.io/projected/6752954c-12b9-4b60-94bb-2f4676de7e6c-kube-api-access-4pv8r\") pod \"mariadb-client\" (UID: \"6752954c-12b9-4b60-94bb-2f4676de7e6c\") " pod="openstack/mariadb-client" Jan 27 16:32:30 crc kubenswrapper[4772]: I0127 16:32:30.423389 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pv8r\" (UniqueName: \"kubernetes.io/projected/6752954c-12b9-4b60-94bb-2f4676de7e6c-kube-api-access-4pv8r\") pod \"mariadb-client\" (UID: \"6752954c-12b9-4b60-94bb-2f4676de7e6c\") " pod="openstack/mariadb-client" Jan 27 16:32:30 crc kubenswrapper[4772]: I0127 16:32:30.536841 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Jan 27 16:32:30 crc kubenswrapper[4772]: I0127 16:32:30.677030 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4adf565d-c266-4781-9107-fd05a16d3a53" path="/var/lib/kubelet/pods/4adf565d-c266-4781-9107-fd05a16d3a53/volumes" Jan 27 16:32:30 crc kubenswrapper[4772]: I0127 16:32:30.725047 4772 scope.go:117] "RemoveContainer" containerID="f861f93267509f86f0bf58147bbab5f93435b373f74dada3332239c3cced3199" Jan 27 16:32:30 crc kubenswrapper[4772]: I0127 16:32:30.725160 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Jan 27 16:32:30 crc kubenswrapper[4772]: W0127 16:32:30.936085 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6752954c_12b9_4b60_94bb_2f4676de7e6c.slice/crio-e9724199e84cd11ca0a68cfbb4deb346f80fb41c1fbd3adf14ea00d4e3521552 WatchSource:0}: Error finding container e9724199e84cd11ca0a68cfbb4deb346f80fb41c1fbd3adf14ea00d4e3521552: Status 404 returned error can't find the container with id e9724199e84cd11ca0a68cfbb4deb346f80fb41c1fbd3adf14ea00d4e3521552 Jan 27 16:32:30 crc kubenswrapper[4772]: I0127 16:32:30.936948 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Jan 27 16:32:31 crc kubenswrapper[4772]: I0127 16:32:31.738039 4772 generic.go:334] "Generic (PLEG): container finished" podID="6752954c-12b9-4b60-94bb-2f4676de7e6c" containerID="f4958ec9744454169fb58baabe20204293a4ff4790174c9c2079b9801fd7028d" exitCode=0 Jan 27 16:32:31 crc kubenswrapper[4772]: I0127 16:32:31.738131 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"6752954c-12b9-4b60-94bb-2f4676de7e6c","Type":"ContainerDied","Data":"f4958ec9744454169fb58baabe20204293a4ff4790174c9c2079b9801fd7028d"} Jan 27 16:32:31 crc kubenswrapper[4772]: I0127 16:32:31.738387 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"6752954c-12b9-4b60-94bb-2f4676de7e6c","Type":"ContainerStarted","Data":"e9724199e84cd11ca0a68cfbb4deb346f80fb41c1fbd3adf14ea00d4e3521552"} Jan 27 16:32:33 crc kubenswrapper[4772]: I0127 16:32:33.164131 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Jan 27 16:32:33 crc kubenswrapper[4772]: I0127 16:32:33.180634 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_6752954c-12b9-4b60-94bb-2f4676de7e6c/mariadb-client/0.log" Jan 27 16:32:33 crc kubenswrapper[4772]: I0127 16:32:33.203068 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Jan 27 16:32:33 crc kubenswrapper[4772]: I0127 16:32:33.208828 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Jan 27 16:32:33 crc kubenswrapper[4772]: I0127 16:32:33.348777 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pv8r\" (UniqueName: \"kubernetes.io/projected/6752954c-12b9-4b60-94bb-2f4676de7e6c-kube-api-access-4pv8r\") pod \"6752954c-12b9-4b60-94bb-2f4676de7e6c\" (UID: \"6752954c-12b9-4b60-94bb-2f4676de7e6c\") " Jan 27 16:32:33 crc kubenswrapper[4772]: I0127 16:32:33.367526 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6752954c-12b9-4b60-94bb-2f4676de7e6c-kube-api-access-4pv8r" (OuterVolumeSpecName: "kube-api-access-4pv8r") pod "6752954c-12b9-4b60-94bb-2f4676de7e6c" (UID: "6752954c-12b9-4b60-94bb-2f4676de7e6c"). InnerVolumeSpecName "kube-api-access-4pv8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:32:33 crc kubenswrapper[4772]: I0127 16:32:33.450469 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pv8r\" (UniqueName: \"kubernetes.io/projected/6752954c-12b9-4b60-94bb-2f4676de7e6c-kube-api-access-4pv8r\") on node \"crc\" DevicePath \"\"" Jan 27 16:32:33 crc kubenswrapper[4772]: I0127 16:32:33.752294 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9724199e84cd11ca0a68cfbb4deb346f80fb41c1fbd3adf14ea00d4e3521552" Jan 27 16:32:33 crc kubenswrapper[4772]: I0127 16:32:33.752365 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Jan 27 16:32:34 crc kubenswrapper[4772]: I0127 16:32:34.674723 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6752954c-12b9-4b60-94bb-2f4676de7e6c" path="/var/lib/kubelet/pods/6752954c-12b9-4b60-94bb-2f4676de7e6c/volumes" Jan 27 16:32:40 crc kubenswrapper[4772]: I0127 16:32:40.663376 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:32:40 crc kubenswrapper[4772]: E0127 16:32:40.664374 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:32:53 crc kubenswrapper[4772]: I0127 16:32:53.662852 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:32:53 crc kubenswrapper[4772]: E0127 16:32:53.663565 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:33:04 crc kubenswrapper[4772]: I0127 16:33:04.668428 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:33:04 crc kubenswrapper[4772]: E0127 16:33:04.669326 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:33:06 crc kubenswrapper[4772]: I0127 16:33:06.605070 4772 scope.go:117] "RemoveContainer" containerID="0bdb0516e5ad0fcd11824f097428db46c1768aa20c8111ee97d7a876d3f00649" Jan 27 16:33:06 crc kubenswrapper[4772]: I0127 16:33:06.629258 4772 scope.go:117] "RemoveContainer" containerID="dbf03a502ec31eeb1c32c76f40edb6250c7c04cd55512165e350a627fdf8e1b7" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.615957 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 27 16:33:07 crc kubenswrapper[4772]: E0127 16:33:07.616608 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6752954c-12b9-4b60-94bb-2f4676de7e6c" containerName="mariadb-client" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.616626 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="6752954c-12b9-4b60-94bb-2f4676de7e6c" containerName="mariadb-client" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.616816 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="6752954c-12b9-4b60-94bb-2f4676de7e6c" containerName="mariadb-client" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.617635 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.620332 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.621248 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.621551 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-m86nf" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.635039 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.657385 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.658762 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.665132 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.665139 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-8751ebf2-d508-441e-bc69-ce17a1a3281f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8751ebf2-d508-441e-bc69-ce17a1a3281f\") pod \"ovsdbserver-nb-0\" (UID: \"c7dba285-1db4-44d8-bdf4-9de6e8d80adb\") " pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.665430 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c7dba285-1db4-44d8-bdf4-9de6e8d80adb-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c7dba285-1db4-44d8-bdf4-9de6e8d80adb\") " pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.665868 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7dba285-1db4-44d8-bdf4-9de6e8d80adb-config\") pod \"ovsdbserver-nb-0\" (UID: \"c7dba285-1db4-44d8-bdf4-9de6e8d80adb\") " pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.665992 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7dba285-1db4-44d8-bdf4-9de6e8d80adb-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c7dba285-1db4-44d8-bdf4-9de6e8d80adb\") " pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.666032 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b49jj\" (UniqueName: \"kubernetes.io/projected/c7dba285-1db4-44d8-bdf4-9de6e8d80adb-kube-api-access-b49jj\") pod \"ovsdbserver-nb-0\" (UID: \"c7dba285-1db4-44d8-bdf4-9de6e8d80adb\") " pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.666076 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c7dba285-1db4-44d8-bdf4-9de6e8d80adb-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c7dba285-1db4-44d8-bdf4-9de6e8d80adb\") " pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.666333 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.678052 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.697449 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.768154 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d6414222-b877-4ed8-964f-8b67b2ad1611\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d6414222-b877-4ed8-964f-8b67b2ad1611\") pod \"ovsdbserver-nb-2\" (UID: \"3e19e84f-6d5e-455b-be78-ae3f04c925b7\") " pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.768235 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbpbb\" (UniqueName: \"kubernetes.io/projected/d647bfb5-69e6-4b10-96ac-5f7fcd72514f-kube-api-access-tbpbb\") pod \"ovsdbserver-nb-1\" (UID: \"d647bfb5-69e6-4b10-96ac-5f7fcd72514f\") " pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.768290 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slcvh\" (UniqueName: \"kubernetes.io/projected/3e19e84f-6d5e-455b-be78-ae3f04c925b7-kube-api-access-slcvh\") pod \"ovsdbserver-nb-2\" (UID: \"3e19e84f-6d5e-455b-be78-ae3f04c925b7\") " pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.768351 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c7dba285-1db4-44d8-bdf4-9de6e8d80adb-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c7dba285-1db4-44d8-bdf4-9de6e8d80adb\") " pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.768376 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7dba285-1db4-44d8-bdf4-9de6e8d80adb-config\") pod \"ovsdbserver-nb-0\" (UID: \"c7dba285-1db4-44d8-bdf4-9de6e8d80adb\") " pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.768404 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b3f26a0e-bac9-4500-b628-aafb6e4b1c41\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b3f26a0e-bac9-4500-b628-aafb6e4b1c41\") pod \"ovsdbserver-nb-1\" (UID: \"d647bfb5-69e6-4b10-96ac-5f7fcd72514f\") " pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.768514 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d647bfb5-69e6-4b10-96ac-5f7fcd72514f-config\") pod \"ovsdbserver-nb-1\" (UID: \"d647bfb5-69e6-4b10-96ac-5f7fcd72514f\") " pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.768537 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e19e84f-6d5e-455b-be78-ae3f04c925b7-config\") pod \"ovsdbserver-nb-2\" (UID: \"3e19e84f-6d5e-455b-be78-ae3f04c925b7\") " pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.768553 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d647bfb5-69e6-4b10-96ac-5f7fcd72514f-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"d647bfb5-69e6-4b10-96ac-5f7fcd72514f\") " pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.768587 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d647bfb5-69e6-4b10-96ac-5f7fcd72514f-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"d647bfb5-69e6-4b10-96ac-5f7fcd72514f\") " pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.768608 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7dba285-1db4-44d8-bdf4-9de6e8d80adb-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c7dba285-1db4-44d8-bdf4-9de6e8d80adb\") " pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.768631 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b49jj\" (UniqueName: \"kubernetes.io/projected/c7dba285-1db4-44d8-bdf4-9de6e8d80adb-kube-api-access-b49jj\") pod \"ovsdbserver-nb-0\" (UID: \"c7dba285-1db4-44d8-bdf4-9de6e8d80adb\") " pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.768664 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d647bfb5-69e6-4b10-96ac-5f7fcd72514f-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"d647bfb5-69e6-4b10-96ac-5f7fcd72514f\") " pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.768699 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c7dba285-1db4-44d8-bdf4-9de6e8d80adb-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c7dba285-1db4-44d8-bdf4-9de6e8d80adb\") " pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.768717 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3e19e84f-6d5e-455b-be78-ae3f04c925b7-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"3e19e84f-6d5e-455b-be78-ae3f04c925b7\") " pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.768776 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3e19e84f-6d5e-455b-be78-ae3f04c925b7-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"3e19e84f-6d5e-455b-be78-ae3f04c925b7\") " pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.768892 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-8751ebf2-d508-441e-bc69-ce17a1a3281f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8751ebf2-d508-441e-bc69-ce17a1a3281f\") pod \"ovsdbserver-nb-0\" (UID: \"c7dba285-1db4-44d8-bdf4-9de6e8d80adb\") " pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.769313 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e19e84f-6d5e-455b-be78-ae3f04c925b7-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"3e19e84f-6d5e-455b-be78-ae3f04c925b7\") " pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.769664 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c7dba285-1db4-44d8-bdf4-9de6e8d80adb-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c7dba285-1db4-44d8-bdf4-9de6e8d80adb\") " pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.769956 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7dba285-1db4-44d8-bdf4-9de6e8d80adb-config\") pod \"ovsdbserver-nb-0\" (UID: \"c7dba285-1db4-44d8-bdf4-9de6e8d80adb\") " pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.770343 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c7dba285-1db4-44d8-bdf4-9de6e8d80adb-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c7dba285-1db4-44d8-bdf4-9de6e8d80adb\") " pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.773196 4772 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.773236 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-8751ebf2-d508-441e-bc69-ce17a1a3281f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8751ebf2-d508-441e-bc69-ce17a1a3281f\") pod \"ovsdbserver-nb-0\" (UID: \"c7dba285-1db4-44d8-bdf4-9de6e8d80adb\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/3337b6a472fbda1da50af1d325792e8abe5f3c077d3bd6ebb24d51801a8d002b/globalmount\"" pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.774613 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7dba285-1db4-44d8-bdf4-9de6e8d80adb-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c7dba285-1db4-44d8-bdf4-9de6e8d80adb\") " pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.785871 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b49jj\" (UniqueName: \"kubernetes.io/projected/c7dba285-1db4-44d8-bdf4-9de6e8d80adb-kube-api-access-b49jj\") pod \"ovsdbserver-nb-0\" (UID: \"c7dba285-1db4-44d8-bdf4-9de6e8d80adb\") " pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.797984 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-8751ebf2-d508-441e-bc69-ce17a1a3281f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8751ebf2-d508-441e-bc69-ce17a1a3281f\") pod \"ovsdbserver-nb-0\" (UID: \"c7dba285-1db4-44d8-bdf4-9de6e8d80adb\") " pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.833317 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.836906 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.839135 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.839538 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-h7s5j" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.839855 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.843656 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.851087 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.853115 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.862576 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.864191 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.870229 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3e19e84f-6d5e-455b-be78-ae3f04c925b7-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"3e19e84f-6d5e-455b-be78-ae3f04c925b7\") " pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.870295 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3e19e84f-6d5e-455b-be78-ae3f04c925b7-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"3e19e84f-6d5e-455b-be78-ae3f04c925b7\") " pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.870347 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e19e84f-6d5e-455b-be78-ae3f04c925b7-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"3e19e84f-6d5e-455b-be78-ae3f04c925b7\") " pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.870378 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d6414222-b877-4ed8-964f-8b67b2ad1611\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d6414222-b877-4ed8-964f-8b67b2ad1611\") pod \"ovsdbserver-nb-2\" (UID: \"3e19e84f-6d5e-455b-be78-ae3f04c925b7\") " pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.870404 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbpbb\" (UniqueName: \"kubernetes.io/projected/d647bfb5-69e6-4b10-96ac-5f7fcd72514f-kube-api-access-tbpbb\") pod \"ovsdbserver-nb-1\" (UID: \"d647bfb5-69e6-4b10-96ac-5f7fcd72514f\") " pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.870429 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slcvh\" (UniqueName: \"kubernetes.io/projected/3e19e84f-6d5e-455b-be78-ae3f04c925b7-kube-api-access-slcvh\") pod \"ovsdbserver-nb-2\" (UID: \"3e19e84f-6d5e-455b-be78-ae3f04c925b7\") " pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.870693 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b3f26a0e-bac9-4500-b628-aafb6e4b1c41\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b3f26a0e-bac9-4500-b628-aafb6e4b1c41\") pod \"ovsdbserver-nb-1\" (UID: \"d647bfb5-69e6-4b10-96ac-5f7fcd72514f\") " pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.870736 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d647bfb5-69e6-4b10-96ac-5f7fcd72514f-config\") pod \"ovsdbserver-nb-1\" (UID: \"d647bfb5-69e6-4b10-96ac-5f7fcd72514f\") " pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.870763 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e19e84f-6d5e-455b-be78-ae3f04c925b7-config\") pod \"ovsdbserver-nb-2\" (UID: \"3e19e84f-6d5e-455b-be78-ae3f04c925b7\") " pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.870787 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d647bfb5-69e6-4b10-96ac-5f7fcd72514f-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"d647bfb5-69e6-4b10-96ac-5f7fcd72514f\") " pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.870811 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d647bfb5-69e6-4b10-96ac-5f7fcd72514f-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"d647bfb5-69e6-4b10-96ac-5f7fcd72514f\") " pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.870844 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d647bfb5-69e6-4b10-96ac-5f7fcd72514f-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"d647bfb5-69e6-4b10-96ac-5f7fcd72514f\") " pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.871704 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3e19e84f-6d5e-455b-be78-ae3f04c925b7-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"3e19e84f-6d5e-455b-be78-ae3f04c925b7\") " pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.871946 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d647bfb5-69e6-4b10-96ac-5f7fcd72514f-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"d647bfb5-69e6-4b10-96ac-5f7fcd72514f\") " pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.873082 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d647bfb5-69e6-4b10-96ac-5f7fcd72514f-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"d647bfb5-69e6-4b10-96ac-5f7fcd72514f\") " pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.873156 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.873311 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d647bfb5-69e6-4b10-96ac-5f7fcd72514f-config\") pod \"ovsdbserver-nb-1\" (UID: \"d647bfb5-69e6-4b10-96ac-5f7fcd72514f\") " pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.873393 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e19e84f-6d5e-455b-be78-ae3f04c925b7-config\") pod \"ovsdbserver-nb-2\" (UID: \"3e19e84f-6d5e-455b-be78-ae3f04c925b7\") " pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.873706 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3e19e84f-6d5e-455b-be78-ae3f04c925b7-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"3e19e84f-6d5e-455b-be78-ae3f04c925b7\") " pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.886061 4772 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.886102 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d6414222-b877-4ed8-964f-8b67b2ad1611\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d6414222-b877-4ed8-964f-8b67b2ad1611\") pod \"ovsdbserver-nb-2\" (UID: \"3e19e84f-6d5e-455b-be78-ae3f04c925b7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1c651a70b282613992325926fead9315e48244190040972e3c56616ccce54862/globalmount\"" pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.886144 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e19e84f-6d5e-455b-be78-ae3f04c925b7-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"3e19e84f-6d5e-455b-be78-ae3f04c925b7\") " pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.886694 4772 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.886730 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b3f26a0e-bac9-4500-b628-aafb6e4b1c41\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b3f26a0e-bac9-4500-b628-aafb6e4b1c41\") pod \"ovsdbserver-nb-1\" (UID: \"d647bfb5-69e6-4b10-96ac-5f7fcd72514f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a098fd0db4cba2f79026a162d263b402d8997356cd854ca8956233b9a2e11413/globalmount\"" pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.888458 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.900237 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d647bfb5-69e6-4b10-96ac-5f7fcd72514f-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"d647bfb5-69e6-4b10-96ac-5f7fcd72514f\") " pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.904887 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slcvh\" (UniqueName: \"kubernetes.io/projected/3e19e84f-6d5e-455b-be78-ae3f04c925b7-kube-api-access-slcvh\") pod \"ovsdbserver-nb-2\" (UID: \"3e19e84f-6d5e-455b-be78-ae3f04c925b7\") " pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.905636 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbpbb\" (UniqueName: \"kubernetes.io/projected/d647bfb5-69e6-4b10-96ac-5f7fcd72514f-kube-api-access-tbpbb\") pod \"ovsdbserver-nb-1\" (UID: \"d647bfb5-69e6-4b10-96ac-5f7fcd72514f\") " pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.925672 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b3f26a0e-bac9-4500-b628-aafb6e4b1c41\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b3f26a0e-bac9-4500-b628-aafb6e4b1c41\") pod \"ovsdbserver-nb-1\" (UID: \"d647bfb5-69e6-4b10-96ac-5f7fcd72514f\") " pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.931578 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d6414222-b877-4ed8-964f-8b67b2ad1611\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d6414222-b877-4ed8-964f-8b67b2ad1611\") pod \"ovsdbserver-nb-2\" (UID: \"3e19e84f-6d5e-455b-be78-ae3f04c925b7\") " pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.942859 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.972489 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f70d2878-d629-4772-b2a4-697fe18a3760-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f70d2878-d629-4772-b2a4-697fe18a3760\") " pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.972547 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l7fw\" (UniqueName: \"kubernetes.io/projected/f70d2878-d629-4772-b2a4-697fe18a3760-kube-api-access-2l7fw\") pod \"ovsdbserver-sb-0\" (UID: \"f70d2878-d629-4772-b2a4-697fe18a3760\") " pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.972570 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6e7e5c6-90b8-4de9-ae6a-11034616734a-config\") pod \"ovsdbserver-sb-1\" (UID: \"e6e7e5c6-90b8-4de9-ae6a-11034616734a\") " pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.972589 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e6e7e5c6-90b8-4de9-ae6a-11034616734a-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"e6e7e5c6-90b8-4de9-ae6a-11034616734a\") " pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.972626 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f70d2878-d629-4772-b2a4-697fe18a3760-config\") pod \"ovsdbserver-sb-0\" (UID: \"f70d2878-d629-4772-b2a4-697fe18a3760\") " pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.972642 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75n4l\" (UniqueName: \"kubernetes.io/projected/e6e7e5c6-90b8-4de9-ae6a-11034616734a-kube-api-access-75n4l\") pod \"ovsdbserver-sb-1\" (UID: \"e6e7e5c6-90b8-4de9-ae6a-11034616734a\") " pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.972661 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-9fb8f666-d05c-49d9-bf4f-89de48d00eee\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9fb8f666-d05c-49d9-bf4f-89de48d00eee\") pod \"ovsdbserver-sb-0\" (UID: \"f70d2878-d629-4772-b2a4-697fe18a3760\") " pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.972684 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f70d2878-d629-4772-b2a4-697fe18a3760-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f70d2878-d629-4772-b2a4-697fe18a3760\") " pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.972703 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f70d2878-d629-4772-b2a4-697fe18a3760-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f70d2878-d629-4772-b2a4-697fe18a3760\") " pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.972744 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6e7e5c6-90b8-4de9-ae6a-11034616734a-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"e6e7e5c6-90b8-4de9-ae6a-11034616734a\") " pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.972772 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d0f7281d-c4e2-4aa0-b339-d80e66a50bb8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d0f7281d-c4e2-4aa0-b339-d80e66a50bb8\") pod \"ovsdbserver-sb-1\" (UID: \"e6e7e5c6-90b8-4de9-ae6a-11034616734a\") " pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.972792 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e6e7e5c6-90b8-4de9-ae6a-11034616734a-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"e6e7e5c6-90b8-4de9-ae6a-11034616734a\") " pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.980844 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:07 crc kubenswrapper[4772]: I0127 16:33:07.995659 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.074151 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2\") " pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.074267 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2-config\") pod \"ovsdbserver-sb-2\" (UID: \"e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2\") " pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.074296 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2\") " pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.074323 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f70d2878-d629-4772-b2a4-697fe18a3760-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f70d2878-d629-4772-b2a4-697fe18a3760\") " pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.074350 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-1a586ac9-f601-4b7a-8500-43d5056dca11\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a586ac9-f601-4b7a-8500-43d5056dca11\") pod \"ovsdbserver-sb-2\" (UID: \"e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2\") " pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.074371 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l7fw\" (UniqueName: \"kubernetes.io/projected/f70d2878-d629-4772-b2a4-697fe18a3760-kube-api-access-2l7fw\") pod \"ovsdbserver-sb-0\" (UID: \"f70d2878-d629-4772-b2a4-697fe18a3760\") " pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.074389 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6e7e5c6-90b8-4de9-ae6a-11034616734a-config\") pod \"ovsdbserver-sb-1\" (UID: \"e6e7e5c6-90b8-4de9-ae6a-11034616734a\") " pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.074409 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e6e7e5c6-90b8-4de9-ae6a-11034616734a-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"e6e7e5c6-90b8-4de9-ae6a-11034616734a\") " pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.074458 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2\") " pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.074482 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2bk4\" (UniqueName: \"kubernetes.io/projected/e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2-kube-api-access-v2bk4\") pod \"ovsdbserver-sb-2\" (UID: \"e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2\") " pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.074512 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f70d2878-d629-4772-b2a4-697fe18a3760-config\") pod \"ovsdbserver-sb-0\" (UID: \"f70d2878-d629-4772-b2a4-697fe18a3760\") " pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.074536 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75n4l\" (UniqueName: \"kubernetes.io/projected/e6e7e5c6-90b8-4de9-ae6a-11034616734a-kube-api-access-75n4l\") pod \"ovsdbserver-sb-1\" (UID: \"e6e7e5c6-90b8-4de9-ae6a-11034616734a\") " pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.074560 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-9fb8f666-d05c-49d9-bf4f-89de48d00eee\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9fb8f666-d05c-49d9-bf4f-89de48d00eee\") pod \"ovsdbserver-sb-0\" (UID: \"f70d2878-d629-4772-b2a4-697fe18a3760\") " pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.074596 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f70d2878-d629-4772-b2a4-697fe18a3760-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f70d2878-d629-4772-b2a4-697fe18a3760\") " pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.074620 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f70d2878-d629-4772-b2a4-697fe18a3760-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f70d2878-d629-4772-b2a4-697fe18a3760\") " pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.075828 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6e7e5c6-90b8-4de9-ae6a-11034616734a-config\") pod \"ovsdbserver-sb-1\" (UID: \"e6e7e5c6-90b8-4de9-ae6a-11034616734a\") " pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.077111 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f70d2878-d629-4772-b2a4-697fe18a3760-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f70d2878-d629-4772-b2a4-697fe18a3760\") " pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.078011 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e6e7e5c6-90b8-4de9-ae6a-11034616734a-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"e6e7e5c6-90b8-4de9-ae6a-11034616734a\") " pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.078472 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f70d2878-d629-4772-b2a4-697fe18a3760-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f70d2878-d629-4772-b2a4-697fe18a3760\") " pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.078640 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6e7e5c6-90b8-4de9-ae6a-11034616734a-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"e6e7e5c6-90b8-4de9-ae6a-11034616734a\") " pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.078756 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d0f7281d-c4e2-4aa0-b339-d80e66a50bb8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d0f7281d-c4e2-4aa0-b339-d80e66a50bb8\") pod \"ovsdbserver-sb-1\" (UID: \"e6e7e5c6-90b8-4de9-ae6a-11034616734a\") " pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.078818 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e6e7e5c6-90b8-4de9-ae6a-11034616734a-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"e6e7e5c6-90b8-4de9-ae6a-11034616734a\") " pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.079087 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f70d2878-d629-4772-b2a4-697fe18a3760-config\") pod \"ovsdbserver-sb-0\" (UID: \"f70d2878-d629-4772-b2a4-697fe18a3760\") " pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.080251 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e6e7e5c6-90b8-4de9-ae6a-11034616734a-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"e6e7e5c6-90b8-4de9-ae6a-11034616734a\") " pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.081720 4772 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.081750 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-9fb8f666-d05c-49d9-bf4f-89de48d00eee\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9fb8f666-d05c-49d9-bf4f-89de48d00eee\") pod \"ovsdbserver-sb-0\" (UID: \"f70d2878-d629-4772-b2a4-697fe18a3760\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/cc749d8993a334b6229d6a39c6b8333d09387f0305fdac2d4a2f13487d621faf/globalmount\"" pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.082199 4772 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.082247 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d0f7281d-c4e2-4aa0-b339-d80e66a50bb8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d0f7281d-c4e2-4aa0-b339-d80e66a50bb8\") pod \"ovsdbserver-sb-1\" (UID: \"e6e7e5c6-90b8-4de9-ae6a-11034616734a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/64e51a673028894bd1a32e587e28872875075ce3051c01bb423c11d2fb7dc74d/globalmount\"" pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.082388 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f70d2878-d629-4772-b2a4-697fe18a3760-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f70d2878-d629-4772-b2a4-697fe18a3760\") " pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.083184 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6e7e5c6-90b8-4de9-ae6a-11034616734a-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"e6e7e5c6-90b8-4de9-ae6a-11034616734a\") " pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.099787 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l7fw\" (UniqueName: \"kubernetes.io/projected/f70d2878-d629-4772-b2a4-697fe18a3760-kube-api-access-2l7fw\") pod \"ovsdbserver-sb-0\" (UID: \"f70d2878-d629-4772-b2a4-697fe18a3760\") " pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.100549 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75n4l\" (UniqueName: \"kubernetes.io/projected/e6e7e5c6-90b8-4de9-ae6a-11034616734a-kube-api-access-75n4l\") pod \"ovsdbserver-sb-1\" (UID: \"e6e7e5c6-90b8-4de9-ae6a-11034616734a\") " pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.128643 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-9fb8f666-d05c-49d9-bf4f-89de48d00eee\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9fb8f666-d05c-49d9-bf4f-89de48d00eee\") pod \"ovsdbserver-sb-0\" (UID: \"f70d2878-d629-4772-b2a4-697fe18a3760\") " pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.129453 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d0f7281d-c4e2-4aa0-b339-d80e66a50bb8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d0f7281d-c4e2-4aa0-b339-d80e66a50bb8\") pod \"ovsdbserver-sb-1\" (UID: \"e6e7e5c6-90b8-4de9-ae6a-11034616734a\") " pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.162237 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.180636 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2\") " pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.180732 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2-config\") pod \"ovsdbserver-sb-2\" (UID: \"e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2\") " pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.180760 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2\") " pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.180795 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-1a586ac9-f601-4b7a-8500-43d5056dca11\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a586ac9-f601-4b7a-8500-43d5056dca11\") pod \"ovsdbserver-sb-2\" (UID: \"e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2\") " pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.180846 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2\") " pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.180875 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2bk4\" (UniqueName: \"kubernetes.io/projected/e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2-kube-api-access-v2bk4\") pod \"ovsdbserver-sb-2\" (UID: \"e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2\") " pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.181437 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2\") " pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.181704 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2-config\") pod \"ovsdbserver-sb-2\" (UID: \"e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2\") " pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.182473 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2\") " pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.183134 4772 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.183183 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-1a586ac9-f601-4b7a-8500-43d5056dca11\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a586ac9-f601-4b7a-8500-43d5056dca11\") pod \"ovsdbserver-sb-2\" (UID: \"e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ecb883243e8fa28e04401fc11c1cc3b48caf5d4b7f163305d47b1cc276850bf3/globalmount\"" pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.185338 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2\") " pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.208254 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2bk4\" (UniqueName: \"kubernetes.io/projected/e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2-kube-api-access-v2bk4\") pod \"ovsdbserver-sb-2\" (UID: \"e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2\") " pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.212855 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-1a586ac9-f601-4b7a-8500-43d5056dca11\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a586ac9-f601-4b7a-8500-43d5056dca11\") pod \"ovsdbserver-sb-2\" (UID: \"e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2\") " pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.248882 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.256396 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.474663 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.583446 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Jan 27 16:33:08 crc kubenswrapper[4772]: W0127 16:33:08.584365 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e19e84f_6d5e_455b_be78_ae3f04c925b7.slice/crio-48f52bd514e908fcf4864fb87e7374d832e8024e0d21ce8b7bcc182f3ed75e05 WatchSource:0}: Error finding container 48f52bd514e908fcf4864fb87e7374d832e8024e0d21ce8b7bcc182f3ed75e05: Status 404 returned error can't find the container with id 48f52bd514e908fcf4864fb87e7374d832e8024e0d21ce8b7bcc182f3ed75e05 Jan 27 16:33:08 crc kubenswrapper[4772]: W0127 16:33:08.674335 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd647bfb5_69e6_4b10_96ac_5f7fcd72514f.slice/crio-074db820aa56e1d3e692305b41061d7a8d773275074c4a730fe63818a3c3370a WatchSource:0}: Error finding container 074db820aa56e1d3e692305b41061d7a8d773275074c4a730fe63818a3c3370a: Status 404 returned error can't find the container with id 074db820aa56e1d3e692305b41061d7a8d773275074c4a730fe63818a3c3370a Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.675795 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.794777 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 27 16:33:08 crc kubenswrapper[4772]: I0127 16:33:08.884214 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Jan 27 16:33:09 crc kubenswrapper[4772]: I0127 16:33:09.040775 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f70d2878-d629-4772-b2a4-697fe18a3760","Type":"ContainerStarted","Data":"bfb61b64ab9af272f753949b50afcd24287fb693d96febe095e3a45ae3c76473"} Jan 27 16:33:09 crc kubenswrapper[4772]: I0127 16:33:09.042307 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"d647bfb5-69e6-4b10-96ac-5f7fcd72514f","Type":"ContainerStarted","Data":"920ffd185ec2c4037d40e133e1096d42c4c2c1a7b8cd18f5f18c96e7c0a0ce1e"} Jan 27 16:33:09 crc kubenswrapper[4772]: I0127 16:33:09.042330 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"d647bfb5-69e6-4b10-96ac-5f7fcd72514f","Type":"ContainerStarted","Data":"074db820aa56e1d3e692305b41061d7a8d773275074c4a730fe63818a3c3370a"} Jan 27 16:33:09 crc kubenswrapper[4772]: I0127 16:33:09.045483 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"3e19e84f-6d5e-455b-be78-ae3f04c925b7","Type":"ContainerStarted","Data":"d1b4adcc8b2d207a74c0d46c315c74c8b626767da359aff95a2d871c59ab047c"} Jan 27 16:33:09 crc kubenswrapper[4772]: I0127 16:33:09.045580 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"3e19e84f-6d5e-455b-be78-ae3f04c925b7","Type":"ContainerStarted","Data":"48f52bd514e908fcf4864fb87e7374d832e8024e0d21ce8b7bcc182f3ed75e05"} Jan 27 16:33:09 crc kubenswrapper[4772]: I0127 16:33:09.047905 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c7dba285-1db4-44d8-bdf4-9de6e8d80adb","Type":"ContainerStarted","Data":"9a393749b3e8c476a8dbf09db1e5516a9dd70ecbfe7b88f62d8f1b2fc7c9ae27"} Jan 27 16:33:09 crc kubenswrapper[4772]: I0127 16:33:09.047935 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c7dba285-1db4-44d8-bdf4-9de6e8d80adb","Type":"ContainerStarted","Data":"0cc7431e3cf8f4c3bcfa0b8a901f57d7a9bb4d5ab838913051b827521f69d36e"} Jan 27 16:33:09 crc kubenswrapper[4772]: I0127 16:33:09.049386 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"e6e7e5c6-90b8-4de9-ae6a-11034616734a","Type":"ContainerStarted","Data":"ad163338ac498171cb6edadfe863d35da50a888044da8366acca9552904d5338"} Jan 27 16:33:09 crc kubenswrapper[4772]: I0127 16:33:09.482377 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Jan 27 16:33:10 crc kubenswrapper[4772]: I0127 16:33:10.059690 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f70d2878-d629-4772-b2a4-697fe18a3760","Type":"ContainerStarted","Data":"b503e0fd0a7c27ff0d459996ac0d39e04861add0c7e9fbea62248cb0d2dc88c6"} Jan 27 16:33:10 crc kubenswrapper[4772]: I0127 16:33:10.060037 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f70d2878-d629-4772-b2a4-697fe18a3760","Type":"ContainerStarted","Data":"5432eb5f67260235f5ffdc38b09f50215a4ec704faf76ba8beca3ca36be7ae8d"} Jan 27 16:33:10 crc kubenswrapper[4772]: I0127 16:33:10.063099 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2","Type":"ContainerStarted","Data":"fa7be627a5545262b52e10c28a8cf3593200f1814d5322934ff1be15d6d478d6"} Jan 27 16:33:10 crc kubenswrapper[4772]: I0127 16:33:10.063143 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2","Type":"ContainerStarted","Data":"526b8505d35eaf62b78e16cc3783b92f6bf3d401410bc5db08769f91af92bd19"} Jan 27 16:33:10 crc kubenswrapper[4772]: I0127 16:33:10.063160 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2","Type":"ContainerStarted","Data":"cedf4ee1f5816248990bbe8ae54ca3ff34b16c8acb95ba79a94f3b775896fb38"} Jan 27 16:33:10 crc kubenswrapper[4772]: I0127 16:33:10.065141 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"d647bfb5-69e6-4b10-96ac-5f7fcd72514f","Type":"ContainerStarted","Data":"482c09bcb742316ddf7d6825e1bc9fe06aff97b0a598c35e0564d7e1e35cad4a"} Jan 27 16:33:10 crc kubenswrapper[4772]: I0127 16:33:10.066694 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"3e19e84f-6d5e-455b-be78-ae3f04c925b7","Type":"ContainerStarted","Data":"9063b11d3bd1855e7851cb3e5775f3995c3aecd5ae985951e817a21513a48b6a"} Jan 27 16:33:10 crc kubenswrapper[4772]: I0127 16:33:10.068422 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c7dba285-1db4-44d8-bdf4-9de6e8d80adb","Type":"ContainerStarted","Data":"887dd01207c102e69b2f99fc63206e19204eda0c285f9027d0cc9857454da907"} Jan 27 16:33:10 crc kubenswrapper[4772]: I0127 16:33:10.071196 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"e6e7e5c6-90b8-4de9-ae6a-11034616734a","Type":"ContainerStarted","Data":"6cd0d8b740bdd89690b3ad2480d1c3a5b7725050dc64d6cead32c124c96648f0"} Jan 27 16:33:10 crc kubenswrapper[4772]: I0127 16:33:10.071226 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"e6e7e5c6-90b8-4de9-ae6a-11034616734a","Type":"ContainerStarted","Data":"9129b568ab6a7ae493d7984fd465da0ccfbb3dc21791f3a4f8c8c37bf88d13fa"} Jan 27 16:33:10 crc kubenswrapper[4772]: I0127 16:33:10.080497 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.080481958 podStartE2EDuration="4.080481958s" podCreationTimestamp="2026-01-27 16:33:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:33:10.075787034 +0000 UTC m=+5176.056396142" watchObservedRunningTime="2026-01-27 16:33:10.080481958 +0000 UTC m=+5176.061091056" Jan 27 16:33:10 crc kubenswrapper[4772]: I0127 16:33:10.104767 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=4.104742029 podStartE2EDuration="4.104742029s" podCreationTimestamp="2026-01-27 16:33:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:33:10.098913433 +0000 UTC m=+5176.079522531" watchObservedRunningTime="2026-01-27 16:33:10.104742029 +0000 UTC m=+5176.085351127" Jan 27 16:33:10 crc kubenswrapper[4772]: I0127 16:33:10.120553 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=4.120532529 podStartE2EDuration="4.120532529s" podCreationTimestamp="2026-01-27 16:33:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:33:10.118308346 +0000 UTC m=+5176.098917444" watchObservedRunningTime="2026-01-27 16:33:10.120532529 +0000 UTC m=+5176.101141627" Jan 27 16:33:10 crc kubenswrapper[4772]: I0127 16:33:10.142966 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=4.142944188 podStartE2EDuration="4.142944188s" podCreationTimestamp="2026-01-27 16:33:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:33:10.13634211 +0000 UTC m=+5176.116951218" watchObservedRunningTime="2026-01-27 16:33:10.142944188 +0000 UTC m=+5176.123553296" Jan 27 16:33:10 crc kubenswrapper[4772]: I0127 16:33:10.156715 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=4.15669592 podStartE2EDuration="4.15669592s" podCreationTimestamp="2026-01-27 16:33:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:33:10.155053433 +0000 UTC m=+5176.135662541" watchObservedRunningTime="2026-01-27 16:33:10.15669592 +0000 UTC m=+5176.137305018" Jan 27 16:33:10 crc kubenswrapper[4772]: I0127 16:33:10.182038 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=4.182017332 podStartE2EDuration="4.182017332s" podCreationTimestamp="2026-01-27 16:33:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:33:10.171701478 +0000 UTC m=+5176.152310576" watchObservedRunningTime="2026-01-27 16:33:10.182017332 +0000 UTC m=+5176.162626440" Jan 27 16:33:10 crc kubenswrapper[4772]: I0127 16:33:10.943811 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:10 crc kubenswrapper[4772]: I0127 16:33:10.981423 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:10 crc kubenswrapper[4772]: I0127 16:33:10.995780 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:11 crc kubenswrapper[4772]: I0127 16:33:11.162923 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:11 crc kubenswrapper[4772]: I0127 16:33:11.250525 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:11 crc kubenswrapper[4772]: I0127 16:33:11.257666 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:12 crc kubenswrapper[4772]: I0127 16:33:12.943071 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:12 crc kubenswrapper[4772]: I0127 16:33:12.981728 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:12 crc kubenswrapper[4772]: I0127 16:33:12.996124 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:13 crc kubenswrapper[4772]: I0127 16:33:13.162470 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:13 crc kubenswrapper[4772]: I0127 16:33:13.250241 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:13 crc kubenswrapper[4772]: I0127 16:33:13.257774 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:13 crc kubenswrapper[4772]: I0127 16:33:13.979589 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.016607 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.131204 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.142615 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.148903 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.177252 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.208833 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.284785 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.308797 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.325688 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.379435 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.413211 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.417103 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f7b485f7-qqlrk"] Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.418964 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f7b485f7-qqlrk" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.423216 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.448049 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f7b485f7-qqlrk"] Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.577928 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f7b485f7-qqlrk"] Jan 27 16:33:14 crc kubenswrapper[4772]: E0127 16:33:14.578713 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-mnwh9 ovsdbserver-nb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-6f7b485f7-qqlrk" podUID="5669174d-af83-4842-9e6f-bf2a32c40bf5" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.598947 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5669174d-af83-4842-9e6f-bf2a32c40bf5-ovsdbserver-nb\") pod \"dnsmasq-dns-6f7b485f7-qqlrk\" (UID: \"5669174d-af83-4842-9e6f-bf2a32c40bf5\") " pod="openstack/dnsmasq-dns-6f7b485f7-qqlrk" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.599015 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5669174d-af83-4842-9e6f-bf2a32c40bf5-config\") pod \"dnsmasq-dns-6f7b485f7-qqlrk\" (UID: \"5669174d-af83-4842-9e6f-bf2a32c40bf5\") " pod="openstack/dnsmasq-dns-6f7b485f7-qqlrk" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.599079 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5669174d-af83-4842-9e6f-bf2a32c40bf5-dns-svc\") pod \"dnsmasq-dns-6f7b485f7-qqlrk\" (UID: \"5669174d-af83-4842-9e6f-bf2a32c40bf5\") " pod="openstack/dnsmasq-dns-6f7b485f7-qqlrk" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.599146 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnwh9\" (UniqueName: \"kubernetes.io/projected/5669174d-af83-4842-9e6f-bf2a32c40bf5-kube-api-access-mnwh9\") pod \"dnsmasq-dns-6f7b485f7-qqlrk\" (UID: \"5669174d-af83-4842-9e6f-bf2a32c40bf5\") " pod="openstack/dnsmasq-dns-6f7b485f7-qqlrk" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.610465 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-746d786469-qv9pq"] Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.612023 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-746d786469-qv9pq" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.618205 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.624141 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-746d786469-qv9pq"] Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.700250 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnwh9\" (UniqueName: \"kubernetes.io/projected/5669174d-af83-4842-9e6f-bf2a32c40bf5-kube-api-access-mnwh9\") pod \"dnsmasq-dns-6f7b485f7-qqlrk\" (UID: \"5669174d-af83-4842-9e6f-bf2a32c40bf5\") " pod="openstack/dnsmasq-dns-6f7b485f7-qqlrk" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.700349 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5669174d-af83-4842-9e6f-bf2a32c40bf5-ovsdbserver-nb\") pod \"dnsmasq-dns-6f7b485f7-qqlrk\" (UID: \"5669174d-af83-4842-9e6f-bf2a32c40bf5\") " pod="openstack/dnsmasq-dns-6f7b485f7-qqlrk" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.700394 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5669174d-af83-4842-9e6f-bf2a32c40bf5-config\") pod \"dnsmasq-dns-6f7b485f7-qqlrk\" (UID: \"5669174d-af83-4842-9e6f-bf2a32c40bf5\") " pod="openstack/dnsmasq-dns-6f7b485f7-qqlrk" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.700462 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5669174d-af83-4842-9e6f-bf2a32c40bf5-dns-svc\") pod \"dnsmasq-dns-6f7b485f7-qqlrk\" (UID: \"5669174d-af83-4842-9e6f-bf2a32c40bf5\") " pod="openstack/dnsmasq-dns-6f7b485f7-qqlrk" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.701496 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5669174d-af83-4842-9e6f-bf2a32c40bf5-dns-svc\") pod \"dnsmasq-dns-6f7b485f7-qqlrk\" (UID: \"5669174d-af83-4842-9e6f-bf2a32c40bf5\") " pod="openstack/dnsmasq-dns-6f7b485f7-qqlrk" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.702954 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5669174d-af83-4842-9e6f-bf2a32c40bf5-config\") pod \"dnsmasq-dns-6f7b485f7-qqlrk\" (UID: \"5669174d-af83-4842-9e6f-bf2a32c40bf5\") " pod="openstack/dnsmasq-dns-6f7b485f7-qqlrk" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.706152 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5669174d-af83-4842-9e6f-bf2a32c40bf5-ovsdbserver-nb\") pod \"dnsmasq-dns-6f7b485f7-qqlrk\" (UID: \"5669174d-af83-4842-9e6f-bf2a32c40bf5\") " pod="openstack/dnsmasq-dns-6f7b485f7-qqlrk" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.721561 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnwh9\" (UniqueName: \"kubernetes.io/projected/5669174d-af83-4842-9e6f-bf2a32c40bf5-kube-api-access-mnwh9\") pod \"dnsmasq-dns-6f7b485f7-qqlrk\" (UID: \"5669174d-af83-4842-9e6f-bf2a32c40bf5\") " pod="openstack/dnsmasq-dns-6f7b485f7-qqlrk" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.802151 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/320b0d26-aa22-4854-ac28-4699f95fb37b-ovsdbserver-sb\") pod \"dnsmasq-dns-746d786469-qv9pq\" (UID: \"320b0d26-aa22-4854-ac28-4699f95fb37b\") " pod="openstack/dnsmasq-dns-746d786469-qv9pq" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.802323 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/320b0d26-aa22-4854-ac28-4699f95fb37b-dns-svc\") pod \"dnsmasq-dns-746d786469-qv9pq\" (UID: \"320b0d26-aa22-4854-ac28-4699f95fb37b\") " pod="openstack/dnsmasq-dns-746d786469-qv9pq" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.802386 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/320b0d26-aa22-4854-ac28-4699f95fb37b-ovsdbserver-nb\") pod \"dnsmasq-dns-746d786469-qv9pq\" (UID: \"320b0d26-aa22-4854-ac28-4699f95fb37b\") " pod="openstack/dnsmasq-dns-746d786469-qv9pq" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.802484 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/320b0d26-aa22-4854-ac28-4699f95fb37b-config\") pod \"dnsmasq-dns-746d786469-qv9pq\" (UID: \"320b0d26-aa22-4854-ac28-4699f95fb37b\") " pod="openstack/dnsmasq-dns-746d786469-qv9pq" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.802542 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjfw7\" (UniqueName: \"kubernetes.io/projected/320b0d26-aa22-4854-ac28-4699f95fb37b-kube-api-access-zjfw7\") pod \"dnsmasq-dns-746d786469-qv9pq\" (UID: \"320b0d26-aa22-4854-ac28-4699f95fb37b\") " pod="openstack/dnsmasq-dns-746d786469-qv9pq" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.903795 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/320b0d26-aa22-4854-ac28-4699f95fb37b-dns-svc\") pod \"dnsmasq-dns-746d786469-qv9pq\" (UID: \"320b0d26-aa22-4854-ac28-4699f95fb37b\") " pod="openstack/dnsmasq-dns-746d786469-qv9pq" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.903866 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/320b0d26-aa22-4854-ac28-4699f95fb37b-ovsdbserver-nb\") pod \"dnsmasq-dns-746d786469-qv9pq\" (UID: \"320b0d26-aa22-4854-ac28-4699f95fb37b\") " pod="openstack/dnsmasq-dns-746d786469-qv9pq" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.903906 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/320b0d26-aa22-4854-ac28-4699f95fb37b-config\") pod \"dnsmasq-dns-746d786469-qv9pq\" (UID: \"320b0d26-aa22-4854-ac28-4699f95fb37b\") " pod="openstack/dnsmasq-dns-746d786469-qv9pq" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.904891 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjfw7\" (UniqueName: \"kubernetes.io/projected/320b0d26-aa22-4854-ac28-4699f95fb37b-kube-api-access-zjfw7\") pod \"dnsmasq-dns-746d786469-qv9pq\" (UID: \"320b0d26-aa22-4854-ac28-4699f95fb37b\") " pod="openstack/dnsmasq-dns-746d786469-qv9pq" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.904847 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/320b0d26-aa22-4854-ac28-4699f95fb37b-config\") pod \"dnsmasq-dns-746d786469-qv9pq\" (UID: \"320b0d26-aa22-4854-ac28-4699f95fb37b\") " pod="openstack/dnsmasq-dns-746d786469-qv9pq" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.904962 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/320b0d26-aa22-4854-ac28-4699f95fb37b-ovsdbserver-nb\") pod \"dnsmasq-dns-746d786469-qv9pq\" (UID: \"320b0d26-aa22-4854-ac28-4699f95fb37b\") " pod="openstack/dnsmasq-dns-746d786469-qv9pq" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.905007 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/320b0d26-aa22-4854-ac28-4699f95fb37b-dns-svc\") pod \"dnsmasq-dns-746d786469-qv9pq\" (UID: \"320b0d26-aa22-4854-ac28-4699f95fb37b\") " pod="openstack/dnsmasq-dns-746d786469-qv9pq" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.905036 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/320b0d26-aa22-4854-ac28-4699f95fb37b-ovsdbserver-sb\") pod \"dnsmasq-dns-746d786469-qv9pq\" (UID: \"320b0d26-aa22-4854-ac28-4699f95fb37b\") " pod="openstack/dnsmasq-dns-746d786469-qv9pq" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.905633 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/320b0d26-aa22-4854-ac28-4699f95fb37b-ovsdbserver-sb\") pod \"dnsmasq-dns-746d786469-qv9pq\" (UID: \"320b0d26-aa22-4854-ac28-4699f95fb37b\") " pod="openstack/dnsmasq-dns-746d786469-qv9pq" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.922464 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjfw7\" (UniqueName: \"kubernetes.io/projected/320b0d26-aa22-4854-ac28-4699f95fb37b-kube-api-access-zjfw7\") pod \"dnsmasq-dns-746d786469-qv9pq\" (UID: \"320b0d26-aa22-4854-ac28-4699f95fb37b\") " pod="openstack/dnsmasq-dns-746d786469-qv9pq" Jan 27 16:33:14 crc kubenswrapper[4772]: I0127 16:33:14.940424 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-746d786469-qv9pq" Jan 27 16:33:15 crc kubenswrapper[4772]: I0127 16:33:15.106433 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f7b485f7-qqlrk" Jan 27 16:33:15 crc kubenswrapper[4772]: I0127 16:33:15.121469 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f7b485f7-qqlrk" Jan 27 16:33:15 crc kubenswrapper[4772]: I0127 16:33:15.209682 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5669174d-af83-4842-9e6f-bf2a32c40bf5-dns-svc\") pod \"5669174d-af83-4842-9e6f-bf2a32c40bf5\" (UID: \"5669174d-af83-4842-9e6f-bf2a32c40bf5\") " Jan 27 16:33:15 crc kubenswrapper[4772]: I0127 16:33:15.209749 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5669174d-af83-4842-9e6f-bf2a32c40bf5-config\") pod \"5669174d-af83-4842-9e6f-bf2a32c40bf5\" (UID: \"5669174d-af83-4842-9e6f-bf2a32c40bf5\") " Jan 27 16:33:15 crc kubenswrapper[4772]: I0127 16:33:15.209780 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5669174d-af83-4842-9e6f-bf2a32c40bf5-ovsdbserver-nb\") pod \"5669174d-af83-4842-9e6f-bf2a32c40bf5\" (UID: \"5669174d-af83-4842-9e6f-bf2a32c40bf5\") " Jan 27 16:33:15 crc kubenswrapper[4772]: I0127 16:33:15.209915 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnwh9\" (UniqueName: \"kubernetes.io/projected/5669174d-af83-4842-9e6f-bf2a32c40bf5-kube-api-access-mnwh9\") pod \"5669174d-af83-4842-9e6f-bf2a32c40bf5\" (UID: \"5669174d-af83-4842-9e6f-bf2a32c40bf5\") " Jan 27 16:33:15 crc kubenswrapper[4772]: I0127 16:33:15.210364 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5669174d-af83-4842-9e6f-bf2a32c40bf5-config" (OuterVolumeSpecName: "config") pod "5669174d-af83-4842-9e6f-bf2a32c40bf5" (UID: "5669174d-af83-4842-9e6f-bf2a32c40bf5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:33:15 crc kubenswrapper[4772]: I0127 16:33:15.210498 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5669174d-af83-4842-9e6f-bf2a32c40bf5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5669174d-af83-4842-9e6f-bf2a32c40bf5" (UID: "5669174d-af83-4842-9e6f-bf2a32c40bf5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:33:15 crc kubenswrapper[4772]: I0127 16:33:15.210949 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5669174d-af83-4842-9e6f-bf2a32c40bf5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5669174d-af83-4842-9e6f-bf2a32c40bf5" (UID: "5669174d-af83-4842-9e6f-bf2a32c40bf5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:33:15 crc kubenswrapper[4772]: I0127 16:33:15.211149 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5669174d-af83-4842-9e6f-bf2a32c40bf5-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:15 crc kubenswrapper[4772]: I0127 16:33:15.211184 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5669174d-af83-4842-9e6f-bf2a32c40bf5-config\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:15 crc kubenswrapper[4772]: I0127 16:33:15.217016 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5669174d-af83-4842-9e6f-bf2a32c40bf5-kube-api-access-mnwh9" (OuterVolumeSpecName: "kube-api-access-mnwh9") pod "5669174d-af83-4842-9e6f-bf2a32c40bf5" (UID: "5669174d-af83-4842-9e6f-bf2a32c40bf5"). InnerVolumeSpecName "kube-api-access-mnwh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:33:15 crc kubenswrapper[4772]: I0127 16:33:15.312891 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnwh9\" (UniqueName: \"kubernetes.io/projected/5669174d-af83-4842-9e6f-bf2a32c40bf5-kube-api-access-mnwh9\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:15 crc kubenswrapper[4772]: I0127 16:33:15.312934 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5669174d-af83-4842-9e6f-bf2a32c40bf5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:15 crc kubenswrapper[4772]: I0127 16:33:15.346879 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-746d786469-qv9pq"] Jan 27 16:33:15 crc kubenswrapper[4772]: I0127 16:33:15.662944 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:33:15 crc kubenswrapper[4772]: E0127 16:33:15.663524 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:33:16 crc kubenswrapper[4772]: I0127 16:33:16.114490 4772 generic.go:334] "Generic (PLEG): container finished" podID="320b0d26-aa22-4854-ac28-4699f95fb37b" containerID="02bb8fc05201bb2286e71696309cca885abf131326f1399ef983ae5f44f953bf" exitCode=0 Jan 27 16:33:16 crc kubenswrapper[4772]: I0127 16:33:16.114584 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746d786469-qv9pq" event={"ID":"320b0d26-aa22-4854-ac28-4699f95fb37b","Type":"ContainerDied","Data":"02bb8fc05201bb2286e71696309cca885abf131326f1399ef983ae5f44f953bf"} Jan 27 16:33:16 crc kubenswrapper[4772]: I0127 16:33:16.114599 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f7b485f7-qqlrk" Jan 27 16:33:16 crc kubenswrapper[4772]: I0127 16:33:16.114619 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746d786469-qv9pq" event={"ID":"320b0d26-aa22-4854-ac28-4699f95fb37b","Type":"ContainerStarted","Data":"551776dfcff4ae5d83241e31aab7653a20a9814c77297414bc41ae9b8bf9cecd"} Jan 27 16:33:16 crc kubenswrapper[4772]: I0127 16:33:16.245921 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f7b485f7-qqlrk"] Jan 27 16:33:16 crc kubenswrapper[4772]: I0127 16:33:16.256696 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f7b485f7-qqlrk"] Jan 27 16:33:16 crc kubenswrapper[4772]: I0127 16:33:16.685003 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5669174d-af83-4842-9e6f-bf2a32c40bf5" path="/var/lib/kubelet/pods/5669174d-af83-4842-9e6f-bf2a32c40bf5/volumes" Jan 27 16:33:17 crc kubenswrapper[4772]: I0127 16:33:17.079553 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Jan 27 16:33:17 crc kubenswrapper[4772]: I0127 16:33:17.082903 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Jan 27 16:33:17 crc kubenswrapper[4772]: I0127 16:33:17.086207 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Jan 27 16:33:17 crc kubenswrapper[4772]: I0127 16:33:17.101691 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Jan 27 16:33:17 crc kubenswrapper[4772]: I0127 16:33:17.126201 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746d786469-qv9pq" event={"ID":"320b0d26-aa22-4854-ac28-4699f95fb37b","Type":"ContainerStarted","Data":"07f4aa1fb5d6e810824b11ccdad9666a53ee8ab3a0bbb958fb4483c8bf5dd691"} Jan 27 16:33:17 crc kubenswrapper[4772]: I0127 16:33:17.126578 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-746d786469-qv9pq" Jan 27 16:33:17 crc kubenswrapper[4772]: I0127 16:33:17.145032 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-746d786469-qv9pq" podStartSLOduration=3.145016254 podStartE2EDuration="3.145016254s" podCreationTimestamp="2026-01-27 16:33:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:33:17.140921987 +0000 UTC m=+5183.121531115" watchObservedRunningTime="2026-01-27 16:33:17.145016254 +0000 UTC m=+5183.125625352" Jan 27 16:33:17 crc kubenswrapper[4772]: I0127 16:33:17.242040 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/6d673b09-a15f-48fc-b399-212dc30fce29-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"6d673b09-a15f-48fc-b399-212dc30fce29\") " pod="openstack/ovn-copy-data" Jan 27 16:33:17 crc kubenswrapper[4772]: I0127 16:33:17.242234 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-4ff4ab2b-2cd2-45d6-b694-fe0a25175676\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4ff4ab2b-2cd2-45d6-b694-fe0a25175676\") pod \"ovn-copy-data\" (UID: \"6d673b09-a15f-48fc-b399-212dc30fce29\") " pod="openstack/ovn-copy-data" Jan 27 16:33:17 crc kubenswrapper[4772]: I0127 16:33:17.242428 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsvvx\" (UniqueName: \"kubernetes.io/projected/6d673b09-a15f-48fc-b399-212dc30fce29-kube-api-access-wsvvx\") pod \"ovn-copy-data\" (UID: \"6d673b09-a15f-48fc-b399-212dc30fce29\") " pod="openstack/ovn-copy-data" Jan 27 16:33:17 crc kubenswrapper[4772]: I0127 16:33:17.344369 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/6d673b09-a15f-48fc-b399-212dc30fce29-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"6d673b09-a15f-48fc-b399-212dc30fce29\") " pod="openstack/ovn-copy-data" Jan 27 16:33:17 crc kubenswrapper[4772]: I0127 16:33:17.344459 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-4ff4ab2b-2cd2-45d6-b694-fe0a25175676\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4ff4ab2b-2cd2-45d6-b694-fe0a25175676\") pod \"ovn-copy-data\" (UID: \"6d673b09-a15f-48fc-b399-212dc30fce29\") " pod="openstack/ovn-copy-data" Jan 27 16:33:17 crc kubenswrapper[4772]: I0127 16:33:17.344525 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsvvx\" (UniqueName: \"kubernetes.io/projected/6d673b09-a15f-48fc-b399-212dc30fce29-kube-api-access-wsvvx\") pod \"ovn-copy-data\" (UID: \"6d673b09-a15f-48fc-b399-212dc30fce29\") " pod="openstack/ovn-copy-data" Jan 27 16:33:17 crc kubenswrapper[4772]: I0127 16:33:17.347867 4772 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 27 16:33:17 crc kubenswrapper[4772]: I0127 16:33:17.347917 4772 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-4ff4ab2b-2cd2-45d6-b694-fe0a25175676\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4ff4ab2b-2cd2-45d6-b694-fe0a25175676\") pod \"ovn-copy-data\" (UID: \"6d673b09-a15f-48fc-b399-212dc30fce29\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8a0d0ab9791e17485b8426753f6dd9038e92be360783b83ca2c6a8c88a22753f/globalmount\"" pod="openstack/ovn-copy-data" Jan 27 16:33:17 crc kubenswrapper[4772]: I0127 16:33:17.358141 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/6d673b09-a15f-48fc-b399-212dc30fce29-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"6d673b09-a15f-48fc-b399-212dc30fce29\") " pod="openstack/ovn-copy-data" Jan 27 16:33:17 crc kubenswrapper[4772]: I0127 16:33:17.363243 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsvvx\" (UniqueName: \"kubernetes.io/projected/6d673b09-a15f-48fc-b399-212dc30fce29-kube-api-access-wsvvx\") pod \"ovn-copy-data\" (UID: \"6d673b09-a15f-48fc-b399-212dc30fce29\") " pod="openstack/ovn-copy-data" Jan 27 16:33:17 crc kubenswrapper[4772]: I0127 16:33:17.378911 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-4ff4ab2b-2cd2-45d6-b694-fe0a25175676\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4ff4ab2b-2cd2-45d6-b694-fe0a25175676\") pod \"ovn-copy-data\" (UID: \"6d673b09-a15f-48fc-b399-212dc30fce29\") " pod="openstack/ovn-copy-data" Jan 27 16:33:17 crc kubenswrapper[4772]: I0127 16:33:17.408999 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Jan 27 16:33:17 crc kubenswrapper[4772]: I0127 16:33:17.908904 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Jan 27 16:33:17 crc kubenswrapper[4772]: W0127 16:33:17.912434 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d673b09_a15f_48fc_b399_212dc30fce29.slice/crio-1357c90249e46be748745d14abaf363cb618d3df528722c24847f6ae9a771835 WatchSource:0}: Error finding container 1357c90249e46be748745d14abaf363cb618d3df528722c24847f6ae9a771835: Status 404 returned error can't find the container with id 1357c90249e46be748745d14abaf363cb618d3df528722c24847f6ae9a771835 Jan 27 16:33:18 crc kubenswrapper[4772]: I0127 16:33:18.133985 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"6d673b09-a15f-48fc-b399-212dc30fce29","Type":"ContainerStarted","Data":"40145951fa81f81ceb443f4e7a103616579dc22c8564338129dc861532e12471"} Jan 27 16:33:18 crc kubenswrapper[4772]: I0127 16:33:18.134408 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"6d673b09-a15f-48fc-b399-212dc30fce29","Type":"ContainerStarted","Data":"1357c90249e46be748745d14abaf363cb618d3df528722c24847f6ae9a771835"} Jan 27 16:33:18 crc kubenswrapper[4772]: I0127 16:33:18.155204 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=2.155159506 podStartE2EDuration="2.155159506s" podCreationTimestamp="2026-01-27 16:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:33:18.151211703 +0000 UTC m=+5184.131820811" watchObservedRunningTime="2026-01-27 16:33:18.155159506 +0000 UTC m=+5184.135768604" Jan 27 16:33:23 crc kubenswrapper[4772]: I0127 16:33:23.028949 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Jan 27 16:33:23 crc kubenswrapper[4772]: I0127 16:33:23.035730 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Jan 27 16:33:23 crc kubenswrapper[4772]: I0127 16:33:23.039003 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Jan 27 16:33:23 crc kubenswrapper[4772]: I0127 16:33:23.039262 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-xbpms" Jan 27 16:33:23 crc kubenswrapper[4772]: I0127 16:33:23.039394 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Jan 27 16:33:23 crc kubenswrapper[4772]: I0127 16:33:23.052922 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Jan 27 16:33:23 crc kubenswrapper[4772]: I0127 16:33:23.153946 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9c221996-e15f-4fe3-bc62-98aac08f546f-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"9c221996-e15f-4fe3-bc62-98aac08f546f\") " pod="openstack/ovn-northd-0" Jan 27 16:33:23 crc kubenswrapper[4772]: I0127 16:33:23.154005 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhfjn\" (UniqueName: \"kubernetes.io/projected/9c221996-e15f-4fe3-bc62-98aac08f546f-kube-api-access-mhfjn\") pod \"ovn-northd-0\" (UID: \"9c221996-e15f-4fe3-bc62-98aac08f546f\") " pod="openstack/ovn-northd-0" Jan 27 16:33:23 crc kubenswrapper[4772]: I0127 16:33:23.154028 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c221996-e15f-4fe3-bc62-98aac08f546f-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"9c221996-e15f-4fe3-bc62-98aac08f546f\") " pod="openstack/ovn-northd-0" Jan 27 16:33:23 crc kubenswrapper[4772]: I0127 16:33:23.154055 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c221996-e15f-4fe3-bc62-98aac08f546f-config\") pod \"ovn-northd-0\" (UID: \"9c221996-e15f-4fe3-bc62-98aac08f546f\") " pod="openstack/ovn-northd-0" Jan 27 16:33:23 crc kubenswrapper[4772]: I0127 16:33:23.154138 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c221996-e15f-4fe3-bc62-98aac08f546f-scripts\") pod \"ovn-northd-0\" (UID: \"9c221996-e15f-4fe3-bc62-98aac08f546f\") " pod="openstack/ovn-northd-0" Jan 27 16:33:23 crc kubenswrapper[4772]: I0127 16:33:23.255472 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c221996-e15f-4fe3-bc62-98aac08f546f-scripts\") pod \"ovn-northd-0\" (UID: \"9c221996-e15f-4fe3-bc62-98aac08f546f\") " pod="openstack/ovn-northd-0" Jan 27 16:33:23 crc kubenswrapper[4772]: I0127 16:33:23.255619 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9c221996-e15f-4fe3-bc62-98aac08f546f-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"9c221996-e15f-4fe3-bc62-98aac08f546f\") " pod="openstack/ovn-northd-0" Jan 27 16:33:23 crc kubenswrapper[4772]: I0127 16:33:23.255650 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhfjn\" (UniqueName: \"kubernetes.io/projected/9c221996-e15f-4fe3-bc62-98aac08f546f-kube-api-access-mhfjn\") pod \"ovn-northd-0\" (UID: \"9c221996-e15f-4fe3-bc62-98aac08f546f\") " pod="openstack/ovn-northd-0" Jan 27 16:33:23 crc kubenswrapper[4772]: I0127 16:33:23.255677 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c221996-e15f-4fe3-bc62-98aac08f546f-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"9c221996-e15f-4fe3-bc62-98aac08f546f\") " pod="openstack/ovn-northd-0" Jan 27 16:33:23 crc kubenswrapper[4772]: I0127 16:33:23.255701 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c221996-e15f-4fe3-bc62-98aac08f546f-config\") pod \"ovn-northd-0\" (UID: \"9c221996-e15f-4fe3-bc62-98aac08f546f\") " pod="openstack/ovn-northd-0" Jan 27 16:33:23 crc kubenswrapper[4772]: I0127 16:33:23.256852 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c221996-e15f-4fe3-bc62-98aac08f546f-config\") pod \"ovn-northd-0\" (UID: \"9c221996-e15f-4fe3-bc62-98aac08f546f\") " pod="openstack/ovn-northd-0" Jan 27 16:33:23 crc kubenswrapper[4772]: I0127 16:33:23.257549 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c221996-e15f-4fe3-bc62-98aac08f546f-scripts\") pod \"ovn-northd-0\" (UID: \"9c221996-e15f-4fe3-bc62-98aac08f546f\") " pod="openstack/ovn-northd-0" Jan 27 16:33:23 crc kubenswrapper[4772]: I0127 16:33:23.258548 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9c221996-e15f-4fe3-bc62-98aac08f546f-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"9c221996-e15f-4fe3-bc62-98aac08f546f\") " pod="openstack/ovn-northd-0" Jan 27 16:33:23 crc kubenswrapper[4772]: I0127 16:33:23.263986 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c221996-e15f-4fe3-bc62-98aac08f546f-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"9c221996-e15f-4fe3-bc62-98aac08f546f\") " pod="openstack/ovn-northd-0" Jan 27 16:33:23 crc kubenswrapper[4772]: I0127 16:33:23.285050 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhfjn\" (UniqueName: \"kubernetes.io/projected/9c221996-e15f-4fe3-bc62-98aac08f546f-kube-api-access-mhfjn\") pod \"ovn-northd-0\" (UID: \"9c221996-e15f-4fe3-bc62-98aac08f546f\") " pod="openstack/ovn-northd-0" Jan 27 16:33:23 crc kubenswrapper[4772]: I0127 16:33:23.374059 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Jan 27 16:33:23 crc kubenswrapper[4772]: I0127 16:33:23.811256 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Jan 27 16:33:23 crc kubenswrapper[4772]: W0127 16:33:23.816789 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c221996_e15f_4fe3_bc62_98aac08f546f.slice/crio-1d9dd40e970c1b7120b85e055948429cd5f75b4a60d029ba58b71540e785fa8f WatchSource:0}: Error finding container 1d9dd40e970c1b7120b85e055948429cd5f75b4a60d029ba58b71540e785fa8f: Status 404 returned error can't find the container with id 1d9dd40e970c1b7120b85e055948429cd5f75b4a60d029ba58b71540e785fa8f Jan 27 16:33:23 crc kubenswrapper[4772]: E0127 16:33:23.953507 4772 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.134:34104->38.129.56.134:35895: write tcp 38.129.56.134:34104->38.129.56.134:35895: write: broken pipe Jan 27 16:33:24 crc kubenswrapper[4772]: I0127 16:33:24.189612 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"9c221996-e15f-4fe3-bc62-98aac08f546f","Type":"ContainerStarted","Data":"d06cea6eaa4e35aa000408c4b69138af9e6765b1225e007fdaeb79b67273c595"} Jan 27 16:33:24 crc kubenswrapper[4772]: I0127 16:33:24.189650 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"9c221996-e15f-4fe3-bc62-98aac08f546f","Type":"ContainerStarted","Data":"922270d0a409360020cdc5184f402a6dc5344a591d52c8f4c4caa5187ff1e91a"} Jan 27 16:33:24 crc kubenswrapper[4772]: I0127 16:33:24.189661 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"9c221996-e15f-4fe3-bc62-98aac08f546f","Type":"ContainerStarted","Data":"1d9dd40e970c1b7120b85e055948429cd5f75b4a60d029ba58b71540e785fa8f"} Jan 27 16:33:24 crc kubenswrapper[4772]: I0127 16:33:24.189763 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Jan 27 16:33:24 crc kubenswrapper[4772]: I0127 16:33:24.207672 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.2076514760000001 podStartE2EDuration="1.207651476s" podCreationTimestamp="2026-01-27 16:33:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:33:24.206426561 +0000 UTC m=+5190.187035659" watchObservedRunningTime="2026-01-27 16:33:24.207651476 +0000 UTC m=+5190.188260574" Jan 27 16:33:24 crc kubenswrapper[4772]: I0127 16:33:24.942498 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-746d786469-qv9pq" Jan 27 16:33:24 crc kubenswrapper[4772]: I0127 16:33:24.994636 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-ndkll"] Jan 27 16:33:24 crc kubenswrapper[4772]: I0127 16:33:24.995121 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b7946d7b9-ndkll" podUID="51a5db5c-8de5-441d-a8e9-7c07acc7df31" containerName="dnsmasq-dns" containerID="cri-o://a355488297e660433fbaef5e701f58797760e94b791d017a807af73346bc3756" gracePeriod=10 Jan 27 16:33:25 crc kubenswrapper[4772]: I0127 16:33:25.109665 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b7946d7b9-ndkll" podUID="51a5db5c-8de5-441d-a8e9-7c07acc7df31" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.244:5353: connect: connection refused" Jan 27 16:33:25 crc kubenswrapper[4772]: I0127 16:33:25.202583 4772 generic.go:334] "Generic (PLEG): container finished" podID="51a5db5c-8de5-441d-a8e9-7c07acc7df31" containerID="a355488297e660433fbaef5e701f58797760e94b791d017a807af73346bc3756" exitCode=0 Jan 27 16:33:25 crc kubenswrapper[4772]: I0127 16:33:25.203520 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-ndkll" event={"ID":"51a5db5c-8de5-441d-a8e9-7c07acc7df31","Type":"ContainerDied","Data":"a355488297e660433fbaef5e701f58797760e94b791d017a807af73346bc3756"} Jan 27 16:33:25 crc kubenswrapper[4772]: I0127 16:33:25.481567 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-ndkll" Jan 27 16:33:25 crc kubenswrapper[4772]: I0127 16:33:25.598643 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7m72l\" (UniqueName: \"kubernetes.io/projected/51a5db5c-8de5-441d-a8e9-7c07acc7df31-kube-api-access-7m72l\") pod \"51a5db5c-8de5-441d-a8e9-7c07acc7df31\" (UID: \"51a5db5c-8de5-441d-a8e9-7c07acc7df31\") " Jan 27 16:33:25 crc kubenswrapper[4772]: I0127 16:33:25.598714 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a5db5c-8de5-441d-a8e9-7c07acc7df31-config\") pod \"51a5db5c-8de5-441d-a8e9-7c07acc7df31\" (UID: \"51a5db5c-8de5-441d-a8e9-7c07acc7df31\") " Jan 27 16:33:25 crc kubenswrapper[4772]: I0127 16:33:25.598854 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a5db5c-8de5-441d-a8e9-7c07acc7df31-dns-svc\") pod \"51a5db5c-8de5-441d-a8e9-7c07acc7df31\" (UID: \"51a5db5c-8de5-441d-a8e9-7c07acc7df31\") " Jan 27 16:33:25 crc kubenswrapper[4772]: I0127 16:33:25.606409 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51a5db5c-8de5-441d-a8e9-7c07acc7df31-kube-api-access-7m72l" (OuterVolumeSpecName: "kube-api-access-7m72l") pod "51a5db5c-8de5-441d-a8e9-7c07acc7df31" (UID: "51a5db5c-8de5-441d-a8e9-7c07acc7df31"). InnerVolumeSpecName "kube-api-access-7m72l". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:33:25 crc kubenswrapper[4772]: I0127 16:33:25.635137 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51a5db5c-8de5-441d-a8e9-7c07acc7df31-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "51a5db5c-8de5-441d-a8e9-7c07acc7df31" (UID: "51a5db5c-8de5-441d-a8e9-7c07acc7df31"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:33:25 crc kubenswrapper[4772]: I0127 16:33:25.635821 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51a5db5c-8de5-441d-a8e9-7c07acc7df31-config" (OuterVolumeSpecName: "config") pod "51a5db5c-8de5-441d-a8e9-7c07acc7df31" (UID: "51a5db5c-8de5-441d-a8e9-7c07acc7df31"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:33:25 crc kubenswrapper[4772]: I0127 16:33:25.700691 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a5db5c-8de5-441d-a8e9-7c07acc7df31-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:25 crc kubenswrapper[4772]: I0127 16:33:25.700724 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7m72l\" (UniqueName: \"kubernetes.io/projected/51a5db5c-8de5-441d-a8e9-7c07acc7df31-kube-api-access-7m72l\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:25 crc kubenswrapper[4772]: I0127 16:33:25.700734 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a5db5c-8de5-441d-a8e9-7c07acc7df31-config\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:26 crc kubenswrapper[4772]: I0127 16:33:26.210774 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-ndkll" event={"ID":"51a5db5c-8de5-441d-a8e9-7c07acc7df31","Type":"ContainerDied","Data":"cd506cf5289047aba524d3325375f9767877e5e9af047a505f68883c25d7ad72"} Jan 27 16:33:26 crc kubenswrapper[4772]: I0127 16:33:26.211111 4772 scope.go:117] "RemoveContainer" containerID="a355488297e660433fbaef5e701f58797760e94b791d017a807af73346bc3756" Jan 27 16:33:26 crc kubenswrapper[4772]: I0127 16:33:26.210866 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-ndkll" Jan 27 16:33:26 crc kubenswrapper[4772]: I0127 16:33:26.242590 4772 scope.go:117] "RemoveContainer" containerID="2b48ddd2c1b4ddcb3e7d2673318ce52a0b89618a27b83de04520bd8160030f43" Jan 27 16:33:26 crc kubenswrapper[4772]: I0127 16:33:26.245891 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-ndkll"] Jan 27 16:33:26 crc kubenswrapper[4772]: I0127 16:33:26.252035 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-ndkll"] Jan 27 16:33:26 crc kubenswrapper[4772]: I0127 16:33:26.671864 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51a5db5c-8de5-441d-a8e9-7c07acc7df31" path="/var/lib/kubelet/pods/51a5db5c-8de5-441d-a8e9-7c07acc7df31/volumes" Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.071215 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-jb5ch"] Jan 27 16:33:28 crc kubenswrapper[4772]: E0127 16:33:28.071967 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51a5db5c-8de5-441d-a8e9-7c07acc7df31" containerName="init" Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.071984 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="51a5db5c-8de5-441d-a8e9-7c07acc7df31" containerName="init" Jan 27 16:33:28 crc kubenswrapper[4772]: E0127 16:33:28.072020 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51a5db5c-8de5-441d-a8e9-7c07acc7df31" containerName="dnsmasq-dns" Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.072028 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="51a5db5c-8de5-441d-a8e9-7c07acc7df31" containerName="dnsmasq-dns" Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.072232 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="51a5db5c-8de5-441d-a8e9-7c07acc7df31" containerName="dnsmasq-dns" Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.072897 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jb5ch" Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.083414 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jb5ch"] Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.137588 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b8sp\" (UniqueName: \"kubernetes.io/projected/5292a043-9ee5-4d14-a991-c50dbf4d136e-kube-api-access-6b8sp\") pod \"keystone-db-create-jb5ch\" (UID: \"5292a043-9ee5-4d14-a991-c50dbf4d136e\") " pod="openstack/keystone-db-create-jb5ch" Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.138450 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5292a043-9ee5-4d14-a991-c50dbf4d136e-operator-scripts\") pod \"keystone-db-create-jb5ch\" (UID: \"5292a043-9ee5-4d14-a991-c50dbf4d136e\") " pod="openstack/keystone-db-create-jb5ch" Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.179633 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-0c73-account-create-update-nz44z"] Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.181630 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-0c73-account-create-update-nz44z" Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.184677 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.189892 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-0c73-account-create-update-nz44z"] Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.240129 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5292a043-9ee5-4d14-a991-c50dbf4d136e-operator-scripts\") pod \"keystone-db-create-jb5ch\" (UID: \"5292a043-9ee5-4d14-a991-c50dbf4d136e\") " pod="openstack/keystone-db-create-jb5ch" Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.240219 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b8sp\" (UniqueName: \"kubernetes.io/projected/5292a043-9ee5-4d14-a991-c50dbf4d136e-kube-api-access-6b8sp\") pod \"keystone-db-create-jb5ch\" (UID: \"5292a043-9ee5-4d14-a991-c50dbf4d136e\") " pod="openstack/keystone-db-create-jb5ch" Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.240961 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5292a043-9ee5-4d14-a991-c50dbf4d136e-operator-scripts\") pod \"keystone-db-create-jb5ch\" (UID: \"5292a043-9ee5-4d14-a991-c50dbf4d136e\") " pod="openstack/keystone-db-create-jb5ch" Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.261627 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b8sp\" (UniqueName: \"kubernetes.io/projected/5292a043-9ee5-4d14-a991-c50dbf4d136e-kube-api-access-6b8sp\") pod \"keystone-db-create-jb5ch\" (UID: \"5292a043-9ee5-4d14-a991-c50dbf4d136e\") " pod="openstack/keystone-db-create-jb5ch" Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.341263 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7e8906f-73a7-4580-81c8-ec81439faea5-operator-scripts\") pod \"keystone-0c73-account-create-update-nz44z\" (UID: \"d7e8906f-73a7-4580-81c8-ec81439faea5\") " pod="openstack/keystone-0c73-account-create-update-nz44z" Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.341601 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khz8d\" (UniqueName: \"kubernetes.io/projected/d7e8906f-73a7-4580-81c8-ec81439faea5-kube-api-access-khz8d\") pod \"keystone-0c73-account-create-update-nz44z\" (UID: \"d7e8906f-73a7-4580-81c8-ec81439faea5\") " pod="openstack/keystone-0c73-account-create-update-nz44z" Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.395697 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jb5ch" Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.443674 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khz8d\" (UniqueName: \"kubernetes.io/projected/d7e8906f-73a7-4580-81c8-ec81439faea5-kube-api-access-khz8d\") pod \"keystone-0c73-account-create-update-nz44z\" (UID: \"d7e8906f-73a7-4580-81c8-ec81439faea5\") " pod="openstack/keystone-0c73-account-create-update-nz44z" Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.443788 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7e8906f-73a7-4580-81c8-ec81439faea5-operator-scripts\") pod \"keystone-0c73-account-create-update-nz44z\" (UID: \"d7e8906f-73a7-4580-81c8-ec81439faea5\") " pod="openstack/keystone-0c73-account-create-update-nz44z" Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.444751 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7e8906f-73a7-4580-81c8-ec81439faea5-operator-scripts\") pod \"keystone-0c73-account-create-update-nz44z\" (UID: \"d7e8906f-73a7-4580-81c8-ec81439faea5\") " pod="openstack/keystone-0c73-account-create-update-nz44z" Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.463133 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khz8d\" (UniqueName: \"kubernetes.io/projected/d7e8906f-73a7-4580-81c8-ec81439faea5-kube-api-access-khz8d\") pod \"keystone-0c73-account-create-update-nz44z\" (UID: \"d7e8906f-73a7-4580-81c8-ec81439faea5\") " pod="openstack/keystone-0c73-account-create-update-nz44z" Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.501018 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-0c73-account-create-update-nz44z" Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.883955 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jb5ch"] Jan 27 16:33:28 crc kubenswrapper[4772]: W0127 16:33:28.885466 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5292a043_9ee5_4d14_a991_c50dbf4d136e.slice/crio-723120bd2af0be8b0958f15b4fca165d3b69967a008c9119ea9f2ca64aa3dad0 WatchSource:0}: Error finding container 723120bd2af0be8b0958f15b4fca165d3b69967a008c9119ea9f2ca64aa3dad0: Status 404 returned error can't find the container with id 723120bd2af0be8b0958f15b4fca165d3b69967a008c9119ea9f2ca64aa3dad0 Jan 27 16:33:28 crc kubenswrapper[4772]: I0127 16:33:28.944705 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-0c73-account-create-update-nz44z"] Jan 27 16:33:29 crc kubenswrapper[4772]: I0127 16:33:29.236448 4772 generic.go:334] "Generic (PLEG): container finished" podID="5292a043-9ee5-4d14-a991-c50dbf4d136e" containerID="6b8b1c4b2ab6a42f7f2fdd3da73f914caa96bb6a3722955fd242601b775105ee" exitCode=0 Jan 27 16:33:29 crc kubenswrapper[4772]: I0127 16:33:29.236504 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jb5ch" event={"ID":"5292a043-9ee5-4d14-a991-c50dbf4d136e","Type":"ContainerDied","Data":"6b8b1c4b2ab6a42f7f2fdd3da73f914caa96bb6a3722955fd242601b775105ee"} Jan 27 16:33:29 crc kubenswrapper[4772]: I0127 16:33:29.236583 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jb5ch" event={"ID":"5292a043-9ee5-4d14-a991-c50dbf4d136e","Type":"ContainerStarted","Data":"723120bd2af0be8b0958f15b4fca165d3b69967a008c9119ea9f2ca64aa3dad0"} Jan 27 16:33:29 crc kubenswrapper[4772]: I0127 16:33:29.238043 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-0c73-account-create-update-nz44z" event={"ID":"d7e8906f-73a7-4580-81c8-ec81439faea5","Type":"ContainerStarted","Data":"058f6eb2f4f5195120a96ef6b4691590bd77114e18b6ed23148cdd3864b329b4"} Jan 27 16:33:29 crc kubenswrapper[4772]: I0127 16:33:29.238072 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-0c73-account-create-update-nz44z" event={"ID":"d7e8906f-73a7-4580-81c8-ec81439faea5","Type":"ContainerStarted","Data":"a51cca167732ea4bcca0a9ad60e145a7ef6c1318026eaf99325988eb86cccb02"} Jan 27 16:33:29 crc kubenswrapper[4772]: I0127 16:33:29.265961 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-0c73-account-create-update-nz44z" podStartSLOduration=1.265943476 podStartE2EDuration="1.265943476s" podCreationTimestamp="2026-01-27 16:33:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:33:29.259629516 +0000 UTC m=+5195.240238614" watchObservedRunningTime="2026-01-27 16:33:29.265943476 +0000 UTC m=+5195.246552564" Jan 27 16:33:30 crc kubenswrapper[4772]: I0127 16:33:30.250532 4772 generic.go:334] "Generic (PLEG): container finished" podID="d7e8906f-73a7-4580-81c8-ec81439faea5" containerID="058f6eb2f4f5195120a96ef6b4691590bd77114e18b6ed23148cdd3864b329b4" exitCode=0 Jan 27 16:33:30 crc kubenswrapper[4772]: I0127 16:33:30.250679 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-0c73-account-create-update-nz44z" event={"ID":"d7e8906f-73a7-4580-81c8-ec81439faea5","Type":"ContainerDied","Data":"058f6eb2f4f5195120a96ef6b4691590bd77114e18b6ed23148cdd3864b329b4"} Jan 27 16:33:30 crc kubenswrapper[4772]: I0127 16:33:30.574617 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jb5ch" Jan 27 16:33:30 crc kubenswrapper[4772]: I0127 16:33:30.663741 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:33:30 crc kubenswrapper[4772]: E0127 16:33:30.664041 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:33:30 crc kubenswrapper[4772]: I0127 16:33:30.688851 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5292a043-9ee5-4d14-a991-c50dbf4d136e-operator-scripts\") pod \"5292a043-9ee5-4d14-a991-c50dbf4d136e\" (UID: \"5292a043-9ee5-4d14-a991-c50dbf4d136e\") " Jan 27 16:33:30 crc kubenswrapper[4772]: I0127 16:33:30.689056 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6b8sp\" (UniqueName: \"kubernetes.io/projected/5292a043-9ee5-4d14-a991-c50dbf4d136e-kube-api-access-6b8sp\") pod \"5292a043-9ee5-4d14-a991-c50dbf4d136e\" (UID: \"5292a043-9ee5-4d14-a991-c50dbf4d136e\") " Jan 27 16:33:30 crc kubenswrapper[4772]: I0127 16:33:30.689696 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5292a043-9ee5-4d14-a991-c50dbf4d136e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5292a043-9ee5-4d14-a991-c50dbf4d136e" (UID: "5292a043-9ee5-4d14-a991-c50dbf4d136e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:33:30 crc kubenswrapper[4772]: I0127 16:33:30.695931 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5292a043-9ee5-4d14-a991-c50dbf4d136e-kube-api-access-6b8sp" (OuterVolumeSpecName: "kube-api-access-6b8sp") pod "5292a043-9ee5-4d14-a991-c50dbf4d136e" (UID: "5292a043-9ee5-4d14-a991-c50dbf4d136e"). InnerVolumeSpecName "kube-api-access-6b8sp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:33:30 crc kubenswrapper[4772]: I0127 16:33:30.791055 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5292a043-9ee5-4d14-a991-c50dbf4d136e-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:30 crc kubenswrapper[4772]: I0127 16:33:30.791088 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6b8sp\" (UniqueName: \"kubernetes.io/projected/5292a043-9ee5-4d14-a991-c50dbf4d136e-kube-api-access-6b8sp\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:31 crc kubenswrapper[4772]: I0127 16:33:31.260629 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jb5ch" Jan 27 16:33:31 crc kubenswrapper[4772]: I0127 16:33:31.260629 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jb5ch" event={"ID":"5292a043-9ee5-4d14-a991-c50dbf4d136e","Type":"ContainerDied","Data":"723120bd2af0be8b0958f15b4fca165d3b69967a008c9119ea9f2ca64aa3dad0"} Jan 27 16:33:31 crc kubenswrapper[4772]: I0127 16:33:31.261086 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="723120bd2af0be8b0958f15b4fca165d3b69967a008c9119ea9f2ca64aa3dad0" Jan 27 16:33:31 crc kubenswrapper[4772]: I0127 16:33:31.623151 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-0c73-account-create-update-nz44z" Jan 27 16:33:31 crc kubenswrapper[4772]: I0127 16:33:31.705101 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7e8906f-73a7-4580-81c8-ec81439faea5-operator-scripts\") pod \"d7e8906f-73a7-4580-81c8-ec81439faea5\" (UID: \"d7e8906f-73a7-4580-81c8-ec81439faea5\") " Jan 27 16:33:31 crc kubenswrapper[4772]: I0127 16:33:31.705212 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khz8d\" (UniqueName: \"kubernetes.io/projected/d7e8906f-73a7-4580-81c8-ec81439faea5-kube-api-access-khz8d\") pod \"d7e8906f-73a7-4580-81c8-ec81439faea5\" (UID: \"d7e8906f-73a7-4580-81c8-ec81439faea5\") " Jan 27 16:33:31 crc kubenswrapper[4772]: I0127 16:33:31.705837 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7e8906f-73a7-4580-81c8-ec81439faea5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d7e8906f-73a7-4580-81c8-ec81439faea5" (UID: "d7e8906f-73a7-4580-81c8-ec81439faea5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:33:31 crc kubenswrapper[4772]: I0127 16:33:31.709410 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7e8906f-73a7-4580-81c8-ec81439faea5-kube-api-access-khz8d" (OuterVolumeSpecName: "kube-api-access-khz8d") pod "d7e8906f-73a7-4580-81c8-ec81439faea5" (UID: "d7e8906f-73a7-4580-81c8-ec81439faea5"). InnerVolumeSpecName "kube-api-access-khz8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:33:31 crc kubenswrapper[4772]: I0127 16:33:31.807279 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7e8906f-73a7-4580-81c8-ec81439faea5-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:31 crc kubenswrapper[4772]: I0127 16:33:31.807323 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khz8d\" (UniqueName: \"kubernetes.io/projected/d7e8906f-73a7-4580-81c8-ec81439faea5-kube-api-access-khz8d\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:32 crc kubenswrapper[4772]: I0127 16:33:32.270097 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-0c73-account-create-update-nz44z" event={"ID":"d7e8906f-73a7-4580-81c8-ec81439faea5","Type":"ContainerDied","Data":"a51cca167732ea4bcca0a9ad60e145a7ef6c1318026eaf99325988eb86cccb02"} Jan 27 16:33:32 crc kubenswrapper[4772]: I0127 16:33:32.270353 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a51cca167732ea4bcca0a9ad60e145a7ef6c1318026eaf99325988eb86cccb02" Jan 27 16:33:32 crc kubenswrapper[4772]: I0127 16:33:32.270153 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-0c73-account-create-update-nz44z" Jan 27 16:33:33 crc kubenswrapper[4772]: I0127 16:33:33.431298 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Jan 27 16:33:33 crc kubenswrapper[4772]: I0127 16:33:33.689922 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-ls87v"] Jan 27 16:33:33 crc kubenswrapper[4772]: E0127 16:33:33.690300 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7e8906f-73a7-4580-81c8-ec81439faea5" containerName="mariadb-account-create-update" Jan 27 16:33:33 crc kubenswrapper[4772]: I0127 16:33:33.690321 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7e8906f-73a7-4580-81c8-ec81439faea5" containerName="mariadb-account-create-update" Jan 27 16:33:33 crc kubenswrapper[4772]: E0127 16:33:33.690345 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5292a043-9ee5-4d14-a991-c50dbf4d136e" containerName="mariadb-database-create" Jan 27 16:33:33 crc kubenswrapper[4772]: I0127 16:33:33.690353 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="5292a043-9ee5-4d14-a991-c50dbf4d136e" containerName="mariadb-database-create" Jan 27 16:33:33 crc kubenswrapper[4772]: I0127 16:33:33.690522 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="5292a043-9ee5-4d14-a991-c50dbf4d136e" containerName="mariadb-database-create" Jan 27 16:33:33 crc kubenswrapper[4772]: I0127 16:33:33.690545 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7e8906f-73a7-4580-81c8-ec81439faea5" containerName="mariadb-account-create-update" Jan 27 16:33:33 crc kubenswrapper[4772]: I0127 16:33:33.691053 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ls87v" Jan 27 16:33:33 crc kubenswrapper[4772]: I0127 16:33:33.693511 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 27 16:33:33 crc kubenswrapper[4772]: I0127 16:33:33.697752 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-p8kvv" Jan 27 16:33:33 crc kubenswrapper[4772]: I0127 16:33:33.697799 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 27 16:33:33 crc kubenswrapper[4772]: I0127 16:33:33.698003 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 27 16:33:33 crc kubenswrapper[4772]: I0127 16:33:33.704317 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-ls87v"] Jan 27 16:33:33 crc kubenswrapper[4772]: I0127 16:33:33.838550 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e179bb8-f9e4-434d-9636-84cc97d632fb-combined-ca-bundle\") pod \"keystone-db-sync-ls87v\" (UID: \"9e179bb8-f9e4-434d-9636-84cc97d632fb\") " pod="openstack/keystone-db-sync-ls87v" Jan 27 16:33:33 crc kubenswrapper[4772]: I0127 16:33:33.838910 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e179bb8-f9e4-434d-9636-84cc97d632fb-config-data\") pod \"keystone-db-sync-ls87v\" (UID: \"9e179bb8-f9e4-434d-9636-84cc97d632fb\") " pod="openstack/keystone-db-sync-ls87v" Jan 27 16:33:33 crc kubenswrapper[4772]: I0127 16:33:33.838978 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24rmv\" (UniqueName: \"kubernetes.io/projected/9e179bb8-f9e4-434d-9636-84cc97d632fb-kube-api-access-24rmv\") pod \"keystone-db-sync-ls87v\" (UID: \"9e179bb8-f9e4-434d-9636-84cc97d632fb\") " pod="openstack/keystone-db-sync-ls87v" Jan 27 16:33:33 crc kubenswrapper[4772]: I0127 16:33:33.940914 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e179bb8-f9e4-434d-9636-84cc97d632fb-combined-ca-bundle\") pod \"keystone-db-sync-ls87v\" (UID: \"9e179bb8-f9e4-434d-9636-84cc97d632fb\") " pod="openstack/keystone-db-sync-ls87v" Jan 27 16:33:33 crc kubenswrapper[4772]: I0127 16:33:33.941065 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e179bb8-f9e4-434d-9636-84cc97d632fb-config-data\") pod \"keystone-db-sync-ls87v\" (UID: \"9e179bb8-f9e4-434d-9636-84cc97d632fb\") " pod="openstack/keystone-db-sync-ls87v" Jan 27 16:33:33 crc kubenswrapper[4772]: I0127 16:33:33.941093 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24rmv\" (UniqueName: \"kubernetes.io/projected/9e179bb8-f9e4-434d-9636-84cc97d632fb-kube-api-access-24rmv\") pod \"keystone-db-sync-ls87v\" (UID: \"9e179bb8-f9e4-434d-9636-84cc97d632fb\") " pod="openstack/keystone-db-sync-ls87v" Jan 27 16:33:33 crc kubenswrapper[4772]: I0127 16:33:33.946853 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e179bb8-f9e4-434d-9636-84cc97d632fb-combined-ca-bundle\") pod \"keystone-db-sync-ls87v\" (UID: \"9e179bb8-f9e4-434d-9636-84cc97d632fb\") " pod="openstack/keystone-db-sync-ls87v" Jan 27 16:33:33 crc kubenswrapper[4772]: I0127 16:33:33.947819 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e179bb8-f9e4-434d-9636-84cc97d632fb-config-data\") pod \"keystone-db-sync-ls87v\" (UID: \"9e179bb8-f9e4-434d-9636-84cc97d632fb\") " pod="openstack/keystone-db-sync-ls87v" Jan 27 16:33:33 crc kubenswrapper[4772]: I0127 16:33:33.987836 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24rmv\" (UniqueName: \"kubernetes.io/projected/9e179bb8-f9e4-434d-9636-84cc97d632fb-kube-api-access-24rmv\") pod \"keystone-db-sync-ls87v\" (UID: \"9e179bb8-f9e4-434d-9636-84cc97d632fb\") " pod="openstack/keystone-db-sync-ls87v" Jan 27 16:33:34 crc kubenswrapper[4772]: I0127 16:33:34.011591 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ls87v" Jan 27 16:33:34 crc kubenswrapper[4772]: I0127 16:33:34.530976 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-ls87v"] Jan 27 16:33:34 crc kubenswrapper[4772]: W0127 16:33:34.533869 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e179bb8_f9e4_434d_9636_84cc97d632fb.slice/crio-6d131a6d7fdbe158c526c8d07c4849e19f1f5996c01dc8bb375bca25b73effc4 WatchSource:0}: Error finding container 6d131a6d7fdbe158c526c8d07c4849e19f1f5996c01dc8bb375bca25b73effc4: Status 404 returned error can't find the container with id 6d131a6d7fdbe158c526c8d07c4849e19f1f5996c01dc8bb375bca25b73effc4 Jan 27 16:33:35 crc kubenswrapper[4772]: I0127 16:33:35.293070 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ls87v" event={"ID":"9e179bb8-f9e4-434d-9636-84cc97d632fb","Type":"ContainerStarted","Data":"5344c5595e4c241d1f3af5be472eb435a5a87e2e74a29b46f8cee0d5b6b1c135"} Jan 27 16:33:35 crc kubenswrapper[4772]: I0127 16:33:35.293433 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ls87v" event={"ID":"9e179bb8-f9e4-434d-9636-84cc97d632fb","Type":"ContainerStarted","Data":"6d131a6d7fdbe158c526c8d07c4849e19f1f5996c01dc8bb375bca25b73effc4"} Jan 27 16:33:35 crc kubenswrapper[4772]: I0127 16:33:35.312802 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-ls87v" podStartSLOduration=2.312782245 podStartE2EDuration="2.312782245s" podCreationTimestamp="2026-01-27 16:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:33:35.307880705 +0000 UTC m=+5201.288489823" watchObservedRunningTime="2026-01-27 16:33:35.312782245 +0000 UTC m=+5201.293391343" Jan 27 16:33:36 crc kubenswrapper[4772]: I0127 16:33:36.300580 4772 generic.go:334] "Generic (PLEG): container finished" podID="9e179bb8-f9e4-434d-9636-84cc97d632fb" containerID="5344c5595e4c241d1f3af5be472eb435a5a87e2e74a29b46f8cee0d5b6b1c135" exitCode=0 Jan 27 16:33:36 crc kubenswrapper[4772]: I0127 16:33:36.300627 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ls87v" event={"ID":"9e179bb8-f9e4-434d-9636-84cc97d632fb","Type":"ContainerDied","Data":"5344c5595e4c241d1f3af5be472eb435a5a87e2e74a29b46f8cee0d5b6b1c135"} Jan 27 16:33:37 crc kubenswrapper[4772]: I0127 16:33:37.668025 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ls87v" Jan 27 16:33:37 crc kubenswrapper[4772]: I0127 16:33:37.799016 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e179bb8-f9e4-434d-9636-84cc97d632fb-combined-ca-bundle\") pod \"9e179bb8-f9e4-434d-9636-84cc97d632fb\" (UID: \"9e179bb8-f9e4-434d-9636-84cc97d632fb\") " Jan 27 16:33:37 crc kubenswrapper[4772]: I0127 16:33:37.799215 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e179bb8-f9e4-434d-9636-84cc97d632fb-config-data\") pod \"9e179bb8-f9e4-434d-9636-84cc97d632fb\" (UID: \"9e179bb8-f9e4-434d-9636-84cc97d632fb\") " Jan 27 16:33:37 crc kubenswrapper[4772]: I0127 16:33:37.799260 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24rmv\" (UniqueName: \"kubernetes.io/projected/9e179bb8-f9e4-434d-9636-84cc97d632fb-kube-api-access-24rmv\") pod \"9e179bb8-f9e4-434d-9636-84cc97d632fb\" (UID: \"9e179bb8-f9e4-434d-9636-84cc97d632fb\") " Jan 27 16:33:37 crc kubenswrapper[4772]: I0127 16:33:37.804694 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e179bb8-f9e4-434d-9636-84cc97d632fb-kube-api-access-24rmv" (OuterVolumeSpecName: "kube-api-access-24rmv") pod "9e179bb8-f9e4-434d-9636-84cc97d632fb" (UID: "9e179bb8-f9e4-434d-9636-84cc97d632fb"). InnerVolumeSpecName "kube-api-access-24rmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:33:37 crc kubenswrapper[4772]: I0127 16:33:37.822321 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e179bb8-f9e4-434d-9636-84cc97d632fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9e179bb8-f9e4-434d-9636-84cc97d632fb" (UID: "9e179bb8-f9e4-434d-9636-84cc97d632fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:33:37 crc kubenswrapper[4772]: I0127 16:33:37.865784 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e179bb8-f9e4-434d-9636-84cc97d632fb-config-data" (OuterVolumeSpecName: "config-data") pod "9e179bb8-f9e4-434d-9636-84cc97d632fb" (UID: "9e179bb8-f9e4-434d-9636-84cc97d632fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:33:37 crc kubenswrapper[4772]: I0127 16:33:37.901066 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e179bb8-f9e4-434d-9636-84cc97d632fb-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:37 crc kubenswrapper[4772]: I0127 16:33:37.901120 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24rmv\" (UniqueName: \"kubernetes.io/projected/9e179bb8-f9e4-434d-9636-84cc97d632fb-kube-api-access-24rmv\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:37 crc kubenswrapper[4772]: I0127 16:33:37.901141 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e179bb8-f9e4-434d-9636-84cc97d632fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:37 crc kubenswrapper[4772]: I0127 16:33:37.984390 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55b86d75d9-f64vw"] Jan 27 16:33:37 crc kubenswrapper[4772]: E0127 16:33:37.984842 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e179bb8-f9e4-434d-9636-84cc97d632fb" containerName="keystone-db-sync" Jan 27 16:33:37 crc kubenswrapper[4772]: I0127 16:33:37.984868 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e179bb8-f9e4-434d-9636-84cc97d632fb" containerName="keystone-db-sync" Jan 27 16:33:37 crc kubenswrapper[4772]: I0127 16:33:37.985061 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e179bb8-f9e4-434d-9636-84cc97d632fb" containerName="keystone-db-sync" Jan 27 16:33:37 crc kubenswrapper[4772]: I0127 16:33:37.990417 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.011590 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-8d7xp"] Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.012763 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8d7xp" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.013398 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55b86d75d9-f64vw"] Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.017938 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.029937 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-8d7xp"] Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.103993 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-fernet-keys\") pod \"keystone-bootstrap-8d7xp\" (UID: \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\") " pod="openstack/keystone-bootstrap-8d7xp" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.104067 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-credential-keys\") pod \"keystone-bootstrap-8d7xp\" (UID: \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\") " pod="openstack/keystone-bootstrap-8d7xp" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.104191 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f5334e8-ad35-41c0-b74f-d7283b625da0-ovsdbserver-nb\") pod \"dnsmasq-dns-55b86d75d9-f64vw\" (UID: \"9f5334e8-ad35-41c0-b74f-d7283b625da0\") " pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.104230 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-combined-ca-bundle\") pod \"keystone-bootstrap-8d7xp\" (UID: \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\") " pod="openstack/keystone-bootstrap-8d7xp" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.104276 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks6jz\" (UniqueName: \"kubernetes.io/projected/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-kube-api-access-ks6jz\") pod \"keystone-bootstrap-8d7xp\" (UID: \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\") " pod="openstack/keystone-bootstrap-8d7xp" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.104306 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-scripts\") pod \"keystone-bootstrap-8d7xp\" (UID: \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\") " pod="openstack/keystone-bootstrap-8d7xp" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.104333 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f5334e8-ad35-41c0-b74f-d7283b625da0-ovsdbserver-sb\") pod \"dnsmasq-dns-55b86d75d9-f64vw\" (UID: \"9f5334e8-ad35-41c0-b74f-d7283b625da0\") " pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.104498 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f5334e8-ad35-41c0-b74f-d7283b625da0-dns-svc\") pod \"dnsmasq-dns-55b86d75d9-f64vw\" (UID: \"9f5334e8-ad35-41c0-b74f-d7283b625da0\") " pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.104594 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f5334e8-ad35-41c0-b74f-d7283b625da0-config\") pod \"dnsmasq-dns-55b86d75d9-f64vw\" (UID: \"9f5334e8-ad35-41c0-b74f-d7283b625da0\") " pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.104674 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-config-data\") pod \"keystone-bootstrap-8d7xp\" (UID: \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\") " pod="openstack/keystone-bootstrap-8d7xp" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.104841 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fqwn\" (UniqueName: \"kubernetes.io/projected/9f5334e8-ad35-41c0-b74f-d7283b625da0-kube-api-access-2fqwn\") pod \"dnsmasq-dns-55b86d75d9-f64vw\" (UID: \"9f5334e8-ad35-41c0-b74f-d7283b625da0\") " pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.206273 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fqwn\" (UniqueName: \"kubernetes.io/projected/9f5334e8-ad35-41c0-b74f-d7283b625da0-kube-api-access-2fqwn\") pod \"dnsmasq-dns-55b86d75d9-f64vw\" (UID: \"9f5334e8-ad35-41c0-b74f-d7283b625da0\") " pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.206371 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-fernet-keys\") pod \"keystone-bootstrap-8d7xp\" (UID: \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\") " pod="openstack/keystone-bootstrap-8d7xp" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.206400 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-credential-keys\") pod \"keystone-bootstrap-8d7xp\" (UID: \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\") " pod="openstack/keystone-bootstrap-8d7xp" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.206453 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f5334e8-ad35-41c0-b74f-d7283b625da0-ovsdbserver-nb\") pod \"dnsmasq-dns-55b86d75d9-f64vw\" (UID: \"9f5334e8-ad35-41c0-b74f-d7283b625da0\") " pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.206479 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-combined-ca-bundle\") pod \"keystone-bootstrap-8d7xp\" (UID: \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\") " pod="openstack/keystone-bootstrap-8d7xp" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.206501 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks6jz\" (UniqueName: \"kubernetes.io/projected/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-kube-api-access-ks6jz\") pod \"keystone-bootstrap-8d7xp\" (UID: \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\") " pod="openstack/keystone-bootstrap-8d7xp" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.206527 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-scripts\") pod \"keystone-bootstrap-8d7xp\" (UID: \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\") " pod="openstack/keystone-bootstrap-8d7xp" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.206549 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f5334e8-ad35-41c0-b74f-d7283b625da0-ovsdbserver-sb\") pod \"dnsmasq-dns-55b86d75d9-f64vw\" (UID: \"9f5334e8-ad35-41c0-b74f-d7283b625da0\") " pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.206591 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f5334e8-ad35-41c0-b74f-d7283b625da0-dns-svc\") pod \"dnsmasq-dns-55b86d75d9-f64vw\" (UID: \"9f5334e8-ad35-41c0-b74f-d7283b625da0\") " pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.206628 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f5334e8-ad35-41c0-b74f-d7283b625da0-config\") pod \"dnsmasq-dns-55b86d75d9-f64vw\" (UID: \"9f5334e8-ad35-41c0-b74f-d7283b625da0\") " pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.206658 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-config-data\") pod \"keystone-bootstrap-8d7xp\" (UID: \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\") " pod="openstack/keystone-bootstrap-8d7xp" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.209305 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f5334e8-ad35-41c0-b74f-d7283b625da0-ovsdbserver-sb\") pod \"dnsmasq-dns-55b86d75d9-f64vw\" (UID: \"9f5334e8-ad35-41c0-b74f-d7283b625da0\") " pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.209330 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f5334e8-ad35-41c0-b74f-d7283b625da0-dns-svc\") pod \"dnsmasq-dns-55b86d75d9-f64vw\" (UID: \"9f5334e8-ad35-41c0-b74f-d7283b625da0\") " pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.209507 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f5334e8-ad35-41c0-b74f-d7283b625da0-config\") pod \"dnsmasq-dns-55b86d75d9-f64vw\" (UID: \"9f5334e8-ad35-41c0-b74f-d7283b625da0\") " pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.209901 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f5334e8-ad35-41c0-b74f-d7283b625da0-ovsdbserver-nb\") pod \"dnsmasq-dns-55b86d75d9-f64vw\" (UID: \"9f5334e8-ad35-41c0-b74f-d7283b625da0\") " pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.212351 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-combined-ca-bundle\") pod \"keystone-bootstrap-8d7xp\" (UID: \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\") " pod="openstack/keystone-bootstrap-8d7xp" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.212472 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-fernet-keys\") pod \"keystone-bootstrap-8d7xp\" (UID: \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\") " pod="openstack/keystone-bootstrap-8d7xp" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.212647 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-config-data\") pod \"keystone-bootstrap-8d7xp\" (UID: \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\") " pod="openstack/keystone-bootstrap-8d7xp" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.212777 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-scripts\") pod \"keystone-bootstrap-8d7xp\" (UID: \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\") " pod="openstack/keystone-bootstrap-8d7xp" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.213610 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-credential-keys\") pod \"keystone-bootstrap-8d7xp\" (UID: \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\") " pod="openstack/keystone-bootstrap-8d7xp" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.225057 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks6jz\" (UniqueName: \"kubernetes.io/projected/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-kube-api-access-ks6jz\") pod \"keystone-bootstrap-8d7xp\" (UID: \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\") " pod="openstack/keystone-bootstrap-8d7xp" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.225239 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fqwn\" (UniqueName: \"kubernetes.io/projected/9f5334e8-ad35-41c0-b74f-d7283b625da0-kube-api-access-2fqwn\") pod \"dnsmasq-dns-55b86d75d9-f64vw\" (UID: \"9f5334e8-ad35-41c0-b74f-d7283b625da0\") " pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.311931 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.318226 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ls87v" event={"ID":"9e179bb8-f9e4-434d-9636-84cc97d632fb","Type":"ContainerDied","Data":"6d131a6d7fdbe158c526c8d07c4849e19f1f5996c01dc8bb375bca25b73effc4"} Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.318273 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d131a6d7fdbe158c526c8d07c4849e19f1f5996c01dc8bb375bca25b73effc4" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.318335 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ls87v" Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.333905 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8d7xp" Jan 27 16:33:38 crc kubenswrapper[4772]: W0127 16:33:38.745211 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podabdf73e5_d2c6_47e5_8408_3ec5cbf1d22a.slice/crio-8defd0de00852d313c592668c916eb4da5093d21c31e3dc3769c1148abb0d104 WatchSource:0}: Error finding container 8defd0de00852d313c592668c916eb4da5093d21c31e3dc3769c1148abb0d104: Status 404 returned error can't find the container with id 8defd0de00852d313c592668c916eb4da5093d21c31e3dc3769c1148abb0d104 Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.746233 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-8d7xp"] Jan 27 16:33:38 crc kubenswrapper[4772]: I0127 16:33:38.817229 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55b86d75d9-f64vw"] Jan 27 16:33:39 crc kubenswrapper[4772]: I0127 16:33:39.328895 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" event={"ID":"9f5334e8-ad35-41c0-b74f-d7283b625da0","Type":"ContainerStarted","Data":"7afe0140c22bc4a9daf1f5cd3e32f3b90e22ae38f75fd3783a129a88072a7fc4"} Jan 27 16:33:39 crc kubenswrapper[4772]: I0127 16:33:39.329263 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" event={"ID":"9f5334e8-ad35-41c0-b74f-d7283b625da0","Type":"ContainerStarted","Data":"f24d4586c71be1b7413cfd65197ec5183ca5b0b700f7ea419468e58b997588e5"} Jan 27 16:33:39 crc kubenswrapper[4772]: I0127 16:33:39.330802 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8d7xp" event={"ID":"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a","Type":"ContainerStarted","Data":"1bbb692ab013dbd6aab71415abc05175b7c980ab72c48771e22a05f7a67573bd"} Jan 27 16:33:39 crc kubenswrapper[4772]: I0127 16:33:39.330845 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8d7xp" event={"ID":"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a","Type":"ContainerStarted","Data":"8defd0de00852d313c592668c916eb4da5093d21c31e3dc3769c1148abb0d104"} Jan 27 16:33:40 crc kubenswrapper[4772]: I0127 16:33:40.341852 4772 generic.go:334] "Generic (PLEG): container finished" podID="9f5334e8-ad35-41c0-b74f-d7283b625da0" containerID="7afe0140c22bc4a9daf1f5cd3e32f3b90e22ae38f75fd3783a129a88072a7fc4" exitCode=0 Jan 27 16:33:40 crc kubenswrapper[4772]: I0127 16:33:40.343591 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" event={"ID":"9f5334e8-ad35-41c0-b74f-d7283b625da0","Type":"ContainerDied","Data":"7afe0140c22bc4a9daf1f5cd3e32f3b90e22ae38f75fd3783a129a88072a7fc4"} Jan 27 16:33:40 crc kubenswrapper[4772]: I0127 16:33:40.390929 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-8d7xp" podStartSLOduration=3.39090932 podStartE2EDuration="3.39090932s" podCreationTimestamp="2026-01-27 16:33:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:33:40.389465479 +0000 UTC m=+5206.370074577" watchObservedRunningTime="2026-01-27 16:33:40.39090932 +0000 UTC m=+5206.371518418" Jan 27 16:33:41 crc kubenswrapper[4772]: I0127 16:33:41.360002 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" event={"ID":"9f5334e8-ad35-41c0-b74f-d7283b625da0","Type":"ContainerStarted","Data":"e2902da210e724572224515decc30f5a1e9fa786701c1868e86464783fea99a9"} Jan 27 16:33:41 crc kubenswrapper[4772]: I0127 16:33:41.360501 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" Jan 27 16:33:42 crc kubenswrapper[4772]: I0127 16:33:42.662566 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:33:42 crc kubenswrapper[4772]: E0127 16:33:42.663223 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:33:44 crc kubenswrapper[4772]: I0127 16:33:44.384666 4772 generic.go:334] "Generic (PLEG): container finished" podID="abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a" containerID="1bbb692ab013dbd6aab71415abc05175b7c980ab72c48771e22a05f7a67573bd" exitCode=0 Jan 27 16:33:44 crc kubenswrapper[4772]: I0127 16:33:44.384717 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8d7xp" event={"ID":"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a","Type":"ContainerDied","Data":"1bbb692ab013dbd6aab71415abc05175b7c980ab72c48771e22a05f7a67573bd"} Jan 27 16:33:44 crc kubenswrapper[4772]: I0127 16:33:44.407246 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" podStartSLOduration=7.407226391 podStartE2EDuration="7.407226391s" podCreationTimestamp="2026-01-27 16:33:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:33:41.386312343 +0000 UTC m=+5207.366921441" watchObservedRunningTime="2026-01-27 16:33:44.407226391 +0000 UTC m=+5210.387835509" Jan 27 16:33:45 crc kubenswrapper[4772]: I0127 16:33:45.792056 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8d7xp" Jan 27 16:33:45 crc kubenswrapper[4772]: I0127 16:33:45.837712 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-scripts\") pod \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\" (UID: \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\") " Jan 27 16:33:45 crc kubenswrapper[4772]: I0127 16:33:45.837827 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-config-data\") pod \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\" (UID: \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\") " Jan 27 16:33:45 crc kubenswrapper[4772]: I0127 16:33:45.837897 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ks6jz\" (UniqueName: \"kubernetes.io/projected/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-kube-api-access-ks6jz\") pod \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\" (UID: \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\") " Jan 27 16:33:45 crc kubenswrapper[4772]: I0127 16:33:45.837962 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-credential-keys\") pod \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\" (UID: \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\") " Jan 27 16:33:45 crc kubenswrapper[4772]: I0127 16:33:45.837989 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-fernet-keys\") pod \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\" (UID: \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\") " Jan 27 16:33:45 crc kubenswrapper[4772]: I0127 16:33:45.838013 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-combined-ca-bundle\") pod \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\" (UID: \"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a\") " Jan 27 16:33:45 crc kubenswrapper[4772]: I0127 16:33:45.843326 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-kube-api-access-ks6jz" (OuterVolumeSpecName: "kube-api-access-ks6jz") pod "abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a" (UID: "abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a"). InnerVolumeSpecName "kube-api-access-ks6jz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:33:45 crc kubenswrapper[4772]: I0127 16:33:45.844790 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a" (UID: "abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:33:45 crc kubenswrapper[4772]: I0127 16:33:45.847382 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-scripts" (OuterVolumeSpecName: "scripts") pod "abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a" (UID: "abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:33:45 crc kubenswrapper[4772]: I0127 16:33:45.857689 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a" (UID: "abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:33:45 crc kubenswrapper[4772]: I0127 16:33:45.877685 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a" (UID: "abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:33:45 crc kubenswrapper[4772]: I0127 16:33:45.880636 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-config-data" (OuterVolumeSpecName: "config-data") pod "abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a" (UID: "abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:33:45 crc kubenswrapper[4772]: I0127 16:33:45.948418 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:45 crc kubenswrapper[4772]: I0127 16:33:45.948462 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:45 crc kubenswrapper[4772]: I0127 16:33:45.948475 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ks6jz\" (UniqueName: \"kubernetes.io/projected/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-kube-api-access-ks6jz\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:45 crc kubenswrapper[4772]: I0127 16:33:45.948485 4772 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:45 crc kubenswrapper[4772]: I0127 16:33:45.948494 4772 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:45 crc kubenswrapper[4772]: I0127 16:33:45.948502 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.414014 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8d7xp" event={"ID":"abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a","Type":"ContainerDied","Data":"8defd0de00852d313c592668c916eb4da5093d21c31e3dc3769c1148abb0d104"} Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.414454 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8defd0de00852d313c592668c916eb4da5093d21c31e3dc3769c1148abb0d104" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.414258 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8d7xp" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.505705 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-8d7xp"] Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.510673 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-8d7xp"] Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.568337 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-8lxch"] Jan 27 16:33:46 crc kubenswrapper[4772]: E0127 16:33:46.568671 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a" containerName="keystone-bootstrap" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.568693 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a" containerName="keystone-bootstrap" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.568850 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a" containerName="keystone-bootstrap" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.569394 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8lxch" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.573710 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.574655 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-p8kvv" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.574850 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.575250 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.576583 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.580886 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-8lxch"] Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.661815 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-fernet-keys\") pod \"keystone-bootstrap-8lxch\" (UID: \"e4b64159-de94-4b31-85ce-b845fdb391b3\") " pod="openstack/keystone-bootstrap-8lxch" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.661883 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-config-data\") pod \"keystone-bootstrap-8lxch\" (UID: \"e4b64159-de94-4b31-85ce-b845fdb391b3\") " pod="openstack/keystone-bootstrap-8lxch" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.661933 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-575p4\" (UniqueName: \"kubernetes.io/projected/e4b64159-de94-4b31-85ce-b845fdb391b3-kube-api-access-575p4\") pod \"keystone-bootstrap-8lxch\" (UID: \"e4b64159-de94-4b31-85ce-b845fdb391b3\") " pod="openstack/keystone-bootstrap-8lxch" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.661987 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-combined-ca-bundle\") pod \"keystone-bootstrap-8lxch\" (UID: \"e4b64159-de94-4b31-85ce-b845fdb391b3\") " pod="openstack/keystone-bootstrap-8lxch" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.662011 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-scripts\") pod \"keystone-bootstrap-8lxch\" (UID: \"e4b64159-de94-4b31-85ce-b845fdb391b3\") " pod="openstack/keystone-bootstrap-8lxch" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.662060 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-credential-keys\") pod \"keystone-bootstrap-8lxch\" (UID: \"e4b64159-de94-4b31-85ce-b845fdb391b3\") " pod="openstack/keystone-bootstrap-8lxch" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.672893 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a" path="/var/lib/kubelet/pods/abdf73e5-d2c6-47e5-8408-3ec5cbf1d22a/volumes" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.764122 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-fernet-keys\") pod \"keystone-bootstrap-8lxch\" (UID: \"e4b64159-de94-4b31-85ce-b845fdb391b3\") " pod="openstack/keystone-bootstrap-8lxch" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.764257 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-config-data\") pod \"keystone-bootstrap-8lxch\" (UID: \"e4b64159-de94-4b31-85ce-b845fdb391b3\") " pod="openstack/keystone-bootstrap-8lxch" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.764342 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-575p4\" (UniqueName: \"kubernetes.io/projected/e4b64159-de94-4b31-85ce-b845fdb391b3-kube-api-access-575p4\") pod \"keystone-bootstrap-8lxch\" (UID: \"e4b64159-de94-4b31-85ce-b845fdb391b3\") " pod="openstack/keystone-bootstrap-8lxch" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.764384 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-combined-ca-bundle\") pod \"keystone-bootstrap-8lxch\" (UID: \"e4b64159-de94-4b31-85ce-b845fdb391b3\") " pod="openstack/keystone-bootstrap-8lxch" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.764413 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-scripts\") pod \"keystone-bootstrap-8lxch\" (UID: \"e4b64159-de94-4b31-85ce-b845fdb391b3\") " pod="openstack/keystone-bootstrap-8lxch" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.764496 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-credential-keys\") pod \"keystone-bootstrap-8lxch\" (UID: \"e4b64159-de94-4b31-85ce-b845fdb391b3\") " pod="openstack/keystone-bootstrap-8lxch" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.768795 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-scripts\") pod \"keystone-bootstrap-8lxch\" (UID: \"e4b64159-de94-4b31-85ce-b845fdb391b3\") " pod="openstack/keystone-bootstrap-8lxch" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.769733 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-credential-keys\") pod \"keystone-bootstrap-8lxch\" (UID: \"e4b64159-de94-4b31-85ce-b845fdb391b3\") " pod="openstack/keystone-bootstrap-8lxch" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.769813 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-combined-ca-bundle\") pod \"keystone-bootstrap-8lxch\" (UID: \"e4b64159-de94-4b31-85ce-b845fdb391b3\") " pod="openstack/keystone-bootstrap-8lxch" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.776732 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-fernet-keys\") pod \"keystone-bootstrap-8lxch\" (UID: \"e4b64159-de94-4b31-85ce-b845fdb391b3\") " pod="openstack/keystone-bootstrap-8lxch" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.777318 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-config-data\") pod \"keystone-bootstrap-8lxch\" (UID: \"e4b64159-de94-4b31-85ce-b845fdb391b3\") " pod="openstack/keystone-bootstrap-8lxch" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.784383 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-575p4\" (UniqueName: \"kubernetes.io/projected/e4b64159-de94-4b31-85ce-b845fdb391b3-kube-api-access-575p4\") pod \"keystone-bootstrap-8lxch\" (UID: \"e4b64159-de94-4b31-85ce-b845fdb391b3\") " pod="openstack/keystone-bootstrap-8lxch" Jan 27 16:33:46 crc kubenswrapper[4772]: I0127 16:33:46.899410 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8lxch" Jan 27 16:33:47 crc kubenswrapper[4772]: I0127 16:33:47.357118 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-8lxch"] Jan 27 16:33:47 crc kubenswrapper[4772]: I0127 16:33:47.427446 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8lxch" event={"ID":"e4b64159-de94-4b31-85ce-b845fdb391b3","Type":"ContainerStarted","Data":"ccf19bff553c261500929d96e5891dcc653cd8e08ae4a1b0c3230f64fe99f92f"} Jan 27 16:33:48 crc kubenswrapper[4772]: I0127 16:33:48.313422 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" Jan 27 16:33:48 crc kubenswrapper[4772]: I0127 16:33:48.388946 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-746d786469-qv9pq"] Jan 27 16:33:48 crc kubenswrapper[4772]: I0127 16:33:48.389220 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-746d786469-qv9pq" podUID="320b0d26-aa22-4854-ac28-4699f95fb37b" containerName="dnsmasq-dns" containerID="cri-o://07f4aa1fb5d6e810824b11ccdad9666a53ee8ab3a0bbb958fb4483c8bf5dd691" gracePeriod=10 Jan 27 16:33:48 crc kubenswrapper[4772]: I0127 16:33:48.460693 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8lxch" event={"ID":"e4b64159-de94-4b31-85ce-b845fdb391b3","Type":"ContainerStarted","Data":"761ce723d19214c2204c8e256e0ee21bf82feb31ce0c9f783c4a5819a97623f7"} Jan 27 16:33:48 crc kubenswrapper[4772]: I0127 16:33:48.520764 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-8lxch" podStartSLOduration=2.520742832 podStartE2EDuration="2.520742832s" podCreationTimestamp="2026-01-27 16:33:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:33:48.501982468 +0000 UTC m=+5214.482591576" watchObservedRunningTime="2026-01-27 16:33:48.520742832 +0000 UTC m=+5214.501351940" Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.348969 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-746d786469-qv9pq" Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.422524 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/320b0d26-aa22-4854-ac28-4699f95fb37b-dns-svc\") pod \"320b0d26-aa22-4854-ac28-4699f95fb37b\" (UID: \"320b0d26-aa22-4854-ac28-4699f95fb37b\") " Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.422783 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/320b0d26-aa22-4854-ac28-4699f95fb37b-ovsdbserver-nb\") pod \"320b0d26-aa22-4854-ac28-4699f95fb37b\" (UID: \"320b0d26-aa22-4854-ac28-4699f95fb37b\") " Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.422947 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/320b0d26-aa22-4854-ac28-4699f95fb37b-config\") pod \"320b0d26-aa22-4854-ac28-4699f95fb37b\" (UID: \"320b0d26-aa22-4854-ac28-4699f95fb37b\") " Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.423086 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjfw7\" (UniqueName: \"kubernetes.io/projected/320b0d26-aa22-4854-ac28-4699f95fb37b-kube-api-access-zjfw7\") pod \"320b0d26-aa22-4854-ac28-4699f95fb37b\" (UID: \"320b0d26-aa22-4854-ac28-4699f95fb37b\") " Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.423179 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/320b0d26-aa22-4854-ac28-4699f95fb37b-ovsdbserver-sb\") pod \"320b0d26-aa22-4854-ac28-4699f95fb37b\" (UID: \"320b0d26-aa22-4854-ac28-4699f95fb37b\") " Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.431478 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/320b0d26-aa22-4854-ac28-4699f95fb37b-kube-api-access-zjfw7" (OuterVolumeSpecName: "kube-api-access-zjfw7") pod "320b0d26-aa22-4854-ac28-4699f95fb37b" (UID: "320b0d26-aa22-4854-ac28-4699f95fb37b"). InnerVolumeSpecName "kube-api-access-zjfw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.460532 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/320b0d26-aa22-4854-ac28-4699f95fb37b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "320b0d26-aa22-4854-ac28-4699f95fb37b" (UID: "320b0d26-aa22-4854-ac28-4699f95fb37b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.461071 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/320b0d26-aa22-4854-ac28-4699f95fb37b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "320b0d26-aa22-4854-ac28-4699f95fb37b" (UID: "320b0d26-aa22-4854-ac28-4699f95fb37b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.473283 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/320b0d26-aa22-4854-ac28-4699f95fb37b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "320b0d26-aa22-4854-ac28-4699f95fb37b" (UID: "320b0d26-aa22-4854-ac28-4699f95fb37b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.473779 4772 generic.go:334] "Generic (PLEG): container finished" podID="320b0d26-aa22-4854-ac28-4699f95fb37b" containerID="07f4aa1fb5d6e810824b11ccdad9666a53ee8ab3a0bbb958fb4483c8bf5dd691" exitCode=0 Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.473836 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-746d786469-qv9pq" Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.473927 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746d786469-qv9pq" event={"ID":"320b0d26-aa22-4854-ac28-4699f95fb37b","Type":"ContainerDied","Data":"07f4aa1fb5d6e810824b11ccdad9666a53ee8ab3a0bbb958fb4483c8bf5dd691"} Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.474142 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746d786469-qv9pq" event={"ID":"320b0d26-aa22-4854-ac28-4699f95fb37b","Type":"ContainerDied","Data":"551776dfcff4ae5d83241e31aab7653a20a9814c77297414bc41ae9b8bf9cecd"} Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.474231 4772 scope.go:117] "RemoveContainer" containerID="07f4aa1fb5d6e810824b11ccdad9666a53ee8ab3a0bbb958fb4483c8bf5dd691" Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.483659 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/320b0d26-aa22-4854-ac28-4699f95fb37b-config" (OuterVolumeSpecName: "config") pod "320b0d26-aa22-4854-ac28-4699f95fb37b" (UID: "320b0d26-aa22-4854-ac28-4699f95fb37b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.526010 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/320b0d26-aa22-4854-ac28-4699f95fb37b-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.526050 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/320b0d26-aa22-4854-ac28-4699f95fb37b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.526063 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/320b0d26-aa22-4854-ac28-4699f95fb37b-config\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.526072 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjfw7\" (UniqueName: \"kubernetes.io/projected/320b0d26-aa22-4854-ac28-4699f95fb37b-kube-api-access-zjfw7\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.526080 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/320b0d26-aa22-4854-ac28-4699f95fb37b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.543730 4772 scope.go:117] "RemoveContainer" containerID="02bb8fc05201bb2286e71696309cca885abf131326f1399ef983ae5f44f953bf" Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.561002 4772 scope.go:117] "RemoveContainer" containerID="07f4aa1fb5d6e810824b11ccdad9666a53ee8ab3a0bbb958fb4483c8bf5dd691" Jan 27 16:33:49 crc kubenswrapper[4772]: E0127 16:33:49.561476 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07f4aa1fb5d6e810824b11ccdad9666a53ee8ab3a0bbb958fb4483c8bf5dd691\": container with ID starting with 07f4aa1fb5d6e810824b11ccdad9666a53ee8ab3a0bbb958fb4483c8bf5dd691 not found: ID does not exist" containerID="07f4aa1fb5d6e810824b11ccdad9666a53ee8ab3a0bbb958fb4483c8bf5dd691" Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.561508 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07f4aa1fb5d6e810824b11ccdad9666a53ee8ab3a0bbb958fb4483c8bf5dd691"} err="failed to get container status \"07f4aa1fb5d6e810824b11ccdad9666a53ee8ab3a0bbb958fb4483c8bf5dd691\": rpc error: code = NotFound desc = could not find container \"07f4aa1fb5d6e810824b11ccdad9666a53ee8ab3a0bbb958fb4483c8bf5dd691\": container with ID starting with 07f4aa1fb5d6e810824b11ccdad9666a53ee8ab3a0bbb958fb4483c8bf5dd691 not found: ID does not exist" Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.561536 4772 scope.go:117] "RemoveContainer" containerID="02bb8fc05201bb2286e71696309cca885abf131326f1399ef983ae5f44f953bf" Jan 27 16:33:49 crc kubenswrapper[4772]: E0127 16:33:49.561932 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02bb8fc05201bb2286e71696309cca885abf131326f1399ef983ae5f44f953bf\": container with ID starting with 02bb8fc05201bb2286e71696309cca885abf131326f1399ef983ae5f44f953bf not found: ID does not exist" containerID="02bb8fc05201bb2286e71696309cca885abf131326f1399ef983ae5f44f953bf" Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.561974 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02bb8fc05201bb2286e71696309cca885abf131326f1399ef983ae5f44f953bf"} err="failed to get container status \"02bb8fc05201bb2286e71696309cca885abf131326f1399ef983ae5f44f953bf\": rpc error: code = NotFound desc = could not find container \"02bb8fc05201bb2286e71696309cca885abf131326f1399ef983ae5f44f953bf\": container with ID starting with 02bb8fc05201bb2286e71696309cca885abf131326f1399ef983ae5f44f953bf not found: ID does not exist" Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.816768 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-746d786469-qv9pq"] Jan 27 16:33:49 crc kubenswrapper[4772]: I0127 16:33:49.824456 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-746d786469-qv9pq"] Jan 27 16:33:50 crc kubenswrapper[4772]: I0127 16:33:50.485212 4772 generic.go:334] "Generic (PLEG): container finished" podID="e4b64159-de94-4b31-85ce-b845fdb391b3" containerID="761ce723d19214c2204c8e256e0ee21bf82feb31ce0c9f783c4a5819a97623f7" exitCode=0 Jan 27 16:33:50 crc kubenswrapper[4772]: I0127 16:33:50.485276 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8lxch" event={"ID":"e4b64159-de94-4b31-85ce-b845fdb391b3","Type":"ContainerDied","Data":"761ce723d19214c2204c8e256e0ee21bf82feb31ce0c9f783c4a5819a97623f7"} Jan 27 16:33:50 crc kubenswrapper[4772]: I0127 16:33:50.672007 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="320b0d26-aa22-4854-ac28-4699f95fb37b" path="/var/lib/kubelet/pods/320b0d26-aa22-4854-ac28-4699f95fb37b/volumes" Jan 27 16:33:51 crc kubenswrapper[4772]: I0127 16:33:51.850907 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8lxch" Jan 27 16:33:51 crc kubenswrapper[4772]: I0127 16:33:51.865569 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-config-data\") pod \"e4b64159-de94-4b31-85ce-b845fdb391b3\" (UID: \"e4b64159-de94-4b31-85ce-b845fdb391b3\") " Jan 27 16:33:51 crc kubenswrapper[4772]: I0127 16:33:51.865661 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-575p4\" (UniqueName: \"kubernetes.io/projected/e4b64159-de94-4b31-85ce-b845fdb391b3-kube-api-access-575p4\") pod \"e4b64159-de94-4b31-85ce-b845fdb391b3\" (UID: \"e4b64159-de94-4b31-85ce-b845fdb391b3\") " Jan 27 16:33:51 crc kubenswrapper[4772]: I0127 16:33:51.865836 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-credential-keys\") pod \"e4b64159-de94-4b31-85ce-b845fdb391b3\" (UID: \"e4b64159-de94-4b31-85ce-b845fdb391b3\") " Jan 27 16:33:51 crc kubenswrapper[4772]: I0127 16:33:51.865904 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-combined-ca-bundle\") pod \"e4b64159-de94-4b31-85ce-b845fdb391b3\" (UID: \"e4b64159-de94-4b31-85ce-b845fdb391b3\") " Jan 27 16:33:51 crc kubenswrapper[4772]: I0127 16:33:51.865991 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-fernet-keys\") pod \"e4b64159-de94-4b31-85ce-b845fdb391b3\" (UID: \"e4b64159-de94-4b31-85ce-b845fdb391b3\") " Jan 27 16:33:51 crc kubenswrapper[4772]: I0127 16:33:51.866051 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-scripts\") pod \"e4b64159-de94-4b31-85ce-b845fdb391b3\" (UID: \"e4b64159-de94-4b31-85ce-b845fdb391b3\") " Jan 27 16:33:51 crc kubenswrapper[4772]: I0127 16:33:51.874632 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4b64159-de94-4b31-85ce-b845fdb391b3-kube-api-access-575p4" (OuterVolumeSpecName: "kube-api-access-575p4") pod "e4b64159-de94-4b31-85ce-b845fdb391b3" (UID: "e4b64159-de94-4b31-85ce-b845fdb391b3"). InnerVolumeSpecName "kube-api-access-575p4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:33:51 crc kubenswrapper[4772]: I0127 16:33:51.875527 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-scripts" (OuterVolumeSpecName: "scripts") pod "e4b64159-de94-4b31-85ce-b845fdb391b3" (UID: "e4b64159-de94-4b31-85ce-b845fdb391b3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:33:51 crc kubenswrapper[4772]: I0127 16:33:51.878305 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "e4b64159-de94-4b31-85ce-b845fdb391b3" (UID: "e4b64159-de94-4b31-85ce-b845fdb391b3"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:33:51 crc kubenswrapper[4772]: I0127 16:33:51.881028 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e4b64159-de94-4b31-85ce-b845fdb391b3" (UID: "e4b64159-de94-4b31-85ce-b845fdb391b3"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:33:51 crc kubenswrapper[4772]: I0127 16:33:51.901458 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e4b64159-de94-4b31-85ce-b845fdb391b3" (UID: "e4b64159-de94-4b31-85ce-b845fdb391b3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:33:51 crc kubenswrapper[4772]: I0127 16:33:51.907583 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-config-data" (OuterVolumeSpecName: "config-data") pod "e4b64159-de94-4b31-85ce-b845fdb391b3" (UID: "e4b64159-de94-4b31-85ce-b845fdb391b3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:33:51 crc kubenswrapper[4772]: I0127 16:33:51.967860 4772 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:51 crc kubenswrapper[4772]: I0127 16:33:51.968096 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:51 crc kubenswrapper[4772]: I0127 16:33:51.968224 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:51 crc kubenswrapper[4772]: I0127 16:33:51.968320 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-575p4\" (UniqueName: \"kubernetes.io/projected/e4b64159-de94-4b31-85ce-b845fdb391b3-kube-api-access-575p4\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:51 crc kubenswrapper[4772]: I0127 16:33:51.968385 4772 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:51 crc kubenswrapper[4772]: I0127 16:33:51.968438 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4b64159-de94-4b31-85ce-b845fdb391b3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:33:52 crc kubenswrapper[4772]: I0127 16:33:52.505384 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8lxch" event={"ID":"e4b64159-de94-4b31-85ce-b845fdb391b3","Type":"ContainerDied","Data":"ccf19bff553c261500929d96e5891dcc653cd8e08ae4a1b0c3230f64fe99f92f"} Jan 27 16:33:52 crc kubenswrapper[4772]: I0127 16:33:52.505428 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccf19bff553c261500929d96e5891dcc653cd8e08ae4a1b0c3230f64fe99f92f" Jan 27 16:33:52 crc kubenswrapper[4772]: I0127 16:33:52.505435 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8lxch" Jan 27 16:33:52 crc kubenswrapper[4772]: I0127 16:33:52.940921 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-56b5f9d6fc-hmz72"] Jan 27 16:33:52 crc kubenswrapper[4772]: E0127 16:33:52.941389 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="320b0d26-aa22-4854-ac28-4699f95fb37b" containerName="dnsmasq-dns" Jan 27 16:33:52 crc kubenswrapper[4772]: I0127 16:33:52.941406 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="320b0d26-aa22-4854-ac28-4699f95fb37b" containerName="dnsmasq-dns" Jan 27 16:33:52 crc kubenswrapper[4772]: E0127 16:33:52.941443 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="320b0d26-aa22-4854-ac28-4699f95fb37b" containerName="init" Jan 27 16:33:52 crc kubenswrapper[4772]: I0127 16:33:52.941450 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="320b0d26-aa22-4854-ac28-4699f95fb37b" containerName="init" Jan 27 16:33:52 crc kubenswrapper[4772]: E0127 16:33:52.941465 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4b64159-de94-4b31-85ce-b845fdb391b3" containerName="keystone-bootstrap" Jan 27 16:33:52 crc kubenswrapper[4772]: I0127 16:33:52.941474 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4b64159-de94-4b31-85ce-b845fdb391b3" containerName="keystone-bootstrap" Jan 27 16:33:52 crc kubenswrapper[4772]: I0127 16:33:52.941666 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="320b0d26-aa22-4854-ac28-4699f95fb37b" containerName="dnsmasq-dns" Jan 27 16:33:52 crc kubenswrapper[4772]: I0127 16:33:52.941679 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4b64159-de94-4b31-85ce-b845fdb391b3" containerName="keystone-bootstrap" Jan 27 16:33:52 crc kubenswrapper[4772]: I0127 16:33:52.942312 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-56b5f9d6fc-hmz72" Jan 27 16:33:52 crc kubenswrapper[4772]: I0127 16:33:52.945307 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 27 16:33:52 crc kubenswrapper[4772]: I0127 16:33:52.945742 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 27 16:33:52 crc kubenswrapper[4772]: I0127 16:33:52.945910 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 27 16:33:52 crc kubenswrapper[4772]: I0127 16:33:52.947812 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-p8kvv" Jan 27 16:33:52 crc kubenswrapper[4772]: I0127 16:33:52.952451 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-56b5f9d6fc-hmz72"] Jan 27 16:33:52 crc kubenswrapper[4772]: I0127 16:33:52.983488 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6255916-357e-42c6-b936-27151f6b2260-config-data\") pod \"keystone-56b5f9d6fc-hmz72\" (UID: \"f6255916-357e-42c6-b936-27151f6b2260\") " pod="openstack/keystone-56b5f9d6fc-hmz72" Jan 27 16:33:52 crc kubenswrapper[4772]: I0127 16:33:52.983748 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f6255916-357e-42c6-b936-27151f6b2260-fernet-keys\") pod \"keystone-56b5f9d6fc-hmz72\" (UID: \"f6255916-357e-42c6-b936-27151f6b2260\") " pod="openstack/keystone-56b5f9d6fc-hmz72" Jan 27 16:33:52 crc kubenswrapper[4772]: I0127 16:33:52.983892 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdxmj\" (UniqueName: \"kubernetes.io/projected/f6255916-357e-42c6-b936-27151f6b2260-kube-api-access-gdxmj\") pod \"keystone-56b5f9d6fc-hmz72\" (UID: \"f6255916-357e-42c6-b936-27151f6b2260\") " pod="openstack/keystone-56b5f9d6fc-hmz72" Jan 27 16:33:52 crc kubenswrapper[4772]: I0127 16:33:52.984006 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f6255916-357e-42c6-b936-27151f6b2260-credential-keys\") pod \"keystone-56b5f9d6fc-hmz72\" (UID: \"f6255916-357e-42c6-b936-27151f6b2260\") " pod="openstack/keystone-56b5f9d6fc-hmz72" Jan 27 16:33:52 crc kubenswrapper[4772]: I0127 16:33:52.984101 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6255916-357e-42c6-b936-27151f6b2260-combined-ca-bundle\") pod \"keystone-56b5f9d6fc-hmz72\" (UID: \"f6255916-357e-42c6-b936-27151f6b2260\") " pod="openstack/keystone-56b5f9d6fc-hmz72" Jan 27 16:33:52 crc kubenswrapper[4772]: I0127 16:33:52.984262 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6255916-357e-42c6-b936-27151f6b2260-scripts\") pod \"keystone-56b5f9d6fc-hmz72\" (UID: \"f6255916-357e-42c6-b936-27151f6b2260\") " pod="openstack/keystone-56b5f9d6fc-hmz72" Jan 27 16:33:53 crc kubenswrapper[4772]: I0127 16:33:53.084845 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6255916-357e-42c6-b936-27151f6b2260-config-data\") pod \"keystone-56b5f9d6fc-hmz72\" (UID: \"f6255916-357e-42c6-b936-27151f6b2260\") " pod="openstack/keystone-56b5f9d6fc-hmz72" Jan 27 16:33:53 crc kubenswrapper[4772]: I0127 16:33:53.084905 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f6255916-357e-42c6-b936-27151f6b2260-fernet-keys\") pod \"keystone-56b5f9d6fc-hmz72\" (UID: \"f6255916-357e-42c6-b936-27151f6b2260\") " pod="openstack/keystone-56b5f9d6fc-hmz72" Jan 27 16:33:53 crc kubenswrapper[4772]: I0127 16:33:53.084933 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdxmj\" (UniqueName: \"kubernetes.io/projected/f6255916-357e-42c6-b936-27151f6b2260-kube-api-access-gdxmj\") pod \"keystone-56b5f9d6fc-hmz72\" (UID: \"f6255916-357e-42c6-b936-27151f6b2260\") " pod="openstack/keystone-56b5f9d6fc-hmz72" Jan 27 16:33:53 crc kubenswrapper[4772]: I0127 16:33:53.084960 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f6255916-357e-42c6-b936-27151f6b2260-credential-keys\") pod \"keystone-56b5f9d6fc-hmz72\" (UID: \"f6255916-357e-42c6-b936-27151f6b2260\") " pod="openstack/keystone-56b5f9d6fc-hmz72" Jan 27 16:33:53 crc kubenswrapper[4772]: I0127 16:33:53.084979 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6255916-357e-42c6-b936-27151f6b2260-combined-ca-bundle\") pod \"keystone-56b5f9d6fc-hmz72\" (UID: \"f6255916-357e-42c6-b936-27151f6b2260\") " pod="openstack/keystone-56b5f9d6fc-hmz72" Jan 27 16:33:53 crc kubenswrapper[4772]: I0127 16:33:53.085023 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6255916-357e-42c6-b936-27151f6b2260-scripts\") pod \"keystone-56b5f9d6fc-hmz72\" (UID: \"f6255916-357e-42c6-b936-27151f6b2260\") " pod="openstack/keystone-56b5f9d6fc-hmz72" Jan 27 16:33:53 crc kubenswrapper[4772]: I0127 16:33:53.090270 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6255916-357e-42c6-b936-27151f6b2260-scripts\") pod \"keystone-56b5f9d6fc-hmz72\" (UID: \"f6255916-357e-42c6-b936-27151f6b2260\") " pod="openstack/keystone-56b5f9d6fc-hmz72" Jan 27 16:33:53 crc kubenswrapper[4772]: I0127 16:33:53.090498 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f6255916-357e-42c6-b936-27151f6b2260-fernet-keys\") pod \"keystone-56b5f9d6fc-hmz72\" (UID: \"f6255916-357e-42c6-b936-27151f6b2260\") " pod="openstack/keystone-56b5f9d6fc-hmz72" Jan 27 16:33:53 crc kubenswrapper[4772]: I0127 16:33:53.091364 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6255916-357e-42c6-b936-27151f6b2260-config-data\") pod \"keystone-56b5f9d6fc-hmz72\" (UID: \"f6255916-357e-42c6-b936-27151f6b2260\") " pod="openstack/keystone-56b5f9d6fc-hmz72" Jan 27 16:33:53 crc kubenswrapper[4772]: I0127 16:33:53.091754 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f6255916-357e-42c6-b936-27151f6b2260-credential-keys\") pod \"keystone-56b5f9d6fc-hmz72\" (UID: \"f6255916-357e-42c6-b936-27151f6b2260\") " pod="openstack/keystone-56b5f9d6fc-hmz72" Jan 27 16:33:53 crc kubenswrapper[4772]: I0127 16:33:53.102239 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdxmj\" (UniqueName: \"kubernetes.io/projected/f6255916-357e-42c6-b936-27151f6b2260-kube-api-access-gdxmj\") pod \"keystone-56b5f9d6fc-hmz72\" (UID: \"f6255916-357e-42c6-b936-27151f6b2260\") " pod="openstack/keystone-56b5f9d6fc-hmz72" Jan 27 16:33:53 crc kubenswrapper[4772]: I0127 16:33:53.102924 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6255916-357e-42c6-b936-27151f6b2260-combined-ca-bundle\") pod \"keystone-56b5f9d6fc-hmz72\" (UID: \"f6255916-357e-42c6-b936-27151f6b2260\") " pod="openstack/keystone-56b5f9d6fc-hmz72" Jan 27 16:33:53 crc kubenswrapper[4772]: I0127 16:33:53.272004 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-56b5f9d6fc-hmz72" Jan 27 16:33:53 crc kubenswrapper[4772]: I0127 16:33:53.744032 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-56b5f9d6fc-hmz72"] Jan 27 16:33:53 crc kubenswrapper[4772]: W0127 16:33:53.752420 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6255916_357e_42c6_b936_27151f6b2260.slice/crio-be66932cfc381c827b2a9362ad58833b6a258a4e7f30abcfb528cc35363e6071 WatchSource:0}: Error finding container be66932cfc381c827b2a9362ad58833b6a258a4e7f30abcfb528cc35363e6071: Status 404 returned error can't find the container with id be66932cfc381c827b2a9362ad58833b6a258a4e7f30abcfb528cc35363e6071 Jan 27 16:33:54 crc kubenswrapper[4772]: I0127 16:33:54.521401 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-56b5f9d6fc-hmz72" event={"ID":"f6255916-357e-42c6-b936-27151f6b2260","Type":"ContainerStarted","Data":"b37f06a0442dbec54e0f8ffb189676e3082888ea33e8970849fb080c590e5548"} Jan 27 16:33:54 crc kubenswrapper[4772]: I0127 16:33:54.521447 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-56b5f9d6fc-hmz72" event={"ID":"f6255916-357e-42c6-b936-27151f6b2260","Type":"ContainerStarted","Data":"be66932cfc381c827b2a9362ad58833b6a258a4e7f30abcfb528cc35363e6071"} Jan 27 16:33:54 crc kubenswrapper[4772]: I0127 16:33:54.521477 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-56b5f9d6fc-hmz72" Jan 27 16:33:54 crc kubenswrapper[4772]: I0127 16:33:54.541351 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-56b5f9d6fc-hmz72" podStartSLOduration=2.541328002 podStartE2EDuration="2.541328002s" podCreationTimestamp="2026-01-27 16:33:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:33:54.536142704 +0000 UTC m=+5220.516751812" watchObservedRunningTime="2026-01-27 16:33:54.541328002 +0000 UTC m=+5220.521937100" Jan 27 16:33:54 crc kubenswrapper[4772]: I0127 16:33:54.672154 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:33:54 crc kubenswrapper[4772]: E0127 16:33:54.672379 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:34:06 crc kubenswrapper[4772]: I0127 16:34:06.663668 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:34:06 crc kubenswrapper[4772]: E0127 16:34:06.664403 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:34:18 crc kubenswrapper[4772]: I0127 16:34:18.663221 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:34:18 crc kubenswrapper[4772]: E0127 16:34:18.664520 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:34:24 crc kubenswrapper[4772]: I0127 16:34:24.714706 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-56b5f9d6fc-hmz72" Jan 27 16:34:28 crc kubenswrapper[4772]: I0127 16:34:28.090118 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Jan 27 16:34:28 crc kubenswrapper[4772]: I0127 16:34:28.091793 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Jan 27 16:34:28 crc kubenswrapper[4772]: I0127 16:34:28.094233 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-jj9xz" Jan 27 16:34:28 crc kubenswrapper[4772]: I0127 16:34:28.095155 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Jan 27 16:34:28 crc kubenswrapper[4772]: I0127 16:34:28.095915 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Jan 27 16:34:28 crc kubenswrapper[4772]: I0127 16:34:28.100719 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Jan 27 16:34:28 crc kubenswrapper[4772]: I0127 16:34:28.177415 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv64z\" (UniqueName: \"kubernetes.io/projected/e05a90a8-dbbb-4e24-ac89-f30360482af9-kube-api-access-sv64z\") pod \"openstackclient\" (UID: \"e05a90a8-dbbb-4e24-ac89-f30360482af9\") " pod="openstack/openstackclient" Jan 27 16:34:28 crc kubenswrapper[4772]: I0127 16:34:28.177528 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e05a90a8-dbbb-4e24-ac89-f30360482af9-openstack-config-secret\") pod \"openstackclient\" (UID: \"e05a90a8-dbbb-4e24-ac89-f30360482af9\") " pod="openstack/openstackclient" Jan 27 16:34:28 crc kubenswrapper[4772]: I0127 16:34:28.177607 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e05a90a8-dbbb-4e24-ac89-f30360482af9-openstack-config\") pod \"openstackclient\" (UID: \"e05a90a8-dbbb-4e24-ac89-f30360482af9\") " pod="openstack/openstackclient" Jan 27 16:34:28 crc kubenswrapper[4772]: I0127 16:34:28.280265 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e05a90a8-dbbb-4e24-ac89-f30360482af9-openstack-config-secret\") pod \"openstackclient\" (UID: \"e05a90a8-dbbb-4e24-ac89-f30360482af9\") " pod="openstack/openstackclient" Jan 27 16:34:28 crc kubenswrapper[4772]: I0127 16:34:28.280426 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e05a90a8-dbbb-4e24-ac89-f30360482af9-openstack-config\") pod \"openstackclient\" (UID: \"e05a90a8-dbbb-4e24-ac89-f30360482af9\") " pod="openstack/openstackclient" Jan 27 16:34:28 crc kubenswrapper[4772]: I0127 16:34:28.280550 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv64z\" (UniqueName: \"kubernetes.io/projected/e05a90a8-dbbb-4e24-ac89-f30360482af9-kube-api-access-sv64z\") pod \"openstackclient\" (UID: \"e05a90a8-dbbb-4e24-ac89-f30360482af9\") " pod="openstack/openstackclient" Jan 27 16:34:28 crc kubenswrapper[4772]: I0127 16:34:28.281460 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e05a90a8-dbbb-4e24-ac89-f30360482af9-openstack-config\") pod \"openstackclient\" (UID: \"e05a90a8-dbbb-4e24-ac89-f30360482af9\") " pod="openstack/openstackclient" Jan 27 16:34:28 crc kubenswrapper[4772]: I0127 16:34:28.286625 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e05a90a8-dbbb-4e24-ac89-f30360482af9-openstack-config-secret\") pod \"openstackclient\" (UID: \"e05a90a8-dbbb-4e24-ac89-f30360482af9\") " pod="openstack/openstackclient" Jan 27 16:34:28 crc kubenswrapper[4772]: I0127 16:34:28.315367 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv64z\" (UniqueName: \"kubernetes.io/projected/e05a90a8-dbbb-4e24-ac89-f30360482af9-kube-api-access-sv64z\") pod \"openstackclient\" (UID: \"e05a90a8-dbbb-4e24-ac89-f30360482af9\") " pod="openstack/openstackclient" Jan 27 16:34:28 crc kubenswrapper[4772]: I0127 16:34:28.428621 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Jan 27 16:34:28 crc kubenswrapper[4772]: I0127 16:34:28.881468 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Jan 27 16:34:29 crc kubenswrapper[4772]: I0127 16:34:29.802478 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"e05a90a8-dbbb-4e24-ac89-f30360482af9","Type":"ContainerStarted","Data":"9b53bc54d8825e655eba77337106d2bf01a57ffe64896ec01ce03446aa5bc757"} Jan 27 16:34:29 crc kubenswrapper[4772]: I0127 16:34:29.802809 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"e05a90a8-dbbb-4e24-ac89-f30360482af9","Type":"ContainerStarted","Data":"0caff9adca834db20dd1547787789d57fb2e7aea65070c5b0255817699f188e8"} Jan 27 16:34:29 crc kubenswrapper[4772]: I0127 16:34:29.818489 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.818468179 podStartE2EDuration="1.818468179s" podCreationTimestamp="2026-01-27 16:34:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:34:29.815914696 +0000 UTC m=+5255.796523794" watchObservedRunningTime="2026-01-27 16:34:29.818468179 +0000 UTC m=+5255.799077277" Jan 27 16:34:31 crc kubenswrapper[4772]: I0127 16:34:31.663043 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:34:31 crc kubenswrapper[4772]: E0127 16:34:31.663829 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:34:45 crc kubenswrapper[4772]: I0127 16:34:45.663444 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:34:46 crc kubenswrapper[4772]: I0127 16:34:46.948928 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"90e27c06727cf113f54cd7c0344565bfa447b15cc343fc7033a04f41dddb22f9"} Jan 27 16:36:05 crc kubenswrapper[4772]: I0127 16:36:05.145054 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-rxmk9"] Jan 27 16:36:05 crc kubenswrapper[4772]: I0127 16:36:05.146459 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rxmk9" Jan 27 16:36:05 crc kubenswrapper[4772]: I0127 16:36:05.154862 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-2bff-account-create-update-nq7qs"] Jan 27 16:36:05 crc kubenswrapper[4772]: I0127 16:36:05.156136 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2bff-account-create-update-nq7qs" Jan 27 16:36:05 crc kubenswrapper[4772]: I0127 16:36:05.158751 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Jan 27 16:36:05 crc kubenswrapper[4772]: I0127 16:36:05.166640 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-rxmk9"] Jan 27 16:36:05 crc kubenswrapper[4772]: I0127 16:36:05.175460 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-2bff-account-create-update-nq7qs"] Jan 27 16:36:05 crc kubenswrapper[4772]: I0127 16:36:05.223545 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lpdq\" (UniqueName: \"kubernetes.io/projected/1fe4269c-3ff7-49b4-82c3-1a419f676f89-kube-api-access-2lpdq\") pod \"barbican-db-create-rxmk9\" (UID: \"1fe4269c-3ff7-49b4-82c3-1a419f676f89\") " pod="openstack/barbican-db-create-rxmk9" Jan 27 16:36:05 crc kubenswrapper[4772]: I0127 16:36:05.223597 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d02ab9dd-317c-4787-aa94-0ad8dff15380-operator-scripts\") pod \"barbican-2bff-account-create-update-nq7qs\" (UID: \"d02ab9dd-317c-4787-aa94-0ad8dff15380\") " pod="openstack/barbican-2bff-account-create-update-nq7qs" Jan 27 16:36:05 crc kubenswrapper[4772]: I0127 16:36:05.223691 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfkt4\" (UniqueName: \"kubernetes.io/projected/d02ab9dd-317c-4787-aa94-0ad8dff15380-kube-api-access-hfkt4\") pod \"barbican-2bff-account-create-update-nq7qs\" (UID: \"d02ab9dd-317c-4787-aa94-0ad8dff15380\") " pod="openstack/barbican-2bff-account-create-update-nq7qs" Jan 27 16:36:05 crc kubenswrapper[4772]: I0127 16:36:05.223757 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1fe4269c-3ff7-49b4-82c3-1a419f676f89-operator-scripts\") pod \"barbican-db-create-rxmk9\" (UID: \"1fe4269c-3ff7-49b4-82c3-1a419f676f89\") " pod="openstack/barbican-db-create-rxmk9" Jan 27 16:36:05 crc kubenswrapper[4772]: I0127 16:36:05.325247 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfkt4\" (UniqueName: \"kubernetes.io/projected/d02ab9dd-317c-4787-aa94-0ad8dff15380-kube-api-access-hfkt4\") pod \"barbican-2bff-account-create-update-nq7qs\" (UID: \"d02ab9dd-317c-4787-aa94-0ad8dff15380\") " pod="openstack/barbican-2bff-account-create-update-nq7qs" Jan 27 16:36:05 crc kubenswrapper[4772]: I0127 16:36:05.325317 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1fe4269c-3ff7-49b4-82c3-1a419f676f89-operator-scripts\") pod \"barbican-db-create-rxmk9\" (UID: \"1fe4269c-3ff7-49b4-82c3-1a419f676f89\") " pod="openstack/barbican-db-create-rxmk9" Jan 27 16:36:05 crc kubenswrapper[4772]: I0127 16:36:05.325378 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lpdq\" (UniqueName: \"kubernetes.io/projected/1fe4269c-3ff7-49b4-82c3-1a419f676f89-kube-api-access-2lpdq\") pod \"barbican-db-create-rxmk9\" (UID: \"1fe4269c-3ff7-49b4-82c3-1a419f676f89\") " pod="openstack/barbican-db-create-rxmk9" Jan 27 16:36:05 crc kubenswrapper[4772]: I0127 16:36:05.325400 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d02ab9dd-317c-4787-aa94-0ad8dff15380-operator-scripts\") pod \"barbican-2bff-account-create-update-nq7qs\" (UID: \"d02ab9dd-317c-4787-aa94-0ad8dff15380\") " pod="openstack/barbican-2bff-account-create-update-nq7qs" Jan 27 16:36:05 crc kubenswrapper[4772]: I0127 16:36:05.326126 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1fe4269c-3ff7-49b4-82c3-1a419f676f89-operator-scripts\") pod \"barbican-db-create-rxmk9\" (UID: \"1fe4269c-3ff7-49b4-82c3-1a419f676f89\") " pod="openstack/barbican-db-create-rxmk9" Jan 27 16:36:05 crc kubenswrapper[4772]: I0127 16:36:05.326324 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d02ab9dd-317c-4787-aa94-0ad8dff15380-operator-scripts\") pod \"barbican-2bff-account-create-update-nq7qs\" (UID: \"d02ab9dd-317c-4787-aa94-0ad8dff15380\") " pod="openstack/barbican-2bff-account-create-update-nq7qs" Jan 27 16:36:05 crc kubenswrapper[4772]: I0127 16:36:05.347483 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lpdq\" (UniqueName: \"kubernetes.io/projected/1fe4269c-3ff7-49b4-82c3-1a419f676f89-kube-api-access-2lpdq\") pod \"barbican-db-create-rxmk9\" (UID: \"1fe4269c-3ff7-49b4-82c3-1a419f676f89\") " pod="openstack/barbican-db-create-rxmk9" Jan 27 16:36:05 crc kubenswrapper[4772]: I0127 16:36:05.348158 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfkt4\" (UniqueName: \"kubernetes.io/projected/d02ab9dd-317c-4787-aa94-0ad8dff15380-kube-api-access-hfkt4\") pod \"barbican-2bff-account-create-update-nq7qs\" (UID: \"d02ab9dd-317c-4787-aa94-0ad8dff15380\") " pod="openstack/barbican-2bff-account-create-update-nq7qs" Jan 27 16:36:05 crc kubenswrapper[4772]: I0127 16:36:05.471906 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rxmk9" Jan 27 16:36:05 crc kubenswrapper[4772]: I0127 16:36:05.486306 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2bff-account-create-update-nq7qs" Jan 27 16:36:05 crc kubenswrapper[4772]: I0127 16:36:05.999206 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-rxmk9"] Jan 27 16:36:06 crc kubenswrapper[4772]: W0127 16:36:06.002415 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1fe4269c_3ff7_49b4_82c3_1a419f676f89.slice/crio-99421eccfe268ad945cb2a25bc785ee2156c9e63300fe417ab74e3ba4e349d22 WatchSource:0}: Error finding container 99421eccfe268ad945cb2a25bc785ee2156c9e63300fe417ab74e3ba4e349d22: Status 404 returned error can't find the container with id 99421eccfe268ad945cb2a25bc785ee2156c9e63300fe417ab74e3ba4e349d22 Jan 27 16:36:06 crc kubenswrapper[4772]: I0127 16:36:06.024572 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-2bff-account-create-update-nq7qs"] Jan 27 16:36:06 crc kubenswrapper[4772]: I0127 16:36:06.782041 4772 generic.go:334] "Generic (PLEG): container finished" podID="1fe4269c-3ff7-49b4-82c3-1a419f676f89" containerID="5a8e2d8cff2b48fb4540fd3d3c996a83107f6a91a01287bf66fdb29b10e52c6b" exitCode=0 Jan 27 16:36:06 crc kubenswrapper[4772]: I0127 16:36:06.782111 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rxmk9" event={"ID":"1fe4269c-3ff7-49b4-82c3-1a419f676f89","Type":"ContainerDied","Data":"5a8e2d8cff2b48fb4540fd3d3c996a83107f6a91a01287bf66fdb29b10e52c6b"} Jan 27 16:36:06 crc kubenswrapper[4772]: I0127 16:36:06.782137 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rxmk9" event={"ID":"1fe4269c-3ff7-49b4-82c3-1a419f676f89","Type":"ContainerStarted","Data":"99421eccfe268ad945cb2a25bc785ee2156c9e63300fe417ab74e3ba4e349d22"} Jan 27 16:36:06 crc kubenswrapper[4772]: I0127 16:36:06.783658 4772 generic.go:334] "Generic (PLEG): container finished" podID="d02ab9dd-317c-4787-aa94-0ad8dff15380" containerID="39e5a537d87039cbc6538b5672d3a7484ab8d88c163808f5672ea54c55115098" exitCode=0 Jan 27 16:36:06 crc kubenswrapper[4772]: I0127 16:36:06.783700 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2bff-account-create-update-nq7qs" event={"ID":"d02ab9dd-317c-4787-aa94-0ad8dff15380","Type":"ContainerDied","Data":"39e5a537d87039cbc6538b5672d3a7484ab8d88c163808f5672ea54c55115098"} Jan 27 16:36:06 crc kubenswrapper[4772]: I0127 16:36:06.783729 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2bff-account-create-update-nq7qs" event={"ID":"d02ab9dd-317c-4787-aa94-0ad8dff15380","Type":"ContainerStarted","Data":"cde59099e96d076103d0ab96b2715ae4a771641a2db8b31a3e4cb2c31ede5713"} Jan 27 16:36:08 crc kubenswrapper[4772]: I0127 16:36:08.355879 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rxmk9" Jan 27 16:36:08 crc kubenswrapper[4772]: I0127 16:36:08.363702 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2bff-account-create-update-nq7qs" Jan 27 16:36:08 crc kubenswrapper[4772]: I0127 16:36:08.409314 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lpdq\" (UniqueName: \"kubernetes.io/projected/1fe4269c-3ff7-49b4-82c3-1a419f676f89-kube-api-access-2lpdq\") pod \"1fe4269c-3ff7-49b4-82c3-1a419f676f89\" (UID: \"1fe4269c-3ff7-49b4-82c3-1a419f676f89\") " Jan 27 16:36:08 crc kubenswrapper[4772]: I0127 16:36:08.409406 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d02ab9dd-317c-4787-aa94-0ad8dff15380-operator-scripts\") pod \"d02ab9dd-317c-4787-aa94-0ad8dff15380\" (UID: \"d02ab9dd-317c-4787-aa94-0ad8dff15380\") " Jan 27 16:36:08 crc kubenswrapper[4772]: I0127 16:36:08.409466 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfkt4\" (UniqueName: \"kubernetes.io/projected/d02ab9dd-317c-4787-aa94-0ad8dff15380-kube-api-access-hfkt4\") pod \"d02ab9dd-317c-4787-aa94-0ad8dff15380\" (UID: \"d02ab9dd-317c-4787-aa94-0ad8dff15380\") " Jan 27 16:36:08 crc kubenswrapper[4772]: I0127 16:36:08.409510 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1fe4269c-3ff7-49b4-82c3-1a419f676f89-operator-scripts\") pod \"1fe4269c-3ff7-49b4-82c3-1a419f676f89\" (UID: \"1fe4269c-3ff7-49b4-82c3-1a419f676f89\") " Jan 27 16:36:08 crc kubenswrapper[4772]: I0127 16:36:08.410281 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fe4269c-3ff7-49b4-82c3-1a419f676f89-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1fe4269c-3ff7-49b4-82c3-1a419f676f89" (UID: "1fe4269c-3ff7-49b4-82c3-1a419f676f89"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:36:08 crc kubenswrapper[4772]: I0127 16:36:08.410431 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d02ab9dd-317c-4787-aa94-0ad8dff15380-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d02ab9dd-317c-4787-aa94-0ad8dff15380" (UID: "d02ab9dd-317c-4787-aa94-0ad8dff15380"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:36:08 crc kubenswrapper[4772]: I0127 16:36:08.417802 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d02ab9dd-317c-4787-aa94-0ad8dff15380-kube-api-access-hfkt4" (OuterVolumeSpecName: "kube-api-access-hfkt4") pod "d02ab9dd-317c-4787-aa94-0ad8dff15380" (UID: "d02ab9dd-317c-4787-aa94-0ad8dff15380"). InnerVolumeSpecName "kube-api-access-hfkt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:36:08 crc kubenswrapper[4772]: I0127 16:36:08.425689 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fe4269c-3ff7-49b4-82c3-1a419f676f89-kube-api-access-2lpdq" (OuterVolumeSpecName: "kube-api-access-2lpdq") pod "1fe4269c-3ff7-49b4-82c3-1a419f676f89" (UID: "1fe4269c-3ff7-49b4-82c3-1a419f676f89"). InnerVolumeSpecName "kube-api-access-2lpdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:36:08 crc kubenswrapper[4772]: I0127 16:36:08.510999 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfkt4\" (UniqueName: \"kubernetes.io/projected/d02ab9dd-317c-4787-aa94-0ad8dff15380-kube-api-access-hfkt4\") on node \"crc\" DevicePath \"\"" Jan 27 16:36:08 crc kubenswrapper[4772]: I0127 16:36:08.511046 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1fe4269c-3ff7-49b4-82c3-1a419f676f89-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:36:08 crc kubenswrapper[4772]: I0127 16:36:08.511060 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lpdq\" (UniqueName: \"kubernetes.io/projected/1fe4269c-3ff7-49b4-82c3-1a419f676f89-kube-api-access-2lpdq\") on node \"crc\" DevicePath \"\"" Jan 27 16:36:08 crc kubenswrapper[4772]: I0127 16:36:08.511071 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d02ab9dd-317c-4787-aa94-0ad8dff15380-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:36:08 crc kubenswrapper[4772]: I0127 16:36:08.798548 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rxmk9" event={"ID":"1fe4269c-3ff7-49b4-82c3-1a419f676f89","Type":"ContainerDied","Data":"99421eccfe268ad945cb2a25bc785ee2156c9e63300fe417ab74e3ba4e349d22"} Jan 27 16:36:08 crc kubenswrapper[4772]: I0127 16:36:08.798588 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99421eccfe268ad945cb2a25bc785ee2156c9e63300fe417ab74e3ba4e349d22" Jan 27 16:36:08 crc kubenswrapper[4772]: I0127 16:36:08.798595 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rxmk9" Jan 27 16:36:08 crc kubenswrapper[4772]: I0127 16:36:08.800362 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-2bff-account-create-update-nq7qs" event={"ID":"d02ab9dd-317c-4787-aa94-0ad8dff15380","Type":"ContainerDied","Data":"cde59099e96d076103d0ab96b2715ae4a771641a2db8b31a3e4cb2c31ede5713"} Jan 27 16:36:08 crc kubenswrapper[4772]: I0127 16:36:08.800389 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cde59099e96d076103d0ab96b2715ae4a771641a2db8b31a3e4cb2c31ede5713" Jan 27 16:36:08 crc kubenswrapper[4772]: I0127 16:36:08.800418 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-2bff-account-create-update-nq7qs" Jan 27 16:36:10 crc kubenswrapper[4772]: I0127 16:36:10.554962 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-2jctn"] Jan 27 16:36:10 crc kubenswrapper[4772]: E0127 16:36:10.555564 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fe4269c-3ff7-49b4-82c3-1a419f676f89" containerName="mariadb-database-create" Jan 27 16:36:10 crc kubenswrapper[4772]: I0127 16:36:10.555577 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fe4269c-3ff7-49b4-82c3-1a419f676f89" containerName="mariadb-database-create" Jan 27 16:36:10 crc kubenswrapper[4772]: E0127 16:36:10.555587 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d02ab9dd-317c-4787-aa94-0ad8dff15380" containerName="mariadb-account-create-update" Jan 27 16:36:10 crc kubenswrapper[4772]: I0127 16:36:10.555594 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="d02ab9dd-317c-4787-aa94-0ad8dff15380" containerName="mariadb-account-create-update" Jan 27 16:36:10 crc kubenswrapper[4772]: I0127 16:36:10.555790 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="d02ab9dd-317c-4787-aa94-0ad8dff15380" containerName="mariadb-account-create-update" Jan 27 16:36:10 crc kubenswrapper[4772]: I0127 16:36:10.555813 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fe4269c-3ff7-49b4-82c3-1a419f676f89" containerName="mariadb-database-create" Jan 27 16:36:10 crc kubenswrapper[4772]: I0127 16:36:10.556364 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2jctn" Jan 27 16:36:10 crc kubenswrapper[4772]: I0127 16:36:10.558113 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-d4g2g" Jan 27 16:36:10 crc kubenswrapper[4772]: I0127 16:36:10.563000 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Jan 27 16:36:10 crc kubenswrapper[4772]: I0127 16:36:10.571458 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-2jctn"] Jan 27 16:36:10 crc kubenswrapper[4772]: I0127 16:36:10.745885 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/370ff587-7186-4f81-83a2-886a15900229-combined-ca-bundle\") pod \"barbican-db-sync-2jctn\" (UID: \"370ff587-7186-4f81-83a2-886a15900229\") " pod="openstack/barbican-db-sync-2jctn" Jan 27 16:36:10 crc kubenswrapper[4772]: I0127 16:36:10.745927 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z55sq\" (UniqueName: \"kubernetes.io/projected/370ff587-7186-4f81-83a2-886a15900229-kube-api-access-z55sq\") pod \"barbican-db-sync-2jctn\" (UID: \"370ff587-7186-4f81-83a2-886a15900229\") " pod="openstack/barbican-db-sync-2jctn" Jan 27 16:36:10 crc kubenswrapper[4772]: I0127 16:36:10.746907 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/370ff587-7186-4f81-83a2-886a15900229-db-sync-config-data\") pod \"barbican-db-sync-2jctn\" (UID: \"370ff587-7186-4f81-83a2-886a15900229\") " pod="openstack/barbican-db-sync-2jctn" Jan 27 16:36:10 crc kubenswrapper[4772]: I0127 16:36:10.847741 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/370ff587-7186-4f81-83a2-886a15900229-db-sync-config-data\") pod \"barbican-db-sync-2jctn\" (UID: \"370ff587-7186-4f81-83a2-886a15900229\") " pod="openstack/barbican-db-sync-2jctn" Jan 27 16:36:10 crc kubenswrapper[4772]: I0127 16:36:10.847832 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/370ff587-7186-4f81-83a2-886a15900229-combined-ca-bundle\") pod \"barbican-db-sync-2jctn\" (UID: \"370ff587-7186-4f81-83a2-886a15900229\") " pod="openstack/barbican-db-sync-2jctn" Jan 27 16:36:10 crc kubenswrapper[4772]: I0127 16:36:10.847848 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z55sq\" (UniqueName: \"kubernetes.io/projected/370ff587-7186-4f81-83a2-886a15900229-kube-api-access-z55sq\") pod \"barbican-db-sync-2jctn\" (UID: \"370ff587-7186-4f81-83a2-886a15900229\") " pod="openstack/barbican-db-sync-2jctn" Jan 27 16:36:10 crc kubenswrapper[4772]: I0127 16:36:10.853446 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/370ff587-7186-4f81-83a2-886a15900229-combined-ca-bundle\") pod \"barbican-db-sync-2jctn\" (UID: \"370ff587-7186-4f81-83a2-886a15900229\") " pod="openstack/barbican-db-sync-2jctn" Jan 27 16:36:10 crc kubenswrapper[4772]: I0127 16:36:10.854699 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/370ff587-7186-4f81-83a2-886a15900229-db-sync-config-data\") pod \"barbican-db-sync-2jctn\" (UID: \"370ff587-7186-4f81-83a2-886a15900229\") " pod="openstack/barbican-db-sync-2jctn" Jan 27 16:36:10 crc kubenswrapper[4772]: I0127 16:36:10.875475 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z55sq\" (UniqueName: \"kubernetes.io/projected/370ff587-7186-4f81-83a2-886a15900229-kube-api-access-z55sq\") pod \"barbican-db-sync-2jctn\" (UID: \"370ff587-7186-4f81-83a2-886a15900229\") " pod="openstack/barbican-db-sync-2jctn" Jan 27 16:36:11 crc kubenswrapper[4772]: I0127 16:36:11.171110 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2jctn" Jan 27 16:36:11 crc kubenswrapper[4772]: I0127 16:36:11.601655 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-2jctn"] Jan 27 16:36:11 crc kubenswrapper[4772]: I0127 16:36:11.824705 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2jctn" event={"ID":"370ff587-7186-4f81-83a2-886a15900229","Type":"ContainerStarted","Data":"ccfecf4ef23c50d400b8ca7553bbf1eff4d9d62e8caff28f5a3562a7630cf393"} Jan 27 16:36:11 crc kubenswrapper[4772]: I0127 16:36:11.825053 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2jctn" event={"ID":"370ff587-7186-4f81-83a2-886a15900229","Type":"ContainerStarted","Data":"66fb01f7e2b25b680204b44292b0be570804c81d694aa738565de3c8455c3ac0"} Jan 27 16:36:11 crc kubenswrapper[4772]: I0127 16:36:11.842066 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-2jctn" podStartSLOduration=1.8420488449999999 podStartE2EDuration="1.842048845s" podCreationTimestamp="2026-01-27 16:36:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:36:11.837809214 +0000 UTC m=+5357.818418322" watchObservedRunningTime="2026-01-27 16:36:11.842048845 +0000 UTC m=+5357.822657943" Jan 27 16:36:13 crc kubenswrapper[4772]: I0127 16:36:13.841113 4772 generic.go:334] "Generic (PLEG): container finished" podID="370ff587-7186-4f81-83a2-886a15900229" containerID="ccfecf4ef23c50d400b8ca7553bbf1eff4d9d62e8caff28f5a3562a7630cf393" exitCode=0 Jan 27 16:36:13 crc kubenswrapper[4772]: I0127 16:36:13.841211 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2jctn" event={"ID":"370ff587-7186-4f81-83a2-886a15900229","Type":"ContainerDied","Data":"ccfecf4ef23c50d400b8ca7553bbf1eff4d9d62e8caff28f5a3562a7630cf393"} Jan 27 16:36:15 crc kubenswrapper[4772]: I0127 16:36:15.149901 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2jctn" Jan 27 16:36:15 crc kubenswrapper[4772]: I0127 16:36:15.327759 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/370ff587-7186-4f81-83a2-886a15900229-combined-ca-bundle\") pod \"370ff587-7186-4f81-83a2-886a15900229\" (UID: \"370ff587-7186-4f81-83a2-886a15900229\") " Jan 27 16:36:15 crc kubenswrapper[4772]: I0127 16:36:15.327992 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z55sq\" (UniqueName: \"kubernetes.io/projected/370ff587-7186-4f81-83a2-886a15900229-kube-api-access-z55sq\") pod \"370ff587-7186-4f81-83a2-886a15900229\" (UID: \"370ff587-7186-4f81-83a2-886a15900229\") " Jan 27 16:36:15 crc kubenswrapper[4772]: I0127 16:36:15.328021 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/370ff587-7186-4f81-83a2-886a15900229-db-sync-config-data\") pod \"370ff587-7186-4f81-83a2-886a15900229\" (UID: \"370ff587-7186-4f81-83a2-886a15900229\") " Jan 27 16:36:15 crc kubenswrapper[4772]: I0127 16:36:15.333353 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/370ff587-7186-4f81-83a2-886a15900229-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "370ff587-7186-4f81-83a2-886a15900229" (UID: "370ff587-7186-4f81-83a2-886a15900229"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:36:15 crc kubenswrapper[4772]: I0127 16:36:15.333683 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/370ff587-7186-4f81-83a2-886a15900229-kube-api-access-z55sq" (OuterVolumeSpecName: "kube-api-access-z55sq") pod "370ff587-7186-4f81-83a2-886a15900229" (UID: "370ff587-7186-4f81-83a2-886a15900229"). InnerVolumeSpecName "kube-api-access-z55sq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:36:15 crc kubenswrapper[4772]: I0127 16:36:15.349050 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/370ff587-7186-4f81-83a2-886a15900229-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "370ff587-7186-4f81-83a2-886a15900229" (UID: "370ff587-7186-4f81-83a2-886a15900229"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:36:15 crc kubenswrapper[4772]: I0127 16:36:15.430959 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z55sq\" (UniqueName: \"kubernetes.io/projected/370ff587-7186-4f81-83a2-886a15900229-kube-api-access-z55sq\") on node \"crc\" DevicePath \"\"" Jan 27 16:36:15 crc kubenswrapper[4772]: I0127 16:36:15.430999 4772 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/370ff587-7186-4f81-83a2-886a15900229-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:36:15 crc kubenswrapper[4772]: I0127 16:36:15.431013 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/370ff587-7186-4f81-83a2-886a15900229-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:36:15 crc kubenswrapper[4772]: I0127 16:36:15.858400 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2jctn" event={"ID":"370ff587-7186-4f81-83a2-886a15900229","Type":"ContainerDied","Data":"66fb01f7e2b25b680204b44292b0be570804c81d694aa738565de3c8455c3ac0"} Jan 27 16:36:15 crc kubenswrapper[4772]: I0127 16:36:15.858439 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66fb01f7e2b25b680204b44292b0be570804c81d694aa738565de3c8455c3ac0" Jan 27 16:36:15 crc kubenswrapper[4772]: I0127 16:36:15.858476 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2jctn" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.088300 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-69b64c5dd7-dj9pw"] Jan 27 16:36:16 crc kubenswrapper[4772]: E0127 16:36:16.088984 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="370ff587-7186-4f81-83a2-886a15900229" containerName="barbican-db-sync" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.089004 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="370ff587-7186-4f81-83a2-886a15900229" containerName="barbican-db-sync" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.089280 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="370ff587-7186-4f81-83a2-886a15900229" containerName="barbican-db-sync" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.090371 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-69b64c5dd7-dj9pw" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.093449 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.093685 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-d4g2g" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.093830 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.116071 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-69b64c5dd7-dj9pw"] Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.136122 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-75df4b6d74-xpp9t"] Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.138279 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-75df4b6d74-xpp9t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.142726 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.152685 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-75df4b6d74-xpp9t"] Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.219890 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55d7c89dc7-g6rsn"] Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.231113 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.246200 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55d7c89dc7-g6rsn"] Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.250846 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d49f4dc-fd69-4e43-9866-87af6da31197-combined-ca-bundle\") pod \"barbican-worker-69b64c5dd7-dj9pw\" (UID: \"0d49f4dc-fd69-4e43-9866-87af6da31197\") " pod="openstack/barbican-worker-69b64c5dd7-dj9pw" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.250910 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vx56\" (UniqueName: \"kubernetes.io/projected/0d49f4dc-fd69-4e43-9866-87af6da31197-kube-api-access-6vx56\") pod \"barbican-worker-69b64c5dd7-dj9pw\" (UID: \"0d49f4dc-fd69-4e43-9866-87af6da31197\") " pod="openstack/barbican-worker-69b64c5dd7-dj9pw" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.250955 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d49f4dc-fd69-4e43-9866-87af6da31197-config-data\") pod \"barbican-worker-69b64c5dd7-dj9pw\" (UID: \"0d49f4dc-fd69-4e43-9866-87af6da31197\") " pod="openstack/barbican-worker-69b64c5dd7-dj9pw" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.250984 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2da34b58-6b43-4e25-bdec-39985c344819-logs\") pod \"barbican-keystone-listener-75df4b6d74-xpp9t\" (UID: \"2da34b58-6b43-4e25-bdec-39985c344819\") " pod="openstack/barbican-keystone-listener-75df4b6d74-xpp9t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.251007 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2da34b58-6b43-4e25-bdec-39985c344819-config-data\") pod \"barbican-keystone-listener-75df4b6d74-xpp9t\" (UID: \"2da34b58-6b43-4e25-bdec-39985c344819\") " pod="openstack/barbican-keystone-listener-75df4b6d74-xpp9t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.251087 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2da34b58-6b43-4e25-bdec-39985c344819-config-data-custom\") pod \"barbican-keystone-listener-75df4b6d74-xpp9t\" (UID: \"2da34b58-6b43-4e25-bdec-39985c344819\") " pod="openstack/barbican-keystone-listener-75df4b6d74-xpp9t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.251119 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmbpb\" (UniqueName: \"kubernetes.io/projected/2da34b58-6b43-4e25-bdec-39985c344819-kube-api-access-nmbpb\") pod \"barbican-keystone-listener-75df4b6d74-xpp9t\" (UID: \"2da34b58-6b43-4e25-bdec-39985c344819\") " pod="openstack/barbican-keystone-listener-75df4b6d74-xpp9t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.251145 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0d49f4dc-fd69-4e43-9866-87af6da31197-config-data-custom\") pod \"barbican-worker-69b64c5dd7-dj9pw\" (UID: \"0d49f4dc-fd69-4e43-9866-87af6da31197\") " pod="openstack/barbican-worker-69b64c5dd7-dj9pw" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.251222 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da34b58-6b43-4e25-bdec-39985c344819-combined-ca-bundle\") pod \"barbican-keystone-listener-75df4b6d74-xpp9t\" (UID: \"2da34b58-6b43-4e25-bdec-39985c344819\") " pod="openstack/barbican-keystone-listener-75df4b6d74-xpp9t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.251245 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d49f4dc-fd69-4e43-9866-87af6da31197-logs\") pod \"barbican-worker-69b64c5dd7-dj9pw\" (UID: \"0d49f4dc-fd69-4e43-9866-87af6da31197\") " pod="openstack/barbican-worker-69b64c5dd7-dj9pw" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.315565 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7dcb7f9846-lrk6t"] Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.319036 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7dcb7f9846-lrk6t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.325603 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.341621 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7dcb7f9846-lrk6t"] Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.352673 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da34b58-6b43-4e25-bdec-39985c344819-combined-ca-bundle\") pod \"barbican-keystone-listener-75df4b6d74-xpp9t\" (UID: \"2da34b58-6b43-4e25-bdec-39985c344819\") " pod="openstack/barbican-keystone-listener-75df4b6d74-xpp9t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.352730 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d49f4dc-fd69-4e43-9866-87af6da31197-logs\") pod \"barbican-worker-69b64c5dd7-dj9pw\" (UID: \"0d49f4dc-fd69-4e43-9866-87af6da31197\") " pod="openstack/barbican-worker-69b64c5dd7-dj9pw" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.352788 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d49f4dc-fd69-4e43-9866-87af6da31197-combined-ca-bundle\") pod \"barbican-worker-69b64c5dd7-dj9pw\" (UID: \"0d49f4dc-fd69-4e43-9866-87af6da31197\") " pod="openstack/barbican-worker-69b64c5dd7-dj9pw" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.352811 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vx56\" (UniqueName: \"kubernetes.io/projected/0d49f4dc-fd69-4e43-9866-87af6da31197-kube-api-access-6vx56\") pod \"barbican-worker-69b64c5dd7-dj9pw\" (UID: \"0d49f4dc-fd69-4e43-9866-87af6da31197\") " pod="openstack/barbican-worker-69b64c5dd7-dj9pw" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.352836 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53c70df7-cdca-4296-af71-b5b002484575-config\") pod \"dnsmasq-dns-55d7c89dc7-g6rsn\" (UID: \"53c70df7-cdca-4296-af71-b5b002484575\") " pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.352863 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d49f4dc-fd69-4e43-9866-87af6da31197-config-data\") pod \"barbican-worker-69b64c5dd7-dj9pw\" (UID: \"0d49f4dc-fd69-4e43-9866-87af6da31197\") " pod="openstack/barbican-worker-69b64c5dd7-dj9pw" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.352883 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53c70df7-cdca-4296-af71-b5b002484575-ovsdbserver-nb\") pod \"dnsmasq-dns-55d7c89dc7-g6rsn\" (UID: \"53c70df7-cdca-4296-af71-b5b002484575\") " pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.352907 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2da34b58-6b43-4e25-bdec-39985c344819-config-data\") pod \"barbican-keystone-listener-75df4b6d74-xpp9t\" (UID: \"2da34b58-6b43-4e25-bdec-39985c344819\") " pod="openstack/barbican-keystone-listener-75df4b6d74-xpp9t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.352928 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2da34b58-6b43-4e25-bdec-39985c344819-logs\") pod \"barbican-keystone-listener-75df4b6d74-xpp9t\" (UID: \"2da34b58-6b43-4e25-bdec-39985c344819\") " pod="openstack/barbican-keystone-listener-75df4b6d74-xpp9t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.352989 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwtxg\" (UniqueName: \"kubernetes.io/projected/53c70df7-cdca-4296-af71-b5b002484575-kube-api-access-qwtxg\") pod \"dnsmasq-dns-55d7c89dc7-g6rsn\" (UID: \"53c70df7-cdca-4296-af71-b5b002484575\") " pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.353029 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2da34b58-6b43-4e25-bdec-39985c344819-config-data-custom\") pod \"barbican-keystone-listener-75df4b6d74-xpp9t\" (UID: \"2da34b58-6b43-4e25-bdec-39985c344819\") " pod="openstack/barbican-keystone-listener-75df4b6d74-xpp9t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.353062 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmbpb\" (UniqueName: \"kubernetes.io/projected/2da34b58-6b43-4e25-bdec-39985c344819-kube-api-access-nmbpb\") pod \"barbican-keystone-listener-75df4b6d74-xpp9t\" (UID: \"2da34b58-6b43-4e25-bdec-39985c344819\") " pod="openstack/barbican-keystone-listener-75df4b6d74-xpp9t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.353090 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0d49f4dc-fd69-4e43-9866-87af6da31197-config-data-custom\") pod \"barbican-worker-69b64c5dd7-dj9pw\" (UID: \"0d49f4dc-fd69-4e43-9866-87af6da31197\") " pod="openstack/barbican-worker-69b64c5dd7-dj9pw" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.353123 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53c70df7-cdca-4296-af71-b5b002484575-dns-svc\") pod \"dnsmasq-dns-55d7c89dc7-g6rsn\" (UID: \"53c70df7-cdca-4296-af71-b5b002484575\") " pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.353146 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53c70df7-cdca-4296-af71-b5b002484575-ovsdbserver-sb\") pod \"dnsmasq-dns-55d7c89dc7-g6rsn\" (UID: \"53c70df7-cdca-4296-af71-b5b002484575\") " pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.355435 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d49f4dc-fd69-4e43-9866-87af6da31197-logs\") pod \"barbican-worker-69b64c5dd7-dj9pw\" (UID: \"0d49f4dc-fd69-4e43-9866-87af6da31197\") " pod="openstack/barbican-worker-69b64c5dd7-dj9pw" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.356955 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2da34b58-6b43-4e25-bdec-39985c344819-logs\") pod \"barbican-keystone-listener-75df4b6d74-xpp9t\" (UID: \"2da34b58-6b43-4e25-bdec-39985c344819\") " pod="openstack/barbican-keystone-listener-75df4b6d74-xpp9t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.357121 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da34b58-6b43-4e25-bdec-39985c344819-combined-ca-bundle\") pod \"barbican-keystone-listener-75df4b6d74-xpp9t\" (UID: \"2da34b58-6b43-4e25-bdec-39985c344819\") " pod="openstack/barbican-keystone-listener-75df4b6d74-xpp9t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.358157 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d49f4dc-fd69-4e43-9866-87af6da31197-config-data\") pod \"barbican-worker-69b64c5dd7-dj9pw\" (UID: \"0d49f4dc-fd69-4e43-9866-87af6da31197\") " pod="openstack/barbican-worker-69b64c5dd7-dj9pw" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.359433 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d49f4dc-fd69-4e43-9866-87af6da31197-combined-ca-bundle\") pod \"barbican-worker-69b64c5dd7-dj9pw\" (UID: \"0d49f4dc-fd69-4e43-9866-87af6da31197\") " pod="openstack/barbican-worker-69b64c5dd7-dj9pw" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.371940 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2da34b58-6b43-4e25-bdec-39985c344819-config-data-custom\") pod \"barbican-keystone-listener-75df4b6d74-xpp9t\" (UID: \"2da34b58-6b43-4e25-bdec-39985c344819\") " pod="openstack/barbican-keystone-listener-75df4b6d74-xpp9t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.373189 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2da34b58-6b43-4e25-bdec-39985c344819-config-data\") pod \"barbican-keystone-listener-75df4b6d74-xpp9t\" (UID: \"2da34b58-6b43-4e25-bdec-39985c344819\") " pod="openstack/barbican-keystone-listener-75df4b6d74-xpp9t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.376266 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0d49f4dc-fd69-4e43-9866-87af6da31197-config-data-custom\") pod \"barbican-worker-69b64c5dd7-dj9pw\" (UID: \"0d49f4dc-fd69-4e43-9866-87af6da31197\") " pod="openstack/barbican-worker-69b64c5dd7-dj9pw" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.377679 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vx56\" (UniqueName: \"kubernetes.io/projected/0d49f4dc-fd69-4e43-9866-87af6da31197-kube-api-access-6vx56\") pod \"barbican-worker-69b64c5dd7-dj9pw\" (UID: \"0d49f4dc-fd69-4e43-9866-87af6da31197\") " pod="openstack/barbican-worker-69b64c5dd7-dj9pw" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.382893 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmbpb\" (UniqueName: \"kubernetes.io/projected/2da34b58-6b43-4e25-bdec-39985c344819-kube-api-access-nmbpb\") pod \"barbican-keystone-listener-75df4b6d74-xpp9t\" (UID: \"2da34b58-6b43-4e25-bdec-39985c344819\") " pod="openstack/barbican-keystone-listener-75df4b6d74-xpp9t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.414996 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-69b64c5dd7-dj9pw" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.455371 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c77e7c3-5320-4fa6-810d-bc819a6f7b03-logs\") pod \"barbican-api-7dcb7f9846-lrk6t\" (UID: \"5c77e7c3-5320-4fa6-810d-bc819a6f7b03\") " pod="openstack/barbican-api-7dcb7f9846-lrk6t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.455424 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c77e7c3-5320-4fa6-810d-bc819a6f7b03-combined-ca-bundle\") pod \"barbican-api-7dcb7f9846-lrk6t\" (UID: \"5c77e7c3-5320-4fa6-810d-bc819a6f7b03\") " pod="openstack/barbican-api-7dcb7f9846-lrk6t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.455460 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53c70df7-cdca-4296-af71-b5b002484575-ovsdbserver-nb\") pod \"dnsmasq-dns-55d7c89dc7-g6rsn\" (UID: \"53c70df7-cdca-4296-af71-b5b002484575\") " pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.455800 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwtxg\" (UniqueName: \"kubernetes.io/projected/53c70df7-cdca-4296-af71-b5b002484575-kube-api-access-qwtxg\") pod \"dnsmasq-dns-55d7c89dc7-g6rsn\" (UID: \"53c70df7-cdca-4296-af71-b5b002484575\") " pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.456017 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53c70df7-cdca-4296-af71-b5b002484575-dns-svc\") pod \"dnsmasq-dns-55d7c89dc7-g6rsn\" (UID: \"53c70df7-cdca-4296-af71-b5b002484575\") " pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.456428 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53c70df7-cdca-4296-af71-b5b002484575-ovsdbserver-nb\") pod \"dnsmasq-dns-55d7c89dc7-g6rsn\" (UID: \"53c70df7-cdca-4296-af71-b5b002484575\") " pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.456926 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53c70df7-cdca-4296-af71-b5b002484575-dns-svc\") pod \"dnsmasq-dns-55d7c89dc7-g6rsn\" (UID: \"53c70df7-cdca-4296-af71-b5b002484575\") " pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.456048 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53c70df7-cdca-4296-af71-b5b002484575-ovsdbserver-sb\") pod \"dnsmasq-dns-55d7c89dc7-g6rsn\" (UID: \"53c70df7-cdca-4296-af71-b5b002484575\") " pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.456981 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53c70df7-cdca-4296-af71-b5b002484575-ovsdbserver-sb\") pod \"dnsmasq-dns-55d7c89dc7-g6rsn\" (UID: \"53c70df7-cdca-4296-af71-b5b002484575\") " pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.457023 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9z7sl\" (UniqueName: \"kubernetes.io/projected/5c77e7c3-5320-4fa6-810d-bc819a6f7b03-kube-api-access-9z7sl\") pod \"barbican-api-7dcb7f9846-lrk6t\" (UID: \"5c77e7c3-5320-4fa6-810d-bc819a6f7b03\") " pod="openstack/barbican-api-7dcb7f9846-lrk6t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.457109 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c77e7c3-5320-4fa6-810d-bc819a6f7b03-config-data\") pod \"barbican-api-7dcb7f9846-lrk6t\" (UID: \"5c77e7c3-5320-4fa6-810d-bc819a6f7b03\") " pod="openstack/barbican-api-7dcb7f9846-lrk6t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.457196 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c77e7c3-5320-4fa6-810d-bc819a6f7b03-config-data-custom\") pod \"barbican-api-7dcb7f9846-lrk6t\" (UID: \"5c77e7c3-5320-4fa6-810d-bc819a6f7b03\") " pod="openstack/barbican-api-7dcb7f9846-lrk6t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.457957 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53c70df7-cdca-4296-af71-b5b002484575-config\") pod \"dnsmasq-dns-55d7c89dc7-g6rsn\" (UID: \"53c70df7-cdca-4296-af71-b5b002484575\") " pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.458008 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53c70df7-cdca-4296-af71-b5b002484575-config\") pod \"dnsmasq-dns-55d7c89dc7-g6rsn\" (UID: \"53c70df7-cdca-4296-af71-b5b002484575\") " pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.478598 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwtxg\" (UniqueName: \"kubernetes.io/projected/53c70df7-cdca-4296-af71-b5b002484575-kube-api-access-qwtxg\") pod \"dnsmasq-dns-55d7c89dc7-g6rsn\" (UID: \"53c70df7-cdca-4296-af71-b5b002484575\") " pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.504667 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-75df4b6d74-xpp9t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.559857 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9z7sl\" (UniqueName: \"kubernetes.io/projected/5c77e7c3-5320-4fa6-810d-bc819a6f7b03-kube-api-access-9z7sl\") pod \"barbican-api-7dcb7f9846-lrk6t\" (UID: \"5c77e7c3-5320-4fa6-810d-bc819a6f7b03\") " pod="openstack/barbican-api-7dcb7f9846-lrk6t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.560252 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c77e7c3-5320-4fa6-810d-bc819a6f7b03-config-data\") pod \"barbican-api-7dcb7f9846-lrk6t\" (UID: \"5c77e7c3-5320-4fa6-810d-bc819a6f7b03\") " pod="openstack/barbican-api-7dcb7f9846-lrk6t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.560276 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c77e7c3-5320-4fa6-810d-bc819a6f7b03-config-data-custom\") pod \"barbican-api-7dcb7f9846-lrk6t\" (UID: \"5c77e7c3-5320-4fa6-810d-bc819a6f7b03\") " pod="openstack/barbican-api-7dcb7f9846-lrk6t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.560323 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c77e7c3-5320-4fa6-810d-bc819a6f7b03-logs\") pod \"barbican-api-7dcb7f9846-lrk6t\" (UID: \"5c77e7c3-5320-4fa6-810d-bc819a6f7b03\") " pod="openstack/barbican-api-7dcb7f9846-lrk6t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.560343 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c77e7c3-5320-4fa6-810d-bc819a6f7b03-combined-ca-bundle\") pod \"barbican-api-7dcb7f9846-lrk6t\" (UID: \"5c77e7c3-5320-4fa6-810d-bc819a6f7b03\") " pod="openstack/barbican-api-7dcb7f9846-lrk6t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.561797 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c77e7c3-5320-4fa6-810d-bc819a6f7b03-logs\") pod \"barbican-api-7dcb7f9846-lrk6t\" (UID: \"5c77e7c3-5320-4fa6-810d-bc819a6f7b03\") " pod="openstack/barbican-api-7dcb7f9846-lrk6t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.562263 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.567467 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c77e7c3-5320-4fa6-810d-bc819a6f7b03-config-data\") pod \"barbican-api-7dcb7f9846-lrk6t\" (UID: \"5c77e7c3-5320-4fa6-810d-bc819a6f7b03\") " pod="openstack/barbican-api-7dcb7f9846-lrk6t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.577131 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c77e7c3-5320-4fa6-810d-bc819a6f7b03-config-data-custom\") pod \"barbican-api-7dcb7f9846-lrk6t\" (UID: \"5c77e7c3-5320-4fa6-810d-bc819a6f7b03\") " pod="openstack/barbican-api-7dcb7f9846-lrk6t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.579614 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c77e7c3-5320-4fa6-810d-bc819a6f7b03-combined-ca-bundle\") pod \"barbican-api-7dcb7f9846-lrk6t\" (UID: \"5c77e7c3-5320-4fa6-810d-bc819a6f7b03\") " pod="openstack/barbican-api-7dcb7f9846-lrk6t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.584373 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9z7sl\" (UniqueName: \"kubernetes.io/projected/5c77e7c3-5320-4fa6-810d-bc819a6f7b03-kube-api-access-9z7sl\") pod \"barbican-api-7dcb7f9846-lrk6t\" (UID: \"5c77e7c3-5320-4fa6-810d-bc819a6f7b03\") " pod="openstack/barbican-api-7dcb7f9846-lrk6t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.644317 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7dcb7f9846-lrk6t" Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.729587 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-69b64c5dd7-dj9pw"] Jan 27 16:36:16 crc kubenswrapper[4772]: I0127 16:36:16.880343 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-69b64c5dd7-dj9pw" event={"ID":"0d49f4dc-fd69-4e43-9866-87af6da31197","Type":"ContainerStarted","Data":"13e2bafea465246f523200b9e1d2507ff9859908f73a49b661a410e4554ffc6a"} Jan 27 16:36:17 crc kubenswrapper[4772]: I0127 16:36:17.081819 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-75df4b6d74-xpp9t"] Jan 27 16:36:17 crc kubenswrapper[4772]: I0127 16:36:17.112888 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7dcb7f9846-lrk6t"] Jan 27 16:36:17 crc kubenswrapper[4772]: W0127 16:36:17.209950 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c77e7c3_5320_4fa6_810d_bc819a6f7b03.slice/crio-74a6a9add23994a7881d8a2122dcd2a1fba5bffb8857d7dcbb30dfad5a2c6e1c WatchSource:0}: Error finding container 74a6a9add23994a7881d8a2122dcd2a1fba5bffb8857d7dcbb30dfad5a2c6e1c: Status 404 returned error can't find the container with id 74a6a9add23994a7881d8a2122dcd2a1fba5bffb8857d7dcbb30dfad5a2c6e1c Jan 27 16:36:17 crc kubenswrapper[4772]: I0127 16:36:17.220562 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55d7c89dc7-g6rsn"] Jan 27 16:36:17 crc kubenswrapper[4772]: I0127 16:36:17.891212 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7dcb7f9846-lrk6t" event={"ID":"5c77e7c3-5320-4fa6-810d-bc819a6f7b03","Type":"ContainerStarted","Data":"f12fb3f9c0d962fbef42c301cc8095cd2f46733470a2bdb51e65accd5f2cd4b5"} Jan 27 16:36:17 crc kubenswrapper[4772]: I0127 16:36:17.891495 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7dcb7f9846-lrk6t" event={"ID":"5c77e7c3-5320-4fa6-810d-bc819a6f7b03","Type":"ContainerStarted","Data":"9eca56474104926e985ccce0312b20cf97a4756f8d45e934237a323350012d2a"} Jan 27 16:36:17 crc kubenswrapper[4772]: I0127 16:36:17.891513 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7dcb7f9846-lrk6t" event={"ID":"5c77e7c3-5320-4fa6-810d-bc819a6f7b03","Type":"ContainerStarted","Data":"74a6a9add23994a7881d8a2122dcd2a1fba5bffb8857d7dcbb30dfad5a2c6e1c"} Jan 27 16:36:17 crc kubenswrapper[4772]: I0127 16:36:17.891530 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7dcb7f9846-lrk6t" Jan 27 16:36:17 crc kubenswrapper[4772]: I0127 16:36:17.892818 4772 generic.go:334] "Generic (PLEG): container finished" podID="53c70df7-cdca-4296-af71-b5b002484575" containerID="d7ad019f98bbdfacb03703ed4de276fec18a82aa27109f1c0ec2b92483868eba" exitCode=0 Jan 27 16:36:17 crc kubenswrapper[4772]: I0127 16:36:17.892910 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" event={"ID":"53c70df7-cdca-4296-af71-b5b002484575","Type":"ContainerDied","Data":"d7ad019f98bbdfacb03703ed4de276fec18a82aa27109f1c0ec2b92483868eba"} Jan 27 16:36:17 crc kubenswrapper[4772]: I0127 16:36:17.892954 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" event={"ID":"53c70df7-cdca-4296-af71-b5b002484575","Type":"ContainerStarted","Data":"ab1dfeb7c7875a02f8dfd72be8da51340b47ef61cc5646a975b8f955d01b7ae0"} Jan 27 16:36:17 crc kubenswrapper[4772]: I0127 16:36:17.895441 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-69b64c5dd7-dj9pw" event={"ID":"0d49f4dc-fd69-4e43-9866-87af6da31197","Type":"ContainerStarted","Data":"fe239a44b812bcf92e470bb7c00a935448d74f7c04681a2145ead95d654c9d60"} Jan 27 16:36:17 crc kubenswrapper[4772]: I0127 16:36:17.895468 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-69b64c5dd7-dj9pw" event={"ID":"0d49f4dc-fd69-4e43-9866-87af6da31197","Type":"ContainerStarted","Data":"c2da4b92862affafff9c7df896dc126644dbea4b5caddda1b9369b24e9864bde"} Jan 27 16:36:17 crc kubenswrapper[4772]: I0127 16:36:17.899145 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-75df4b6d74-xpp9t" event={"ID":"2da34b58-6b43-4e25-bdec-39985c344819","Type":"ContainerStarted","Data":"715392760b58db6867b76332ca5ce68acb4f57a47c4a2d79c281bad50a940b93"} Jan 27 16:36:17 crc kubenswrapper[4772]: I0127 16:36:17.899218 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-75df4b6d74-xpp9t" event={"ID":"2da34b58-6b43-4e25-bdec-39985c344819","Type":"ContainerStarted","Data":"50d65912201083457d8e9653e4542690d4ae602716af39ac11b4a87b0d075aae"} Jan 27 16:36:17 crc kubenswrapper[4772]: I0127 16:36:17.899237 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-75df4b6d74-xpp9t" event={"ID":"2da34b58-6b43-4e25-bdec-39985c344819","Type":"ContainerStarted","Data":"906d0db148f7dd6d703853a32e42bfc3bd01bb165a51a1d14e445ba30a5155fa"} Jan 27 16:36:17 crc kubenswrapper[4772]: I0127 16:36:17.918102 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7dcb7f9846-lrk6t" podStartSLOduration=1.918085255 podStartE2EDuration="1.918085255s" podCreationTimestamp="2026-01-27 16:36:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:36:17.912415223 +0000 UTC m=+5363.893024331" watchObservedRunningTime="2026-01-27 16:36:17.918085255 +0000 UTC m=+5363.898694353" Jan 27 16:36:17 crc kubenswrapper[4772]: I0127 16:36:17.945995 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-75df4b6d74-xpp9t" podStartSLOduration=1.945953439 podStartE2EDuration="1.945953439s" podCreationTimestamp="2026-01-27 16:36:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:36:17.942360967 +0000 UTC m=+5363.922970065" watchObservedRunningTime="2026-01-27 16:36:17.945953439 +0000 UTC m=+5363.926562537" Jan 27 16:36:17 crc kubenswrapper[4772]: I0127 16:36:17.993038 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-69b64c5dd7-dj9pw" podStartSLOduration=1.993021261 podStartE2EDuration="1.993021261s" podCreationTimestamp="2026-01-27 16:36:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:36:17.990302683 +0000 UTC m=+5363.970911781" watchObservedRunningTime="2026-01-27 16:36:17.993021261 +0000 UTC m=+5363.973630359" Jan 27 16:36:18 crc kubenswrapper[4772]: I0127 16:36:18.910467 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" event={"ID":"53c70df7-cdca-4296-af71-b5b002484575","Type":"ContainerStarted","Data":"6271d29d3114164d534a9aed9b1a74dcb46dca81dd4e99dc8ee2c848b0770a70"} Jan 27 16:36:18 crc kubenswrapper[4772]: I0127 16:36:18.910894 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7dcb7f9846-lrk6t" Jan 27 16:36:18 crc kubenswrapper[4772]: I0127 16:36:18.911180 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" Jan 27 16:36:18 crc kubenswrapper[4772]: I0127 16:36:18.942452 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" podStartSLOduration=2.942430413 podStartE2EDuration="2.942430413s" podCreationTimestamp="2026-01-27 16:36:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:36:18.934117786 +0000 UTC m=+5364.914726894" watchObservedRunningTime="2026-01-27 16:36:18.942430413 +0000 UTC m=+5364.923039501" Jan 27 16:36:26 crc kubenswrapper[4772]: I0127 16:36:26.563313 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" Jan 27 16:36:26 crc kubenswrapper[4772]: I0127 16:36:26.637545 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55b86d75d9-f64vw"] Jan 27 16:36:26 crc kubenswrapper[4772]: I0127 16:36:26.639032 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" podUID="9f5334e8-ad35-41c0-b74f-d7283b625da0" containerName="dnsmasq-dns" containerID="cri-o://e2902da210e724572224515decc30f5a1e9fa786701c1868e86464783fea99a9" gracePeriod=10 Jan 27 16:36:26 crc kubenswrapper[4772]: I0127 16:36:26.987487 4772 generic.go:334] "Generic (PLEG): container finished" podID="9f5334e8-ad35-41c0-b74f-d7283b625da0" containerID="e2902da210e724572224515decc30f5a1e9fa786701c1868e86464783fea99a9" exitCode=0 Jan 27 16:36:26 crc kubenswrapper[4772]: I0127 16:36:26.987532 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" event={"ID":"9f5334e8-ad35-41c0-b74f-d7283b625da0","Type":"ContainerDied","Data":"e2902da210e724572224515decc30f5a1e9fa786701c1868e86464783fea99a9"} Jan 27 16:36:27 crc kubenswrapper[4772]: I0127 16:36:27.169983 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" Jan 27 16:36:27 crc kubenswrapper[4772]: I0127 16:36:27.257713 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f5334e8-ad35-41c0-b74f-d7283b625da0-dns-svc\") pod \"9f5334e8-ad35-41c0-b74f-d7283b625da0\" (UID: \"9f5334e8-ad35-41c0-b74f-d7283b625da0\") " Jan 27 16:36:27 crc kubenswrapper[4772]: I0127 16:36:27.258071 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f5334e8-ad35-41c0-b74f-d7283b625da0-config\") pod \"9f5334e8-ad35-41c0-b74f-d7283b625da0\" (UID: \"9f5334e8-ad35-41c0-b74f-d7283b625da0\") " Jan 27 16:36:27 crc kubenswrapper[4772]: I0127 16:36:27.258130 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f5334e8-ad35-41c0-b74f-d7283b625da0-ovsdbserver-sb\") pod \"9f5334e8-ad35-41c0-b74f-d7283b625da0\" (UID: \"9f5334e8-ad35-41c0-b74f-d7283b625da0\") " Jan 27 16:36:27 crc kubenswrapper[4772]: I0127 16:36:27.258220 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fqwn\" (UniqueName: \"kubernetes.io/projected/9f5334e8-ad35-41c0-b74f-d7283b625da0-kube-api-access-2fqwn\") pod \"9f5334e8-ad35-41c0-b74f-d7283b625da0\" (UID: \"9f5334e8-ad35-41c0-b74f-d7283b625da0\") " Jan 27 16:36:27 crc kubenswrapper[4772]: I0127 16:36:27.258286 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f5334e8-ad35-41c0-b74f-d7283b625da0-ovsdbserver-nb\") pod \"9f5334e8-ad35-41c0-b74f-d7283b625da0\" (UID: \"9f5334e8-ad35-41c0-b74f-d7283b625da0\") " Jan 27 16:36:27 crc kubenswrapper[4772]: I0127 16:36:27.269499 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f5334e8-ad35-41c0-b74f-d7283b625da0-kube-api-access-2fqwn" (OuterVolumeSpecName: "kube-api-access-2fqwn") pod "9f5334e8-ad35-41c0-b74f-d7283b625da0" (UID: "9f5334e8-ad35-41c0-b74f-d7283b625da0"). InnerVolumeSpecName "kube-api-access-2fqwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:36:27 crc kubenswrapper[4772]: I0127 16:36:27.304756 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f5334e8-ad35-41c0-b74f-d7283b625da0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9f5334e8-ad35-41c0-b74f-d7283b625da0" (UID: "9f5334e8-ad35-41c0-b74f-d7283b625da0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:36:27 crc kubenswrapper[4772]: I0127 16:36:27.308183 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f5334e8-ad35-41c0-b74f-d7283b625da0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9f5334e8-ad35-41c0-b74f-d7283b625da0" (UID: "9f5334e8-ad35-41c0-b74f-d7283b625da0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:36:27 crc kubenswrapper[4772]: I0127 16:36:27.309044 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f5334e8-ad35-41c0-b74f-d7283b625da0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9f5334e8-ad35-41c0-b74f-d7283b625da0" (UID: "9f5334e8-ad35-41c0-b74f-d7283b625da0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:36:27 crc kubenswrapper[4772]: I0127 16:36:27.316465 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f5334e8-ad35-41c0-b74f-d7283b625da0-config" (OuterVolumeSpecName: "config") pod "9f5334e8-ad35-41c0-b74f-d7283b625da0" (UID: "9f5334e8-ad35-41c0-b74f-d7283b625da0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:36:27 crc kubenswrapper[4772]: I0127 16:36:27.361075 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f5334e8-ad35-41c0-b74f-d7283b625da0-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 16:36:27 crc kubenswrapper[4772]: I0127 16:36:27.361118 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f5334e8-ad35-41c0-b74f-d7283b625da0-config\") on node \"crc\" DevicePath \"\"" Jan 27 16:36:27 crc kubenswrapper[4772]: I0127 16:36:27.361133 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f5334e8-ad35-41c0-b74f-d7283b625da0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 27 16:36:27 crc kubenswrapper[4772]: I0127 16:36:27.361147 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fqwn\" (UniqueName: \"kubernetes.io/projected/9f5334e8-ad35-41c0-b74f-d7283b625da0-kube-api-access-2fqwn\") on node \"crc\" DevicePath \"\"" Jan 27 16:36:27 crc kubenswrapper[4772]: I0127 16:36:27.361156 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f5334e8-ad35-41c0-b74f-d7283b625da0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.000442 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" event={"ID":"9f5334e8-ad35-41c0-b74f-d7283b625da0","Type":"ContainerDied","Data":"f24d4586c71be1b7413cfd65197ec5183ca5b0b700f7ea419468e58b997588e5"} Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.000505 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55b86d75d9-f64vw" Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.001210 4772 scope.go:117] "RemoveContainer" containerID="e2902da210e724572224515decc30f5a1e9fa786701c1868e86464783fea99a9" Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.052885 4772 scope.go:117] "RemoveContainer" containerID="7afe0140c22bc4a9daf1f5cd3e32f3b90e22ae38f75fd3783a129a88072a7fc4" Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.053001 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55b86d75d9-f64vw"] Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.060806 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55b86d75d9-f64vw"] Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.091362 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-74dnd"] Jan 27 16:36:28 crc kubenswrapper[4772]: E0127 16:36:28.092700 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f5334e8-ad35-41c0-b74f-d7283b625da0" containerName="init" Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.092725 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f5334e8-ad35-41c0-b74f-d7283b625da0" containerName="init" Jan 27 16:36:28 crc kubenswrapper[4772]: E0127 16:36:28.092755 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f5334e8-ad35-41c0-b74f-d7283b625da0" containerName="dnsmasq-dns" Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.092763 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f5334e8-ad35-41c0-b74f-d7283b625da0" containerName="dnsmasq-dns" Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.092965 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f5334e8-ad35-41c0-b74f-d7283b625da0" containerName="dnsmasq-dns" Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.097750 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-74dnd" Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.107201 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-74dnd"] Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.118953 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7dcb7f9846-lrk6t" Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.178453 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3141ae0-d9c6-4eb9-8ada-8dd2454da297-utilities\") pod \"redhat-operators-74dnd\" (UID: \"a3141ae0-d9c6-4eb9-8ada-8dd2454da297\") " pod="openshift-marketplace/redhat-operators-74dnd" Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.178506 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cr8tc\" (UniqueName: \"kubernetes.io/projected/a3141ae0-d9c6-4eb9-8ada-8dd2454da297-kube-api-access-cr8tc\") pod \"redhat-operators-74dnd\" (UID: \"a3141ae0-d9c6-4eb9-8ada-8dd2454da297\") " pod="openshift-marketplace/redhat-operators-74dnd" Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.178531 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3141ae0-d9c6-4eb9-8ada-8dd2454da297-catalog-content\") pod \"redhat-operators-74dnd\" (UID: \"a3141ae0-d9c6-4eb9-8ada-8dd2454da297\") " pod="openshift-marketplace/redhat-operators-74dnd" Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.265070 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7dcb7f9846-lrk6t" Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.282087 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3141ae0-d9c6-4eb9-8ada-8dd2454da297-utilities\") pod \"redhat-operators-74dnd\" (UID: \"a3141ae0-d9c6-4eb9-8ada-8dd2454da297\") " pod="openshift-marketplace/redhat-operators-74dnd" Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.282129 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cr8tc\" (UniqueName: \"kubernetes.io/projected/a3141ae0-d9c6-4eb9-8ada-8dd2454da297-kube-api-access-cr8tc\") pod \"redhat-operators-74dnd\" (UID: \"a3141ae0-d9c6-4eb9-8ada-8dd2454da297\") " pod="openshift-marketplace/redhat-operators-74dnd" Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.282179 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3141ae0-d9c6-4eb9-8ada-8dd2454da297-catalog-content\") pod \"redhat-operators-74dnd\" (UID: \"a3141ae0-d9c6-4eb9-8ada-8dd2454da297\") " pod="openshift-marketplace/redhat-operators-74dnd" Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.282644 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3141ae0-d9c6-4eb9-8ada-8dd2454da297-catalog-content\") pod \"redhat-operators-74dnd\" (UID: \"a3141ae0-d9c6-4eb9-8ada-8dd2454da297\") " pod="openshift-marketplace/redhat-operators-74dnd" Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.284859 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3141ae0-d9c6-4eb9-8ada-8dd2454da297-utilities\") pod \"redhat-operators-74dnd\" (UID: \"a3141ae0-d9c6-4eb9-8ada-8dd2454da297\") " pod="openshift-marketplace/redhat-operators-74dnd" Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.312336 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cr8tc\" (UniqueName: \"kubernetes.io/projected/a3141ae0-d9c6-4eb9-8ada-8dd2454da297-kube-api-access-cr8tc\") pod \"redhat-operators-74dnd\" (UID: \"a3141ae0-d9c6-4eb9-8ada-8dd2454da297\") " pod="openshift-marketplace/redhat-operators-74dnd" Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.430118 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-74dnd" Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.681049 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f5334e8-ad35-41c0-b74f-d7283b625da0" path="/var/lib/kubelet/pods/9f5334e8-ad35-41c0-b74f-d7283b625da0/volumes" Jan 27 16:36:28 crc kubenswrapper[4772]: I0127 16:36:28.685018 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-74dnd"] Jan 27 16:36:29 crc kubenswrapper[4772]: I0127 16:36:29.022822 4772 generic.go:334] "Generic (PLEG): container finished" podID="a3141ae0-d9c6-4eb9-8ada-8dd2454da297" containerID="4823b71d7a398dbe52d74d93991da82e8680383fcd12420daf8b869332d53333" exitCode=0 Jan 27 16:36:29 crc kubenswrapper[4772]: I0127 16:36:29.022923 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74dnd" event={"ID":"a3141ae0-d9c6-4eb9-8ada-8dd2454da297","Type":"ContainerDied","Data":"4823b71d7a398dbe52d74d93991da82e8680383fcd12420daf8b869332d53333"} Jan 27 16:36:29 crc kubenswrapper[4772]: I0127 16:36:29.023266 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74dnd" event={"ID":"a3141ae0-d9c6-4eb9-8ada-8dd2454da297","Type":"ContainerStarted","Data":"edeeb734fb1a3ced8ebfb84b4f3f42b1d212bd81615ff209859424b50ce18695"} Jan 27 16:36:30 crc kubenswrapper[4772]: I0127 16:36:30.032805 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74dnd" event={"ID":"a3141ae0-d9c6-4eb9-8ada-8dd2454da297","Type":"ContainerStarted","Data":"b27ff19e88ae10072e54a035b5eeff3472d319b04facf31b91170a2a03f287d0"} Jan 27 16:36:31 crc kubenswrapper[4772]: I0127 16:36:31.047473 4772 generic.go:334] "Generic (PLEG): container finished" podID="a3141ae0-d9c6-4eb9-8ada-8dd2454da297" containerID="b27ff19e88ae10072e54a035b5eeff3472d319b04facf31b91170a2a03f287d0" exitCode=0 Jan 27 16:36:31 crc kubenswrapper[4772]: I0127 16:36:31.047530 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74dnd" event={"ID":"a3141ae0-d9c6-4eb9-8ada-8dd2454da297","Type":"ContainerDied","Data":"b27ff19e88ae10072e54a035b5eeff3472d319b04facf31b91170a2a03f287d0"} Jan 27 16:36:32 crc kubenswrapper[4772]: I0127 16:36:32.061216 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74dnd" event={"ID":"a3141ae0-d9c6-4eb9-8ada-8dd2454da297","Type":"ContainerStarted","Data":"ce0e7196a0a64497f42a021e97d6f099be19a6b7a8db931509a25e0302440168"} Jan 27 16:36:32 crc kubenswrapper[4772]: I0127 16:36:32.083653 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-74dnd" podStartSLOduration=1.66403699 podStartE2EDuration="4.083633477s" podCreationTimestamp="2026-01-27 16:36:28 +0000 UTC" firstStartedPulling="2026-01-27 16:36:29.027427004 +0000 UTC m=+5375.008036102" lastFinishedPulling="2026-01-27 16:36:31.447023481 +0000 UTC m=+5377.427632589" observedRunningTime="2026-01-27 16:36:32.082488835 +0000 UTC m=+5378.063097933" watchObservedRunningTime="2026-01-27 16:36:32.083633477 +0000 UTC m=+5378.064242575" Jan 27 16:36:38 crc kubenswrapper[4772]: I0127 16:36:38.430300 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-74dnd" Jan 27 16:36:38 crc kubenswrapper[4772]: I0127 16:36:38.430812 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-74dnd" Jan 27 16:36:38 crc kubenswrapper[4772]: I0127 16:36:38.485180 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-74dnd" Jan 27 16:36:39 crc kubenswrapper[4772]: I0127 16:36:39.156885 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-74dnd" Jan 27 16:36:39 crc kubenswrapper[4772]: I0127 16:36:39.209413 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-74dnd"] Jan 27 16:36:39 crc kubenswrapper[4772]: I0127 16:36:39.908115 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-49fxh"] Jan 27 16:36:39 crc kubenswrapper[4772]: I0127 16:36:39.910110 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-49fxh" Jan 27 16:36:39 crc kubenswrapper[4772]: I0127 16:36:39.917743 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-49fxh"] Jan 27 16:36:40 crc kubenswrapper[4772]: I0127 16:36:40.013747 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-fbd6-account-create-update-kvjsf"] Jan 27 16:36:40 crc kubenswrapper[4772]: I0127 16:36:40.014953 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-fbd6-account-create-update-kvjsf" Jan 27 16:36:40 crc kubenswrapper[4772]: I0127 16:36:40.024746 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Jan 27 16:36:40 crc kubenswrapper[4772]: I0127 16:36:40.055444 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-fbd6-account-create-update-kvjsf"] Jan 27 16:36:40 crc kubenswrapper[4772]: I0127 16:36:40.081419 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf2sr\" (UniqueName: \"kubernetes.io/projected/6dd740e1-b6ee-444d-b71e-e18d4837ef8a-kube-api-access-hf2sr\") pod \"neutron-db-create-49fxh\" (UID: \"6dd740e1-b6ee-444d-b71e-e18d4837ef8a\") " pod="openstack/neutron-db-create-49fxh" Jan 27 16:36:40 crc kubenswrapper[4772]: I0127 16:36:40.081510 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dd740e1-b6ee-444d-b71e-e18d4837ef8a-operator-scripts\") pod \"neutron-db-create-49fxh\" (UID: \"6dd740e1-b6ee-444d-b71e-e18d4837ef8a\") " pod="openstack/neutron-db-create-49fxh" Jan 27 16:36:40 crc kubenswrapper[4772]: I0127 16:36:40.183463 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b-operator-scripts\") pod \"neutron-fbd6-account-create-update-kvjsf\" (UID: \"0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b\") " pod="openstack/neutron-fbd6-account-create-update-kvjsf" Jan 27 16:36:40 crc kubenswrapper[4772]: I0127 16:36:40.183724 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf2sr\" (UniqueName: \"kubernetes.io/projected/6dd740e1-b6ee-444d-b71e-e18d4837ef8a-kube-api-access-hf2sr\") pod \"neutron-db-create-49fxh\" (UID: \"6dd740e1-b6ee-444d-b71e-e18d4837ef8a\") " pod="openstack/neutron-db-create-49fxh" Jan 27 16:36:40 crc kubenswrapper[4772]: I0127 16:36:40.183822 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dd740e1-b6ee-444d-b71e-e18d4837ef8a-operator-scripts\") pod \"neutron-db-create-49fxh\" (UID: \"6dd740e1-b6ee-444d-b71e-e18d4837ef8a\") " pod="openstack/neutron-db-create-49fxh" Jan 27 16:36:40 crc kubenswrapper[4772]: I0127 16:36:40.184011 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkkbh\" (UniqueName: \"kubernetes.io/projected/0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b-kube-api-access-wkkbh\") pod \"neutron-fbd6-account-create-update-kvjsf\" (UID: \"0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b\") " pod="openstack/neutron-fbd6-account-create-update-kvjsf" Jan 27 16:36:40 crc kubenswrapper[4772]: I0127 16:36:40.184680 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dd740e1-b6ee-444d-b71e-e18d4837ef8a-operator-scripts\") pod \"neutron-db-create-49fxh\" (UID: \"6dd740e1-b6ee-444d-b71e-e18d4837ef8a\") " pod="openstack/neutron-db-create-49fxh" Jan 27 16:36:40 crc kubenswrapper[4772]: I0127 16:36:40.207807 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf2sr\" (UniqueName: \"kubernetes.io/projected/6dd740e1-b6ee-444d-b71e-e18d4837ef8a-kube-api-access-hf2sr\") pod \"neutron-db-create-49fxh\" (UID: \"6dd740e1-b6ee-444d-b71e-e18d4837ef8a\") " pod="openstack/neutron-db-create-49fxh" Jan 27 16:36:40 crc kubenswrapper[4772]: I0127 16:36:40.227096 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-49fxh" Jan 27 16:36:40 crc kubenswrapper[4772]: I0127 16:36:40.285624 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b-operator-scripts\") pod \"neutron-fbd6-account-create-update-kvjsf\" (UID: \"0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b\") " pod="openstack/neutron-fbd6-account-create-update-kvjsf" Jan 27 16:36:40 crc kubenswrapper[4772]: I0127 16:36:40.285787 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkkbh\" (UniqueName: \"kubernetes.io/projected/0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b-kube-api-access-wkkbh\") pod \"neutron-fbd6-account-create-update-kvjsf\" (UID: \"0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b\") " pod="openstack/neutron-fbd6-account-create-update-kvjsf" Jan 27 16:36:40 crc kubenswrapper[4772]: I0127 16:36:40.286649 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b-operator-scripts\") pod \"neutron-fbd6-account-create-update-kvjsf\" (UID: \"0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b\") " pod="openstack/neutron-fbd6-account-create-update-kvjsf" Jan 27 16:36:40 crc kubenswrapper[4772]: I0127 16:36:40.309600 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkkbh\" (UniqueName: \"kubernetes.io/projected/0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b-kube-api-access-wkkbh\") pod \"neutron-fbd6-account-create-update-kvjsf\" (UID: \"0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b\") " pod="openstack/neutron-fbd6-account-create-update-kvjsf" Jan 27 16:36:40 crc kubenswrapper[4772]: I0127 16:36:40.346383 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-fbd6-account-create-update-kvjsf" Jan 27 16:36:40 crc kubenswrapper[4772]: I0127 16:36:40.677123 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-49fxh"] Jan 27 16:36:40 crc kubenswrapper[4772]: I0127 16:36:40.800019 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-fbd6-account-create-update-kvjsf"] Jan 27 16:36:41 crc kubenswrapper[4772]: I0127 16:36:41.134519 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-fbd6-account-create-update-kvjsf" event={"ID":"0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b","Type":"ContainerStarted","Data":"7820202ed4fe89d2fb42f752ab53ea23a8554f7e3772282ffa645c3878d8acde"} Jan 27 16:36:41 crc kubenswrapper[4772]: I0127 16:36:41.134570 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-fbd6-account-create-update-kvjsf" event={"ID":"0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b","Type":"ContainerStarted","Data":"98049675b35bf9aa5f55e27edcefac02abeb383c73491edfaf8dc4f2d0cc9061"} Jan 27 16:36:41 crc kubenswrapper[4772]: I0127 16:36:41.135873 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-49fxh" event={"ID":"6dd740e1-b6ee-444d-b71e-e18d4837ef8a","Type":"ContainerStarted","Data":"0eb151bf9a1bfafe986b27f8d85f2a3fcbd2a1f6be731bcd2ce95359e6a6e136"} Jan 27 16:36:41 crc kubenswrapper[4772]: I0127 16:36:41.135902 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-49fxh" event={"ID":"6dd740e1-b6ee-444d-b71e-e18d4837ef8a","Type":"ContainerStarted","Data":"63032a2c7b9511cf6dd4c145bad9bed20775a86fc21f664cdf358270e2e03ec9"} Jan 27 16:36:41 crc kubenswrapper[4772]: I0127 16:36:41.136216 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-74dnd" podUID="a3141ae0-d9c6-4eb9-8ada-8dd2454da297" containerName="registry-server" containerID="cri-o://ce0e7196a0a64497f42a021e97d6f099be19a6b7a8db931509a25e0302440168" gracePeriod=2 Jan 27 16:36:41 crc kubenswrapper[4772]: I0127 16:36:41.147347 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-fbd6-account-create-update-kvjsf" podStartSLOduration=2.147329898 podStartE2EDuration="2.147329898s" podCreationTimestamp="2026-01-27 16:36:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:36:41.147213485 +0000 UTC m=+5387.127822593" watchObservedRunningTime="2026-01-27 16:36:41.147329898 +0000 UTC m=+5387.127938996" Jan 27 16:36:41 crc kubenswrapper[4772]: I0127 16:36:41.169069 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-49fxh" podStartSLOduration=2.169050107 podStartE2EDuration="2.169050107s" podCreationTimestamp="2026-01-27 16:36:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:36:41.163271212 +0000 UTC m=+5387.143880300" watchObservedRunningTime="2026-01-27 16:36:41.169050107 +0000 UTC m=+5387.149659205" Jan 27 16:36:42 crc kubenswrapper[4772]: I0127 16:36:42.842074 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-74dnd" Jan 27 16:36:42 crc kubenswrapper[4772]: I0127 16:36:42.926843 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3141ae0-d9c6-4eb9-8ada-8dd2454da297-utilities\") pod \"a3141ae0-d9c6-4eb9-8ada-8dd2454da297\" (UID: \"a3141ae0-d9c6-4eb9-8ada-8dd2454da297\") " Jan 27 16:36:42 crc kubenswrapper[4772]: I0127 16:36:42.927199 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3141ae0-d9c6-4eb9-8ada-8dd2454da297-catalog-content\") pod \"a3141ae0-d9c6-4eb9-8ada-8dd2454da297\" (UID: \"a3141ae0-d9c6-4eb9-8ada-8dd2454da297\") " Jan 27 16:36:42 crc kubenswrapper[4772]: I0127 16:36:42.927309 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cr8tc\" (UniqueName: \"kubernetes.io/projected/a3141ae0-d9c6-4eb9-8ada-8dd2454da297-kube-api-access-cr8tc\") pod \"a3141ae0-d9c6-4eb9-8ada-8dd2454da297\" (UID: \"a3141ae0-d9c6-4eb9-8ada-8dd2454da297\") " Jan 27 16:36:42 crc kubenswrapper[4772]: I0127 16:36:42.927891 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3141ae0-d9c6-4eb9-8ada-8dd2454da297-utilities" (OuterVolumeSpecName: "utilities") pod "a3141ae0-d9c6-4eb9-8ada-8dd2454da297" (UID: "a3141ae0-d9c6-4eb9-8ada-8dd2454da297"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:36:42 crc kubenswrapper[4772]: I0127 16:36:42.934629 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3141ae0-d9c6-4eb9-8ada-8dd2454da297-kube-api-access-cr8tc" (OuterVolumeSpecName: "kube-api-access-cr8tc") pod "a3141ae0-d9c6-4eb9-8ada-8dd2454da297" (UID: "a3141ae0-d9c6-4eb9-8ada-8dd2454da297"). InnerVolumeSpecName "kube-api-access-cr8tc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:36:43 crc kubenswrapper[4772]: I0127 16:36:43.029782 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3141ae0-d9c6-4eb9-8ada-8dd2454da297-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 16:36:43 crc kubenswrapper[4772]: I0127 16:36:43.029825 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cr8tc\" (UniqueName: \"kubernetes.io/projected/a3141ae0-d9c6-4eb9-8ada-8dd2454da297-kube-api-access-cr8tc\") on node \"crc\" DevicePath \"\"" Jan 27 16:36:43 crc kubenswrapper[4772]: I0127 16:36:43.041529 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3141ae0-d9c6-4eb9-8ada-8dd2454da297-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a3141ae0-d9c6-4eb9-8ada-8dd2454da297" (UID: "a3141ae0-d9c6-4eb9-8ada-8dd2454da297"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:36:43 crc kubenswrapper[4772]: I0127 16:36:43.131716 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3141ae0-d9c6-4eb9-8ada-8dd2454da297-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 16:36:43 crc kubenswrapper[4772]: I0127 16:36:43.156070 4772 generic.go:334] "Generic (PLEG): container finished" podID="0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b" containerID="7820202ed4fe89d2fb42f752ab53ea23a8554f7e3772282ffa645c3878d8acde" exitCode=0 Jan 27 16:36:43 crc kubenswrapper[4772]: I0127 16:36:43.156143 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-fbd6-account-create-update-kvjsf" event={"ID":"0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b","Type":"ContainerDied","Data":"7820202ed4fe89d2fb42f752ab53ea23a8554f7e3772282ffa645c3878d8acde"} Jan 27 16:36:43 crc kubenswrapper[4772]: I0127 16:36:43.159306 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-74dnd" Jan 27 16:36:43 crc kubenswrapper[4772]: I0127 16:36:43.159348 4772 generic.go:334] "Generic (PLEG): container finished" podID="a3141ae0-d9c6-4eb9-8ada-8dd2454da297" containerID="ce0e7196a0a64497f42a021e97d6f099be19a6b7a8db931509a25e0302440168" exitCode=0 Jan 27 16:36:43 crc kubenswrapper[4772]: I0127 16:36:43.159425 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74dnd" event={"ID":"a3141ae0-d9c6-4eb9-8ada-8dd2454da297","Type":"ContainerDied","Data":"ce0e7196a0a64497f42a021e97d6f099be19a6b7a8db931509a25e0302440168"} Jan 27 16:36:43 crc kubenswrapper[4772]: I0127 16:36:43.159570 4772 scope.go:117] "RemoveContainer" containerID="ce0e7196a0a64497f42a021e97d6f099be19a6b7a8db931509a25e0302440168" Jan 27 16:36:43 crc kubenswrapper[4772]: I0127 16:36:43.159454 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74dnd" event={"ID":"a3141ae0-d9c6-4eb9-8ada-8dd2454da297","Type":"ContainerDied","Data":"edeeb734fb1a3ced8ebfb84b4f3f42b1d212bd81615ff209859424b50ce18695"} Jan 27 16:36:43 crc kubenswrapper[4772]: I0127 16:36:43.161299 4772 generic.go:334] "Generic (PLEG): container finished" podID="6dd740e1-b6ee-444d-b71e-e18d4837ef8a" containerID="0eb151bf9a1bfafe986b27f8d85f2a3fcbd2a1f6be731bcd2ce95359e6a6e136" exitCode=0 Jan 27 16:36:43 crc kubenswrapper[4772]: I0127 16:36:43.161332 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-49fxh" event={"ID":"6dd740e1-b6ee-444d-b71e-e18d4837ef8a","Type":"ContainerDied","Data":"0eb151bf9a1bfafe986b27f8d85f2a3fcbd2a1f6be731bcd2ce95359e6a6e136"} Jan 27 16:36:43 crc kubenswrapper[4772]: I0127 16:36:43.185598 4772 scope.go:117] "RemoveContainer" containerID="b27ff19e88ae10072e54a035b5eeff3472d319b04facf31b91170a2a03f287d0" Jan 27 16:36:43 crc kubenswrapper[4772]: I0127 16:36:43.203118 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-74dnd"] Jan 27 16:36:43 crc kubenswrapper[4772]: I0127 16:36:43.213696 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-74dnd"] Jan 27 16:36:43 crc kubenswrapper[4772]: I0127 16:36:43.221612 4772 scope.go:117] "RemoveContainer" containerID="4823b71d7a398dbe52d74d93991da82e8680383fcd12420daf8b869332d53333" Jan 27 16:36:43 crc kubenswrapper[4772]: I0127 16:36:43.246805 4772 scope.go:117] "RemoveContainer" containerID="ce0e7196a0a64497f42a021e97d6f099be19a6b7a8db931509a25e0302440168" Jan 27 16:36:43 crc kubenswrapper[4772]: E0127 16:36:43.247280 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce0e7196a0a64497f42a021e97d6f099be19a6b7a8db931509a25e0302440168\": container with ID starting with ce0e7196a0a64497f42a021e97d6f099be19a6b7a8db931509a25e0302440168 not found: ID does not exist" containerID="ce0e7196a0a64497f42a021e97d6f099be19a6b7a8db931509a25e0302440168" Jan 27 16:36:43 crc kubenswrapper[4772]: I0127 16:36:43.247343 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce0e7196a0a64497f42a021e97d6f099be19a6b7a8db931509a25e0302440168"} err="failed to get container status \"ce0e7196a0a64497f42a021e97d6f099be19a6b7a8db931509a25e0302440168\": rpc error: code = NotFound desc = could not find container \"ce0e7196a0a64497f42a021e97d6f099be19a6b7a8db931509a25e0302440168\": container with ID starting with ce0e7196a0a64497f42a021e97d6f099be19a6b7a8db931509a25e0302440168 not found: ID does not exist" Jan 27 16:36:43 crc kubenswrapper[4772]: I0127 16:36:43.247372 4772 scope.go:117] "RemoveContainer" containerID="b27ff19e88ae10072e54a035b5eeff3472d319b04facf31b91170a2a03f287d0" Jan 27 16:36:43 crc kubenswrapper[4772]: E0127 16:36:43.247692 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b27ff19e88ae10072e54a035b5eeff3472d319b04facf31b91170a2a03f287d0\": container with ID starting with b27ff19e88ae10072e54a035b5eeff3472d319b04facf31b91170a2a03f287d0 not found: ID does not exist" containerID="b27ff19e88ae10072e54a035b5eeff3472d319b04facf31b91170a2a03f287d0" Jan 27 16:36:43 crc kubenswrapper[4772]: I0127 16:36:43.247747 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b27ff19e88ae10072e54a035b5eeff3472d319b04facf31b91170a2a03f287d0"} err="failed to get container status \"b27ff19e88ae10072e54a035b5eeff3472d319b04facf31b91170a2a03f287d0\": rpc error: code = NotFound desc = could not find container \"b27ff19e88ae10072e54a035b5eeff3472d319b04facf31b91170a2a03f287d0\": container with ID starting with b27ff19e88ae10072e54a035b5eeff3472d319b04facf31b91170a2a03f287d0 not found: ID does not exist" Jan 27 16:36:43 crc kubenswrapper[4772]: I0127 16:36:43.247781 4772 scope.go:117] "RemoveContainer" containerID="4823b71d7a398dbe52d74d93991da82e8680383fcd12420daf8b869332d53333" Jan 27 16:36:43 crc kubenswrapper[4772]: E0127 16:36:43.248089 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4823b71d7a398dbe52d74d93991da82e8680383fcd12420daf8b869332d53333\": container with ID starting with 4823b71d7a398dbe52d74d93991da82e8680383fcd12420daf8b869332d53333 not found: ID does not exist" containerID="4823b71d7a398dbe52d74d93991da82e8680383fcd12420daf8b869332d53333" Jan 27 16:36:43 crc kubenswrapper[4772]: I0127 16:36:43.248132 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4823b71d7a398dbe52d74d93991da82e8680383fcd12420daf8b869332d53333"} err="failed to get container status \"4823b71d7a398dbe52d74d93991da82e8680383fcd12420daf8b869332d53333\": rpc error: code = NotFound desc = could not find container \"4823b71d7a398dbe52d74d93991da82e8680383fcd12420daf8b869332d53333\": container with ID starting with 4823b71d7a398dbe52d74d93991da82e8680383fcd12420daf8b869332d53333 not found: ID does not exist" Jan 27 16:36:44 crc kubenswrapper[4772]: I0127 16:36:44.050194 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-bxrqf"] Jan 27 16:36:44 crc kubenswrapper[4772]: I0127 16:36:44.057415 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-bxrqf"] Jan 27 16:36:44 crc kubenswrapper[4772]: I0127 16:36:44.522801 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-49fxh" Jan 27 16:36:44 crc kubenswrapper[4772]: I0127 16:36:44.528199 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-fbd6-account-create-update-kvjsf" Jan 27 16:36:44 crc kubenswrapper[4772]: I0127 16:36:44.660076 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dd740e1-b6ee-444d-b71e-e18d4837ef8a-operator-scripts\") pod \"6dd740e1-b6ee-444d-b71e-e18d4837ef8a\" (UID: \"6dd740e1-b6ee-444d-b71e-e18d4837ef8a\") " Jan 27 16:36:44 crc kubenswrapper[4772]: I0127 16:36:44.660223 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkkbh\" (UniqueName: \"kubernetes.io/projected/0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b-kube-api-access-wkkbh\") pod \"0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b\" (UID: \"0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b\") " Jan 27 16:36:44 crc kubenswrapper[4772]: I0127 16:36:44.660322 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hf2sr\" (UniqueName: \"kubernetes.io/projected/6dd740e1-b6ee-444d-b71e-e18d4837ef8a-kube-api-access-hf2sr\") pod \"6dd740e1-b6ee-444d-b71e-e18d4837ef8a\" (UID: \"6dd740e1-b6ee-444d-b71e-e18d4837ef8a\") " Jan 27 16:36:44 crc kubenswrapper[4772]: I0127 16:36:44.660399 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b-operator-scripts\") pod \"0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b\" (UID: \"0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b\") " Jan 27 16:36:44 crc kubenswrapper[4772]: I0127 16:36:44.660997 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6dd740e1-b6ee-444d-b71e-e18d4837ef8a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6dd740e1-b6ee-444d-b71e-e18d4837ef8a" (UID: "6dd740e1-b6ee-444d-b71e-e18d4837ef8a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:36:44 crc kubenswrapper[4772]: I0127 16:36:44.661054 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b" (UID: "0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:36:44 crc kubenswrapper[4772]: I0127 16:36:44.670448 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b-kube-api-access-wkkbh" (OuterVolumeSpecName: "kube-api-access-wkkbh") pod "0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b" (UID: "0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b"). InnerVolumeSpecName "kube-api-access-wkkbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:36:44 crc kubenswrapper[4772]: I0127 16:36:44.670498 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dd740e1-b6ee-444d-b71e-e18d4837ef8a-kube-api-access-hf2sr" (OuterVolumeSpecName: "kube-api-access-hf2sr") pod "6dd740e1-b6ee-444d-b71e-e18d4837ef8a" (UID: "6dd740e1-b6ee-444d-b71e-e18d4837ef8a"). InnerVolumeSpecName "kube-api-access-hf2sr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:36:44 crc kubenswrapper[4772]: I0127 16:36:44.674622 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3141ae0-d9c6-4eb9-8ada-8dd2454da297" path="/var/lib/kubelet/pods/a3141ae0-d9c6-4eb9-8ada-8dd2454da297/volumes" Jan 27 16:36:44 crc kubenswrapper[4772]: I0127 16:36:44.675499 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb558533-27c2-4249-9beb-e01d5b918c58" path="/var/lib/kubelet/pods/cb558533-27c2-4249-9beb-e01d5b918c58/volumes" Jan 27 16:36:44 crc kubenswrapper[4772]: I0127 16:36:44.762568 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:36:44 crc kubenswrapper[4772]: I0127 16:36:44.762598 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dd740e1-b6ee-444d-b71e-e18d4837ef8a-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:36:44 crc kubenswrapper[4772]: I0127 16:36:44.762609 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkkbh\" (UniqueName: \"kubernetes.io/projected/0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b-kube-api-access-wkkbh\") on node \"crc\" DevicePath \"\"" Jan 27 16:36:44 crc kubenswrapper[4772]: I0127 16:36:44.762618 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hf2sr\" (UniqueName: \"kubernetes.io/projected/6dd740e1-b6ee-444d-b71e-e18d4837ef8a-kube-api-access-hf2sr\") on node \"crc\" DevicePath \"\"" Jan 27 16:36:45 crc kubenswrapper[4772]: I0127 16:36:45.179404 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-fbd6-account-create-update-kvjsf" Jan 27 16:36:45 crc kubenswrapper[4772]: I0127 16:36:45.179402 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-fbd6-account-create-update-kvjsf" event={"ID":"0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b","Type":"ContainerDied","Data":"98049675b35bf9aa5f55e27edcefac02abeb383c73491edfaf8dc4f2d0cc9061"} Jan 27 16:36:45 crc kubenswrapper[4772]: I0127 16:36:45.179520 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98049675b35bf9aa5f55e27edcefac02abeb383c73491edfaf8dc4f2d0cc9061" Jan 27 16:36:45 crc kubenswrapper[4772]: I0127 16:36:45.181245 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-49fxh" event={"ID":"6dd740e1-b6ee-444d-b71e-e18d4837ef8a","Type":"ContainerDied","Data":"63032a2c7b9511cf6dd4c145bad9bed20775a86fc21f664cdf358270e2e03ec9"} Jan 27 16:36:45 crc kubenswrapper[4772]: I0127 16:36:45.181275 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63032a2c7b9511cf6dd4c145bad9bed20775a86fc21f664cdf358270e2e03ec9" Jan 27 16:36:45 crc kubenswrapper[4772]: I0127 16:36:45.181307 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-49fxh" Jan 27 16:36:50 crc kubenswrapper[4772]: I0127 16:36:50.254801 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-xvkmn"] Jan 27 16:36:50 crc kubenswrapper[4772]: E0127 16:36:50.256708 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3141ae0-d9c6-4eb9-8ada-8dd2454da297" containerName="extract-utilities" Jan 27 16:36:50 crc kubenswrapper[4772]: I0127 16:36:50.256801 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3141ae0-d9c6-4eb9-8ada-8dd2454da297" containerName="extract-utilities" Jan 27 16:36:50 crc kubenswrapper[4772]: E0127 16:36:50.256876 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3141ae0-d9c6-4eb9-8ada-8dd2454da297" containerName="extract-content" Jan 27 16:36:50 crc kubenswrapper[4772]: I0127 16:36:50.256928 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3141ae0-d9c6-4eb9-8ada-8dd2454da297" containerName="extract-content" Jan 27 16:36:50 crc kubenswrapper[4772]: E0127 16:36:50.256978 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3141ae0-d9c6-4eb9-8ada-8dd2454da297" containerName="registry-server" Jan 27 16:36:50 crc kubenswrapper[4772]: I0127 16:36:50.257027 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3141ae0-d9c6-4eb9-8ada-8dd2454da297" containerName="registry-server" Jan 27 16:36:50 crc kubenswrapper[4772]: E0127 16:36:50.257084 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b" containerName="mariadb-account-create-update" Jan 27 16:36:50 crc kubenswrapper[4772]: I0127 16:36:50.257132 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b" containerName="mariadb-account-create-update" Jan 27 16:36:50 crc kubenswrapper[4772]: E0127 16:36:50.257225 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dd740e1-b6ee-444d-b71e-e18d4837ef8a" containerName="mariadb-database-create" Jan 27 16:36:50 crc kubenswrapper[4772]: I0127 16:36:50.257295 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dd740e1-b6ee-444d-b71e-e18d4837ef8a" containerName="mariadb-database-create" Jan 27 16:36:50 crc kubenswrapper[4772]: I0127 16:36:50.257491 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b" containerName="mariadb-account-create-update" Jan 27 16:36:50 crc kubenswrapper[4772]: I0127 16:36:50.257571 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dd740e1-b6ee-444d-b71e-e18d4837ef8a" containerName="mariadb-database-create" Jan 27 16:36:50 crc kubenswrapper[4772]: I0127 16:36:50.257636 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3141ae0-d9c6-4eb9-8ada-8dd2454da297" containerName="registry-server" Jan 27 16:36:50 crc kubenswrapper[4772]: I0127 16:36:50.258277 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-xvkmn" Jan 27 16:36:50 crc kubenswrapper[4772]: I0127 16:36:50.260714 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Jan 27 16:36:50 crc kubenswrapper[4772]: I0127 16:36:50.260843 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-4g84q" Jan 27 16:36:50 crc kubenswrapper[4772]: I0127 16:36:50.261747 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Jan 27 16:36:50 crc kubenswrapper[4772]: I0127 16:36:50.268771 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-xvkmn"] Jan 27 16:36:50 crc kubenswrapper[4772]: I0127 16:36:50.353146 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2-config\") pod \"neutron-db-sync-xvkmn\" (UID: \"fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2\") " pod="openstack/neutron-db-sync-xvkmn" Jan 27 16:36:50 crc kubenswrapper[4772]: I0127 16:36:50.353793 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqvpw\" (UniqueName: \"kubernetes.io/projected/fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2-kube-api-access-tqvpw\") pod \"neutron-db-sync-xvkmn\" (UID: \"fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2\") " pod="openstack/neutron-db-sync-xvkmn" Jan 27 16:36:50 crc kubenswrapper[4772]: I0127 16:36:50.354082 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2-combined-ca-bundle\") pod \"neutron-db-sync-xvkmn\" (UID: \"fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2\") " pod="openstack/neutron-db-sync-xvkmn" Jan 27 16:36:50 crc kubenswrapper[4772]: I0127 16:36:50.455563 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2-config\") pod \"neutron-db-sync-xvkmn\" (UID: \"fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2\") " pod="openstack/neutron-db-sync-xvkmn" Jan 27 16:36:50 crc kubenswrapper[4772]: I0127 16:36:50.455617 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqvpw\" (UniqueName: \"kubernetes.io/projected/fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2-kube-api-access-tqvpw\") pod \"neutron-db-sync-xvkmn\" (UID: \"fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2\") " pod="openstack/neutron-db-sync-xvkmn" Jan 27 16:36:50 crc kubenswrapper[4772]: I0127 16:36:50.455750 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2-combined-ca-bundle\") pod \"neutron-db-sync-xvkmn\" (UID: \"fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2\") " pod="openstack/neutron-db-sync-xvkmn" Jan 27 16:36:50 crc kubenswrapper[4772]: I0127 16:36:50.464565 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2-combined-ca-bundle\") pod \"neutron-db-sync-xvkmn\" (UID: \"fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2\") " pod="openstack/neutron-db-sync-xvkmn" Jan 27 16:36:50 crc kubenswrapper[4772]: I0127 16:36:50.466922 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2-config\") pod \"neutron-db-sync-xvkmn\" (UID: \"fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2\") " pod="openstack/neutron-db-sync-xvkmn" Jan 27 16:36:50 crc kubenswrapper[4772]: I0127 16:36:50.472585 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqvpw\" (UniqueName: \"kubernetes.io/projected/fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2-kube-api-access-tqvpw\") pod \"neutron-db-sync-xvkmn\" (UID: \"fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2\") " pod="openstack/neutron-db-sync-xvkmn" Jan 27 16:36:50 crc kubenswrapper[4772]: I0127 16:36:50.583680 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-xvkmn" Jan 27 16:36:51 crc kubenswrapper[4772]: I0127 16:36:51.014804 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-xvkmn"] Jan 27 16:36:51 crc kubenswrapper[4772]: I0127 16:36:51.224982 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-xvkmn" event={"ID":"fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2","Type":"ContainerStarted","Data":"47404bfd4ce996befd817152654b651725e0f51b6b3b55ee5ee110296d25e0c8"} Jan 27 16:36:51 crc kubenswrapper[4772]: I0127 16:36:51.225037 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-xvkmn" event={"ID":"fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2","Type":"ContainerStarted","Data":"53062de60772780f98beb1805b741d9fbfee4d6d4c0d97ad27dba9fd1047fbfc"} Jan 27 16:36:51 crc kubenswrapper[4772]: I0127 16:36:51.249494 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-xvkmn" podStartSLOduration=1.249475248 podStartE2EDuration="1.249475248s" podCreationTimestamp="2026-01-27 16:36:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:36:51.24007387 +0000 UTC m=+5397.220682978" watchObservedRunningTime="2026-01-27 16:36:51.249475248 +0000 UTC m=+5397.230084346" Jan 27 16:36:56 crc kubenswrapper[4772]: I0127 16:36:56.276540 4772 generic.go:334] "Generic (PLEG): container finished" podID="fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2" containerID="47404bfd4ce996befd817152654b651725e0f51b6b3b55ee5ee110296d25e0c8" exitCode=0 Jan 27 16:36:56 crc kubenswrapper[4772]: I0127 16:36:56.276646 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-xvkmn" event={"ID":"fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2","Type":"ContainerDied","Data":"47404bfd4ce996befd817152654b651725e0f51b6b3b55ee5ee110296d25e0c8"} Jan 27 16:36:57 crc kubenswrapper[4772]: I0127 16:36:57.620356 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-xvkmn" Jan 27 16:36:57 crc kubenswrapper[4772]: I0127 16:36:57.706904 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2-config\") pod \"fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2\" (UID: \"fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2\") " Jan 27 16:36:57 crc kubenswrapper[4772]: I0127 16:36:57.707270 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqvpw\" (UniqueName: \"kubernetes.io/projected/fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2-kube-api-access-tqvpw\") pod \"fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2\" (UID: \"fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2\") " Jan 27 16:36:57 crc kubenswrapper[4772]: I0127 16:36:57.707320 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2-combined-ca-bundle\") pod \"fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2\" (UID: \"fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2\") " Jan 27 16:36:57 crc kubenswrapper[4772]: I0127 16:36:57.712367 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2-kube-api-access-tqvpw" (OuterVolumeSpecName: "kube-api-access-tqvpw") pod "fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2" (UID: "fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2"). InnerVolumeSpecName "kube-api-access-tqvpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:36:57 crc kubenswrapper[4772]: I0127 16:36:57.728468 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2" (UID: "fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:36:57 crc kubenswrapper[4772]: I0127 16:36:57.733679 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2-config" (OuterVolumeSpecName: "config") pod "fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2" (UID: "fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:36:57 crc kubenswrapper[4772]: I0127 16:36:57.809470 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqvpw\" (UniqueName: \"kubernetes.io/projected/fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2-kube-api-access-tqvpw\") on node \"crc\" DevicePath \"\"" Jan 27 16:36:57 crc kubenswrapper[4772]: I0127 16:36:57.809507 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:36:57 crc kubenswrapper[4772]: I0127 16:36:57.809519 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2-config\") on node \"crc\" DevicePath \"\"" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.292617 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-xvkmn" event={"ID":"fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2","Type":"ContainerDied","Data":"53062de60772780f98beb1805b741d9fbfee4d6d4c0d97ad27dba9fd1047fbfc"} Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.292973 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53062de60772780f98beb1805b741d9fbfee4d6d4c0d97ad27dba9fd1047fbfc" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.292687 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-xvkmn" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.533093 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ccbf7777-mh9xn"] Jan 27 16:36:58 crc kubenswrapper[4772]: E0127 16:36:58.533533 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2" containerName="neutron-db-sync" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.533557 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2" containerName="neutron-db-sync" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.533805 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2" containerName="neutron-db-sync" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.534937 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.547152 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ccbf7777-mh9xn"] Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.626262 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-ovsdbserver-sb\") pod \"dnsmasq-dns-85ccbf7777-mh9xn\" (UID: \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\") " pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.626399 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-ovsdbserver-nb\") pod \"dnsmasq-dns-85ccbf7777-mh9xn\" (UID: \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\") " pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.626438 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sl6sk\" (UniqueName: \"kubernetes.io/projected/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-kube-api-access-sl6sk\") pod \"dnsmasq-dns-85ccbf7777-mh9xn\" (UID: \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\") " pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.626455 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-config\") pod \"dnsmasq-dns-85ccbf7777-mh9xn\" (UID: \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\") " pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.626473 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-dns-svc\") pod \"dnsmasq-dns-85ccbf7777-mh9xn\" (UID: \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\") " pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.637378 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-8bf5d4b7c-bfg78"] Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.639376 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8bf5d4b7c-bfg78" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.645001 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.646453 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-4g84q" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.646575 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.676978 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8bf5d4b7c-bfg78"] Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.727901 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rj5m\" (UniqueName: \"kubernetes.io/projected/b5a89957-107d-449b-b438-2215fd4ed522-kube-api-access-8rj5m\") pod \"neutron-8bf5d4b7c-bfg78\" (UID: \"b5a89957-107d-449b-b438-2215fd4ed522\") " pod="openstack/neutron-8bf5d4b7c-bfg78" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.727990 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5a89957-107d-449b-b438-2215fd4ed522-combined-ca-bundle\") pod \"neutron-8bf5d4b7c-bfg78\" (UID: \"b5a89957-107d-449b-b438-2215fd4ed522\") " pod="openstack/neutron-8bf5d4b7c-bfg78" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.728111 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b5a89957-107d-449b-b438-2215fd4ed522-config\") pod \"neutron-8bf5d4b7c-bfg78\" (UID: \"b5a89957-107d-449b-b438-2215fd4ed522\") " pod="openstack/neutron-8bf5d4b7c-bfg78" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.728136 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b5a89957-107d-449b-b438-2215fd4ed522-httpd-config\") pod \"neutron-8bf5d4b7c-bfg78\" (UID: \"b5a89957-107d-449b-b438-2215fd4ed522\") " pod="openstack/neutron-8bf5d4b7c-bfg78" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.728284 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-ovsdbserver-nb\") pod \"dnsmasq-dns-85ccbf7777-mh9xn\" (UID: \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\") " pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.728342 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sl6sk\" (UniqueName: \"kubernetes.io/projected/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-kube-api-access-sl6sk\") pod \"dnsmasq-dns-85ccbf7777-mh9xn\" (UID: \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\") " pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.728368 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-config\") pod \"dnsmasq-dns-85ccbf7777-mh9xn\" (UID: \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\") " pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.728391 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-dns-svc\") pod \"dnsmasq-dns-85ccbf7777-mh9xn\" (UID: \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\") " pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.728420 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-ovsdbserver-sb\") pod \"dnsmasq-dns-85ccbf7777-mh9xn\" (UID: \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\") " pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.729322 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-ovsdbserver-nb\") pod \"dnsmasq-dns-85ccbf7777-mh9xn\" (UID: \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\") " pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.729389 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-config\") pod \"dnsmasq-dns-85ccbf7777-mh9xn\" (UID: \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\") " pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.729425 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-ovsdbserver-sb\") pod \"dnsmasq-dns-85ccbf7777-mh9xn\" (UID: \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\") " pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.729476 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-dns-svc\") pod \"dnsmasq-dns-85ccbf7777-mh9xn\" (UID: \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\") " pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.746936 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sl6sk\" (UniqueName: \"kubernetes.io/projected/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-kube-api-access-sl6sk\") pod \"dnsmasq-dns-85ccbf7777-mh9xn\" (UID: \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\") " pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.831447 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rj5m\" (UniqueName: \"kubernetes.io/projected/b5a89957-107d-449b-b438-2215fd4ed522-kube-api-access-8rj5m\") pod \"neutron-8bf5d4b7c-bfg78\" (UID: \"b5a89957-107d-449b-b438-2215fd4ed522\") " pod="openstack/neutron-8bf5d4b7c-bfg78" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.831704 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5a89957-107d-449b-b438-2215fd4ed522-combined-ca-bundle\") pod \"neutron-8bf5d4b7c-bfg78\" (UID: \"b5a89957-107d-449b-b438-2215fd4ed522\") " pod="openstack/neutron-8bf5d4b7c-bfg78" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.831816 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b5a89957-107d-449b-b438-2215fd4ed522-config\") pod \"neutron-8bf5d4b7c-bfg78\" (UID: \"b5a89957-107d-449b-b438-2215fd4ed522\") " pod="openstack/neutron-8bf5d4b7c-bfg78" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.831870 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b5a89957-107d-449b-b438-2215fd4ed522-httpd-config\") pod \"neutron-8bf5d4b7c-bfg78\" (UID: \"b5a89957-107d-449b-b438-2215fd4ed522\") " pod="openstack/neutron-8bf5d4b7c-bfg78" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.835884 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b5a89957-107d-449b-b438-2215fd4ed522-config\") pod \"neutron-8bf5d4b7c-bfg78\" (UID: \"b5a89957-107d-449b-b438-2215fd4ed522\") " pod="openstack/neutron-8bf5d4b7c-bfg78" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.835921 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5a89957-107d-449b-b438-2215fd4ed522-combined-ca-bundle\") pod \"neutron-8bf5d4b7c-bfg78\" (UID: \"b5a89957-107d-449b-b438-2215fd4ed522\") " pod="openstack/neutron-8bf5d4b7c-bfg78" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.840184 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b5a89957-107d-449b-b438-2215fd4ed522-httpd-config\") pod \"neutron-8bf5d4b7c-bfg78\" (UID: \"b5a89957-107d-449b-b438-2215fd4ed522\") " pod="openstack/neutron-8bf5d4b7c-bfg78" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.850872 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rj5m\" (UniqueName: \"kubernetes.io/projected/b5a89957-107d-449b-b438-2215fd4ed522-kube-api-access-8rj5m\") pod \"neutron-8bf5d4b7c-bfg78\" (UID: \"b5a89957-107d-449b-b438-2215fd4ed522\") " pod="openstack/neutron-8bf5d4b7c-bfg78" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.864834 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" Jan 27 16:36:58 crc kubenswrapper[4772]: I0127 16:36:58.958630 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8bf5d4b7c-bfg78" Jan 27 16:36:59 crc kubenswrapper[4772]: I0127 16:36:59.374258 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ccbf7777-mh9xn"] Jan 27 16:36:59 crc kubenswrapper[4772]: W0127 16:36:59.381677 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6449727_ae23_4ae0_b6e6_4c1cef43ef53.slice/crio-0286cc31499d6fb5d457987c1e11aad0e7f2bc6e4ccd62d6d0d7e8570d1f3e2e WatchSource:0}: Error finding container 0286cc31499d6fb5d457987c1e11aad0e7f2bc6e4ccd62d6d0d7e8570d1f3e2e: Status 404 returned error can't find the container with id 0286cc31499d6fb5d457987c1e11aad0e7f2bc6e4ccd62d6d0d7e8570d1f3e2e Jan 27 16:36:59 crc kubenswrapper[4772]: I0127 16:36:59.523071 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8bf5d4b7c-bfg78"] Jan 27 16:36:59 crc kubenswrapper[4772]: W0127 16:36:59.525405 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5a89957_107d_449b_b438_2215fd4ed522.slice/crio-b962fcddf62956fb0c0d0b689222c008baa58e3667e11f9643a61f281a7a1dba WatchSource:0}: Error finding container b962fcddf62956fb0c0d0b689222c008baa58e3667e11f9643a61f281a7a1dba: Status 404 returned error can't find the container with id b962fcddf62956fb0c0d0b689222c008baa58e3667e11f9643a61f281a7a1dba Jan 27 16:37:00 crc kubenswrapper[4772]: I0127 16:37:00.306764 4772 generic.go:334] "Generic (PLEG): container finished" podID="b6449727-ae23-4ae0-b6e6-4c1cef43ef53" containerID="71552776ddecc62c8f2d51f59b6ec4b6ba66501bb9fcfb2256ecea66da9231d2" exitCode=0 Jan 27 16:37:00 crc kubenswrapper[4772]: I0127 16:37:00.306870 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" event={"ID":"b6449727-ae23-4ae0-b6e6-4c1cef43ef53","Type":"ContainerDied","Data":"71552776ddecc62c8f2d51f59b6ec4b6ba66501bb9fcfb2256ecea66da9231d2"} Jan 27 16:37:00 crc kubenswrapper[4772]: I0127 16:37:00.307067 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" event={"ID":"b6449727-ae23-4ae0-b6e6-4c1cef43ef53","Type":"ContainerStarted","Data":"0286cc31499d6fb5d457987c1e11aad0e7f2bc6e4ccd62d6d0d7e8570d1f3e2e"} Jan 27 16:37:00 crc kubenswrapper[4772]: I0127 16:37:00.309936 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8bf5d4b7c-bfg78" event={"ID":"b5a89957-107d-449b-b438-2215fd4ed522","Type":"ContainerStarted","Data":"e7a293400012c2920a2a67640d49f3635b735a4b47247efa0afa1b93866b0de5"} Jan 27 16:37:00 crc kubenswrapper[4772]: I0127 16:37:00.309976 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8bf5d4b7c-bfg78" event={"ID":"b5a89957-107d-449b-b438-2215fd4ed522","Type":"ContainerStarted","Data":"414006ec43fc2848a611209d333ad38eb189a587359c2196aa3950825367faba"} Jan 27 16:37:00 crc kubenswrapper[4772]: I0127 16:37:00.309987 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8bf5d4b7c-bfg78" event={"ID":"b5a89957-107d-449b-b438-2215fd4ed522","Type":"ContainerStarted","Data":"b962fcddf62956fb0c0d0b689222c008baa58e3667e11f9643a61f281a7a1dba"} Jan 27 16:37:00 crc kubenswrapper[4772]: I0127 16:37:00.310107 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-8bf5d4b7c-bfg78" Jan 27 16:37:01 crc kubenswrapper[4772]: I0127 16:37:01.323098 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" event={"ID":"b6449727-ae23-4ae0-b6e6-4c1cef43ef53","Type":"ContainerStarted","Data":"f3387679ac491bfa95cee52f7df561b5ae6817e29d93d8687efd8b7b43af3938"} Jan 27 16:37:01 crc kubenswrapper[4772]: I0127 16:37:01.323569 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" Jan 27 16:37:01 crc kubenswrapper[4772]: I0127 16:37:01.352741 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-8bf5d4b7c-bfg78" podStartSLOduration=3.352718699 podStartE2EDuration="3.352718699s" podCreationTimestamp="2026-01-27 16:36:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:37:00.360755504 +0000 UTC m=+5406.341364602" watchObservedRunningTime="2026-01-27 16:37:01.352718699 +0000 UTC m=+5407.333327797" Jan 27 16:37:06 crc kubenswrapper[4772]: I0127 16:37:06.838784 4772 scope.go:117] "RemoveContainer" containerID="65777ccce3cb931b879ebb264390f1a957ffbebf2e9446690c38c79d1e3ddb7c" Jan 27 16:37:08 crc kubenswrapper[4772]: I0127 16:37:08.867140 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" Jan 27 16:37:08 crc kubenswrapper[4772]: I0127 16:37:08.898598 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" podStartSLOduration=10.898579426 podStartE2EDuration="10.898579426s" podCreationTimestamp="2026-01-27 16:36:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:37:01.346933974 +0000 UTC m=+5407.327543072" watchObservedRunningTime="2026-01-27 16:37:08.898579426 +0000 UTC m=+5414.879188524" Jan 27 16:37:08 crc kubenswrapper[4772]: I0127 16:37:08.944731 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55d7c89dc7-g6rsn"] Jan 27 16:37:08 crc kubenswrapper[4772]: I0127 16:37:08.945079 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" podUID="53c70df7-cdca-4296-af71-b5b002484575" containerName="dnsmasq-dns" containerID="cri-o://6271d29d3114164d534a9aed9b1a74dcb46dca81dd4e99dc8ee2c848b0770a70" gracePeriod=10 Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.384746 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.388270 4772 generic.go:334] "Generic (PLEG): container finished" podID="53c70df7-cdca-4296-af71-b5b002484575" containerID="6271d29d3114164d534a9aed9b1a74dcb46dca81dd4e99dc8ee2c848b0770a70" exitCode=0 Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.388312 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.388315 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" event={"ID":"53c70df7-cdca-4296-af71-b5b002484575","Type":"ContainerDied","Data":"6271d29d3114164d534a9aed9b1a74dcb46dca81dd4e99dc8ee2c848b0770a70"} Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.388485 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55d7c89dc7-g6rsn" event={"ID":"53c70df7-cdca-4296-af71-b5b002484575","Type":"ContainerDied","Data":"ab1dfeb7c7875a02f8dfd72be8da51340b47ef61cc5646a975b8f955d01b7ae0"} Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.388530 4772 scope.go:117] "RemoveContainer" containerID="6271d29d3114164d534a9aed9b1a74dcb46dca81dd4e99dc8ee2c848b0770a70" Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.409971 4772 scope.go:117] "RemoveContainer" containerID="d7ad019f98bbdfacb03703ed4de276fec18a82aa27109f1c0ec2b92483868eba" Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.461229 4772 scope.go:117] "RemoveContainer" containerID="6271d29d3114164d534a9aed9b1a74dcb46dca81dd4e99dc8ee2c848b0770a70" Jan 27 16:37:09 crc kubenswrapper[4772]: E0127 16:37:09.465860 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6271d29d3114164d534a9aed9b1a74dcb46dca81dd4e99dc8ee2c848b0770a70\": container with ID starting with 6271d29d3114164d534a9aed9b1a74dcb46dca81dd4e99dc8ee2c848b0770a70 not found: ID does not exist" containerID="6271d29d3114164d534a9aed9b1a74dcb46dca81dd4e99dc8ee2c848b0770a70" Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.465912 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6271d29d3114164d534a9aed9b1a74dcb46dca81dd4e99dc8ee2c848b0770a70"} err="failed to get container status \"6271d29d3114164d534a9aed9b1a74dcb46dca81dd4e99dc8ee2c848b0770a70\": rpc error: code = NotFound desc = could not find container \"6271d29d3114164d534a9aed9b1a74dcb46dca81dd4e99dc8ee2c848b0770a70\": container with ID starting with 6271d29d3114164d534a9aed9b1a74dcb46dca81dd4e99dc8ee2c848b0770a70 not found: ID does not exist" Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.465940 4772 scope.go:117] "RemoveContainer" containerID="d7ad019f98bbdfacb03703ed4de276fec18a82aa27109f1c0ec2b92483868eba" Jan 27 16:37:09 crc kubenswrapper[4772]: E0127 16:37:09.467411 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7ad019f98bbdfacb03703ed4de276fec18a82aa27109f1c0ec2b92483868eba\": container with ID starting with d7ad019f98bbdfacb03703ed4de276fec18a82aa27109f1c0ec2b92483868eba not found: ID does not exist" containerID="d7ad019f98bbdfacb03703ed4de276fec18a82aa27109f1c0ec2b92483868eba" Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.467467 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7ad019f98bbdfacb03703ed4de276fec18a82aa27109f1c0ec2b92483868eba"} err="failed to get container status \"d7ad019f98bbdfacb03703ed4de276fec18a82aa27109f1c0ec2b92483868eba\": rpc error: code = NotFound desc = could not find container \"d7ad019f98bbdfacb03703ed4de276fec18a82aa27109f1c0ec2b92483868eba\": container with ID starting with d7ad019f98bbdfacb03703ed4de276fec18a82aa27109f1c0ec2b92483868eba not found: ID does not exist" Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.529894 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53c70df7-cdca-4296-af71-b5b002484575-ovsdbserver-nb\") pod \"53c70df7-cdca-4296-af71-b5b002484575\" (UID: \"53c70df7-cdca-4296-af71-b5b002484575\") " Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.530048 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53c70df7-cdca-4296-af71-b5b002484575-ovsdbserver-sb\") pod \"53c70df7-cdca-4296-af71-b5b002484575\" (UID: \"53c70df7-cdca-4296-af71-b5b002484575\") " Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.530096 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwtxg\" (UniqueName: \"kubernetes.io/projected/53c70df7-cdca-4296-af71-b5b002484575-kube-api-access-qwtxg\") pod \"53c70df7-cdca-4296-af71-b5b002484575\" (UID: \"53c70df7-cdca-4296-af71-b5b002484575\") " Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.530149 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53c70df7-cdca-4296-af71-b5b002484575-config\") pod \"53c70df7-cdca-4296-af71-b5b002484575\" (UID: \"53c70df7-cdca-4296-af71-b5b002484575\") " Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.530200 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53c70df7-cdca-4296-af71-b5b002484575-dns-svc\") pod \"53c70df7-cdca-4296-af71-b5b002484575\" (UID: \"53c70df7-cdca-4296-af71-b5b002484575\") " Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.539127 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53c70df7-cdca-4296-af71-b5b002484575-kube-api-access-qwtxg" (OuterVolumeSpecName: "kube-api-access-qwtxg") pod "53c70df7-cdca-4296-af71-b5b002484575" (UID: "53c70df7-cdca-4296-af71-b5b002484575"). InnerVolumeSpecName "kube-api-access-qwtxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.570895 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53c70df7-cdca-4296-af71-b5b002484575-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "53c70df7-cdca-4296-af71-b5b002484575" (UID: "53c70df7-cdca-4296-af71-b5b002484575"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.579904 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53c70df7-cdca-4296-af71-b5b002484575-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "53c70df7-cdca-4296-af71-b5b002484575" (UID: "53c70df7-cdca-4296-af71-b5b002484575"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.580673 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53c70df7-cdca-4296-af71-b5b002484575-config" (OuterVolumeSpecName: "config") pod "53c70df7-cdca-4296-af71-b5b002484575" (UID: "53c70df7-cdca-4296-af71-b5b002484575"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.583060 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53c70df7-cdca-4296-af71-b5b002484575-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "53c70df7-cdca-4296-af71-b5b002484575" (UID: "53c70df7-cdca-4296-af71-b5b002484575"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.632345 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53c70df7-cdca-4296-af71-b5b002484575-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.632377 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53c70df7-cdca-4296-af71-b5b002484575-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.632388 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwtxg\" (UniqueName: \"kubernetes.io/projected/53c70df7-cdca-4296-af71-b5b002484575-kube-api-access-qwtxg\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.632399 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53c70df7-cdca-4296-af71-b5b002484575-config\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.632409 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53c70df7-cdca-4296-af71-b5b002484575-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.719374 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55d7c89dc7-g6rsn"] Jan 27 16:37:09 crc kubenswrapper[4772]: I0127 16:37:09.725592 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55d7c89dc7-g6rsn"] Jan 27 16:37:10 crc kubenswrapper[4772]: I0127 16:37:10.680730 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53c70df7-cdca-4296-af71-b5b002484575" path="/var/lib/kubelet/pods/53c70df7-cdca-4296-af71-b5b002484575/volumes" Jan 27 16:37:12 crc kubenswrapper[4772]: I0127 16:37:12.058504 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:37:12 crc kubenswrapper[4772]: I0127 16:37:12.058932 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:37:28 crc kubenswrapper[4772]: I0127 16:37:28.968037 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-8bf5d4b7c-bfg78" Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.082464 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-ng4h8"] Jan 27 16:37:36 crc kubenswrapper[4772]: E0127 16:37:36.083472 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53c70df7-cdca-4296-af71-b5b002484575" containerName="dnsmasq-dns" Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.083490 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="53c70df7-cdca-4296-af71-b5b002484575" containerName="dnsmasq-dns" Jan 27 16:37:36 crc kubenswrapper[4772]: E0127 16:37:36.083502 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53c70df7-cdca-4296-af71-b5b002484575" containerName="init" Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.083509 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="53c70df7-cdca-4296-af71-b5b002484575" containerName="init" Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.083709 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="53c70df7-cdca-4296-af71-b5b002484575" containerName="dnsmasq-dns" Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.084467 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ng4h8" Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.092563 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-ng4h8"] Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.169020 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-ab59-account-create-update-p9dsf"] Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.170475 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ab59-account-create-update-p9dsf" Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.174488 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.186085 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-ab59-account-create-update-p9dsf"] Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.197017 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mzj8\" (UniqueName: \"kubernetes.io/projected/2b150c73-f6eb-4193-81ef-84941ff1abef-kube-api-access-6mzj8\") pod \"glance-db-create-ng4h8\" (UID: \"2b150c73-f6eb-4193-81ef-84941ff1abef\") " pod="openstack/glance-db-create-ng4h8" Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.197473 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b150c73-f6eb-4193-81ef-84941ff1abef-operator-scripts\") pod \"glance-db-create-ng4h8\" (UID: \"2b150c73-f6eb-4193-81ef-84941ff1abef\") " pod="openstack/glance-db-create-ng4h8" Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.298939 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b150c73-f6eb-4193-81ef-84941ff1abef-operator-scripts\") pod \"glance-db-create-ng4h8\" (UID: \"2b150c73-f6eb-4193-81ef-84941ff1abef\") " pod="openstack/glance-db-create-ng4h8" Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.299041 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/615446cc-6fba-46f2-aad9-434f11519be9-operator-scripts\") pod \"glance-ab59-account-create-update-p9dsf\" (UID: \"615446cc-6fba-46f2-aad9-434f11519be9\") " pod="openstack/glance-ab59-account-create-update-p9dsf" Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.299078 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mzj8\" (UniqueName: \"kubernetes.io/projected/2b150c73-f6eb-4193-81ef-84941ff1abef-kube-api-access-6mzj8\") pod \"glance-db-create-ng4h8\" (UID: \"2b150c73-f6eb-4193-81ef-84941ff1abef\") " pod="openstack/glance-db-create-ng4h8" Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.299245 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5s7w\" (UniqueName: \"kubernetes.io/projected/615446cc-6fba-46f2-aad9-434f11519be9-kube-api-access-l5s7w\") pod \"glance-ab59-account-create-update-p9dsf\" (UID: \"615446cc-6fba-46f2-aad9-434f11519be9\") " pod="openstack/glance-ab59-account-create-update-p9dsf" Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.299679 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b150c73-f6eb-4193-81ef-84941ff1abef-operator-scripts\") pod \"glance-db-create-ng4h8\" (UID: \"2b150c73-f6eb-4193-81ef-84941ff1abef\") " pod="openstack/glance-db-create-ng4h8" Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.318303 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mzj8\" (UniqueName: \"kubernetes.io/projected/2b150c73-f6eb-4193-81ef-84941ff1abef-kube-api-access-6mzj8\") pod \"glance-db-create-ng4h8\" (UID: \"2b150c73-f6eb-4193-81ef-84941ff1abef\") " pod="openstack/glance-db-create-ng4h8" Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.401119 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/615446cc-6fba-46f2-aad9-434f11519be9-operator-scripts\") pod \"glance-ab59-account-create-update-p9dsf\" (UID: \"615446cc-6fba-46f2-aad9-434f11519be9\") " pod="openstack/glance-ab59-account-create-update-p9dsf" Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.401226 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5s7w\" (UniqueName: \"kubernetes.io/projected/615446cc-6fba-46f2-aad9-434f11519be9-kube-api-access-l5s7w\") pod \"glance-ab59-account-create-update-p9dsf\" (UID: \"615446cc-6fba-46f2-aad9-434f11519be9\") " pod="openstack/glance-ab59-account-create-update-p9dsf" Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.402331 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/615446cc-6fba-46f2-aad9-434f11519be9-operator-scripts\") pod \"glance-ab59-account-create-update-p9dsf\" (UID: \"615446cc-6fba-46f2-aad9-434f11519be9\") " pod="openstack/glance-ab59-account-create-update-p9dsf" Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.412819 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ng4h8" Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.419584 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5s7w\" (UniqueName: \"kubernetes.io/projected/615446cc-6fba-46f2-aad9-434f11519be9-kube-api-access-l5s7w\") pod \"glance-ab59-account-create-update-p9dsf\" (UID: \"615446cc-6fba-46f2-aad9-434f11519be9\") " pod="openstack/glance-ab59-account-create-update-p9dsf" Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.485135 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ab59-account-create-update-p9dsf" Jan 27 16:37:36 crc kubenswrapper[4772]: I0127 16:37:36.897781 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-ng4h8"] Jan 27 16:37:37 crc kubenswrapper[4772]: I0127 16:37:37.018092 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-ab59-account-create-update-p9dsf"] Jan 27 16:37:37 crc kubenswrapper[4772]: W0127 16:37:37.032976 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod615446cc_6fba_46f2_aad9_434f11519be9.slice/crio-f776daf862a05563cb80d9a9a1825c7258a4908f0d54212641976908fafb9c63 WatchSource:0}: Error finding container f776daf862a05563cb80d9a9a1825c7258a4908f0d54212641976908fafb9c63: Status 404 returned error can't find the container with id f776daf862a05563cb80d9a9a1825c7258a4908f0d54212641976908fafb9c63 Jan 27 16:37:37 crc kubenswrapper[4772]: I0127 16:37:37.625935 4772 generic.go:334] "Generic (PLEG): container finished" podID="615446cc-6fba-46f2-aad9-434f11519be9" containerID="04b6871605c5a7ec7b4197615c41edaa0e9453b396fcff037ce590df9243c6a0" exitCode=0 Jan 27 16:37:37 crc kubenswrapper[4772]: I0127 16:37:37.626318 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ab59-account-create-update-p9dsf" event={"ID":"615446cc-6fba-46f2-aad9-434f11519be9","Type":"ContainerDied","Data":"04b6871605c5a7ec7b4197615c41edaa0e9453b396fcff037ce590df9243c6a0"} Jan 27 16:37:37 crc kubenswrapper[4772]: I0127 16:37:37.626839 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ab59-account-create-update-p9dsf" event={"ID":"615446cc-6fba-46f2-aad9-434f11519be9","Type":"ContainerStarted","Data":"f776daf862a05563cb80d9a9a1825c7258a4908f0d54212641976908fafb9c63"} Jan 27 16:37:37 crc kubenswrapper[4772]: I0127 16:37:37.629355 4772 generic.go:334] "Generic (PLEG): container finished" podID="2b150c73-f6eb-4193-81ef-84941ff1abef" containerID="39b253bea060fd0ef46113186002866f36924e03fe711afec1f871af51f40edf" exitCode=0 Jan 27 16:37:37 crc kubenswrapper[4772]: I0127 16:37:37.629397 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ng4h8" event={"ID":"2b150c73-f6eb-4193-81ef-84941ff1abef","Type":"ContainerDied","Data":"39b253bea060fd0ef46113186002866f36924e03fe711afec1f871af51f40edf"} Jan 27 16:37:37 crc kubenswrapper[4772]: I0127 16:37:37.629421 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ng4h8" event={"ID":"2b150c73-f6eb-4193-81ef-84941ff1abef","Type":"ContainerStarted","Data":"c88e5ee63c5b143c29f95eb7a9163107e38a5ec7e21af15fb8f049bff95deca3"} Jan 27 16:37:39 crc kubenswrapper[4772]: I0127 16:37:39.006091 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ab59-account-create-update-p9dsf" Jan 27 16:37:39 crc kubenswrapper[4772]: I0127 16:37:39.013100 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ng4h8" Jan 27 16:37:39 crc kubenswrapper[4772]: I0127 16:37:39.040237 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mzj8\" (UniqueName: \"kubernetes.io/projected/2b150c73-f6eb-4193-81ef-84941ff1abef-kube-api-access-6mzj8\") pod \"2b150c73-f6eb-4193-81ef-84941ff1abef\" (UID: \"2b150c73-f6eb-4193-81ef-84941ff1abef\") " Jan 27 16:37:39 crc kubenswrapper[4772]: I0127 16:37:39.040343 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/615446cc-6fba-46f2-aad9-434f11519be9-operator-scripts\") pod \"615446cc-6fba-46f2-aad9-434f11519be9\" (UID: \"615446cc-6fba-46f2-aad9-434f11519be9\") " Jan 27 16:37:39 crc kubenswrapper[4772]: I0127 16:37:39.040461 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b150c73-f6eb-4193-81ef-84941ff1abef-operator-scripts\") pod \"2b150c73-f6eb-4193-81ef-84941ff1abef\" (UID: \"2b150c73-f6eb-4193-81ef-84941ff1abef\") " Jan 27 16:37:39 crc kubenswrapper[4772]: I0127 16:37:39.040621 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5s7w\" (UniqueName: \"kubernetes.io/projected/615446cc-6fba-46f2-aad9-434f11519be9-kube-api-access-l5s7w\") pod \"615446cc-6fba-46f2-aad9-434f11519be9\" (UID: \"615446cc-6fba-46f2-aad9-434f11519be9\") " Jan 27 16:37:39 crc kubenswrapper[4772]: I0127 16:37:39.043104 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/615446cc-6fba-46f2-aad9-434f11519be9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "615446cc-6fba-46f2-aad9-434f11519be9" (UID: "615446cc-6fba-46f2-aad9-434f11519be9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:37:39 crc kubenswrapper[4772]: I0127 16:37:39.043968 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b150c73-f6eb-4193-81ef-84941ff1abef-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2b150c73-f6eb-4193-81ef-84941ff1abef" (UID: "2b150c73-f6eb-4193-81ef-84941ff1abef"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:37:39 crc kubenswrapper[4772]: I0127 16:37:39.055939 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b150c73-f6eb-4193-81ef-84941ff1abef-kube-api-access-6mzj8" (OuterVolumeSpecName: "kube-api-access-6mzj8") pod "2b150c73-f6eb-4193-81ef-84941ff1abef" (UID: "2b150c73-f6eb-4193-81ef-84941ff1abef"). InnerVolumeSpecName "kube-api-access-6mzj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:37:39 crc kubenswrapper[4772]: I0127 16:37:39.058868 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/615446cc-6fba-46f2-aad9-434f11519be9-kube-api-access-l5s7w" (OuterVolumeSpecName: "kube-api-access-l5s7w") pod "615446cc-6fba-46f2-aad9-434f11519be9" (UID: "615446cc-6fba-46f2-aad9-434f11519be9"). InnerVolumeSpecName "kube-api-access-l5s7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:37:39 crc kubenswrapper[4772]: I0127 16:37:39.142935 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b150c73-f6eb-4193-81ef-84941ff1abef-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:39 crc kubenswrapper[4772]: I0127 16:37:39.142975 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5s7w\" (UniqueName: \"kubernetes.io/projected/615446cc-6fba-46f2-aad9-434f11519be9-kube-api-access-l5s7w\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:39 crc kubenswrapper[4772]: I0127 16:37:39.142992 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mzj8\" (UniqueName: \"kubernetes.io/projected/2b150c73-f6eb-4193-81ef-84941ff1abef-kube-api-access-6mzj8\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:39 crc kubenswrapper[4772]: I0127 16:37:39.143000 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/615446cc-6fba-46f2-aad9-434f11519be9-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:39 crc kubenswrapper[4772]: I0127 16:37:39.645072 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ab59-account-create-update-p9dsf" event={"ID":"615446cc-6fba-46f2-aad9-434f11519be9","Type":"ContainerDied","Data":"f776daf862a05563cb80d9a9a1825c7258a4908f0d54212641976908fafb9c63"} Jan 27 16:37:39 crc kubenswrapper[4772]: I0127 16:37:39.645119 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f776daf862a05563cb80d9a9a1825c7258a4908f0d54212641976908fafb9c63" Jan 27 16:37:39 crc kubenswrapper[4772]: I0127 16:37:39.645095 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ab59-account-create-update-p9dsf" Jan 27 16:37:39 crc kubenswrapper[4772]: I0127 16:37:39.646266 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ng4h8" event={"ID":"2b150c73-f6eb-4193-81ef-84941ff1abef","Type":"ContainerDied","Data":"c88e5ee63c5b143c29f95eb7a9163107e38a5ec7e21af15fb8f049bff95deca3"} Jan 27 16:37:39 crc kubenswrapper[4772]: I0127 16:37:39.646329 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ng4h8" Jan 27 16:37:39 crc kubenswrapper[4772]: I0127 16:37:39.646330 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c88e5ee63c5b143c29f95eb7a9163107e38a5ec7e21af15fb8f049bff95deca3" Jan 27 16:37:41 crc kubenswrapper[4772]: I0127 16:37:41.429718 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-wlbbm"] Jan 27 16:37:41 crc kubenswrapper[4772]: E0127 16:37:41.430363 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b150c73-f6eb-4193-81ef-84941ff1abef" containerName="mariadb-database-create" Jan 27 16:37:41 crc kubenswrapper[4772]: I0127 16:37:41.430376 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b150c73-f6eb-4193-81ef-84941ff1abef" containerName="mariadb-database-create" Jan 27 16:37:41 crc kubenswrapper[4772]: E0127 16:37:41.430396 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="615446cc-6fba-46f2-aad9-434f11519be9" containerName="mariadb-account-create-update" Jan 27 16:37:41 crc kubenswrapper[4772]: I0127 16:37:41.430402 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="615446cc-6fba-46f2-aad9-434f11519be9" containerName="mariadb-account-create-update" Jan 27 16:37:41 crc kubenswrapper[4772]: I0127 16:37:41.430574 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b150c73-f6eb-4193-81ef-84941ff1abef" containerName="mariadb-database-create" Jan 27 16:37:41 crc kubenswrapper[4772]: I0127 16:37:41.430590 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="615446cc-6fba-46f2-aad9-434f11519be9" containerName="mariadb-account-create-update" Jan 27 16:37:41 crc kubenswrapper[4772]: I0127 16:37:41.431317 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-wlbbm" Jan 27 16:37:41 crc kubenswrapper[4772]: I0127 16:37:41.433466 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Jan 27 16:37:41 crc kubenswrapper[4772]: I0127 16:37:41.434177 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-7r5xp" Jan 27 16:37:41 crc kubenswrapper[4772]: I0127 16:37:41.438193 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-wlbbm"] Jan 27 16:37:41 crc kubenswrapper[4772]: I0127 16:37:41.483789 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac290494-b5ad-4d85-9f14-daf092e3a6ed-config-data\") pod \"glance-db-sync-wlbbm\" (UID: \"ac290494-b5ad-4d85-9f14-daf092e3a6ed\") " pod="openstack/glance-db-sync-wlbbm" Jan 27 16:37:41 crc kubenswrapper[4772]: I0127 16:37:41.483852 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac290494-b5ad-4d85-9f14-daf092e3a6ed-combined-ca-bundle\") pod \"glance-db-sync-wlbbm\" (UID: \"ac290494-b5ad-4d85-9f14-daf092e3a6ed\") " pod="openstack/glance-db-sync-wlbbm" Jan 27 16:37:41 crc kubenswrapper[4772]: I0127 16:37:41.483892 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4b7k\" (UniqueName: \"kubernetes.io/projected/ac290494-b5ad-4d85-9f14-daf092e3a6ed-kube-api-access-w4b7k\") pod \"glance-db-sync-wlbbm\" (UID: \"ac290494-b5ad-4d85-9f14-daf092e3a6ed\") " pod="openstack/glance-db-sync-wlbbm" Jan 27 16:37:41 crc kubenswrapper[4772]: I0127 16:37:41.483971 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ac290494-b5ad-4d85-9f14-daf092e3a6ed-db-sync-config-data\") pod \"glance-db-sync-wlbbm\" (UID: \"ac290494-b5ad-4d85-9f14-daf092e3a6ed\") " pod="openstack/glance-db-sync-wlbbm" Jan 27 16:37:41 crc kubenswrapper[4772]: I0127 16:37:41.585621 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ac290494-b5ad-4d85-9f14-daf092e3a6ed-db-sync-config-data\") pod \"glance-db-sync-wlbbm\" (UID: \"ac290494-b5ad-4d85-9f14-daf092e3a6ed\") " pod="openstack/glance-db-sync-wlbbm" Jan 27 16:37:41 crc kubenswrapper[4772]: I0127 16:37:41.585743 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac290494-b5ad-4d85-9f14-daf092e3a6ed-config-data\") pod \"glance-db-sync-wlbbm\" (UID: \"ac290494-b5ad-4d85-9f14-daf092e3a6ed\") " pod="openstack/glance-db-sync-wlbbm" Jan 27 16:37:41 crc kubenswrapper[4772]: I0127 16:37:41.585781 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac290494-b5ad-4d85-9f14-daf092e3a6ed-combined-ca-bundle\") pod \"glance-db-sync-wlbbm\" (UID: \"ac290494-b5ad-4d85-9f14-daf092e3a6ed\") " pod="openstack/glance-db-sync-wlbbm" Jan 27 16:37:41 crc kubenswrapper[4772]: I0127 16:37:41.585810 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4b7k\" (UniqueName: \"kubernetes.io/projected/ac290494-b5ad-4d85-9f14-daf092e3a6ed-kube-api-access-w4b7k\") pod \"glance-db-sync-wlbbm\" (UID: \"ac290494-b5ad-4d85-9f14-daf092e3a6ed\") " pod="openstack/glance-db-sync-wlbbm" Jan 27 16:37:41 crc kubenswrapper[4772]: I0127 16:37:41.592094 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ac290494-b5ad-4d85-9f14-daf092e3a6ed-db-sync-config-data\") pod \"glance-db-sync-wlbbm\" (UID: \"ac290494-b5ad-4d85-9f14-daf092e3a6ed\") " pod="openstack/glance-db-sync-wlbbm" Jan 27 16:37:41 crc kubenswrapper[4772]: I0127 16:37:41.592913 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac290494-b5ad-4d85-9f14-daf092e3a6ed-config-data\") pod \"glance-db-sync-wlbbm\" (UID: \"ac290494-b5ad-4d85-9f14-daf092e3a6ed\") " pod="openstack/glance-db-sync-wlbbm" Jan 27 16:37:41 crc kubenswrapper[4772]: I0127 16:37:41.596095 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac290494-b5ad-4d85-9f14-daf092e3a6ed-combined-ca-bundle\") pod \"glance-db-sync-wlbbm\" (UID: \"ac290494-b5ad-4d85-9f14-daf092e3a6ed\") " pod="openstack/glance-db-sync-wlbbm" Jan 27 16:37:41 crc kubenswrapper[4772]: I0127 16:37:41.607017 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4b7k\" (UniqueName: \"kubernetes.io/projected/ac290494-b5ad-4d85-9f14-daf092e3a6ed-kube-api-access-w4b7k\") pod \"glance-db-sync-wlbbm\" (UID: \"ac290494-b5ad-4d85-9f14-daf092e3a6ed\") " pod="openstack/glance-db-sync-wlbbm" Jan 27 16:37:41 crc kubenswrapper[4772]: I0127 16:37:41.757356 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-wlbbm" Jan 27 16:37:42 crc kubenswrapper[4772]: I0127 16:37:42.058304 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:37:42 crc kubenswrapper[4772]: I0127 16:37:42.058641 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:37:42 crc kubenswrapper[4772]: I0127 16:37:42.428755 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-wlbbm"] Jan 27 16:37:42 crc kubenswrapper[4772]: I0127 16:37:42.673780 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-wlbbm" event={"ID":"ac290494-b5ad-4d85-9f14-daf092e3a6ed","Type":"ContainerStarted","Data":"e22dc408be2c6bac485917f551c4608cbccf8ac7f37c8b56903873b8fa5f6e6e"} Jan 27 16:37:43 crc kubenswrapper[4772]: I0127 16:37:43.677737 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-wlbbm" event={"ID":"ac290494-b5ad-4d85-9f14-daf092e3a6ed","Type":"ContainerStarted","Data":"972404acaa773e237766879c66e6f19a8b3951bb52780f66333fe8405eb0ccb2"} Jan 27 16:37:46 crc kubenswrapper[4772]: I0127 16:37:46.716573 4772 generic.go:334] "Generic (PLEG): container finished" podID="ac290494-b5ad-4d85-9f14-daf092e3a6ed" containerID="972404acaa773e237766879c66e6f19a8b3951bb52780f66333fe8405eb0ccb2" exitCode=0 Jan 27 16:37:46 crc kubenswrapper[4772]: I0127 16:37:46.716660 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-wlbbm" event={"ID":"ac290494-b5ad-4d85-9f14-daf092e3a6ed","Type":"ContainerDied","Data":"972404acaa773e237766879c66e6f19a8b3951bb52780f66333fe8405eb0ccb2"} Jan 27 16:37:48 crc kubenswrapper[4772]: I0127 16:37:48.097885 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-wlbbm" Jan 27 16:37:48 crc kubenswrapper[4772]: I0127 16:37:48.289570 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac290494-b5ad-4d85-9f14-daf092e3a6ed-config-data\") pod \"ac290494-b5ad-4d85-9f14-daf092e3a6ed\" (UID: \"ac290494-b5ad-4d85-9f14-daf092e3a6ed\") " Jan 27 16:37:48 crc kubenswrapper[4772]: I0127 16:37:48.289945 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4b7k\" (UniqueName: \"kubernetes.io/projected/ac290494-b5ad-4d85-9f14-daf092e3a6ed-kube-api-access-w4b7k\") pod \"ac290494-b5ad-4d85-9f14-daf092e3a6ed\" (UID: \"ac290494-b5ad-4d85-9f14-daf092e3a6ed\") " Jan 27 16:37:48 crc kubenswrapper[4772]: I0127 16:37:48.289989 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac290494-b5ad-4d85-9f14-daf092e3a6ed-combined-ca-bundle\") pod \"ac290494-b5ad-4d85-9f14-daf092e3a6ed\" (UID: \"ac290494-b5ad-4d85-9f14-daf092e3a6ed\") " Jan 27 16:37:48 crc kubenswrapper[4772]: I0127 16:37:48.290092 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ac290494-b5ad-4d85-9f14-daf092e3a6ed-db-sync-config-data\") pod \"ac290494-b5ad-4d85-9f14-daf092e3a6ed\" (UID: \"ac290494-b5ad-4d85-9f14-daf092e3a6ed\") " Jan 27 16:37:48 crc kubenswrapper[4772]: I0127 16:37:48.295591 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac290494-b5ad-4d85-9f14-daf092e3a6ed-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ac290494-b5ad-4d85-9f14-daf092e3a6ed" (UID: "ac290494-b5ad-4d85-9f14-daf092e3a6ed"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:37:48 crc kubenswrapper[4772]: I0127 16:37:48.301870 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac290494-b5ad-4d85-9f14-daf092e3a6ed-kube-api-access-w4b7k" (OuterVolumeSpecName: "kube-api-access-w4b7k") pod "ac290494-b5ad-4d85-9f14-daf092e3a6ed" (UID: "ac290494-b5ad-4d85-9f14-daf092e3a6ed"). InnerVolumeSpecName "kube-api-access-w4b7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:37:48 crc kubenswrapper[4772]: I0127 16:37:48.312877 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac290494-b5ad-4d85-9f14-daf092e3a6ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac290494-b5ad-4d85-9f14-daf092e3a6ed" (UID: "ac290494-b5ad-4d85-9f14-daf092e3a6ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:37:48 crc kubenswrapper[4772]: I0127 16:37:48.335203 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac290494-b5ad-4d85-9f14-daf092e3a6ed-config-data" (OuterVolumeSpecName: "config-data") pod "ac290494-b5ad-4d85-9f14-daf092e3a6ed" (UID: "ac290494-b5ad-4d85-9f14-daf092e3a6ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:37:48 crc kubenswrapper[4772]: I0127 16:37:48.392271 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac290494-b5ad-4d85-9f14-daf092e3a6ed-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:48 crc kubenswrapper[4772]: I0127 16:37:48.392307 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4b7k\" (UniqueName: \"kubernetes.io/projected/ac290494-b5ad-4d85-9f14-daf092e3a6ed-kube-api-access-w4b7k\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:48 crc kubenswrapper[4772]: I0127 16:37:48.392321 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac290494-b5ad-4d85-9f14-daf092e3a6ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:48 crc kubenswrapper[4772]: I0127 16:37:48.392335 4772 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ac290494-b5ad-4d85-9f14-daf092e3a6ed-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:48 crc kubenswrapper[4772]: I0127 16:37:48.749734 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-wlbbm" event={"ID":"ac290494-b5ad-4d85-9f14-daf092e3a6ed","Type":"ContainerDied","Data":"e22dc408be2c6bac485917f551c4608cbccf8ac7f37c8b56903873b8fa5f6e6e"} Jan 27 16:37:48 crc kubenswrapper[4772]: I0127 16:37:48.749782 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e22dc408be2c6bac485917f551c4608cbccf8ac7f37c8b56903873b8fa5f6e6e" Jan 27 16:37:48 crc kubenswrapper[4772]: I0127 16:37:48.749831 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-wlbbm" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.025755 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 16:37:49 crc kubenswrapper[4772]: E0127 16:37:49.026080 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac290494-b5ad-4d85-9f14-daf092e3a6ed" containerName="glance-db-sync" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.026093 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac290494-b5ad-4d85-9f14-daf092e3a6ed" containerName="glance-db-sync" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.026289 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac290494-b5ad-4d85-9f14-daf092e3a6ed" containerName="glance-db-sync" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.028380 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.030601 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.030912 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-7r5xp" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.031083 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.033976 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.045577 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.142097 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d645dd9d5-2pwb9"] Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.152747 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.161065 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d645dd9d5-2pwb9"] Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.206749 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cf8dce7-53a4-4a54-baad-3787346773ae-logs\") pod \"glance-default-external-api-0\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.206797 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cf8dce7-53a4-4a54-baad-3787346773ae-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.206848 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7cf8dce7-53a4-4a54-baad-3787346773ae-ceph\") pod \"glance-default-external-api-0\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.206867 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7cf8dce7-53a4-4a54-baad-3787346773ae-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.206885 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwn2d\" (UniqueName: \"kubernetes.io/projected/7cf8dce7-53a4-4a54-baad-3787346773ae-kube-api-access-jwn2d\") pod \"glance-default-external-api-0\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.206952 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cf8dce7-53a4-4a54-baad-3787346773ae-scripts\") pod \"glance-default-external-api-0\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.206984 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cf8dce7-53a4-4a54-baad-3787346773ae-config-data\") pod \"glance-default-external-api-0\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.280454 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.282143 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.284289 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.298146 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.308855 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e289d3f6-26ba-4306-a7f0-bf95513c9068-dns-svc\") pod \"dnsmasq-dns-6d645dd9d5-2pwb9\" (UID: \"e289d3f6-26ba-4306-a7f0-bf95513c9068\") " pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.308910 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e289d3f6-26ba-4306-a7f0-bf95513c9068-config\") pod \"dnsmasq-dns-6d645dd9d5-2pwb9\" (UID: \"e289d3f6-26ba-4306-a7f0-bf95513c9068\") " pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.308949 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cf8dce7-53a4-4a54-baad-3787346773ae-scripts\") pod \"glance-default-external-api-0\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.308976 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e289d3f6-26ba-4306-a7f0-bf95513c9068-ovsdbserver-nb\") pod \"dnsmasq-dns-6d645dd9d5-2pwb9\" (UID: \"e289d3f6-26ba-4306-a7f0-bf95513c9068\") " pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.309020 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7grf\" (UniqueName: \"kubernetes.io/projected/e289d3f6-26ba-4306-a7f0-bf95513c9068-kube-api-access-q7grf\") pod \"dnsmasq-dns-6d645dd9d5-2pwb9\" (UID: \"e289d3f6-26ba-4306-a7f0-bf95513c9068\") " pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.309044 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cf8dce7-53a4-4a54-baad-3787346773ae-config-data\") pod \"glance-default-external-api-0\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.309087 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cf8dce7-53a4-4a54-baad-3787346773ae-logs\") pod \"glance-default-external-api-0\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.309115 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cf8dce7-53a4-4a54-baad-3787346773ae-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.309153 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e289d3f6-26ba-4306-a7f0-bf95513c9068-ovsdbserver-sb\") pod \"dnsmasq-dns-6d645dd9d5-2pwb9\" (UID: \"e289d3f6-26ba-4306-a7f0-bf95513c9068\") " pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.309264 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7cf8dce7-53a4-4a54-baad-3787346773ae-ceph\") pod \"glance-default-external-api-0\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.309293 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7cf8dce7-53a4-4a54-baad-3787346773ae-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.309316 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwn2d\" (UniqueName: \"kubernetes.io/projected/7cf8dce7-53a4-4a54-baad-3787346773ae-kube-api-access-jwn2d\") pod \"glance-default-external-api-0\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.309993 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cf8dce7-53a4-4a54-baad-3787346773ae-logs\") pod \"glance-default-external-api-0\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.310294 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7cf8dce7-53a4-4a54-baad-3787346773ae-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.314230 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7cf8dce7-53a4-4a54-baad-3787346773ae-ceph\") pod \"glance-default-external-api-0\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.318916 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cf8dce7-53a4-4a54-baad-3787346773ae-scripts\") pod \"glance-default-external-api-0\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.320273 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cf8dce7-53a4-4a54-baad-3787346773ae-config-data\") pod \"glance-default-external-api-0\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.325818 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cf8dce7-53a4-4a54-baad-3787346773ae-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.339233 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwn2d\" (UniqueName: \"kubernetes.io/projected/7cf8dce7-53a4-4a54-baad-3787346773ae-kube-api-access-jwn2d\") pod \"glance-default-external-api-0\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.352627 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.410798 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e289d3f6-26ba-4306-a7f0-bf95513c9068-dns-svc\") pod \"dnsmasq-dns-6d645dd9d5-2pwb9\" (UID: \"e289d3f6-26ba-4306-a7f0-bf95513c9068\") " pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.411075 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/04330651-2770-404a-a5ff-66c7ce91b3e7-ceph\") pod \"glance-default-internal-api-0\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.411114 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e289d3f6-26ba-4306-a7f0-bf95513c9068-config\") pod \"dnsmasq-dns-6d645dd9d5-2pwb9\" (UID: \"e289d3f6-26ba-4306-a7f0-bf95513c9068\") " pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.411144 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04330651-2770-404a-a5ff-66c7ce91b3e7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.411186 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e289d3f6-26ba-4306-a7f0-bf95513c9068-ovsdbserver-nb\") pod \"dnsmasq-dns-6d645dd9d5-2pwb9\" (UID: \"e289d3f6-26ba-4306-a7f0-bf95513c9068\") " pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.411234 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7grf\" (UniqueName: \"kubernetes.io/projected/e289d3f6-26ba-4306-a7f0-bf95513c9068-kube-api-access-q7grf\") pod \"dnsmasq-dns-6d645dd9d5-2pwb9\" (UID: \"e289d3f6-26ba-4306-a7f0-bf95513c9068\") " pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.411260 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04330651-2770-404a-a5ff-66c7ce91b3e7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.411333 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e289d3f6-26ba-4306-a7f0-bf95513c9068-ovsdbserver-sb\") pod \"dnsmasq-dns-6d645dd9d5-2pwb9\" (UID: \"e289d3f6-26ba-4306-a7f0-bf95513c9068\") " pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.411381 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04330651-2770-404a-a5ff-66c7ce91b3e7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.411402 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/04330651-2770-404a-a5ff-66c7ce91b3e7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.411439 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wpmx\" (UniqueName: \"kubernetes.io/projected/04330651-2770-404a-a5ff-66c7ce91b3e7-kube-api-access-7wpmx\") pod \"glance-default-internal-api-0\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.411476 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04330651-2770-404a-a5ff-66c7ce91b3e7-logs\") pod \"glance-default-internal-api-0\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.411675 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e289d3f6-26ba-4306-a7f0-bf95513c9068-dns-svc\") pod \"dnsmasq-dns-6d645dd9d5-2pwb9\" (UID: \"e289d3f6-26ba-4306-a7f0-bf95513c9068\") " pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.412283 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e289d3f6-26ba-4306-a7f0-bf95513c9068-config\") pod \"dnsmasq-dns-6d645dd9d5-2pwb9\" (UID: \"e289d3f6-26ba-4306-a7f0-bf95513c9068\") " pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.412655 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e289d3f6-26ba-4306-a7f0-bf95513c9068-ovsdbserver-nb\") pod \"dnsmasq-dns-6d645dd9d5-2pwb9\" (UID: \"e289d3f6-26ba-4306-a7f0-bf95513c9068\") " pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.412914 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e289d3f6-26ba-4306-a7f0-bf95513c9068-ovsdbserver-sb\") pod \"dnsmasq-dns-6d645dd9d5-2pwb9\" (UID: \"e289d3f6-26ba-4306-a7f0-bf95513c9068\") " pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.440671 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7grf\" (UniqueName: \"kubernetes.io/projected/e289d3f6-26ba-4306-a7f0-bf95513c9068-kube-api-access-q7grf\") pod \"dnsmasq-dns-6d645dd9d5-2pwb9\" (UID: \"e289d3f6-26ba-4306-a7f0-bf95513c9068\") " pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.472593 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.513490 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04330651-2770-404a-a5ff-66c7ce91b3e7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.513539 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/04330651-2770-404a-a5ff-66c7ce91b3e7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.513580 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wpmx\" (UniqueName: \"kubernetes.io/projected/04330651-2770-404a-a5ff-66c7ce91b3e7-kube-api-access-7wpmx\") pod \"glance-default-internal-api-0\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.513621 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04330651-2770-404a-a5ff-66c7ce91b3e7-logs\") pod \"glance-default-internal-api-0\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.513683 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/04330651-2770-404a-a5ff-66c7ce91b3e7-ceph\") pod \"glance-default-internal-api-0\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.513722 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04330651-2770-404a-a5ff-66c7ce91b3e7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.513766 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04330651-2770-404a-a5ff-66c7ce91b3e7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.514474 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04330651-2770-404a-a5ff-66c7ce91b3e7-logs\") pod \"glance-default-internal-api-0\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.514716 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/04330651-2770-404a-a5ff-66c7ce91b3e7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.519667 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04330651-2770-404a-a5ff-66c7ce91b3e7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.522150 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04330651-2770-404a-a5ff-66c7ce91b3e7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.535819 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04330651-2770-404a-a5ff-66c7ce91b3e7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.538784 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/04330651-2770-404a-a5ff-66c7ce91b3e7-ceph\") pod \"glance-default-internal-api-0\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.538926 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wpmx\" (UniqueName: \"kubernetes.io/projected/04330651-2770-404a-a5ff-66c7ce91b3e7-kube-api-access-7wpmx\") pod \"glance-default-internal-api-0\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.599382 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 27 16:37:49 crc kubenswrapper[4772]: I0127 16:37:49.969471 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 16:37:50 crc kubenswrapper[4772]: I0127 16:37:50.183077 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d645dd9d5-2pwb9"] Jan 27 16:37:50 crc kubenswrapper[4772]: I0127 16:37:50.423660 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 16:37:50 crc kubenswrapper[4772]: I0127 16:37:50.750006 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 16:37:50 crc kubenswrapper[4772]: I0127 16:37:50.850499 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"04330651-2770-404a-a5ff-66c7ce91b3e7","Type":"ContainerStarted","Data":"e9209abb013b605df859331d891ccbf91ecc4ce083f4e3c8a98e078222c67e1b"} Jan 27 16:37:50 crc kubenswrapper[4772]: I0127 16:37:50.851874 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7cf8dce7-53a4-4a54-baad-3787346773ae","Type":"ContainerStarted","Data":"a2465bbbe9a52b9b42ecc99a8f99f0479ff0f8665e48c3d8d198379279a15b56"} Jan 27 16:37:50 crc kubenswrapper[4772]: I0127 16:37:50.851901 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7cf8dce7-53a4-4a54-baad-3787346773ae","Type":"ContainerStarted","Data":"516d152c267168193958fb213e9f1682cf2f22937a080abe30eadfb6c7582832"} Jan 27 16:37:50 crc kubenswrapper[4772]: I0127 16:37:50.853564 4772 generic.go:334] "Generic (PLEG): container finished" podID="e289d3f6-26ba-4306-a7f0-bf95513c9068" containerID="cfd61d60b4611b222a3dcb5c92c6682ed0c719a75f85f6d5730b053534f9b1b0" exitCode=0 Jan 27 16:37:50 crc kubenswrapper[4772]: I0127 16:37:50.853605 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" event={"ID":"e289d3f6-26ba-4306-a7f0-bf95513c9068","Type":"ContainerDied","Data":"cfd61d60b4611b222a3dcb5c92c6682ed0c719a75f85f6d5730b053534f9b1b0"} Jan 27 16:37:50 crc kubenswrapper[4772]: I0127 16:37:50.853620 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" event={"ID":"e289d3f6-26ba-4306-a7f0-bf95513c9068","Type":"ContainerStarted","Data":"5dc3b25da49a88c2e2d810e005946006b5e97f2c6fb476e48c50681bedaf4daf"} Jan 27 16:37:51 crc kubenswrapper[4772]: I0127 16:37:51.863692 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"04330651-2770-404a-a5ff-66c7ce91b3e7","Type":"ContainerStarted","Data":"91a840ae8df2a102ddfe506d79b6aa47edece57d2e8cac22e3c8b0eb6eb724a0"} Jan 27 16:37:51 crc kubenswrapper[4772]: I0127 16:37:51.864325 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"04330651-2770-404a-a5ff-66c7ce91b3e7","Type":"ContainerStarted","Data":"5ee5dd0b1e52995f524745c5dd7bbd5d75f473825ad1a4ec87c3a7016be76889"} Jan 27 16:37:51 crc kubenswrapper[4772]: I0127 16:37:51.865634 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7cf8dce7-53a4-4a54-baad-3787346773ae","Type":"ContainerStarted","Data":"12839ace5fe2bccfe75af24d40677d5809cc38fc3bfddb8ee3117ee32088743c"} Jan 27 16:37:51 crc kubenswrapper[4772]: I0127 16:37:51.865916 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="7cf8dce7-53a4-4a54-baad-3787346773ae" containerName="glance-log" containerID="cri-o://a2465bbbe9a52b9b42ecc99a8f99f0479ff0f8665e48c3d8d198379279a15b56" gracePeriod=30 Jan 27 16:37:51 crc kubenswrapper[4772]: I0127 16:37:51.865956 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="7cf8dce7-53a4-4a54-baad-3787346773ae" containerName="glance-httpd" containerID="cri-o://12839ace5fe2bccfe75af24d40677d5809cc38fc3bfddb8ee3117ee32088743c" gracePeriod=30 Jan 27 16:37:51 crc kubenswrapper[4772]: I0127 16:37:51.867790 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" event={"ID":"e289d3f6-26ba-4306-a7f0-bf95513c9068","Type":"ContainerStarted","Data":"de4a1855c8f97732a1ffd2229841c1e180e5f299f24cbbca78a154e8db2ccc0d"} Jan 27 16:37:51 crc kubenswrapper[4772]: I0127 16:37:51.868245 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" Jan 27 16:37:51 crc kubenswrapper[4772]: I0127 16:37:51.890134 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.890113881 podStartE2EDuration="2.890113881s" podCreationTimestamp="2026-01-27 16:37:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:37:51.884571633 +0000 UTC m=+5457.865180741" watchObservedRunningTime="2026-01-27 16:37:51.890113881 +0000 UTC m=+5457.870722969" Jan 27 16:37:51 crc kubenswrapper[4772]: I0127 16:37:51.915451 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" podStartSLOduration=2.915427912 podStartE2EDuration="2.915427912s" podCreationTimestamp="2026-01-27 16:37:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:37:51.908126344 +0000 UTC m=+5457.888735462" watchObservedRunningTime="2026-01-27 16:37:51.915427912 +0000 UTC m=+5457.896037010" Jan 27 16:37:51 crc kubenswrapper[4772]: I0127 16:37:51.930055 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.930021258 podStartE2EDuration="3.930021258s" podCreationTimestamp="2026-01-27 16:37:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:37:51.929944726 +0000 UTC m=+5457.910553854" watchObservedRunningTime="2026-01-27 16:37:51.930021258 +0000 UTC m=+5457.910630356" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.484909 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.647803 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cf8dce7-53a4-4a54-baad-3787346773ae-logs\") pod \"7cf8dce7-53a4-4a54-baad-3787346773ae\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.647857 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cf8dce7-53a4-4a54-baad-3787346773ae-config-data\") pod \"7cf8dce7-53a4-4a54-baad-3787346773ae\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.647949 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7cf8dce7-53a4-4a54-baad-3787346773ae-httpd-run\") pod \"7cf8dce7-53a4-4a54-baad-3787346773ae\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.647990 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cf8dce7-53a4-4a54-baad-3787346773ae-scripts\") pod \"7cf8dce7-53a4-4a54-baad-3787346773ae\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.648014 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7cf8dce7-53a4-4a54-baad-3787346773ae-ceph\") pod \"7cf8dce7-53a4-4a54-baad-3787346773ae\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.648058 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwn2d\" (UniqueName: \"kubernetes.io/projected/7cf8dce7-53a4-4a54-baad-3787346773ae-kube-api-access-jwn2d\") pod \"7cf8dce7-53a4-4a54-baad-3787346773ae\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.648110 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cf8dce7-53a4-4a54-baad-3787346773ae-combined-ca-bundle\") pod \"7cf8dce7-53a4-4a54-baad-3787346773ae\" (UID: \"7cf8dce7-53a4-4a54-baad-3787346773ae\") " Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.648575 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cf8dce7-53a4-4a54-baad-3787346773ae-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7cf8dce7-53a4-4a54-baad-3787346773ae" (UID: "7cf8dce7-53a4-4a54-baad-3787346773ae"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.648681 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cf8dce7-53a4-4a54-baad-3787346773ae-logs" (OuterVolumeSpecName: "logs") pod "7cf8dce7-53a4-4a54-baad-3787346773ae" (UID: "7cf8dce7-53a4-4a54-baad-3787346773ae"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.649282 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cf8dce7-53a4-4a54-baad-3787346773ae-logs\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.649309 4772 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7cf8dce7-53a4-4a54-baad-3787346773ae-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.655036 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cf8dce7-53a4-4a54-baad-3787346773ae-ceph" (OuterVolumeSpecName: "ceph") pod "7cf8dce7-53a4-4a54-baad-3787346773ae" (UID: "7cf8dce7-53a4-4a54-baad-3787346773ae"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.655178 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cf8dce7-53a4-4a54-baad-3787346773ae-scripts" (OuterVolumeSpecName: "scripts") pod "7cf8dce7-53a4-4a54-baad-3787346773ae" (UID: "7cf8dce7-53a4-4a54-baad-3787346773ae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.656373 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cf8dce7-53a4-4a54-baad-3787346773ae-kube-api-access-jwn2d" (OuterVolumeSpecName: "kube-api-access-jwn2d") pod "7cf8dce7-53a4-4a54-baad-3787346773ae" (UID: "7cf8dce7-53a4-4a54-baad-3787346773ae"). InnerVolumeSpecName "kube-api-access-jwn2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.689304 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cf8dce7-53a4-4a54-baad-3787346773ae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7cf8dce7-53a4-4a54-baad-3787346773ae" (UID: "7cf8dce7-53a4-4a54-baad-3787346773ae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.705065 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cf8dce7-53a4-4a54-baad-3787346773ae-config-data" (OuterVolumeSpecName: "config-data") pod "7cf8dce7-53a4-4a54-baad-3787346773ae" (UID: "7cf8dce7-53a4-4a54-baad-3787346773ae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.750093 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cf8dce7-53a4-4a54-baad-3787346773ae-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.750137 4772 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7cf8dce7-53a4-4a54-baad-3787346773ae-ceph\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.750153 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwn2d\" (UniqueName: \"kubernetes.io/projected/7cf8dce7-53a4-4a54-baad-3787346773ae-kube-api-access-jwn2d\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.750185 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cf8dce7-53a4-4a54-baad-3787346773ae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.750199 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cf8dce7-53a4-4a54-baad-3787346773ae-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.755977 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.888037 4772 generic.go:334] "Generic (PLEG): container finished" podID="7cf8dce7-53a4-4a54-baad-3787346773ae" containerID="12839ace5fe2bccfe75af24d40677d5809cc38fc3bfddb8ee3117ee32088743c" exitCode=0 Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.888070 4772 generic.go:334] "Generic (PLEG): container finished" podID="7cf8dce7-53a4-4a54-baad-3787346773ae" containerID="a2465bbbe9a52b9b42ecc99a8f99f0479ff0f8665e48c3d8d198379279a15b56" exitCode=143 Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.888090 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.888144 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7cf8dce7-53a4-4a54-baad-3787346773ae","Type":"ContainerDied","Data":"12839ace5fe2bccfe75af24d40677d5809cc38fc3bfddb8ee3117ee32088743c"} Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.888212 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7cf8dce7-53a4-4a54-baad-3787346773ae","Type":"ContainerDied","Data":"a2465bbbe9a52b9b42ecc99a8f99f0479ff0f8665e48c3d8d198379279a15b56"} Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.888227 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7cf8dce7-53a4-4a54-baad-3787346773ae","Type":"ContainerDied","Data":"516d152c267168193958fb213e9f1682cf2f22937a080abe30eadfb6c7582832"} Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.888246 4772 scope.go:117] "RemoveContainer" containerID="12839ace5fe2bccfe75af24d40677d5809cc38fc3bfddb8ee3117ee32088743c" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.923010 4772 scope.go:117] "RemoveContainer" containerID="a2465bbbe9a52b9b42ecc99a8f99f0479ff0f8665e48c3d8d198379279a15b56" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.925138 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.942370 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.953012 4772 scope.go:117] "RemoveContainer" containerID="12839ace5fe2bccfe75af24d40677d5809cc38fc3bfddb8ee3117ee32088743c" Jan 27 16:37:52 crc kubenswrapper[4772]: E0127 16:37:52.953976 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12839ace5fe2bccfe75af24d40677d5809cc38fc3bfddb8ee3117ee32088743c\": container with ID starting with 12839ace5fe2bccfe75af24d40677d5809cc38fc3bfddb8ee3117ee32088743c not found: ID does not exist" containerID="12839ace5fe2bccfe75af24d40677d5809cc38fc3bfddb8ee3117ee32088743c" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.954003 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12839ace5fe2bccfe75af24d40677d5809cc38fc3bfddb8ee3117ee32088743c"} err="failed to get container status \"12839ace5fe2bccfe75af24d40677d5809cc38fc3bfddb8ee3117ee32088743c\": rpc error: code = NotFound desc = could not find container \"12839ace5fe2bccfe75af24d40677d5809cc38fc3bfddb8ee3117ee32088743c\": container with ID starting with 12839ace5fe2bccfe75af24d40677d5809cc38fc3bfddb8ee3117ee32088743c not found: ID does not exist" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.954022 4772 scope.go:117] "RemoveContainer" containerID="a2465bbbe9a52b9b42ecc99a8f99f0479ff0f8665e48c3d8d198379279a15b56" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.955092 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 16:37:52 crc kubenswrapper[4772]: E0127 16:37:52.955483 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cf8dce7-53a4-4a54-baad-3787346773ae" containerName="glance-httpd" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.955496 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cf8dce7-53a4-4a54-baad-3787346773ae" containerName="glance-httpd" Jan 27 16:37:52 crc kubenswrapper[4772]: E0127 16:37:52.955517 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cf8dce7-53a4-4a54-baad-3787346773ae" containerName="glance-log" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.955524 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cf8dce7-53a4-4a54-baad-3787346773ae" containerName="glance-log" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.955678 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cf8dce7-53a4-4a54-baad-3787346773ae" containerName="glance-httpd" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.955698 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cf8dce7-53a4-4a54-baad-3787346773ae" containerName="glance-log" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.958065 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 27 16:37:52 crc kubenswrapper[4772]: E0127 16:37:52.962902 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2465bbbe9a52b9b42ecc99a8f99f0479ff0f8665e48c3d8d198379279a15b56\": container with ID starting with a2465bbbe9a52b9b42ecc99a8f99f0479ff0f8665e48c3d8d198379279a15b56 not found: ID does not exist" containerID="a2465bbbe9a52b9b42ecc99a8f99f0479ff0f8665e48c3d8d198379279a15b56" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.983782 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2465bbbe9a52b9b42ecc99a8f99f0479ff0f8665e48c3d8d198379279a15b56"} err="failed to get container status \"a2465bbbe9a52b9b42ecc99a8f99f0479ff0f8665e48c3d8d198379279a15b56\": rpc error: code = NotFound desc = could not find container \"a2465bbbe9a52b9b42ecc99a8f99f0479ff0f8665e48c3d8d198379279a15b56\": container with ID starting with a2465bbbe9a52b9b42ecc99a8f99f0479ff0f8665e48c3d8d198379279a15b56 not found: ID does not exist" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.992039 4772 scope.go:117] "RemoveContainer" containerID="12839ace5fe2bccfe75af24d40677d5809cc38fc3bfddb8ee3117ee32088743c" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.994236 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12839ace5fe2bccfe75af24d40677d5809cc38fc3bfddb8ee3117ee32088743c"} err="failed to get container status \"12839ace5fe2bccfe75af24d40677d5809cc38fc3bfddb8ee3117ee32088743c\": rpc error: code = NotFound desc = could not find container \"12839ace5fe2bccfe75af24d40677d5809cc38fc3bfddb8ee3117ee32088743c\": container with ID starting with 12839ace5fe2bccfe75af24d40677d5809cc38fc3bfddb8ee3117ee32088743c not found: ID does not exist" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.994284 4772 scope.go:117] "RemoveContainer" containerID="a2465bbbe9a52b9b42ecc99a8f99f0479ff0f8665e48c3d8d198379279a15b56" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.995894 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.999015 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2465bbbe9a52b9b42ecc99a8f99f0479ff0f8665e48c3d8d198379279a15b56"} err="failed to get container status \"a2465bbbe9a52b9b42ecc99a8f99f0479ff0f8665e48c3d8d198379279a15b56\": rpc error: code = NotFound desc = could not find container \"a2465bbbe9a52b9b42ecc99a8f99f0479ff0f8665e48c3d8d198379279a15b56\": container with ID starting with a2465bbbe9a52b9b42ecc99a8f99f0479ff0f8665e48c3d8d198379279a15b56 not found: ID does not exist" Jan 27 16:37:52 crc kubenswrapper[4772]: I0127 16:37:52.999853 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.156872 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a7917063-9e04-41e8-8fb9-e8383f839bd6-ceph\") pod \"glance-default-external-api-0\" (UID: \"a7917063-9e04-41e8-8fb9-e8383f839bd6\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.157047 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7917063-9e04-41e8-8fb9-e8383f839bd6-config-data\") pod \"glance-default-external-api-0\" (UID: \"a7917063-9e04-41e8-8fb9-e8383f839bd6\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.157094 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7917063-9e04-41e8-8fb9-e8383f839bd6-scripts\") pod \"glance-default-external-api-0\" (UID: \"a7917063-9e04-41e8-8fb9-e8383f839bd6\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.157139 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl9ff\" (UniqueName: \"kubernetes.io/projected/a7917063-9e04-41e8-8fb9-e8383f839bd6-kube-api-access-bl9ff\") pod \"glance-default-external-api-0\" (UID: \"a7917063-9e04-41e8-8fb9-e8383f839bd6\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.157278 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7917063-9e04-41e8-8fb9-e8383f839bd6-logs\") pod \"glance-default-external-api-0\" (UID: \"a7917063-9e04-41e8-8fb9-e8383f839bd6\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.157351 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a7917063-9e04-41e8-8fb9-e8383f839bd6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a7917063-9e04-41e8-8fb9-e8383f839bd6\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.157428 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7917063-9e04-41e8-8fb9-e8383f839bd6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a7917063-9e04-41e8-8fb9-e8383f839bd6\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.258483 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl9ff\" (UniqueName: \"kubernetes.io/projected/a7917063-9e04-41e8-8fb9-e8383f839bd6-kube-api-access-bl9ff\") pod \"glance-default-external-api-0\" (UID: \"a7917063-9e04-41e8-8fb9-e8383f839bd6\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.258554 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7917063-9e04-41e8-8fb9-e8383f839bd6-logs\") pod \"glance-default-external-api-0\" (UID: \"a7917063-9e04-41e8-8fb9-e8383f839bd6\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.258598 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a7917063-9e04-41e8-8fb9-e8383f839bd6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a7917063-9e04-41e8-8fb9-e8383f839bd6\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.258629 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7917063-9e04-41e8-8fb9-e8383f839bd6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a7917063-9e04-41e8-8fb9-e8383f839bd6\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.258703 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a7917063-9e04-41e8-8fb9-e8383f839bd6-ceph\") pod \"glance-default-external-api-0\" (UID: \"a7917063-9e04-41e8-8fb9-e8383f839bd6\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.258777 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7917063-9e04-41e8-8fb9-e8383f839bd6-config-data\") pod \"glance-default-external-api-0\" (UID: \"a7917063-9e04-41e8-8fb9-e8383f839bd6\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.258806 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7917063-9e04-41e8-8fb9-e8383f839bd6-scripts\") pod \"glance-default-external-api-0\" (UID: \"a7917063-9e04-41e8-8fb9-e8383f839bd6\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.259064 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7917063-9e04-41e8-8fb9-e8383f839bd6-logs\") pod \"glance-default-external-api-0\" (UID: \"a7917063-9e04-41e8-8fb9-e8383f839bd6\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.261205 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a7917063-9e04-41e8-8fb9-e8383f839bd6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a7917063-9e04-41e8-8fb9-e8383f839bd6\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.264847 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a7917063-9e04-41e8-8fb9-e8383f839bd6-ceph\") pod \"glance-default-external-api-0\" (UID: \"a7917063-9e04-41e8-8fb9-e8383f839bd6\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.264880 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7917063-9e04-41e8-8fb9-e8383f839bd6-scripts\") pod \"glance-default-external-api-0\" (UID: \"a7917063-9e04-41e8-8fb9-e8383f839bd6\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.265534 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7917063-9e04-41e8-8fb9-e8383f839bd6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a7917063-9e04-41e8-8fb9-e8383f839bd6\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.269410 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7917063-9e04-41e8-8fb9-e8383f839bd6-config-data\") pod \"glance-default-external-api-0\" (UID: \"a7917063-9e04-41e8-8fb9-e8383f839bd6\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.278159 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl9ff\" (UniqueName: \"kubernetes.io/projected/a7917063-9e04-41e8-8fb9-e8383f839bd6-kube-api-access-bl9ff\") pod \"glance-default-external-api-0\" (UID: \"a7917063-9e04-41e8-8fb9-e8383f839bd6\") " pod="openstack/glance-default-external-api-0" Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.297962 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.881523 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 27 16:37:53 crc kubenswrapper[4772]: W0127 16:37:53.885038 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7917063_9e04_41e8_8fb9_e8383f839bd6.slice/crio-3ebb35abcf6caaaad30b633f23eadd07d995c99ccfbf81457ba471e6fe0cbc23 WatchSource:0}: Error finding container 3ebb35abcf6caaaad30b633f23eadd07d995c99ccfbf81457ba471e6fe0cbc23: Status 404 returned error can't find the container with id 3ebb35abcf6caaaad30b633f23eadd07d995c99ccfbf81457ba471e6fe0cbc23 Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.900280 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a7917063-9e04-41e8-8fb9-e8383f839bd6","Type":"ContainerStarted","Data":"3ebb35abcf6caaaad30b633f23eadd07d995c99ccfbf81457ba471e6fe0cbc23"} Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.900658 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="04330651-2770-404a-a5ff-66c7ce91b3e7" containerName="glance-httpd" containerID="cri-o://91a840ae8df2a102ddfe506d79b6aa47edece57d2e8cac22e3c8b0eb6eb724a0" gracePeriod=30 Jan 27 16:37:53 crc kubenswrapper[4772]: I0127 16:37:53.900455 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="04330651-2770-404a-a5ff-66c7ce91b3e7" containerName="glance-log" containerID="cri-o://5ee5dd0b1e52995f524745c5dd7bbd5d75f473825ad1a4ec87c3a7016be76889" gracePeriod=30 Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.552122 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.589815 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04330651-2770-404a-a5ff-66c7ce91b3e7-config-data\") pod \"04330651-2770-404a-a5ff-66c7ce91b3e7\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.589935 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04330651-2770-404a-a5ff-66c7ce91b3e7-scripts\") pod \"04330651-2770-404a-a5ff-66c7ce91b3e7\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.590590 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wpmx\" (UniqueName: \"kubernetes.io/projected/04330651-2770-404a-a5ff-66c7ce91b3e7-kube-api-access-7wpmx\") pod \"04330651-2770-404a-a5ff-66c7ce91b3e7\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.590637 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/04330651-2770-404a-a5ff-66c7ce91b3e7-ceph\") pod \"04330651-2770-404a-a5ff-66c7ce91b3e7\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.590666 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04330651-2770-404a-a5ff-66c7ce91b3e7-combined-ca-bundle\") pod \"04330651-2770-404a-a5ff-66c7ce91b3e7\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.590700 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/04330651-2770-404a-a5ff-66c7ce91b3e7-httpd-run\") pod \"04330651-2770-404a-a5ff-66c7ce91b3e7\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.591094 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04330651-2770-404a-a5ff-66c7ce91b3e7-logs\") pod \"04330651-2770-404a-a5ff-66c7ce91b3e7\" (UID: \"04330651-2770-404a-a5ff-66c7ce91b3e7\") " Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.591793 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04330651-2770-404a-a5ff-66c7ce91b3e7-logs" (OuterVolumeSpecName: "logs") pod "04330651-2770-404a-a5ff-66c7ce91b3e7" (UID: "04330651-2770-404a-a5ff-66c7ce91b3e7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.592040 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04330651-2770-404a-a5ff-66c7ce91b3e7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "04330651-2770-404a-a5ff-66c7ce91b3e7" (UID: "04330651-2770-404a-a5ff-66c7ce91b3e7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.593360 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04330651-2770-404a-a5ff-66c7ce91b3e7-kube-api-access-7wpmx" (OuterVolumeSpecName: "kube-api-access-7wpmx") pod "04330651-2770-404a-a5ff-66c7ce91b3e7" (UID: "04330651-2770-404a-a5ff-66c7ce91b3e7"). InnerVolumeSpecName "kube-api-access-7wpmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.597076 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04330651-2770-404a-a5ff-66c7ce91b3e7-ceph" (OuterVolumeSpecName: "ceph") pod "04330651-2770-404a-a5ff-66c7ce91b3e7" (UID: "04330651-2770-404a-a5ff-66c7ce91b3e7"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.609222 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04330651-2770-404a-a5ff-66c7ce91b3e7-scripts" (OuterVolumeSpecName: "scripts") pod "04330651-2770-404a-a5ff-66c7ce91b3e7" (UID: "04330651-2770-404a-a5ff-66c7ce91b3e7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.650605 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04330651-2770-404a-a5ff-66c7ce91b3e7-config-data" (OuterVolumeSpecName: "config-data") pod "04330651-2770-404a-a5ff-66c7ce91b3e7" (UID: "04330651-2770-404a-a5ff-66c7ce91b3e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.652264 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04330651-2770-404a-a5ff-66c7ce91b3e7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04330651-2770-404a-a5ff-66c7ce91b3e7" (UID: "04330651-2770-404a-a5ff-66c7ce91b3e7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.680510 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cf8dce7-53a4-4a54-baad-3787346773ae" path="/var/lib/kubelet/pods/7cf8dce7-53a4-4a54-baad-3787346773ae/volumes" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.693090 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04330651-2770-404a-a5ff-66c7ce91b3e7-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.693127 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04330651-2770-404a-a5ff-66c7ce91b3e7-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.693140 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wpmx\" (UniqueName: \"kubernetes.io/projected/04330651-2770-404a-a5ff-66c7ce91b3e7-kube-api-access-7wpmx\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.693155 4772 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/04330651-2770-404a-a5ff-66c7ce91b3e7-ceph\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.693183 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04330651-2770-404a-a5ff-66c7ce91b3e7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.693197 4772 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/04330651-2770-404a-a5ff-66c7ce91b3e7-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.693207 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04330651-2770-404a-a5ff-66c7ce91b3e7-logs\") on node \"crc\" DevicePath \"\"" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.916019 4772 generic.go:334] "Generic (PLEG): container finished" podID="04330651-2770-404a-a5ff-66c7ce91b3e7" containerID="91a840ae8df2a102ddfe506d79b6aa47edece57d2e8cac22e3c8b0eb6eb724a0" exitCode=0 Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.916054 4772 generic.go:334] "Generic (PLEG): container finished" podID="04330651-2770-404a-a5ff-66c7ce91b3e7" containerID="5ee5dd0b1e52995f524745c5dd7bbd5d75f473825ad1a4ec87c3a7016be76889" exitCode=143 Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.916080 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.916103 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"04330651-2770-404a-a5ff-66c7ce91b3e7","Type":"ContainerDied","Data":"91a840ae8df2a102ddfe506d79b6aa47edece57d2e8cac22e3c8b0eb6eb724a0"} Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.916145 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"04330651-2770-404a-a5ff-66c7ce91b3e7","Type":"ContainerDied","Data":"5ee5dd0b1e52995f524745c5dd7bbd5d75f473825ad1a4ec87c3a7016be76889"} Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.916178 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"04330651-2770-404a-a5ff-66c7ce91b3e7","Type":"ContainerDied","Data":"e9209abb013b605df859331d891ccbf91ecc4ce083f4e3c8a98e078222c67e1b"} Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.916206 4772 scope.go:117] "RemoveContainer" containerID="91a840ae8df2a102ddfe506d79b6aa47edece57d2e8cac22e3c8b0eb6eb724a0" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.919115 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a7917063-9e04-41e8-8fb9-e8383f839bd6","Type":"ContainerStarted","Data":"1367a12b2c0bd35eb3d620c4a253cd5dac72c6b345a3f6be430247d7c052b1c4"} Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.941297 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.942963 4772 scope.go:117] "RemoveContainer" containerID="5ee5dd0b1e52995f524745c5dd7bbd5d75f473825ad1a4ec87c3a7016be76889" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.958801 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.965070 4772 scope.go:117] "RemoveContainer" containerID="91a840ae8df2a102ddfe506d79b6aa47edece57d2e8cac22e3c8b0eb6eb724a0" Jan 27 16:37:54 crc kubenswrapper[4772]: E0127 16:37:54.965658 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91a840ae8df2a102ddfe506d79b6aa47edece57d2e8cac22e3c8b0eb6eb724a0\": container with ID starting with 91a840ae8df2a102ddfe506d79b6aa47edece57d2e8cac22e3c8b0eb6eb724a0 not found: ID does not exist" containerID="91a840ae8df2a102ddfe506d79b6aa47edece57d2e8cac22e3c8b0eb6eb724a0" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.965707 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91a840ae8df2a102ddfe506d79b6aa47edece57d2e8cac22e3c8b0eb6eb724a0"} err="failed to get container status \"91a840ae8df2a102ddfe506d79b6aa47edece57d2e8cac22e3c8b0eb6eb724a0\": rpc error: code = NotFound desc = could not find container \"91a840ae8df2a102ddfe506d79b6aa47edece57d2e8cac22e3c8b0eb6eb724a0\": container with ID starting with 91a840ae8df2a102ddfe506d79b6aa47edece57d2e8cac22e3c8b0eb6eb724a0 not found: ID does not exist" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.965740 4772 scope.go:117] "RemoveContainer" containerID="5ee5dd0b1e52995f524745c5dd7bbd5d75f473825ad1a4ec87c3a7016be76889" Jan 27 16:37:54 crc kubenswrapper[4772]: E0127 16:37:54.966002 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ee5dd0b1e52995f524745c5dd7bbd5d75f473825ad1a4ec87c3a7016be76889\": container with ID starting with 5ee5dd0b1e52995f524745c5dd7bbd5d75f473825ad1a4ec87c3a7016be76889 not found: ID does not exist" containerID="5ee5dd0b1e52995f524745c5dd7bbd5d75f473825ad1a4ec87c3a7016be76889" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.966025 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ee5dd0b1e52995f524745c5dd7bbd5d75f473825ad1a4ec87c3a7016be76889"} err="failed to get container status \"5ee5dd0b1e52995f524745c5dd7bbd5d75f473825ad1a4ec87c3a7016be76889\": rpc error: code = NotFound desc = could not find container \"5ee5dd0b1e52995f524745c5dd7bbd5d75f473825ad1a4ec87c3a7016be76889\": container with ID starting with 5ee5dd0b1e52995f524745c5dd7bbd5d75f473825ad1a4ec87c3a7016be76889 not found: ID does not exist" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.966041 4772 scope.go:117] "RemoveContainer" containerID="91a840ae8df2a102ddfe506d79b6aa47edece57d2e8cac22e3c8b0eb6eb724a0" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.966264 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91a840ae8df2a102ddfe506d79b6aa47edece57d2e8cac22e3c8b0eb6eb724a0"} err="failed to get container status \"91a840ae8df2a102ddfe506d79b6aa47edece57d2e8cac22e3c8b0eb6eb724a0\": rpc error: code = NotFound desc = could not find container \"91a840ae8df2a102ddfe506d79b6aa47edece57d2e8cac22e3c8b0eb6eb724a0\": container with ID starting with 91a840ae8df2a102ddfe506d79b6aa47edece57d2e8cac22e3c8b0eb6eb724a0 not found: ID does not exist" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.966283 4772 scope.go:117] "RemoveContainer" containerID="5ee5dd0b1e52995f524745c5dd7bbd5d75f473825ad1a4ec87c3a7016be76889" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.966477 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ee5dd0b1e52995f524745c5dd7bbd5d75f473825ad1a4ec87c3a7016be76889"} err="failed to get container status \"5ee5dd0b1e52995f524745c5dd7bbd5d75f473825ad1a4ec87c3a7016be76889\": rpc error: code = NotFound desc = could not find container \"5ee5dd0b1e52995f524745c5dd7bbd5d75f473825ad1a4ec87c3a7016be76889\": container with ID starting with 5ee5dd0b1e52995f524745c5dd7bbd5d75f473825ad1a4ec87c3a7016be76889 not found: ID does not exist" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.968693 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 16:37:54 crc kubenswrapper[4772]: E0127 16:37:54.969008 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04330651-2770-404a-a5ff-66c7ce91b3e7" containerName="glance-httpd" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.969021 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="04330651-2770-404a-a5ff-66c7ce91b3e7" containerName="glance-httpd" Jan 27 16:37:54 crc kubenswrapper[4772]: E0127 16:37:54.969060 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04330651-2770-404a-a5ff-66c7ce91b3e7" containerName="glance-log" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.969069 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="04330651-2770-404a-a5ff-66c7ce91b3e7" containerName="glance-log" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.969264 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="04330651-2770-404a-a5ff-66c7ce91b3e7" containerName="glance-httpd" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.969280 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="04330651-2770-404a-a5ff-66c7ce91b3e7" containerName="glance-log" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.970186 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.979128 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Jan 27 16:37:54 crc kubenswrapper[4772]: I0127 16:37:54.991852 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.004005 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1535f57-0540-45ea-b53c-1b4cac461cf3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a1535f57-0540-45ea-b53c-1b4cac461cf3\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.004092 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1535f57-0540-45ea-b53c-1b4cac461cf3-logs\") pod \"glance-default-internal-api-0\" (UID: \"a1535f57-0540-45ea-b53c-1b4cac461cf3\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.004134 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d88rx\" (UniqueName: \"kubernetes.io/projected/a1535f57-0540-45ea-b53c-1b4cac461cf3-kube-api-access-d88rx\") pod \"glance-default-internal-api-0\" (UID: \"a1535f57-0540-45ea-b53c-1b4cac461cf3\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.004222 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1535f57-0540-45ea-b53c-1b4cac461cf3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a1535f57-0540-45ea-b53c-1b4cac461cf3\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.004305 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1535f57-0540-45ea-b53c-1b4cac461cf3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a1535f57-0540-45ea-b53c-1b4cac461cf3\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.004341 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1535f57-0540-45ea-b53c-1b4cac461cf3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a1535f57-0540-45ea-b53c-1b4cac461cf3\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.004363 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a1535f57-0540-45ea-b53c-1b4cac461cf3-ceph\") pod \"glance-default-internal-api-0\" (UID: \"a1535f57-0540-45ea-b53c-1b4cac461cf3\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.105566 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1535f57-0540-45ea-b53c-1b4cac461cf3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a1535f57-0540-45ea-b53c-1b4cac461cf3\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.105606 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a1535f57-0540-45ea-b53c-1b4cac461cf3-ceph\") pod \"glance-default-internal-api-0\" (UID: \"a1535f57-0540-45ea-b53c-1b4cac461cf3\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.105658 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1535f57-0540-45ea-b53c-1b4cac461cf3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a1535f57-0540-45ea-b53c-1b4cac461cf3\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.105692 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1535f57-0540-45ea-b53c-1b4cac461cf3-logs\") pod \"glance-default-internal-api-0\" (UID: \"a1535f57-0540-45ea-b53c-1b4cac461cf3\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.105715 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d88rx\" (UniqueName: \"kubernetes.io/projected/a1535f57-0540-45ea-b53c-1b4cac461cf3-kube-api-access-d88rx\") pod \"glance-default-internal-api-0\" (UID: \"a1535f57-0540-45ea-b53c-1b4cac461cf3\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.105985 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1535f57-0540-45ea-b53c-1b4cac461cf3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a1535f57-0540-45ea-b53c-1b4cac461cf3\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.106073 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1535f57-0540-45ea-b53c-1b4cac461cf3-logs\") pod \"glance-default-internal-api-0\" (UID: \"a1535f57-0540-45ea-b53c-1b4cac461cf3\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.106738 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1535f57-0540-45ea-b53c-1b4cac461cf3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a1535f57-0540-45ea-b53c-1b4cac461cf3\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.106811 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1535f57-0540-45ea-b53c-1b4cac461cf3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a1535f57-0540-45ea-b53c-1b4cac461cf3\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.111814 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1535f57-0540-45ea-b53c-1b4cac461cf3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a1535f57-0540-45ea-b53c-1b4cac461cf3\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.111854 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1535f57-0540-45ea-b53c-1b4cac461cf3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a1535f57-0540-45ea-b53c-1b4cac461cf3\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.112128 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1535f57-0540-45ea-b53c-1b4cac461cf3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a1535f57-0540-45ea-b53c-1b4cac461cf3\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.112146 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a1535f57-0540-45ea-b53c-1b4cac461cf3-ceph\") pod \"glance-default-internal-api-0\" (UID: \"a1535f57-0540-45ea-b53c-1b4cac461cf3\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.122987 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d88rx\" (UniqueName: \"kubernetes.io/projected/a1535f57-0540-45ea-b53c-1b4cac461cf3-kube-api-access-d88rx\") pod \"glance-default-internal-api-0\" (UID: \"a1535f57-0540-45ea-b53c-1b4cac461cf3\") " pod="openstack/glance-default-internal-api-0" Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.288370 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.769758 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.942023 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a1535f57-0540-45ea-b53c-1b4cac461cf3","Type":"ContainerStarted","Data":"bdd1e6d352ccc9ac313ef81a1a9d4729caed4c9ef3252c7f8ccf561487a8dddf"} Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.951611 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a7917063-9e04-41e8-8fb9-e8383f839bd6","Type":"ContainerStarted","Data":"4ef258baec7de4dd434ebb599831e99ae07b1cc3d174d18617f6f85f4baa1b33"} Jan 27 16:37:55 crc kubenswrapper[4772]: I0127 16:37:55.977217 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.977195878 podStartE2EDuration="3.977195878s" podCreationTimestamp="2026-01-27 16:37:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:37:55.969308824 +0000 UTC m=+5461.949917932" watchObservedRunningTime="2026-01-27 16:37:55.977195878 +0000 UTC m=+5461.957804976" Jan 27 16:37:56 crc kubenswrapper[4772]: I0127 16:37:56.674705 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04330651-2770-404a-a5ff-66c7ce91b3e7" path="/var/lib/kubelet/pods/04330651-2770-404a-a5ff-66c7ce91b3e7/volumes" Jan 27 16:37:56 crc kubenswrapper[4772]: I0127 16:37:56.962994 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a1535f57-0540-45ea-b53c-1b4cac461cf3","Type":"ContainerStarted","Data":"59949bbfc428bded0d6c0d47e4e1ba36374017de77611ef43497f757582df6cb"} Jan 27 16:37:56 crc kubenswrapper[4772]: I0127 16:37:56.963047 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a1535f57-0540-45ea-b53c-1b4cac461cf3","Type":"ContainerStarted","Data":"beef63a5f0beb83017ebac7eba7f0abb95542a31b3a6d3d3ab0cffdfa906e99a"} Jan 27 16:37:56 crc kubenswrapper[4772]: I0127 16:37:56.989427 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.989406201 podStartE2EDuration="2.989406201s" podCreationTimestamp="2026-01-27 16:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:37:56.988242927 +0000 UTC m=+5462.968852045" watchObservedRunningTime="2026-01-27 16:37:56.989406201 +0000 UTC m=+5462.970015289" Jan 27 16:37:59 crc kubenswrapper[4772]: I0127 16:37:59.474337 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" Jan 27 16:37:59 crc kubenswrapper[4772]: I0127 16:37:59.550578 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ccbf7777-mh9xn"] Jan 27 16:37:59 crc kubenswrapper[4772]: I0127 16:37:59.550861 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" podUID="b6449727-ae23-4ae0-b6e6-4c1cef43ef53" containerName="dnsmasq-dns" containerID="cri-o://f3387679ac491bfa95cee52f7df561b5ae6817e29d93d8687efd8b7b43af3938" gracePeriod=10 Jan 27 16:37:59 crc kubenswrapper[4772]: I0127 16:37:59.996157 4772 generic.go:334] "Generic (PLEG): container finished" podID="b6449727-ae23-4ae0-b6e6-4c1cef43ef53" containerID="f3387679ac491bfa95cee52f7df561b5ae6817e29d93d8687efd8b7b43af3938" exitCode=0 Jan 27 16:37:59 crc kubenswrapper[4772]: I0127 16:37:59.996200 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" event={"ID":"b6449727-ae23-4ae0-b6e6-4c1cef43ef53","Type":"ContainerDied","Data":"f3387679ac491bfa95cee52f7df561b5ae6817e29d93d8687efd8b7b43af3938"} Jan 27 16:37:59 crc kubenswrapper[4772]: I0127 16:37:59.996572 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" event={"ID":"b6449727-ae23-4ae0-b6e6-4c1cef43ef53","Type":"ContainerDied","Data":"0286cc31499d6fb5d457987c1e11aad0e7f2bc6e4ccd62d6d0d7e8570d1f3e2e"} Jan 27 16:37:59 crc kubenswrapper[4772]: I0127 16:37:59.996606 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0286cc31499d6fb5d457987c1e11aad0e7f2bc6e4ccd62d6d0d7e8570d1f3e2e" Jan 27 16:38:00 crc kubenswrapper[4772]: I0127 16:38:00.047552 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" Jan 27 16:38:00 crc kubenswrapper[4772]: I0127 16:38:00.103441 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-ovsdbserver-nb\") pod \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\" (UID: \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\") " Jan 27 16:38:00 crc kubenswrapper[4772]: I0127 16:38:00.103539 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-dns-svc\") pod \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\" (UID: \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\") " Jan 27 16:38:00 crc kubenswrapper[4772]: I0127 16:38:00.103754 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-ovsdbserver-sb\") pod \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\" (UID: \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\") " Jan 27 16:38:00 crc kubenswrapper[4772]: I0127 16:38:00.103816 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sl6sk\" (UniqueName: \"kubernetes.io/projected/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-kube-api-access-sl6sk\") pod \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\" (UID: \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\") " Jan 27 16:38:00 crc kubenswrapper[4772]: I0127 16:38:00.103838 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-config\") pod \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\" (UID: \"b6449727-ae23-4ae0-b6e6-4c1cef43ef53\") " Jan 27 16:38:00 crc kubenswrapper[4772]: I0127 16:38:00.117965 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-kube-api-access-sl6sk" (OuterVolumeSpecName: "kube-api-access-sl6sk") pod "b6449727-ae23-4ae0-b6e6-4c1cef43ef53" (UID: "b6449727-ae23-4ae0-b6e6-4c1cef43ef53"). InnerVolumeSpecName "kube-api-access-sl6sk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:38:00 crc kubenswrapper[4772]: I0127 16:38:00.171115 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b6449727-ae23-4ae0-b6e6-4c1cef43ef53" (UID: "b6449727-ae23-4ae0-b6e6-4c1cef43ef53"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:38:00 crc kubenswrapper[4772]: I0127 16:38:00.171708 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-config" (OuterVolumeSpecName: "config") pod "b6449727-ae23-4ae0-b6e6-4c1cef43ef53" (UID: "b6449727-ae23-4ae0-b6e6-4c1cef43ef53"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:38:00 crc kubenswrapper[4772]: I0127 16:38:00.174811 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b6449727-ae23-4ae0-b6e6-4c1cef43ef53" (UID: "b6449727-ae23-4ae0-b6e6-4c1cef43ef53"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:38:00 crc kubenswrapper[4772]: I0127 16:38:00.203947 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b6449727-ae23-4ae0-b6e6-4c1cef43ef53" (UID: "b6449727-ae23-4ae0-b6e6-4c1cef43ef53"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:38:00 crc kubenswrapper[4772]: I0127 16:38:00.206829 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 27 16:38:00 crc kubenswrapper[4772]: I0127 16:38:00.206889 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sl6sk\" (UniqueName: \"kubernetes.io/projected/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-kube-api-access-sl6sk\") on node \"crc\" DevicePath \"\"" Jan 27 16:38:00 crc kubenswrapper[4772]: I0127 16:38:00.206907 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-config\") on node \"crc\" DevicePath \"\"" Jan 27 16:38:00 crc kubenswrapper[4772]: I0127 16:38:00.206919 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 27 16:38:00 crc kubenswrapper[4772]: I0127 16:38:00.206949 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6449727-ae23-4ae0-b6e6-4c1cef43ef53-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 16:38:01 crc kubenswrapper[4772]: I0127 16:38:01.005728 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ccbf7777-mh9xn" Jan 27 16:38:01 crc kubenswrapper[4772]: I0127 16:38:01.027371 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ccbf7777-mh9xn"] Jan 27 16:38:01 crc kubenswrapper[4772]: I0127 16:38:01.034147 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ccbf7777-mh9xn"] Jan 27 16:38:02 crc kubenswrapper[4772]: I0127 16:38:02.672481 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6449727-ae23-4ae0-b6e6-4c1cef43ef53" path="/var/lib/kubelet/pods/b6449727-ae23-4ae0-b6e6-4c1cef43ef53/volumes" Jan 27 16:38:03 crc kubenswrapper[4772]: I0127 16:38:03.298642 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 27 16:38:03 crc kubenswrapper[4772]: I0127 16:38:03.298690 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 27 16:38:03 crc kubenswrapper[4772]: I0127 16:38:03.343916 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 27 16:38:03 crc kubenswrapper[4772]: I0127 16:38:03.343975 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 27 16:38:04 crc kubenswrapper[4772]: I0127 16:38:04.029404 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 27 16:38:04 crc kubenswrapper[4772]: I0127 16:38:04.029693 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 27 16:38:05 crc kubenswrapper[4772]: I0127 16:38:05.289038 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 27 16:38:05 crc kubenswrapper[4772]: I0127 16:38:05.290042 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 27 16:38:05 crc kubenswrapper[4772]: I0127 16:38:05.314223 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 27 16:38:05 crc kubenswrapper[4772]: I0127 16:38:05.337065 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 27 16:38:06 crc kubenswrapper[4772]: I0127 16:38:06.043525 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 27 16:38:06 crc kubenswrapper[4772]: I0127 16:38:06.043565 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 27 16:38:06 crc kubenswrapper[4772]: I0127 16:38:06.226315 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 27 16:38:06 crc kubenswrapper[4772]: I0127 16:38:06.226453 4772 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 27 16:38:06 crc kubenswrapper[4772]: I0127 16:38:06.228447 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 27 16:38:08 crc kubenswrapper[4772]: I0127 16:38:08.118920 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 27 16:38:08 crc kubenswrapper[4772]: I0127 16:38:08.119345 4772 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 27 16:38:08 crc kubenswrapper[4772]: I0127 16:38:08.372579 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 27 16:38:12 crc kubenswrapper[4772]: I0127 16:38:12.058350 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:38:12 crc kubenswrapper[4772]: I0127 16:38:12.058902 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:38:12 crc kubenswrapper[4772]: I0127 16:38:12.058953 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 16:38:12 crc kubenswrapper[4772]: I0127 16:38:12.059695 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"90e27c06727cf113f54cd7c0344565bfa447b15cc343fc7033a04f41dddb22f9"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 16:38:12 crc kubenswrapper[4772]: I0127 16:38:12.059750 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://90e27c06727cf113f54cd7c0344565bfa447b15cc343fc7033a04f41dddb22f9" gracePeriod=600 Jan 27 16:38:13 crc kubenswrapper[4772]: I0127 16:38:13.102726 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="90e27c06727cf113f54cd7c0344565bfa447b15cc343fc7033a04f41dddb22f9" exitCode=0 Jan 27 16:38:13 crc kubenswrapper[4772]: I0127 16:38:13.102902 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"90e27c06727cf113f54cd7c0344565bfa447b15cc343fc7033a04f41dddb22f9"} Jan 27 16:38:13 crc kubenswrapper[4772]: I0127 16:38:13.103330 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69"} Jan 27 16:38:13 crc kubenswrapper[4772]: I0127 16:38:13.103363 4772 scope.go:117] "RemoveContainer" containerID="beb82f81f96be589cf221c90702e405768d59833a36f70e2929085c7b622f86b" Jan 27 16:38:18 crc kubenswrapper[4772]: I0127 16:38:18.272305 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-wfhq4"] Jan 27 16:38:18 crc kubenswrapper[4772]: E0127 16:38:18.273256 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6449727-ae23-4ae0-b6e6-4c1cef43ef53" containerName="dnsmasq-dns" Jan 27 16:38:18 crc kubenswrapper[4772]: I0127 16:38:18.273272 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6449727-ae23-4ae0-b6e6-4c1cef43ef53" containerName="dnsmasq-dns" Jan 27 16:38:18 crc kubenswrapper[4772]: E0127 16:38:18.273285 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6449727-ae23-4ae0-b6e6-4c1cef43ef53" containerName="init" Jan 27 16:38:18 crc kubenswrapper[4772]: I0127 16:38:18.273294 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6449727-ae23-4ae0-b6e6-4c1cef43ef53" containerName="init" Jan 27 16:38:18 crc kubenswrapper[4772]: I0127 16:38:18.273458 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6449727-ae23-4ae0-b6e6-4c1cef43ef53" containerName="dnsmasq-dns" Jan 27 16:38:18 crc kubenswrapper[4772]: I0127 16:38:18.274084 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wfhq4" Jan 27 16:38:18 crc kubenswrapper[4772]: I0127 16:38:18.281257 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-wfhq4"] Jan 27 16:38:18 crc kubenswrapper[4772]: I0127 16:38:18.377920 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-c674-account-create-update-zr9fr"] Jan 27 16:38:18 crc kubenswrapper[4772]: I0127 16:38:18.378961 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c674-account-create-update-zr9fr" Jan 27 16:38:18 crc kubenswrapper[4772]: I0127 16:38:18.381154 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Jan 27 16:38:18 crc kubenswrapper[4772]: I0127 16:38:18.390235 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c674-account-create-update-zr9fr"] Jan 27 16:38:18 crc kubenswrapper[4772]: I0127 16:38:18.413075 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmzw5\" (UniqueName: \"kubernetes.io/projected/5a6c1c65-36ca-4017-a8e2-5e22a550d601-kube-api-access-lmzw5\") pod \"placement-db-create-wfhq4\" (UID: \"5a6c1c65-36ca-4017-a8e2-5e22a550d601\") " pod="openstack/placement-db-create-wfhq4" Jan 27 16:38:18 crc kubenswrapper[4772]: I0127 16:38:18.413232 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a6c1c65-36ca-4017-a8e2-5e22a550d601-operator-scripts\") pod \"placement-db-create-wfhq4\" (UID: \"5a6c1c65-36ca-4017-a8e2-5e22a550d601\") " pod="openstack/placement-db-create-wfhq4" Jan 27 16:38:18 crc kubenswrapper[4772]: I0127 16:38:18.515254 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9bcr\" (UniqueName: \"kubernetes.io/projected/68c87e84-0237-41a3-b248-59f0e0156b81-kube-api-access-h9bcr\") pod \"placement-c674-account-create-update-zr9fr\" (UID: \"68c87e84-0237-41a3-b248-59f0e0156b81\") " pod="openstack/placement-c674-account-create-update-zr9fr" Jan 27 16:38:18 crc kubenswrapper[4772]: I0127 16:38:18.515587 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a6c1c65-36ca-4017-a8e2-5e22a550d601-operator-scripts\") pod \"placement-db-create-wfhq4\" (UID: \"5a6c1c65-36ca-4017-a8e2-5e22a550d601\") " pod="openstack/placement-db-create-wfhq4" Jan 27 16:38:18 crc kubenswrapper[4772]: I0127 16:38:18.515646 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68c87e84-0237-41a3-b248-59f0e0156b81-operator-scripts\") pod \"placement-c674-account-create-update-zr9fr\" (UID: \"68c87e84-0237-41a3-b248-59f0e0156b81\") " pod="openstack/placement-c674-account-create-update-zr9fr" Jan 27 16:38:18 crc kubenswrapper[4772]: I0127 16:38:18.515809 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmzw5\" (UniqueName: \"kubernetes.io/projected/5a6c1c65-36ca-4017-a8e2-5e22a550d601-kube-api-access-lmzw5\") pod \"placement-db-create-wfhq4\" (UID: \"5a6c1c65-36ca-4017-a8e2-5e22a550d601\") " pod="openstack/placement-db-create-wfhq4" Jan 27 16:38:18 crc kubenswrapper[4772]: I0127 16:38:18.516513 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a6c1c65-36ca-4017-a8e2-5e22a550d601-operator-scripts\") pod \"placement-db-create-wfhq4\" (UID: \"5a6c1c65-36ca-4017-a8e2-5e22a550d601\") " pod="openstack/placement-db-create-wfhq4" Jan 27 16:38:18 crc kubenswrapper[4772]: I0127 16:38:18.541238 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmzw5\" (UniqueName: \"kubernetes.io/projected/5a6c1c65-36ca-4017-a8e2-5e22a550d601-kube-api-access-lmzw5\") pod \"placement-db-create-wfhq4\" (UID: \"5a6c1c65-36ca-4017-a8e2-5e22a550d601\") " pod="openstack/placement-db-create-wfhq4" Jan 27 16:38:18 crc kubenswrapper[4772]: I0127 16:38:18.594644 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wfhq4" Jan 27 16:38:18 crc kubenswrapper[4772]: I0127 16:38:18.616949 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9bcr\" (UniqueName: \"kubernetes.io/projected/68c87e84-0237-41a3-b248-59f0e0156b81-kube-api-access-h9bcr\") pod \"placement-c674-account-create-update-zr9fr\" (UID: \"68c87e84-0237-41a3-b248-59f0e0156b81\") " pod="openstack/placement-c674-account-create-update-zr9fr" Jan 27 16:38:18 crc kubenswrapper[4772]: I0127 16:38:18.617034 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68c87e84-0237-41a3-b248-59f0e0156b81-operator-scripts\") pod \"placement-c674-account-create-update-zr9fr\" (UID: \"68c87e84-0237-41a3-b248-59f0e0156b81\") " pod="openstack/placement-c674-account-create-update-zr9fr" Jan 27 16:38:18 crc kubenswrapper[4772]: I0127 16:38:18.617735 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68c87e84-0237-41a3-b248-59f0e0156b81-operator-scripts\") pod \"placement-c674-account-create-update-zr9fr\" (UID: \"68c87e84-0237-41a3-b248-59f0e0156b81\") " pod="openstack/placement-c674-account-create-update-zr9fr" Jan 27 16:38:18 crc kubenswrapper[4772]: I0127 16:38:18.636478 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9bcr\" (UniqueName: \"kubernetes.io/projected/68c87e84-0237-41a3-b248-59f0e0156b81-kube-api-access-h9bcr\") pod \"placement-c674-account-create-update-zr9fr\" (UID: \"68c87e84-0237-41a3-b248-59f0e0156b81\") " pod="openstack/placement-c674-account-create-update-zr9fr" Jan 27 16:38:18 crc kubenswrapper[4772]: I0127 16:38:18.693582 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c674-account-create-update-zr9fr" Jan 27 16:38:19 crc kubenswrapper[4772]: I0127 16:38:19.030762 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-wfhq4"] Jan 27 16:38:19 crc kubenswrapper[4772]: I0127 16:38:19.145259 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c674-account-create-update-zr9fr"] Jan 27 16:38:19 crc kubenswrapper[4772]: W0127 16:38:19.148965 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68c87e84_0237_41a3_b248_59f0e0156b81.slice/crio-d877cfb8558a44192f82348a36e723a383dfeb73c183d53028045e743c3eb819 WatchSource:0}: Error finding container d877cfb8558a44192f82348a36e723a383dfeb73c183d53028045e743c3eb819: Status 404 returned error can't find the container with id d877cfb8558a44192f82348a36e723a383dfeb73c183d53028045e743c3eb819 Jan 27 16:38:19 crc kubenswrapper[4772]: I0127 16:38:19.175830 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wfhq4" event={"ID":"5a6c1c65-36ca-4017-a8e2-5e22a550d601","Type":"ContainerStarted","Data":"159f9ff911847523ab0387be1212efb17cae848a6dcdc1e80961565a39d1eac9"} Jan 27 16:38:19 crc kubenswrapper[4772]: I0127 16:38:19.175886 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wfhq4" event={"ID":"5a6c1c65-36ca-4017-a8e2-5e22a550d601","Type":"ContainerStarted","Data":"206780997b1e028010072b79de7396fbb1a2e61dfc93554e5d59fbd376679321"} Jan 27 16:38:19 crc kubenswrapper[4772]: I0127 16:38:19.177922 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c674-account-create-update-zr9fr" event={"ID":"68c87e84-0237-41a3-b248-59f0e0156b81","Type":"ContainerStarted","Data":"d877cfb8558a44192f82348a36e723a383dfeb73c183d53028045e743c3eb819"} Jan 27 16:38:19 crc kubenswrapper[4772]: I0127 16:38:19.196771 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-wfhq4" podStartSLOduration=1.196750614 podStartE2EDuration="1.196750614s" podCreationTimestamp="2026-01-27 16:38:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:38:19.189506128 +0000 UTC m=+5485.170115246" watchObservedRunningTime="2026-01-27 16:38:19.196750614 +0000 UTC m=+5485.177359722" Jan 27 16:38:20 crc kubenswrapper[4772]: I0127 16:38:20.186472 4772 generic.go:334] "Generic (PLEG): container finished" podID="5a6c1c65-36ca-4017-a8e2-5e22a550d601" containerID="159f9ff911847523ab0387be1212efb17cae848a6dcdc1e80961565a39d1eac9" exitCode=0 Jan 27 16:38:20 crc kubenswrapper[4772]: I0127 16:38:20.186657 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wfhq4" event={"ID":"5a6c1c65-36ca-4017-a8e2-5e22a550d601","Type":"ContainerDied","Data":"159f9ff911847523ab0387be1212efb17cae848a6dcdc1e80961565a39d1eac9"} Jan 27 16:38:20 crc kubenswrapper[4772]: I0127 16:38:20.188735 4772 generic.go:334] "Generic (PLEG): container finished" podID="68c87e84-0237-41a3-b248-59f0e0156b81" containerID="fffc3d88e7cbd76e4f8c55e9d4f80e7dfb6325460b6429ed67f89d060f3c380c" exitCode=0 Jan 27 16:38:20 crc kubenswrapper[4772]: I0127 16:38:20.188790 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c674-account-create-update-zr9fr" event={"ID":"68c87e84-0237-41a3-b248-59f0e0156b81","Type":"ContainerDied","Data":"fffc3d88e7cbd76e4f8c55e9d4f80e7dfb6325460b6429ed67f89d060f3c380c"} Jan 27 16:38:21 crc kubenswrapper[4772]: I0127 16:38:21.561717 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c674-account-create-update-zr9fr" Jan 27 16:38:21 crc kubenswrapper[4772]: I0127 16:38:21.566459 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wfhq4" Jan 27 16:38:21 crc kubenswrapper[4772]: I0127 16:38:21.669917 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9bcr\" (UniqueName: \"kubernetes.io/projected/68c87e84-0237-41a3-b248-59f0e0156b81-kube-api-access-h9bcr\") pod \"68c87e84-0237-41a3-b248-59f0e0156b81\" (UID: \"68c87e84-0237-41a3-b248-59f0e0156b81\") " Jan 27 16:38:21 crc kubenswrapper[4772]: I0127 16:38:21.669980 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a6c1c65-36ca-4017-a8e2-5e22a550d601-operator-scripts\") pod \"5a6c1c65-36ca-4017-a8e2-5e22a550d601\" (UID: \"5a6c1c65-36ca-4017-a8e2-5e22a550d601\") " Jan 27 16:38:21 crc kubenswrapper[4772]: I0127 16:38:21.670089 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmzw5\" (UniqueName: \"kubernetes.io/projected/5a6c1c65-36ca-4017-a8e2-5e22a550d601-kube-api-access-lmzw5\") pod \"5a6c1c65-36ca-4017-a8e2-5e22a550d601\" (UID: \"5a6c1c65-36ca-4017-a8e2-5e22a550d601\") " Jan 27 16:38:21 crc kubenswrapper[4772]: I0127 16:38:21.670192 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68c87e84-0237-41a3-b248-59f0e0156b81-operator-scripts\") pod \"68c87e84-0237-41a3-b248-59f0e0156b81\" (UID: \"68c87e84-0237-41a3-b248-59f0e0156b81\") " Jan 27 16:38:21 crc kubenswrapper[4772]: I0127 16:38:21.671388 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68c87e84-0237-41a3-b248-59f0e0156b81-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "68c87e84-0237-41a3-b248-59f0e0156b81" (UID: "68c87e84-0237-41a3-b248-59f0e0156b81"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:38:21 crc kubenswrapper[4772]: I0127 16:38:21.671442 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a6c1c65-36ca-4017-a8e2-5e22a550d601-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5a6c1c65-36ca-4017-a8e2-5e22a550d601" (UID: "5a6c1c65-36ca-4017-a8e2-5e22a550d601"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:38:21 crc kubenswrapper[4772]: I0127 16:38:21.676688 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68c87e84-0237-41a3-b248-59f0e0156b81-kube-api-access-h9bcr" (OuterVolumeSpecName: "kube-api-access-h9bcr") pod "68c87e84-0237-41a3-b248-59f0e0156b81" (UID: "68c87e84-0237-41a3-b248-59f0e0156b81"). InnerVolumeSpecName "kube-api-access-h9bcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:38:21 crc kubenswrapper[4772]: I0127 16:38:21.677091 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a6c1c65-36ca-4017-a8e2-5e22a550d601-kube-api-access-lmzw5" (OuterVolumeSpecName: "kube-api-access-lmzw5") pod "5a6c1c65-36ca-4017-a8e2-5e22a550d601" (UID: "5a6c1c65-36ca-4017-a8e2-5e22a550d601"). InnerVolumeSpecName "kube-api-access-lmzw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:38:21 crc kubenswrapper[4772]: I0127 16:38:21.772402 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a6c1c65-36ca-4017-a8e2-5e22a550d601-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:38:21 crc kubenswrapper[4772]: I0127 16:38:21.772435 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmzw5\" (UniqueName: \"kubernetes.io/projected/5a6c1c65-36ca-4017-a8e2-5e22a550d601-kube-api-access-lmzw5\") on node \"crc\" DevicePath \"\"" Jan 27 16:38:21 crc kubenswrapper[4772]: I0127 16:38:21.772444 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68c87e84-0237-41a3-b248-59f0e0156b81-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:38:21 crc kubenswrapper[4772]: I0127 16:38:21.772453 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9bcr\" (UniqueName: \"kubernetes.io/projected/68c87e84-0237-41a3-b248-59f0e0156b81-kube-api-access-h9bcr\") on node \"crc\" DevicePath \"\"" Jan 27 16:38:22 crc kubenswrapper[4772]: I0127 16:38:22.206965 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c674-account-create-update-zr9fr" event={"ID":"68c87e84-0237-41a3-b248-59f0e0156b81","Type":"ContainerDied","Data":"d877cfb8558a44192f82348a36e723a383dfeb73c183d53028045e743c3eb819"} Jan 27 16:38:22 crc kubenswrapper[4772]: I0127 16:38:22.207320 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d877cfb8558a44192f82348a36e723a383dfeb73c183d53028045e743c3eb819" Jan 27 16:38:22 crc kubenswrapper[4772]: I0127 16:38:22.206979 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c674-account-create-update-zr9fr" Jan 27 16:38:22 crc kubenswrapper[4772]: I0127 16:38:22.208796 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wfhq4" event={"ID":"5a6c1c65-36ca-4017-a8e2-5e22a550d601","Type":"ContainerDied","Data":"206780997b1e028010072b79de7396fbb1a2e61dfc93554e5d59fbd376679321"} Jan 27 16:38:22 crc kubenswrapper[4772]: I0127 16:38:22.208823 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wfhq4" Jan 27 16:38:22 crc kubenswrapper[4772]: I0127 16:38:22.208840 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="206780997b1e028010072b79de7396fbb1a2e61dfc93554e5d59fbd376679321" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.691993 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-545c956d45-h25qs"] Jan 27 16:38:23 crc kubenswrapper[4772]: E0127 16:38:23.692637 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68c87e84-0237-41a3-b248-59f0e0156b81" containerName="mariadb-account-create-update" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.692651 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="68c87e84-0237-41a3-b248-59f0e0156b81" containerName="mariadb-account-create-update" Jan 27 16:38:23 crc kubenswrapper[4772]: E0127 16:38:23.692669 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a6c1c65-36ca-4017-a8e2-5e22a550d601" containerName="mariadb-database-create" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.692675 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a6c1c65-36ca-4017-a8e2-5e22a550d601" containerName="mariadb-database-create" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.692849 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a6c1c65-36ca-4017-a8e2-5e22a550d601" containerName="mariadb-database-create" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.692866 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="68c87e84-0237-41a3-b248-59f0e0156b81" containerName="mariadb-account-create-update" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.693785 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-545c956d45-h25qs" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.721688 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-545c956d45-h25qs"] Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.771526 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-4pvhb"] Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.772761 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4pvhb" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.778619 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.778685 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-h5gqt" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.778700 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.791959 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-4pvhb"] Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.815369 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6da65d05-29e3-4d97-869f-d3386a45a38e-ovsdbserver-nb\") pod \"dnsmasq-dns-545c956d45-h25qs\" (UID: \"6da65d05-29e3-4d97-869f-d3386a45a38e\") " pod="openstack/dnsmasq-dns-545c956d45-h25qs" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.815573 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6da65d05-29e3-4d97-869f-d3386a45a38e-ovsdbserver-sb\") pod \"dnsmasq-dns-545c956d45-h25qs\" (UID: \"6da65d05-29e3-4d97-869f-d3386a45a38e\") " pod="openstack/dnsmasq-dns-545c956d45-h25qs" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.815700 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh8vx\" (UniqueName: \"kubernetes.io/projected/6da65d05-29e3-4d97-869f-d3386a45a38e-kube-api-access-sh8vx\") pod \"dnsmasq-dns-545c956d45-h25qs\" (UID: \"6da65d05-29e3-4d97-869f-d3386a45a38e\") " pod="openstack/dnsmasq-dns-545c956d45-h25qs" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.815759 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6da65d05-29e3-4d97-869f-d3386a45a38e-dns-svc\") pod \"dnsmasq-dns-545c956d45-h25qs\" (UID: \"6da65d05-29e3-4d97-869f-d3386a45a38e\") " pod="openstack/dnsmasq-dns-545c956d45-h25qs" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.816027 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6da65d05-29e3-4d97-869f-d3386a45a38e-config\") pod \"dnsmasq-dns-545c956d45-h25qs\" (UID: \"6da65d05-29e3-4d97-869f-d3386a45a38e\") " pod="openstack/dnsmasq-dns-545c956d45-h25qs" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.917801 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6da65d05-29e3-4d97-869f-d3386a45a38e-config\") pod \"dnsmasq-dns-545c956d45-h25qs\" (UID: \"6da65d05-29e3-4d97-869f-d3386a45a38e\") " pod="openstack/dnsmasq-dns-545c956d45-h25qs" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.917858 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b972e003-d915-4c6e-b84e-00d1f53740c1-scripts\") pod \"placement-db-sync-4pvhb\" (UID: \"b972e003-d915-4c6e-b84e-00d1f53740c1\") " pod="openstack/placement-db-sync-4pvhb" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.917915 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b972e003-d915-4c6e-b84e-00d1f53740c1-config-data\") pod \"placement-db-sync-4pvhb\" (UID: \"b972e003-d915-4c6e-b84e-00d1f53740c1\") " pod="openstack/placement-db-sync-4pvhb" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.917934 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6da65d05-29e3-4d97-869f-d3386a45a38e-ovsdbserver-nb\") pod \"dnsmasq-dns-545c956d45-h25qs\" (UID: \"6da65d05-29e3-4d97-869f-d3386a45a38e\") " pod="openstack/dnsmasq-dns-545c956d45-h25qs" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.918234 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b972e003-d915-4c6e-b84e-00d1f53740c1-combined-ca-bundle\") pod \"placement-db-sync-4pvhb\" (UID: \"b972e003-d915-4c6e-b84e-00d1f53740c1\") " pod="openstack/placement-db-sync-4pvhb" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.918419 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6da65d05-29e3-4d97-869f-d3386a45a38e-ovsdbserver-sb\") pod \"dnsmasq-dns-545c956d45-h25qs\" (UID: \"6da65d05-29e3-4d97-869f-d3386a45a38e\") " pod="openstack/dnsmasq-dns-545c956d45-h25qs" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.918491 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b972e003-d915-4c6e-b84e-00d1f53740c1-logs\") pod \"placement-db-sync-4pvhb\" (UID: \"b972e003-d915-4c6e-b84e-00d1f53740c1\") " pod="openstack/placement-db-sync-4pvhb" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.918595 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh8vx\" (UniqueName: \"kubernetes.io/projected/6da65d05-29e3-4d97-869f-d3386a45a38e-kube-api-access-sh8vx\") pod \"dnsmasq-dns-545c956d45-h25qs\" (UID: \"6da65d05-29e3-4d97-869f-d3386a45a38e\") " pod="openstack/dnsmasq-dns-545c956d45-h25qs" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.918659 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6da65d05-29e3-4d97-869f-d3386a45a38e-dns-svc\") pod \"dnsmasq-dns-545c956d45-h25qs\" (UID: \"6da65d05-29e3-4d97-869f-d3386a45a38e\") " pod="openstack/dnsmasq-dns-545c956d45-h25qs" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.918755 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj2tt\" (UniqueName: \"kubernetes.io/projected/b972e003-d915-4c6e-b84e-00d1f53740c1-kube-api-access-dj2tt\") pod \"placement-db-sync-4pvhb\" (UID: \"b972e003-d915-4c6e-b84e-00d1f53740c1\") " pod="openstack/placement-db-sync-4pvhb" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.918898 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6da65d05-29e3-4d97-869f-d3386a45a38e-config\") pod \"dnsmasq-dns-545c956d45-h25qs\" (UID: \"6da65d05-29e3-4d97-869f-d3386a45a38e\") " pod="openstack/dnsmasq-dns-545c956d45-h25qs" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.918967 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6da65d05-29e3-4d97-869f-d3386a45a38e-ovsdbserver-nb\") pod \"dnsmasq-dns-545c956d45-h25qs\" (UID: \"6da65d05-29e3-4d97-869f-d3386a45a38e\") " pod="openstack/dnsmasq-dns-545c956d45-h25qs" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.919476 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6da65d05-29e3-4d97-869f-d3386a45a38e-ovsdbserver-sb\") pod \"dnsmasq-dns-545c956d45-h25qs\" (UID: \"6da65d05-29e3-4d97-869f-d3386a45a38e\") " pod="openstack/dnsmasq-dns-545c956d45-h25qs" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.919817 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6da65d05-29e3-4d97-869f-d3386a45a38e-dns-svc\") pod \"dnsmasq-dns-545c956d45-h25qs\" (UID: \"6da65d05-29e3-4d97-869f-d3386a45a38e\") " pod="openstack/dnsmasq-dns-545c956d45-h25qs" Jan 27 16:38:23 crc kubenswrapper[4772]: I0127 16:38:23.953821 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh8vx\" (UniqueName: \"kubernetes.io/projected/6da65d05-29e3-4d97-869f-d3386a45a38e-kube-api-access-sh8vx\") pod \"dnsmasq-dns-545c956d45-h25qs\" (UID: \"6da65d05-29e3-4d97-869f-d3386a45a38e\") " pod="openstack/dnsmasq-dns-545c956d45-h25qs" Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.019033 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-545c956d45-h25qs" Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.020021 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj2tt\" (UniqueName: \"kubernetes.io/projected/b972e003-d915-4c6e-b84e-00d1f53740c1-kube-api-access-dj2tt\") pod \"placement-db-sync-4pvhb\" (UID: \"b972e003-d915-4c6e-b84e-00d1f53740c1\") " pod="openstack/placement-db-sync-4pvhb" Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.020107 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b972e003-d915-4c6e-b84e-00d1f53740c1-scripts\") pod \"placement-db-sync-4pvhb\" (UID: \"b972e003-d915-4c6e-b84e-00d1f53740c1\") " pod="openstack/placement-db-sync-4pvhb" Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.020154 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b972e003-d915-4c6e-b84e-00d1f53740c1-config-data\") pod \"placement-db-sync-4pvhb\" (UID: \"b972e003-d915-4c6e-b84e-00d1f53740c1\") " pod="openstack/placement-db-sync-4pvhb" Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.020232 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b972e003-d915-4c6e-b84e-00d1f53740c1-combined-ca-bundle\") pod \"placement-db-sync-4pvhb\" (UID: \"b972e003-d915-4c6e-b84e-00d1f53740c1\") " pod="openstack/placement-db-sync-4pvhb" Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.020282 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b972e003-d915-4c6e-b84e-00d1f53740c1-logs\") pod \"placement-db-sync-4pvhb\" (UID: \"b972e003-d915-4c6e-b84e-00d1f53740c1\") " pod="openstack/placement-db-sync-4pvhb" Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.020756 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b972e003-d915-4c6e-b84e-00d1f53740c1-logs\") pod \"placement-db-sync-4pvhb\" (UID: \"b972e003-d915-4c6e-b84e-00d1f53740c1\") " pod="openstack/placement-db-sync-4pvhb" Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.025090 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b972e003-d915-4c6e-b84e-00d1f53740c1-config-data\") pod \"placement-db-sync-4pvhb\" (UID: \"b972e003-d915-4c6e-b84e-00d1f53740c1\") " pod="openstack/placement-db-sync-4pvhb" Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.034914 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b972e003-d915-4c6e-b84e-00d1f53740c1-combined-ca-bundle\") pod \"placement-db-sync-4pvhb\" (UID: \"b972e003-d915-4c6e-b84e-00d1f53740c1\") " pod="openstack/placement-db-sync-4pvhb" Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.042597 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b972e003-d915-4c6e-b84e-00d1f53740c1-scripts\") pod \"placement-db-sync-4pvhb\" (UID: \"b972e003-d915-4c6e-b84e-00d1f53740c1\") " pod="openstack/placement-db-sync-4pvhb" Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.048724 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj2tt\" (UniqueName: \"kubernetes.io/projected/b972e003-d915-4c6e-b84e-00d1f53740c1-kube-api-access-dj2tt\") pod \"placement-db-sync-4pvhb\" (UID: \"b972e003-d915-4c6e-b84e-00d1f53740c1\") " pod="openstack/placement-db-sync-4pvhb" Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.096236 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p5mjl"] Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.097119 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4pvhb" Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.098284 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p5mjl" Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.117424 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p5mjl"] Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.225504 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7637c27-38bb-4544-a948-040122a7d526-utilities\") pod \"community-operators-p5mjl\" (UID: \"d7637c27-38bb-4544-a948-040122a7d526\") " pod="openshift-marketplace/community-operators-p5mjl" Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.225576 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7637c27-38bb-4544-a948-040122a7d526-catalog-content\") pod \"community-operators-p5mjl\" (UID: \"d7637c27-38bb-4544-a948-040122a7d526\") " pod="openshift-marketplace/community-operators-p5mjl" Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.225614 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfw6t\" (UniqueName: \"kubernetes.io/projected/d7637c27-38bb-4544-a948-040122a7d526-kube-api-access-zfw6t\") pod \"community-operators-p5mjl\" (UID: \"d7637c27-38bb-4544-a948-040122a7d526\") " pod="openshift-marketplace/community-operators-p5mjl" Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.330301 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7637c27-38bb-4544-a948-040122a7d526-utilities\") pod \"community-operators-p5mjl\" (UID: \"d7637c27-38bb-4544-a948-040122a7d526\") " pod="openshift-marketplace/community-operators-p5mjl" Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.330738 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7637c27-38bb-4544-a948-040122a7d526-catalog-content\") pod \"community-operators-p5mjl\" (UID: \"d7637c27-38bb-4544-a948-040122a7d526\") " pod="openshift-marketplace/community-operators-p5mjl" Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.330796 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfw6t\" (UniqueName: \"kubernetes.io/projected/d7637c27-38bb-4544-a948-040122a7d526-kube-api-access-zfw6t\") pod \"community-operators-p5mjl\" (UID: \"d7637c27-38bb-4544-a948-040122a7d526\") " pod="openshift-marketplace/community-operators-p5mjl" Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.331008 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7637c27-38bb-4544-a948-040122a7d526-utilities\") pod \"community-operators-p5mjl\" (UID: \"d7637c27-38bb-4544-a948-040122a7d526\") " pod="openshift-marketplace/community-operators-p5mjl" Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.331327 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7637c27-38bb-4544-a948-040122a7d526-catalog-content\") pod \"community-operators-p5mjl\" (UID: \"d7637c27-38bb-4544-a948-040122a7d526\") " pod="openshift-marketplace/community-operators-p5mjl" Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.352402 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfw6t\" (UniqueName: \"kubernetes.io/projected/d7637c27-38bb-4544-a948-040122a7d526-kube-api-access-zfw6t\") pod \"community-operators-p5mjl\" (UID: \"d7637c27-38bb-4544-a948-040122a7d526\") " pod="openshift-marketplace/community-operators-p5mjl" Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.509590 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p5mjl" Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.642815 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-4pvhb"] Jan 27 16:38:24 crc kubenswrapper[4772]: I0127 16:38:24.797527 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-545c956d45-h25qs"] Jan 27 16:38:25 crc kubenswrapper[4772]: I0127 16:38:25.052613 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p5mjl"] Jan 27 16:38:25 crc kubenswrapper[4772]: W0127 16:38:25.072819 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7637c27_38bb_4544_a948_040122a7d526.slice/crio-bde9c232190c4ff499588da71f6a6c3fb3565fcb2c79805b8745d1730e782d30 WatchSource:0}: Error finding container bde9c232190c4ff499588da71f6a6c3fb3565fcb2c79805b8745d1730e782d30: Status 404 returned error can't find the container with id bde9c232190c4ff499588da71f6a6c3fb3565fcb2c79805b8745d1730e782d30 Jan 27 16:38:25 crc kubenswrapper[4772]: E0127 16:38:25.079823 4772 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6da65d05_29e3_4d97_869f_d3386a45a38e.slice/crio-b8551d53299337e281772b00e27dddd2da7a39a65b37574b860847ac8a5c90ba.scope\": RecentStats: unable to find data in memory cache]" Jan 27 16:38:25 crc kubenswrapper[4772]: I0127 16:38:25.241570 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4pvhb" event={"ID":"b972e003-d915-4c6e-b84e-00d1f53740c1","Type":"ContainerStarted","Data":"897a6f6215480fb2b302f194b142ae62f3461ba459140cef6dbb5530febc39e7"} Jan 27 16:38:25 crc kubenswrapper[4772]: I0127 16:38:25.241635 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4pvhb" event={"ID":"b972e003-d915-4c6e-b84e-00d1f53740c1","Type":"ContainerStarted","Data":"392d0c24ac730d47db73a8d7dc9563348e7fe293fddf40e791b09a461f9aa862"} Jan 27 16:38:25 crc kubenswrapper[4772]: I0127 16:38:25.251764 4772 generic.go:334] "Generic (PLEG): container finished" podID="6da65d05-29e3-4d97-869f-d3386a45a38e" containerID="b8551d53299337e281772b00e27dddd2da7a39a65b37574b860847ac8a5c90ba" exitCode=0 Jan 27 16:38:25 crc kubenswrapper[4772]: I0127 16:38:25.251820 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-545c956d45-h25qs" event={"ID":"6da65d05-29e3-4d97-869f-d3386a45a38e","Type":"ContainerDied","Data":"b8551d53299337e281772b00e27dddd2da7a39a65b37574b860847ac8a5c90ba"} Jan 27 16:38:25 crc kubenswrapper[4772]: I0127 16:38:25.251878 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-545c956d45-h25qs" event={"ID":"6da65d05-29e3-4d97-869f-d3386a45a38e","Type":"ContainerStarted","Data":"1b2cc021d3453413ef029602e151c491355c94bca54c6c90c56cd7685dc93518"} Jan 27 16:38:25 crc kubenswrapper[4772]: I0127 16:38:25.256475 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5mjl" event={"ID":"d7637c27-38bb-4544-a948-040122a7d526","Type":"ContainerStarted","Data":"bde9c232190c4ff499588da71f6a6c3fb3565fcb2c79805b8745d1730e782d30"} Jan 27 16:38:25 crc kubenswrapper[4772]: I0127 16:38:25.261025 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-4pvhb" podStartSLOduration=2.26100833 podStartE2EDuration="2.26100833s" podCreationTimestamp="2026-01-27 16:38:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:38:25.25540107 +0000 UTC m=+5491.236010208" watchObservedRunningTime="2026-01-27 16:38:25.26100833 +0000 UTC m=+5491.241617418" Jan 27 16:38:26 crc kubenswrapper[4772]: I0127 16:38:26.270953 4772 generic.go:334] "Generic (PLEG): container finished" podID="d7637c27-38bb-4544-a948-040122a7d526" containerID="5f3e19efa8f94154ea817a07d57fc0bef91de6b3d9985a454f4b799359042687" exitCode=0 Jan 27 16:38:26 crc kubenswrapper[4772]: I0127 16:38:26.271019 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5mjl" event={"ID":"d7637c27-38bb-4544-a948-040122a7d526","Type":"ContainerDied","Data":"5f3e19efa8f94154ea817a07d57fc0bef91de6b3d9985a454f4b799359042687"} Jan 27 16:38:26 crc kubenswrapper[4772]: I0127 16:38:26.275746 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 16:38:26 crc kubenswrapper[4772]: I0127 16:38:26.279034 4772 generic.go:334] "Generic (PLEG): container finished" podID="b972e003-d915-4c6e-b84e-00d1f53740c1" containerID="897a6f6215480fb2b302f194b142ae62f3461ba459140cef6dbb5530febc39e7" exitCode=0 Jan 27 16:38:26 crc kubenswrapper[4772]: I0127 16:38:26.279194 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4pvhb" event={"ID":"b972e003-d915-4c6e-b84e-00d1f53740c1","Type":"ContainerDied","Data":"897a6f6215480fb2b302f194b142ae62f3461ba459140cef6dbb5530febc39e7"} Jan 27 16:38:26 crc kubenswrapper[4772]: I0127 16:38:26.281231 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-545c956d45-h25qs" event={"ID":"6da65d05-29e3-4d97-869f-d3386a45a38e","Type":"ContainerStarted","Data":"54f0c375865810e4cb819d2019e6c8ab7827175c7d633ed00deb75c9600fb4b9"} Jan 27 16:38:26 crc kubenswrapper[4772]: I0127 16:38:26.281419 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-545c956d45-h25qs" Jan 27 16:38:26 crc kubenswrapper[4772]: I0127 16:38:26.326406 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-545c956d45-h25qs" podStartSLOduration=3.326387646 podStartE2EDuration="3.326387646s" podCreationTimestamp="2026-01-27 16:38:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:38:26.322799564 +0000 UTC m=+5492.303408682" watchObservedRunningTime="2026-01-27 16:38:26.326387646 +0000 UTC m=+5492.306996744" Jan 27 16:38:27 crc kubenswrapper[4772]: I0127 16:38:27.673082 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4pvhb" Jan 27 16:38:27 crc kubenswrapper[4772]: I0127 16:38:27.835041 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b972e003-d915-4c6e-b84e-00d1f53740c1-config-data\") pod \"b972e003-d915-4c6e-b84e-00d1f53740c1\" (UID: \"b972e003-d915-4c6e-b84e-00d1f53740c1\") " Jan 27 16:38:27 crc kubenswrapper[4772]: I0127 16:38:27.835102 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b972e003-d915-4c6e-b84e-00d1f53740c1-logs\") pod \"b972e003-d915-4c6e-b84e-00d1f53740c1\" (UID: \"b972e003-d915-4c6e-b84e-00d1f53740c1\") " Jan 27 16:38:27 crc kubenswrapper[4772]: I0127 16:38:27.835225 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dj2tt\" (UniqueName: \"kubernetes.io/projected/b972e003-d915-4c6e-b84e-00d1f53740c1-kube-api-access-dj2tt\") pod \"b972e003-d915-4c6e-b84e-00d1f53740c1\" (UID: \"b972e003-d915-4c6e-b84e-00d1f53740c1\") " Jan 27 16:38:27 crc kubenswrapper[4772]: I0127 16:38:27.835412 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b972e003-d915-4c6e-b84e-00d1f53740c1-combined-ca-bundle\") pod \"b972e003-d915-4c6e-b84e-00d1f53740c1\" (UID: \"b972e003-d915-4c6e-b84e-00d1f53740c1\") " Jan 27 16:38:27 crc kubenswrapper[4772]: I0127 16:38:27.835458 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b972e003-d915-4c6e-b84e-00d1f53740c1-scripts\") pod \"b972e003-d915-4c6e-b84e-00d1f53740c1\" (UID: \"b972e003-d915-4c6e-b84e-00d1f53740c1\") " Jan 27 16:38:27 crc kubenswrapper[4772]: I0127 16:38:27.835510 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b972e003-d915-4c6e-b84e-00d1f53740c1-logs" (OuterVolumeSpecName: "logs") pod "b972e003-d915-4c6e-b84e-00d1f53740c1" (UID: "b972e003-d915-4c6e-b84e-00d1f53740c1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:38:27 crc kubenswrapper[4772]: I0127 16:38:27.835962 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b972e003-d915-4c6e-b84e-00d1f53740c1-logs\") on node \"crc\" DevicePath \"\"" Jan 27 16:38:27 crc kubenswrapper[4772]: I0127 16:38:27.846290 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b972e003-d915-4c6e-b84e-00d1f53740c1-kube-api-access-dj2tt" (OuterVolumeSpecName: "kube-api-access-dj2tt") pod "b972e003-d915-4c6e-b84e-00d1f53740c1" (UID: "b972e003-d915-4c6e-b84e-00d1f53740c1"). InnerVolumeSpecName "kube-api-access-dj2tt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:38:27 crc kubenswrapper[4772]: I0127 16:38:27.846363 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b972e003-d915-4c6e-b84e-00d1f53740c1-scripts" (OuterVolumeSpecName: "scripts") pod "b972e003-d915-4c6e-b84e-00d1f53740c1" (UID: "b972e003-d915-4c6e-b84e-00d1f53740c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:38:27 crc kubenswrapper[4772]: I0127 16:38:27.862510 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b972e003-d915-4c6e-b84e-00d1f53740c1-config-data" (OuterVolumeSpecName: "config-data") pod "b972e003-d915-4c6e-b84e-00d1f53740c1" (UID: "b972e003-d915-4c6e-b84e-00d1f53740c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:38:27 crc kubenswrapper[4772]: I0127 16:38:27.865365 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b972e003-d915-4c6e-b84e-00d1f53740c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b972e003-d915-4c6e-b84e-00d1f53740c1" (UID: "b972e003-d915-4c6e-b84e-00d1f53740c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:38:27 crc kubenswrapper[4772]: I0127 16:38:27.912895 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6ffd9fc5c6-99g52"] Jan 27 16:38:27 crc kubenswrapper[4772]: E0127 16:38:27.913323 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b972e003-d915-4c6e-b84e-00d1f53740c1" containerName="placement-db-sync" Jan 27 16:38:27 crc kubenswrapper[4772]: I0127 16:38:27.913340 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="b972e003-d915-4c6e-b84e-00d1f53740c1" containerName="placement-db-sync" Jan 27 16:38:27 crc kubenswrapper[4772]: I0127 16:38:27.913707 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="b972e003-d915-4c6e-b84e-00d1f53740c1" containerName="placement-db-sync" Jan 27 16:38:27 crc kubenswrapper[4772]: I0127 16:38:27.914684 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6ffd9fc5c6-99g52" Jan 27 16:38:27 crc kubenswrapper[4772]: I0127 16:38:27.929056 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6ffd9fc5c6-99g52"] Jan 27 16:38:27 crc kubenswrapper[4772]: I0127 16:38:27.940490 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b972e003-d915-4c6e-b84e-00d1f53740c1-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:38:27 crc kubenswrapper[4772]: I0127 16:38:27.940534 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b972e003-d915-4c6e-b84e-00d1f53740c1-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:38:27 crc kubenswrapper[4772]: I0127 16:38:27.940547 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dj2tt\" (UniqueName: \"kubernetes.io/projected/b972e003-d915-4c6e-b84e-00d1f53740c1-kube-api-access-dj2tt\") on node \"crc\" DevicePath \"\"" Jan 27 16:38:27 crc kubenswrapper[4772]: I0127 16:38:27.940559 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b972e003-d915-4c6e-b84e-00d1f53740c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:38:28 crc kubenswrapper[4772]: I0127 16:38:28.041986 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8daf690-d375-4d0a-b763-4b610aaeac45-config-data\") pod \"placement-6ffd9fc5c6-99g52\" (UID: \"c8daf690-d375-4d0a-b763-4b610aaeac45\") " pod="openstack/placement-6ffd9fc5c6-99g52" Jan 27 16:38:28 crc kubenswrapper[4772]: I0127 16:38:28.042044 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8daf690-d375-4d0a-b763-4b610aaeac45-logs\") pod \"placement-6ffd9fc5c6-99g52\" (UID: \"c8daf690-d375-4d0a-b763-4b610aaeac45\") " pod="openstack/placement-6ffd9fc5c6-99g52" Jan 27 16:38:28 crc kubenswrapper[4772]: I0127 16:38:28.042138 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kx9v\" (UniqueName: \"kubernetes.io/projected/c8daf690-d375-4d0a-b763-4b610aaeac45-kube-api-access-7kx9v\") pod \"placement-6ffd9fc5c6-99g52\" (UID: \"c8daf690-d375-4d0a-b763-4b610aaeac45\") " pod="openstack/placement-6ffd9fc5c6-99g52" Jan 27 16:38:28 crc kubenswrapper[4772]: I0127 16:38:28.042496 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8daf690-d375-4d0a-b763-4b610aaeac45-scripts\") pod \"placement-6ffd9fc5c6-99g52\" (UID: \"c8daf690-d375-4d0a-b763-4b610aaeac45\") " pod="openstack/placement-6ffd9fc5c6-99g52" Jan 27 16:38:28 crc kubenswrapper[4772]: I0127 16:38:28.042564 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8daf690-d375-4d0a-b763-4b610aaeac45-combined-ca-bundle\") pod \"placement-6ffd9fc5c6-99g52\" (UID: \"c8daf690-d375-4d0a-b763-4b610aaeac45\") " pod="openstack/placement-6ffd9fc5c6-99g52" Jan 27 16:38:28 crc kubenswrapper[4772]: I0127 16:38:28.144763 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8daf690-d375-4d0a-b763-4b610aaeac45-scripts\") pod \"placement-6ffd9fc5c6-99g52\" (UID: \"c8daf690-d375-4d0a-b763-4b610aaeac45\") " pod="openstack/placement-6ffd9fc5c6-99g52" Jan 27 16:38:28 crc kubenswrapper[4772]: I0127 16:38:28.144876 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8daf690-d375-4d0a-b763-4b610aaeac45-combined-ca-bundle\") pod \"placement-6ffd9fc5c6-99g52\" (UID: \"c8daf690-d375-4d0a-b763-4b610aaeac45\") " pod="openstack/placement-6ffd9fc5c6-99g52" Jan 27 16:38:28 crc kubenswrapper[4772]: I0127 16:38:28.144930 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8daf690-d375-4d0a-b763-4b610aaeac45-config-data\") pod \"placement-6ffd9fc5c6-99g52\" (UID: \"c8daf690-d375-4d0a-b763-4b610aaeac45\") " pod="openstack/placement-6ffd9fc5c6-99g52" Jan 27 16:38:28 crc kubenswrapper[4772]: I0127 16:38:28.144961 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8daf690-d375-4d0a-b763-4b610aaeac45-logs\") pod \"placement-6ffd9fc5c6-99g52\" (UID: \"c8daf690-d375-4d0a-b763-4b610aaeac45\") " pod="openstack/placement-6ffd9fc5c6-99g52" Jan 27 16:38:28 crc kubenswrapper[4772]: I0127 16:38:28.144990 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kx9v\" (UniqueName: \"kubernetes.io/projected/c8daf690-d375-4d0a-b763-4b610aaeac45-kube-api-access-7kx9v\") pod \"placement-6ffd9fc5c6-99g52\" (UID: \"c8daf690-d375-4d0a-b763-4b610aaeac45\") " pod="openstack/placement-6ffd9fc5c6-99g52" Jan 27 16:38:28 crc kubenswrapper[4772]: I0127 16:38:28.145557 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8daf690-d375-4d0a-b763-4b610aaeac45-logs\") pod \"placement-6ffd9fc5c6-99g52\" (UID: \"c8daf690-d375-4d0a-b763-4b610aaeac45\") " pod="openstack/placement-6ffd9fc5c6-99g52" Jan 27 16:38:28 crc kubenswrapper[4772]: I0127 16:38:28.148445 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8daf690-d375-4d0a-b763-4b610aaeac45-scripts\") pod \"placement-6ffd9fc5c6-99g52\" (UID: \"c8daf690-d375-4d0a-b763-4b610aaeac45\") " pod="openstack/placement-6ffd9fc5c6-99g52" Jan 27 16:38:28 crc kubenswrapper[4772]: I0127 16:38:28.148636 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8daf690-d375-4d0a-b763-4b610aaeac45-config-data\") pod \"placement-6ffd9fc5c6-99g52\" (UID: \"c8daf690-d375-4d0a-b763-4b610aaeac45\") " pod="openstack/placement-6ffd9fc5c6-99g52" Jan 27 16:38:28 crc kubenswrapper[4772]: I0127 16:38:28.148909 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8daf690-d375-4d0a-b763-4b610aaeac45-combined-ca-bundle\") pod \"placement-6ffd9fc5c6-99g52\" (UID: \"c8daf690-d375-4d0a-b763-4b610aaeac45\") " pod="openstack/placement-6ffd9fc5c6-99g52" Jan 27 16:38:28 crc kubenswrapper[4772]: I0127 16:38:28.161739 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kx9v\" (UniqueName: \"kubernetes.io/projected/c8daf690-d375-4d0a-b763-4b610aaeac45-kube-api-access-7kx9v\") pod \"placement-6ffd9fc5c6-99g52\" (UID: \"c8daf690-d375-4d0a-b763-4b610aaeac45\") " pod="openstack/placement-6ffd9fc5c6-99g52" Jan 27 16:38:28 crc kubenswrapper[4772]: I0127 16:38:28.250389 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6ffd9fc5c6-99g52" Jan 27 16:38:28 crc kubenswrapper[4772]: I0127 16:38:28.298044 4772 generic.go:334] "Generic (PLEG): container finished" podID="d7637c27-38bb-4544-a948-040122a7d526" containerID="73a1bf579e1cad299ee8dcf75b2e0f5d995e618d33eecb56a1a42a2021cd5d48" exitCode=0 Jan 27 16:38:28 crc kubenswrapper[4772]: I0127 16:38:28.298115 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5mjl" event={"ID":"d7637c27-38bb-4544-a948-040122a7d526","Type":"ContainerDied","Data":"73a1bf579e1cad299ee8dcf75b2e0f5d995e618d33eecb56a1a42a2021cd5d48"} Jan 27 16:38:28 crc kubenswrapper[4772]: I0127 16:38:28.302427 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4pvhb" event={"ID":"b972e003-d915-4c6e-b84e-00d1f53740c1","Type":"ContainerDied","Data":"392d0c24ac730d47db73a8d7dc9563348e7fe293fddf40e791b09a461f9aa862"} Jan 27 16:38:28 crc kubenswrapper[4772]: I0127 16:38:28.302464 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="392d0c24ac730d47db73a8d7dc9563348e7fe293fddf40e791b09a461f9aa862" Jan 27 16:38:28 crc kubenswrapper[4772]: I0127 16:38:28.302516 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4pvhb" Jan 27 16:38:28 crc kubenswrapper[4772]: I0127 16:38:28.715459 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6ffd9fc5c6-99g52"] Jan 27 16:38:28 crc kubenswrapper[4772]: W0127 16:38:28.723904 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8daf690_d375_4d0a_b763_4b610aaeac45.slice/crio-cede86ef2b62c284020dea92755cd8f1c45ccf853891c9353a1aaaba262f274e WatchSource:0}: Error finding container cede86ef2b62c284020dea92755cd8f1c45ccf853891c9353a1aaaba262f274e: Status 404 returned error can't find the container with id cede86ef2b62c284020dea92755cd8f1c45ccf853891c9353a1aaaba262f274e Jan 27 16:38:29 crc kubenswrapper[4772]: I0127 16:38:29.314628 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6ffd9fc5c6-99g52" event={"ID":"c8daf690-d375-4d0a-b763-4b610aaeac45","Type":"ContainerStarted","Data":"a4f7397ce195b3ef684fdb7e4e7bdf987f0aad1b8524c25545d425f4bc248bbd"} Jan 27 16:38:29 crc kubenswrapper[4772]: I0127 16:38:29.314926 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6ffd9fc5c6-99g52" Jan 27 16:38:29 crc kubenswrapper[4772]: I0127 16:38:29.314940 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6ffd9fc5c6-99g52" event={"ID":"c8daf690-d375-4d0a-b763-4b610aaeac45","Type":"ContainerStarted","Data":"2a9baa311ee399537353526ea6225df9a78dbebd8cc28695cf4db8dba9e9bfb8"} Jan 27 16:38:29 crc kubenswrapper[4772]: I0127 16:38:29.314951 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6ffd9fc5c6-99g52" event={"ID":"c8daf690-d375-4d0a-b763-4b610aaeac45","Type":"ContainerStarted","Data":"cede86ef2b62c284020dea92755cd8f1c45ccf853891c9353a1aaaba262f274e"} Jan 27 16:38:29 crc kubenswrapper[4772]: I0127 16:38:29.318114 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5mjl" event={"ID":"d7637c27-38bb-4544-a948-040122a7d526","Type":"ContainerStarted","Data":"1c35a579a5c75625296e16ce590ed79bef653fbfe286f6422ef23d2e956388a7"} Jan 27 16:38:29 crc kubenswrapper[4772]: I0127 16:38:29.342936 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6ffd9fc5c6-99g52" podStartSLOduration=2.342894988 podStartE2EDuration="2.342894988s" podCreationTimestamp="2026-01-27 16:38:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:38:29.334486028 +0000 UTC m=+5495.315095146" watchObservedRunningTime="2026-01-27 16:38:29.342894988 +0000 UTC m=+5495.323504086" Jan 27 16:38:29 crc kubenswrapper[4772]: I0127 16:38:29.361923 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p5mjl" podStartSLOduration=2.916434584 podStartE2EDuration="5.36190654s" podCreationTimestamp="2026-01-27 16:38:24 +0000 UTC" firstStartedPulling="2026-01-27 16:38:26.275488445 +0000 UTC m=+5492.256097543" lastFinishedPulling="2026-01-27 16:38:28.720960391 +0000 UTC m=+5494.701569499" observedRunningTime="2026-01-27 16:38:29.359366647 +0000 UTC m=+5495.339975765" watchObservedRunningTime="2026-01-27 16:38:29.36190654 +0000 UTC m=+5495.342515638" Jan 27 16:38:30 crc kubenswrapper[4772]: I0127 16:38:30.335801 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6ffd9fc5c6-99g52" Jan 27 16:38:34 crc kubenswrapper[4772]: I0127 16:38:34.020377 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-545c956d45-h25qs" Jan 27 16:38:34 crc kubenswrapper[4772]: I0127 16:38:34.094035 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d645dd9d5-2pwb9"] Jan 27 16:38:34 crc kubenswrapper[4772]: I0127 16:38:34.094342 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" podUID="e289d3f6-26ba-4306-a7f0-bf95513c9068" containerName="dnsmasq-dns" containerID="cri-o://de4a1855c8f97732a1ffd2229841c1e180e5f299f24cbbca78a154e8db2ccc0d" gracePeriod=10 Jan 27 16:38:34 crc kubenswrapper[4772]: I0127 16:38:34.389090 4772 generic.go:334] "Generic (PLEG): container finished" podID="e289d3f6-26ba-4306-a7f0-bf95513c9068" containerID="de4a1855c8f97732a1ffd2229841c1e180e5f299f24cbbca78a154e8db2ccc0d" exitCode=0 Jan 27 16:38:34 crc kubenswrapper[4772]: I0127 16:38:34.389136 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" event={"ID":"e289d3f6-26ba-4306-a7f0-bf95513c9068","Type":"ContainerDied","Data":"de4a1855c8f97732a1ffd2229841c1e180e5f299f24cbbca78a154e8db2ccc0d"} Jan 27 16:38:34 crc kubenswrapper[4772]: I0127 16:38:34.510652 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p5mjl" Jan 27 16:38:34 crc kubenswrapper[4772]: I0127 16:38:34.510722 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p5mjl" Jan 27 16:38:34 crc kubenswrapper[4772]: I0127 16:38:34.554717 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p5mjl" Jan 27 16:38:34 crc kubenswrapper[4772]: I0127 16:38:34.582024 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" Jan 27 16:38:34 crc kubenswrapper[4772]: I0127 16:38:34.766826 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e289d3f6-26ba-4306-a7f0-bf95513c9068-dns-svc\") pod \"e289d3f6-26ba-4306-a7f0-bf95513c9068\" (UID: \"e289d3f6-26ba-4306-a7f0-bf95513c9068\") " Jan 27 16:38:34 crc kubenswrapper[4772]: I0127 16:38:34.766935 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e289d3f6-26ba-4306-a7f0-bf95513c9068-ovsdbserver-nb\") pod \"e289d3f6-26ba-4306-a7f0-bf95513c9068\" (UID: \"e289d3f6-26ba-4306-a7f0-bf95513c9068\") " Jan 27 16:38:34 crc kubenswrapper[4772]: I0127 16:38:34.766985 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e289d3f6-26ba-4306-a7f0-bf95513c9068-ovsdbserver-sb\") pod \"e289d3f6-26ba-4306-a7f0-bf95513c9068\" (UID: \"e289d3f6-26ba-4306-a7f0-bf95513c9068\") " Jan 27 16:38:34 crc kubenswrapper[4772]: I0127 16:38:34.767003 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e289d3f6-26ba-4306-a7f0-bf95513c9068-config\") pod \"e289d3f6-26ba-4306-a7f0-bf95513c9068\" (UID: \"e289d3f6-26ba-4306-a7f0-bf95513c9068\") " Jan 27 16:38:34 crc kubenswrapper[4772]: I0127 16:38:34.767027 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7grf\" (UniqueName: \"kubernetes.io/projected/e289d3f6-26ba-4306-a7f0-bf95513c9068-kube-api-access-q7grf\") pod \"e289d3f6-26ba-4306-a7f0-bf95513c9068\" (UID: \"e289d3f6-26ba-4306-a7f0-bf95513c9068\") " Jan 27 16:38:34 crc kubenswrapper[4772]: I0127 16:38:34.774716 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e289d3f6-26ba-4306-a7f0-bf95513c9068-kube-api-access-q7grf" (OuterVolumeSpecName: "kube-api-access-q7grf") pod "e289d3f6-26ba-4306-a7f0-bf95513c9068" (UID: "e289d3f6-26ba-4306-a7f0-bf95513c9068"). InnerVolumeSpecName "kube-api-access-q7grf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:38:34 crc kubenswrapper[4772]: I0127 16:38:34.812079 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e289d3f6-26ba-4306-a7f0-bf95513c9068-config" (OuterVolumeSpecName: "config") pod "e289d3f6-26ba-4306-a7f0-bf95513c9068" (UID: "e289d3f6-26ba-4306-a7f0-bf95513c9068"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:38:34 crc kubenswrapper[4772]: I0127 16:38:34.813371 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e289d3f6-26ba-4306-a7f0-bf95513c9068-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e289d3f6-26ba-4306-a7f0-bf95513c9068" (UID: "e289d3f6-26ba-4306-a7f0-bf95513c9068"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:38:34 crc kubenswrapper[4772]: I0127 16:38:34.816226 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e289d3f6-26ba-4306-a7f0-bf95513c9068-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e289d3f6-26ba-4306-a7f0-bf95513c9068" (UID: "e289d3f6-26ba-4306-a7f0-bf95513c9068"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:38:34 crc kubenswrapper[4772]: I0127 16:38:34.824481 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e289d3f6-26ba-4306-a7f0-bf95513c9068-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e289d3f6-26ba-4306-a7f0-bf95513c9068" (UID: "e289d3f6-26ba-4306-a7f0-bf95513c9068"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:38:34 crc kubenswrapper[4772]: I0127 16:38:34.870275 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e289d3f6-26ba-4306-a7f0-bf95513c9068-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 16:38:34 crc kubenswrapper[4772]: I0127 16:38:34.870321 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e289d3f6-26ba-4306-a7f0-bf95513c9068-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 27 16:38:34 crc kubenswrapper[4772]: I0127 16:38:34.870336 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e289d3f6-26ba-4306-a7f0-bf95513c9068-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 27 16:38:34 crc kubenswrapper[4772]: I0127 16:38:34.870348 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e289d3f6-26ba-4306-a7f0-bf95513c9068-config\") on node \"crc\" DevicePath \"\"" Jan 27 16:38:34 crc kubenswrapper[4772]: I0127 16:38:34.870361 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7grf\" (UniqueName: \"kubernetes.io/projected/e289d3f6-26ba-4306-a7f0-bf95513c9068-kube-api-access-q7grf\") on node \"crc\" DevicePath \"\"" Jan 27 16:38:35 crc kubenswrapper[4772]: I0127 16:38:35.399637 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" event={"ID":"e289d3f6-26ba-4306-a7f0-bf95513c9068","Type":"ContainerDied","Data":"5dc3b25da49a88c2e2d810e005946006b5e97f2c6fb476e48c50681bedaf4daf"} Jan 27 16:38:35 crc kubenswrapper[4772]: I0127 16:38:35.399660 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" Jan 27 16:38:35 crc kubenswrapper[4772]: I0127 16:38:35.399998 4772 scope.go:117] "RemoveContainer" containerID="de4a1855c8f97732a1ffd2229841c1e180e5f299f24cbbca78a154e8db2ccc0d" Jan 27 16:38:35 crc kubenswrapper[4772]: I0127 16:38:35.420355 4772 scope.go:117] "RemoveContainer" containerID="cfd61d60b4611b222a3dcb5c92c6682ed0c719a75f85f6d5730b053534f9b1b0" Jan 27 16:38:35 crc kubenswrapper[4772]: I0127 16:38:35.442502 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d645dd9d5-2pwb9"] Jan 27 16:38:35 crc kubenswrapper[4772]: I0127 16:38:35.452488 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d645dd9d5-2pwb9"] Jan 27 16:38:35 crc kubenswrapper[4772]: I0127 16:38:35.476572 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p5mjl" Jan 27 16:38:35 crc kubenswrapper[4772]: I0127 16:38:35.538138 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p5mjl"] Jan 27 16:38:36 crc kubenswrapper[4772]: I0127 16:38:36.682488 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e289d3f6-26ba-4306-a7f0-bf95513c9068" path="/var/lib/kubelet/pods/e289d3f6-26ba-4306-a7f0-bf95513c9068/volumes" Jan 27 16:38:37 crc kubenswrapper[4772]: I0127 16:38:37.423930 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p5mjl" podUID="d7637c27-38bb-4544-a948-040122a7d526" containerName="registry-server" containerID="cri-o://1c35a579a5c75625296e16ce590ed79bef653fbfe286f6422ef23d2e956388a7" gracePeriod=2 Jan 27 16:38:37 crc kubenswrapper[4772]: I0127 16:38:37.883260 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p5mjl" Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.036642 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7637c27-38bb-4544-a948-040122a7d526-catalog-content\") pod \"d7637c27-38bb-4544-a948-040122a7d526\" (UID: \"d7637c27-38bb-4544-a948-040122a7d526\") " Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.036767 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfw6t\" (UniqueName: \"kubernetes.io/projected/d7637c27-38bb-4544-a948-040122a7d526-kube-api-access-zfw6t\") pod \"d7637c27-38bb-4544-a948-040122a7d526\" (UID: \"d7637c27-38bb-4544-a948-040122a7d526\") " Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.036914 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7637c27-38bb-4544-a948-040122a7d526-utilities\") pod \"d7637c27-38bb-4544-a948-040122a7d526\" (UID: \"d7637c27-38bb-4544-a948-040122a7d526\") " Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.037670 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7637c27-38bb-4544-a948-040122a7d526-utilities" (OuterVolumeSpecName: "utilities") pod "d7637c27-38bb-4544-a948-040122a7d526" (UID: "d7637c27-38bb-4544-a948-040122a7d526"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.045416 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7637c27-38bb-4544-a948-040122a7d526-kube-api-access-zfw6t" (OuterVolumeSpecName: "kube-api-access-zfw6t") pod "d7637c27-38bb-4544-a948-040122a7d526" (UID: "d7637c27-38bb-4544-a948-040122a7d526"). InnerVolumeSpecName "kube-api-access-zfw6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.086197 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7637c27-38bb-4544-a948-040122a7d526-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d7637c27-38bb-4544-a948-040122a7d526" (UID: "d7637c27-38bb-4544-a948-040122a7d526"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.138806 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7637c27-38bb-4544-a948-040122a7d526-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.138849 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7637c27-38bb-4544-a948-040122a7d526-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.138862 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfw6t\" (UniqueName: \"kubernetes.io/projected/d7637c27-38bb-4544-a948-040122a7d526-kube-api-access-zfw6t\") on node \"crc\" DevicePath \"\"" Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.451740 4772 generic.go:334] "Generic (PLEG): container finished" podID="d7637c27-38bb-4544-a948-040122a7d526" containerID="1c35a579a5c75625296e16ce590ed79bef653fbfe286f6422ef23d2e956388a7" exitCode=0 Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.451788 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5mjl" event={"ID":"d7637c27-38bb-4544-a948-040122a7d526","Type":"ContainerDied","Data":"1c35a579a5c75625296e16ce590ed79bef653fbfe286f6422ef23d2e956388a7"} Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.451821 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5mjl" event={"ID":"d7637c27-38bb-4544-a948-040122a7d526","Type":"ContainerDied","Data":"bde9c232190c4ff499588da71f6a6c3fb3565fcb2c79805b8745d1730e782d30"} Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.451838 4772 scope.go:117] "RemoveContainer" containerID="1c35a579a5c75625296e16ce590ed79bef653fbfe286f6422ef23d2e956388a7" Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.451845 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p5mjl" Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.475538 4772 scope.go:117] "RemoveContainer" containerID="73a1bf579e1cad299ee8dcf75b2e0f5d995e618d33eecb56a1a42a2021cd5d48" Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.492922 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p5mjl"] Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.501052 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p5mjl"] Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.506336 4772 scope.go:117] "RemoveContainer" containerID="5f3e19efa8f94154ea817a07d57fc0bef91de6b3d9985a454f4b799359042687" Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.552857 4772 scope.go:117] "RemoveContainer" containerID="1c35a579a5c75625296e16ce590ed79bef653fbfe286f6422ef23d2e956388a7" Jan 27 16:38:38 crc kubenswrapper[4772]: E0127 16:38:38.553432 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c35a579a5c75625296e16ce590ed79bef653fbfe286f6422ef23d2e956388a7\": container with ID starting with 1c35a579a5c75625296e16ce590ed79bef653fbfe286f6422ef23d2e956388a7 not found: ID does not exist" containerID="1c35a579a5c75625296e16ce590ed79bef653fbfe286f6422ef23d2e956388a7" Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.553528 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c35a579a5c75625296e16ce590ed79bef653fbfe286f6422ef23d2e956388a7"} err="failed to get container status \"1c35a579a5c75625296e16ce590ed79bef653fbfe286f6422ef23d2e956388a7\": rpc error: code = NotFound desc = could not find container \"1c35a579a5c75625296e16ce590ed79bef653fbfe286f6422ef23d2e956388a7\": container with ID starting with 1c35a579a5c75625296e16ce590ed79bef653fbfe286f6422ef23d2e956388a7 not found: ID does not exist" Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.553556 4772 scope.go:117] "RemoveContainer" containerID="73a1bf579e1cad299ee8dcf75b2e0f5d995e618d33eecb56a1a42a2021cd5d48" Jan 27 16:38:38 crc kubenswrapper[4772]: E0127 16:38:38.553965 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73a1bf579e1cad299ee8dcf75b2e0f5d995e618d33eecb56a1a42a2021cd5d48\": container with ID starting with 73a1bf579e1cad299ee8dcf75b2e0f5d995e618d33eecb56a1a42a2021cd5d48 not found: ID does not exist" containerID="73a1bf579e1cad299ee8dcf75b2e0f5d995e618d33eecb56a1a42a2021cd5d48" Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.554001 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73a1bf579e1cad299ee8dcf75b2e0f5d995e618d33eecb56a1a42a2021cd5d48"} err="failed to get container status \"73a1bf579e1cad299ee8dcf75b2e0f5d995e618d33eecb56a1a42a2021cd5d48\": rpc error: code = NotFound desc = could not find container \"73a1bf579e1cad299ee8dcf75b2e0f5d995e618d33eecb56a1a42a2021cd5d48\": container with ID starting with 73a1bf579e1cad299ee8dcf75b2e0f5d995e618d33eecb56a1a42a2021cd5d48 not found: ID does not exist" Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.554022 4772 scope.go:117] "RemoveContainer" containerID="5f3e19efa8f94154ea817a07d57fc0bef91de6b3d9985a454f4b799359042687" Jan 27 16:38:38 crc kubenswrapper[4772]: E0127 16:38:38.554471 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f3e19efa8f94154ea817a07d57fc0bef91de6b3d9985a454f4b799359042687\": container with ID starting with 5f3e19efa8f94154ea817a07d57fc0bef91de6b3d9985a454f4b799359042687 not found: ID does not exist" containerID="5f3e19efa8f94154ea817a07d57fc0bef91de6b3d9985a454f4b799359042687" Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.554524 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f3e19efa8f94154ea817a07d57fc0bef91de6b3d9985a454f4b799359042687"} err="failed to get container status \"5f3e19efa8f94154ea817a07d57fc0bef91de6b3d9985a454f4b799359042687\": rpc error: code = NotFound desc = could not find container \"5f3e19efa8f94154ea817a07d57fc0bef91de6b3d9985a454f4b799359042687\": container with ID starting with 5f3e19efa8f94154ea817a07d57fc0bef91de6b3d9985a454f4b799359042687 not found: ID does not exist" Jan 27 16:38:38 crc kubenswrapper[4772]: I0127 16:38:38.674509 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7637c27-38bb-4544-a948-040122a7d526" path="/var/lib/kubelet/pods/d7637c27-38bb-4544-a948-040122a7d526/volumes" Jan 27 16:38:39 crc kubenswrapper[4772]: I0127 16:38:39.473877 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d645dd9d5-2pwb9" podUID="e289d3f6-26ba-4306-a7f0-bf95513c9068" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.34:5353: i/o timeout" Jan 27 16:38:59 crc kubenswrapper[4772]: I0127 16:38:59.319329 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6ffd9fc5c6-99g52" Jan 27 16:39:00 crc kubenswrapper[4772]: I0127 16:39:00.319398 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6ffd9fc5c6-99g52" Jan 27 16:39:06 crc kubenswrapper[4772]: I0127 16:39:06.992132 4772 scope.go:117] "RemoveContainer" containerID="f4958ec9744454169fb58baabe20204293a4ff4790174c9c2079b9801fd7028d" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.568455 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-5g8c9"] Jan 27 16:39:20 crc kubenswrapper[4772]: E0127 16:39:20.569324 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e289d3f6-26ba-4306-a7f0-bf95513c9068" containerName="dnsmasq-dns" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.569338 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e289d3f6-26ba-4306-a7f0-bf95513c9068" containerName="dnsmasq-dns" Jan 27 16:39:20 crc kubenswrapper[4772]: E0127 16:39:20.569369 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7637c27-38bb-4544-a948-040122a7d526" containerName="extract-utilities" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.569378 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7637c27-38bb-4544-a948-040122a7d526" containerName="extract-utilities" Jan 27 16:39:20 crc kubenswrapper[4772]: E0127 16:39:20.569391 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e289d3f6-26ba-4306-a7f0-bf95513c9068" containerName="init" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.569399 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e289d3f6-26ba-4306-a7f0-bf95513c9068" containerName="init" Jan 27 16:39:20 crc kubenswrapper[4772]: E0127 16:39:20.569418 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7637c27-38bb-4544-a948-040122a7d526" containerName="extract-content" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.569426 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7637c27-38bb-4544-a948-040122a7d526" containerName="extract-content" Jan 27 16:39:20 crc kubenswrapper[4772]: E0127 16:39:20.569437 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7637c27-38bb-4544-a948-040122a7d526" containerName="registry-server" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.569445 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7637c27-38bb-4544-a948-040122a7d526" containerName="registry-server" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.569636 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7637c27-38bb-4544-a948-040122a7d526" containerName="registry-server" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.569663 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="e289d3f6-26ba-4306-a7f0-bf95513c9068" containerName="dnsmasq-dns" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.570400 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5g8c9" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.581765 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-5g8c9"] Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.649468 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-wmvjd"] Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.650541 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wmvjd" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.684417 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-wmvjd"] Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.739057 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfb8z\" (UniqueName: \"kubernetes.io/projected/952d9a1e-efbf-4617-94af-b5ad42cce494-kube-api-access-bfb8z\") pod \"nova-api-db-create-5g8c9\" (UID: \"952d9a1e-efbf-4617-94af-b5ad42cce494\") " pod="openstack/nova-api-db-create-5g8c9" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.739141 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgrrr\" (UniqueName: \"kubernetes.io/projected/81b3d773-720e-42c5-af9e-abddc2180ac7-kube-api-access-bgrrr\") pod \"nova-cell0-db-create-wmvjd\" (UID: \"81b3d773-720e-42c5-af9e-abddc2180ac7\") " pod="openstack/nova-cell0-db-create-wmvjd" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.739293 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81b3d773-720e-42c5-af9e-abddc2180ac7-operator-scripts\") pod \"nova-cell0-db-create-wmvjd\" (UID: \"81b3d773-720e-42c5-af9e-abddc2180ac7\") " pod="openstack/nova-cell0-db-create-wmvjd" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.739475 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/952d9a1e-efbf-4617-94af-b5ad42cce494-operator-scripts\") pod \"nova-api-db-create-5g8c9\" (UID: \"952d9a1e-efbf-4617-94af-b5ad42cce494\") " pod="openstack/nova-api-db-create-5g8c9" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.758494 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-fxhzl"] Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.759683 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fxhzl" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.772084 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-fxhzl"] Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.786238 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-e098-account-create-update-s8qnw"] Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.787531 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e098-account-create-update-s8qnw" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.789748 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.795056 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-e098-account-create-update-s8qnw"] Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.841064 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81b3d773-720e-42c5-af9e-abddc2180ac7-operator-scripts\") pod \"nova-cell0-db-create-wmvjd\" (UID: \"81b3d773-720e-42c5-af9e-abddc2180ac7\") " pod="openstack/nova-cell0-db-create-wmvjd" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.841191 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/952d9a1e-efbf-4617-94af-b5ad42cce494-operator-scripts\") pod \"nova-api-db-create-5g8c9\" (UID: \"952d9a1e-efbf-4617-94af-b5ad42cce494\") " pod="openstack/nova-api-db-create-5g8c9" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.841914 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfb8z\" (UniqueName: \"kubernetes.io/projected/952d9a1e-efbf-4617-94af-b5ad42cce494-kube-api-access-bfb8z\") pod \"nova-api-db-create-5g8c9\" (UID: \"952d9a1e-efbf-4617-94af-b5ad42cce494\") " pod="openstack/nova-api-db-create-5g8c9" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.841990 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/952d9a1e-efbf-4617-94af-b5ad42cce494-operator-scripts\") pod \"nova-api-db-create-5g8c9\" (UID: \"952d9a1e-efbf-4617-94af-b5ad42cce494\") " pod="openstack/nova-api-db-create-5g8c9" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.842005 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgrrr\" (UniqueName: \"kubernetes.io/projected/81b3d773-720e-42c5-af9e-abddc2180ac7-kube-api-access-bgrrr\") pod \"nova-cell0-db-create-wmvjd\" (UID: \"81b3d773-720e-42c5-af9e-abddc2180ac7\") " pod="openstack/nova-cell0-db-create-wmvjd" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.842457 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81b3d773-720e-42c5-af9e-abddc2180ac7-operator-scripts\") pod \"nova-cell0-db-create-wmvjd\" (UID: \"81b3d773-720e-42c5-af9e-abddc2180ac7\") " pod="openstack/nova-cell0-db-create-wmvjd" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.860616 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfb8z\" (UniqueName: \"kubernetes.io/projected/952d9a1e-efbf-4617-94af-b5ad42cce494-kube-api-access-bfb8z\") pod \"nova-api-db-create-5g8c9\" (UID: \"952d9a1e-efbf-4617-94af-b5ad42cce494\") " pod="openstack/nova-api-db-create-5g8c9" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.864282 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgrrr\" (UniqueName: \"kubernetes.io/projected/81b3d773-720e-42c5-af9e-abddc2180ac7-kube-api-access-bgrrr\") pod \"nova-cell0-db-create-wmvjd\" (UID: \"81b3d773-720e-42c5-af9e-abddc2180ac7\") " pod="openstack/nova-cell0-db-create-wmvjd" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.893055 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5g8c9" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.943594 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e808813a-e588-4fb9-a15d-588d94a4cd59-operator-scripts\") pod \"nova-api-e098-account-create-update-s8qnw\" (UID: \"e808813a-e588-4fb9-a15d-588d94a4cd59\") " pod="openstack/nova-api-e098-account-create-update-s8qnw" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.943699 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzxl7\" (UniqueName: \"kubernetes.io/projected/84c86af5-fd1f-4c53-8978-2b436db59b2a-kube-api-access-pzxl7\") pod \"nova-cell1-db-create-fxhzl\" (UID: \"84c86af5-fd1f-4c53-8978-2b436db59b2a\") " pod="openstack/nova-cell1-db-create-fxhzl" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.943727 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kb9n\" (UniqueName: \"kubernetes.io/projected/e808813a-e588-4fb9-a15d-588d94a4cd59-kube-api-access-7kb9n\") pod \"nova-api-e098-account-create-update-s8qnw\" (UID: \"e808813a-e588-4fb9-a15d-588d94a4cd59\") " pod="openstack/nova-api-e098-account-create-update-s8qnw" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.943813 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84c86af5-fd1f-4c53-8978-2b436db59b2a-operator-scripts\") pod \"nova-cell1-db-create-fxhzl\" (UID: \"84c86af5-fd1f-4c53-8978-2b436db59b2a\") " pod="openstack/nova-cell1-db-create-fxhzl" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.962718 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-a20c-account-create-update-7pf8l"] Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.964097 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a20c-account-create-update-7pf8l" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.967648 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.968150 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wmvjd" Jan 27 16:39:20 crc kubenswrapper[4772]: I0127 16:39:20.980069 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-a20c-account-create-update-7pf8l"] Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.045286 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-267vz\" (UniqueName: \"kubernetes.io/projected/86f6e427-99ce-4873-bacc-697edca3d34e-kube-api-access-267vz\") pod \"nova-cell0-a20c-account-create-update-7pf8l\" (UID: \"86f6e427-99ce-4873-bacc-697edca3d34e\") " pod="openstack/nova-cell0-a20c-account-create-update-7pf8l" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.045551 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzxl7\" (UniqueName: \"kubernetes.io/projected/84c86af5-fd1f-4c53-8978-2b436db59b2a-kube-api-access-pzxl7\") pod \"nova-cell1-db-create-fxhzl\" (UID: \"84c86af5-fd1f-4c53-8978-2b436db59b2a\") " pod="openstack/nova-cell1-db-create-fxhzl" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.045573 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kb9n\" (UniqueName: \"kubernetes.io/projected/e808813a-e588-4fb9-a15d-588d94a4cd59-kube-api-access-7kb9n\") pod \"nova-api-e098-account-create-update-s8qnw\" (UID: \"e808813a-e588-4fb9-a15d-588d94a4cd59\") " pod="openstack/nova-api-e098-account-create-update-s8qnw" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.045654 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86f6e427-99ce-4873-bacc-697edca3d34e-operator-scripts\") pod \"nova-cell0-a20c-account-create-update-7pf8l\" (UID: \"86f6e427-99ce-4873-bacc-697edca3d34e\") " pod="openstack/nova-cell0-a20c-account-create-update-7pf8l" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.045674 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84c86af5-fd1f-4c53-8978-2b436db59b2a-operator-scripts\") pod \"nova-cell1-db-create-fxhzl\" (UID: \"84c86af5-fd1f-4c53-8978-2b436db59b2a\") " pod="openstack/nova-cell1-db-create-fxhzl" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.045734 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e808813a-e588-4fb9-a15d-588d94a4cd59-operator-scripts\") pod \"nova-api-e098-account-create-update-s8qnw\" (UID: \"e808813a-e588-4fb9-a15d-588d94a4cd59\") " pod="openstack/nova-api-e098-account-create-update-s8qnw" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.046541 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e808813a-e588-4fb9-a15d-588d94a4cd59-operator-scripts\") pod \"nova-api-e098-account-create-update-s8qnw\" (UID: \"e808813a-e588-4fb9-a15d-588d94a4cd59\") " pod="openstack/nova-api-e098-account-create-update-s8qnw" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.046606 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84c86af5-fd1f-4c53-8978-2b436db59b2a-operator-scripts\") pod \"nova-cell1-db-create-fxhzl\" (UID: \"84c86af5-fd1f-4c53-8978-2b436db59b2a\") " pod="openstack/nova-cell1-db-create-fxhzl" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.066778 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kb9n\" (UniqueName: \"kubernetes.io/projected/e808813a-e588-4fb9-a15d-588d94a4cd59-kube-api-access-7kb9n\") pod \"nova-api-e098-account-create-update-s8qnw\" (UID: \"e808813a-e588-4fb9-a15d-588d94a4cd59\") " pod="openstack/nova-api-e098-account-create-update-s8qnw" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.067194 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzxl7\" (UniqueName: \"kubernetes.io/projected/84c86af5-fd1f-4c53-8978-2b436db59b2a-kube-api-access-pzxl7\") pod \"nova-cell1-db-create-fxhzl\" (UID: \"84c86af5-fd1f-4c53-8978-2b436db59b2a\") " pod="openstack/nova-cell1-db-create-fxhzl" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.079887 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fxhzl" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.106158 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e098-account-create-update-s8qnw" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.148148 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-267vz\" (UniqueName: \"kubernetes.io/projected/86f6e427-99ce-4873-bacc-697edca3d34e-kube-api-access-267vz\") pod \"nova-cell0-a20c-account-create-update-7pf8l\" (UID: \"86f6e427-99ce-4873-bacc-697edca3d34e\") " pod="openstack/nova-cell0-a20c-account-create-update-7pf8l" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.148315 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86f6e427-99ce-4873-bacc-697edca3d34e-operator-scripts\") pod \"nova-cell0-a20c-account-create-update-7pf8l\" (UID: \"86f6e427-99ce-4873-bacc-697edca3d34e\") " pod="openstack/nova-cell0-a20c-account-create-update-7pf8l" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.149280 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86f6e427-99ce-4873-bacc-697edca3d34e-operator-scripts\") pod \"nova-cell0-a20c-account-create-update-7pf8l\" (UID: \"86f6e427-99ce-4873-bacc-697edca3d34e\") " pod="openstack/nova-cell0-a20c-account-create-update-7pf8l" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.176662 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-267vz\" (UniqueName: \"kubernetes.io/projected/86f6e427-99ce-4873-bacc-697edca3d34e-kube-api-access-267vz\") pod \"nova-cell0-a20c-account-create-update-7pf8l\" (UID: \"86f6e427-99ce-4873-bacc-697edca3d34e\") " pod="openstack/nova-cell0-a20c-account-create-update-7pf8l" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.176667 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-a831-account-create-update-bq6jl"] Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.178145 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-a831-account-create-update-bq6jl" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.181583 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.185916 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-a831-account-create-update-bq6jl"] Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.352729 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4545581d-5f56-406d-938f-c3b073fdcbce-operator-scripts\") pod \"nova-cell1-a831-account-create-update-bq6jl\" (UID: \"4545581d-5f56-406d-938f-c3b073fdcbce\") " pod="openstack/nova-cell1-a831-account-create-update-bq6jl" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.352773 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rddx\" (UniqueName: \"kubernetes.io/projected/4545581d-5f56-406d-938f-c3b073fdcbce-kube-api-access-2rddx\") pod \"nova-cell1-a831-account-create-update-bq6jl\" (UID: \"4545581d-5f56-406d-938f-c3b073fdcbce\") " pod="openstack/nova-cell1-a831-account-create-update-bq6jl" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.370641 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a20c-account-create-update-7pf8l" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.386494 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-5g8c9"] Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.457822 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4545581d-5f56-406d-938f-c3b073fdcbce-operator-scripts\") pod \"nova-cell1-a831-account-create-update-bq6jl\" (UID: \"4545581d-5f56-406d-938f-c3b073fdcbce\") " pod="openstack/nova-cell1-a831-account-create-update-bq6jl" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.458199 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rddx\" (UniqueName: \"kubernetes.io/projected/4545581d-5f56-406d-938f-c3b073fdcbce-kube-api-access-2rddx\") pod \"nova-cell1-a831-account-create-update-bq6jl\" (UID: \"4545581d-5f56-406d-938f-c3b073fdcbce\") " pod="openstack/nova-cell1-a831-account-create-update-bq6jl" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.458512 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4545581d-5f56-406d-938f-c3b073fdcbce-operator-scripts\") pod \"nova-cell1-a831-account-create-update-bq6jl\" (UID: \"4545581d-5f56-406d-938f-c3b073fdcbce\") " pod="openstack/nova-cell1-a831-account-create-update-bq6jl" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.485565 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rddx\" (UniqueName: \"kubernetes.io/projected/4545581d-5f56-406d-938f-c3b073fdcbce-kube-api-access-2rddx\") pod \"nova-cell1-a831-account-create-update-bq6jl\" (UID: \"4545581d-5f56-406d-938f-c3b073fdcbce\") " pod="openstack/nova-cell1-a831-account-create-update-bq6jl" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.491422 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-wmvjd"] Jan 27 16:39:21 crc kubenswrapper[4772]: W0127 16:39:21.500615 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81b3d773_720e_42c5_af9e_abddc2180ac7.slice/crio-74aafe8a75a36a2526a4ba6482968b03120e31e15b066c2c8da6074effe77189 WatchSource:0}: Error finding container 74aafe8a75a36a2526a4ba6482968b03120e31e15b066c2c8da6074effe77189: Status 404 returned error can't find the container with id 74aafe8a75a36a2526a4ba6482968b03120e31e15b066c2c8da6074effe77189 Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.503150 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-a831-account-create-update-bq6jl" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.663369 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-e098-account-create-update-s8qnw"] Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.713030 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-fxhzl"] Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.817330 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-a20c-account-create-update-7pf8l"] Jan 27 16:39:21 crc kubenswrapper[4772]: W0127 16:39:21.818106 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86f6e427_99ce_4873_bacc_697edca3d34e.slice/crio-84670f9b7280d1f9e799307995f1eb490a345f26043746f3bece904916c6717d WatchSource:0}: Error finding container 84670f9b7280d1f9e799307995f1eb490a345f26043746f3bece904916c6717d: Status 404 returned error can't find the container with id 84670f9b7280d1f9e799307995f1eb490a345f26043746f3bece904916c6717d Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.819791 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wmvjd" event={"ID":"81b3d773-720e-42c5-af9e-abddc2180ac7","Type":"ContainerStarted","Data":"92b70acb37a2142424102ba84a18ba6908a56b707c2a540be2311cd899ea872a"} Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.819831 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wmvjd" event={"ID":"81b3d773-720e-42c5-af9e-abddc2180ac7","Type":"ContainerStarted","Data":"74aafe8a75a36a2526a4ba6482968b03120e31e15b066c2c8da6074effe77189"} Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.824632 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5g8c9" event={"ID":"952d9a1e-efbf-4617-94af-b5ad42cce494","Type":"ContainerStarted","Data":"cba54a4b30bd539d4b80fc206c2394f0309c6182e722300037332e6e39b8dedb"} Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.824683 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5g8c9" event={"ID":"952d9a1e-efbf-4617-94af-b5ad42cce494","Type":"ContainerStarted","Data":"681ac82fdd84aa4e5d105dd78003928a30e49c90e7fd8b4edd049cfc6b685eff"} Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.826896 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-e098-account-create-update-s8qnw" event={"ID":"e808813a-e588-4fb9-a15d-588d94a4cd59","Type":"ContainerStarted","Data":"2a8731d36142a485e01f80917d9b4cd3eee05d2c848cd6dc5346e696b8c548cb"} Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.834643 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-fxhzl" event={"ID":"84c86af5-fd1f-4c53-8978-2b436db59b2a","Type":"ContainerStarted","Data":"9e71adf9731dfffc02261215475cf37fa6421ef0d218583ccfa458e659a17ae9"} Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.844150 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-wmvjd" podStartSLOduration=1.844123566 podStartE2EDuration="1.844123566s" podCreationTimestamp="2026-01-27 16:39:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:39:21.833981367 +0000 UTC m=+5547.814590475" watchObservedRunningTime="2026-01-27 16:39:21.844123566 +0000 UTC m=+5547.824732674" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.852998 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-5g8c9" podStartSLOduration=1.852975878 podStartE2EDuration="1.852975878s" podCreationTimestamp="2026-01-27 16:39:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:39:21.84811532 +0000 UTC m=+5547.828724428" watchObservedRunningTime="2026-01-27 16:39:21.852975878 +0000 UTC m=+5547.833584976" Jan 27 16:39:21 crc kubenswrapper[4772]: I0127 16:39:21.960097 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-a831-account-create-update-bq6jl"] Jan 27 16:39:22 crc kubenswrapper[4772]: W0127 16:39:22.035667 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4545581d_5f56_406d_938f_c3b073fdcbce.slice/crio-e185de8351d3e108267a03aa84e999cda5068ec389181389ffbe49bdab27c598 WatchSource:0}: Error finding container e185de8351d3e108267a03aa84e999cda5068ec389181389ffbe49bdab27c598: Status 404 returned error can't find the container with id e185de8351d3e108267a03aa84e999cda5068ec389181389ffbe49bdab27c598 Jan 27 16:39:22 crc kubenswrapper[4772]: I0127 16:39:22.854812 4772 generic.go:334] "Generic (PLEG): container finished" podID="81b3d773-720e-42c5-af9e-abddc2180ac7" containerID="92b70acb37a2142424102ba84a18ba6908a56b707c2a540be2311cd899ea872a" exitCode=0 Jan 27 16:39:22 crc kubenswrapper[4772]: I0127 16:39:22.855198 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wmvjd" event={"ID":"81b3d773-720e-42c5-af9e-abddc2180ac7","Type":"ContainerDied","Data":"92b70acb37a2142424102ba84a18ba6908a56b707c2a540be2311cd899ea872a"} Jan 27 16:39:22 crc kubenswrapper[4772]: I0127 16:39:22.857803 4772 generic.go:334] "Generic (PLEG): container finished" podID="86f6e427-99ce-4873-bacc-697edca3d34e" containerID="83a6d65ba439c93f6ff663d0a68308cf85e00e3f86836413a3077a6bf72f351a" exitCode=0 Jan 27 16:39:22 crc kubenswrapper[4772]: I0127 16:39:22.857866 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a20c-account-create-update-7pf8l" event={"ID":"86f6e427-99ce-4873-bacc-697edca3d34e","Type":"ContainerDied","Data":"83a6d65ba439c93f6ff663d0a68308cf85e00e3f86836413a3077a6bf72f351a"} Jan 27 16:39:22 crc kubenswrapper[4772]: I0127 16:39:22.857885 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a20c-account-create-update-7pf8l" event={"ID":"86f6e427-99ce-4873-bacc-697edca3d34e","Type":"ContainerStarted","Data":"84670f9b7280d1f9e799307995f1eb490a345f26043746f3bece904916c6717d"} Jan 27 16:39:22 crc kubenswrapper[4772]: I0127 16:39:22.859658 4772 generic.go:334] "Generic (PLEG): container finished" podID="4545581d-5f56-406d-938f-c3b073fdcbce" containerID="0f3d5e3b05300094485e382bc00ab51b2c741fade4c1a775474788ecee8633d6" exitCode=0 Jan 27 16:39:22 crc kubenswrapper[4772]: I0127 16:39:22.859697 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-a831-account-create-update-bq6jl" event={"ID":"4545581d-5f56-406d-938f-c3b073fdcbce","Type":"ContainerDied","Data":"0f3d5e3b05300094485e382bc00ab51b2c741fade4c1a775474788ecee8633d6"} Jan 27 16:39:22 crc kubenswrapper[4772]: I0127 16:39:22.859725 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-a831-account-create-update-bq6jl" event={"ID":"4545581d-5f56-406d-938f-c3b073fdcbce","Type":"ContainerStarted","Data":"e185de8351d3e108267a03aa84e999cda5068ec389181389ffbe49bdab27c598"} Jan 27 16:39:22 crc kubenswrapper[4772]: I0127 16:39:22.861293 4772 generic.go:334] "Generic (PLEG): container finished" podID="952d9a1e-efbf-4617-94af-b5ad42cce494" containerID="cba54a4b30bd539d4b80fc206c2394f0309c6182e722300037332e6e39b8dedb" exitCode=0 Jan 27 16:39:22 crc kubenswrapper[4772]: I0127 16:39:22.861358 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5g8c9" event={"ID":"952d9a1e-efbf-4617-94af-b5ad42cce494","Type":"ContainerDied","Data":"cba54a4b30bd539d4b80fc206c2394f0309c6182e722300037332e6e39b8dedb"} Jan 27 16:39:22 crc kubenswrapper[4772]: I0127 16:39:22.863053 4772 generic.go:334] "Generic (PLEG): container finished" podID="e808813a-e588-4fb9-a15d-588d94a4cd59" containerID="b2dc1b1a95dcd405b12ae74c124c7458f4aa13c48525ad20496373458b83b670" exitCode=0 Jan 27 16:39:22 crc kubenswrapper[4772]: I0127 16:39:22.863148 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-e098-account-create-update-s8qnw" event={"ID":"e808813a-e588-4fb9-a15d-588d94a4cd59","Type":"ContainerDied","Data":"b2dc1b1a95dcd405b12ae74c124c7458f4aa13c48525ad20496373458b83b670"} Jan 27 16:39:22 crc kubenswrapper[4772]: I0127 16:39:22.869059 4772 generic.go:334] "Generic (PLEG): container finished" podID="84c86af5-fd1f-4c53-8978-2b436db59b2a" containerID="f05adee94e87980f08389d2716dd0d5a92a148aff23a1b26c057f06fd19c6f9b" exitCode=0 Jan 27 16:39:22 crc kubenswrapper[4772]: I0127 16:39:22.869105 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-fxhzl" event={"ID":"84c86af5-fd1f-4c53-8978-2b436db59b2a","Type":"ContainerDied","Data":"f05adee94e87980f08389d2716dd0d5a92a148aff23a1b26c057f06fd19c6f9b"} Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.229267 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e098-account-create-update-s8qnw" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.423734 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e808813a-e588-4fb9-a15d-588d94a4cd59-operator-scripts\") pod \"e808813a-e588-4fb9-a15d-588d94a4cd59\" (UID: \"e808813a-e588-4fb9-a15d-588d94a4cd59\") " Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.423912 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kb9n\" (UniqueName: \"kubernetes.io/projected/e808813a-e588-4fb9-a15d-588d94a4cd59-kube-api-access-7kb9n\") pod \"e808813a-e588-4fb9-a15d-588d94a4cd59\" (UID: \"e808813a-e588-4fb9-a15d-588d94a4cd59\") " Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.424268 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e808813a-e588-4fb9-a15d-588d94a4cd59-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e808813a-e588-4fb9-a15d-588d94a4cd59" (UID: "e808813a-e588-4fb9-a15d-588d94a4cd59"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.424538 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e808813a-e588-4fb9-a15d-588d94a4cd59-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.425993 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5g8c9" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.429464 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e808813a-e588-4fb9-a15d-588d94a4cd59-kube-api-access-7kb9n" (OuterVolumeSpecName: "kube-api-access-7kb9n") pod "e808813a-e588-4fb9-a15d-588d94a4cd59" (UID: "e808813a-e588-4fb9-a15d-588d94a4cd59"). InnerVolumeSpecName "kube-api-access-7kb9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.430503 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a20c-account-create-update-7pf8l" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.444758 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wmvjd" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.448603 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-a831-account-create-update-bq6jl" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.462051 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fxhzl" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.525930 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kb9n\" (UniqueName: \"kubernetes.io/projected/e808813a-e588-4fb9-a15d-588d94a4cd59-kube-api-access-7kb9n\") on node \"crc\" DevicePath \"\"" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.626974 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-267vz\" (UniqueName: \"kubernetes.io/projected/86f6e427-99ce-4873-bacc-697edca3d34e-kube-api-access-267vz\") pod \"86f6e427-99ce-4873-bacc-697edca3d34e\" (UID: \"86f6e427-99ce-4873-bacc-697edca3d34e\") " Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.627040 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84c86af5-fd1f-4c53-8978-2b436db59b2a-operator-scripts\") pod \"84c86af5-fd1f-4c53-8978-2b436db59b2a\" (UID: \"84c86af5-fd1f-4c53-8978-2b436db59b2a\") " Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.627068 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzxl7\" (UniqueName: \"kubernetes.io/projected/84c86af5-fd1f-4c53-8978-2b436db59b2a-kube-api-access-pzxl7\") pod \"84c86af5-fd1f-4c53-8978-2b436db59b2a\" (UID: \"84c86af5-fd1f-4c53-8978-2b436db59b2a\") " Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.627127 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgrrr\" (UniqueName: \"kubernetes.io/projected/81b3d773-720e-42c5-af9e-abddc2180ac7-kube-api-access-bgrrr\") pod \"81b3d773-720e-42c5-af9e-abddc2180ac7\" (UID: \"81b3d773-720e-42c5-af9e-abddc2180ac7\") " Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.627186 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfb8z\" (UniqueName: \"kubernetes.io/projected/952d9a1e-efbf-4617-94af-b5ad42cce494-kube-api-access-bfb8z\") pod \"952d9a1e-efbf-4617-94af-b5ad42cce494\" (UID: \"952d9a1e-efbf-4617-94af-b5ad42cce494\") " Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.627211 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81b3d773-720e-42c5-af9e-abddc2180ac7-operator-scripts\") pod \"81b3d773-720e-42c5-af9e-abddc2180ac7\" (UID: \"81b3d773-720e-42c5-af9e-abddc2180ac7\") " Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.627244 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4545581d-5f56-406d-938f-c3b073fdcbce-operator-scripts\") pod \"4545581d-5f56-406d-938f-c3b073fdcbce\" (UID: \"4545581d-5f56-406d-938f-c3b073fdcbce\") " Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.627263 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/952d9a1e-efbf-4617-94af-b5ad42cce494-operator-scripts\") pod \"952d9a1e-efbf-4617-94af-b5ad42cce494\" (UID: \"952d9a1e-efbf-4617-94af-b5ad42cce494\") " Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.627323 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rddx\" (UniqueName: \"kubernetes.io/projected/4545581d-5f56-406d-938f-c3b073fdcbce-kube-api-access-2rddx\") pod \"4545581d-5f56-406d-938f-c3b073fdcbce\" (UID: \"4545581d-5f56-406d-938f-c3b073fdcbce\") " Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.627358 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86f6e427-99ce-4873-bacc-697edca3d34e-operator-scripts\") pod \"86f6e427-99ce-4873-bacc-697edca3d34e\" (UID: \"86f6e427-99ce-4873-bacc-697edca3d34e\") " Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.627914 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84c86af5-fd1f-4c53-8978-2b436db59b2a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "84c86af5-fd1f-4c53-8978-2b436db59b2a" (UID: "84c86af5-fd1f-4c53-8978-2b436db59b2a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.628012 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86f6e427-99ce-4873-bacc-697edca3d34e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "86f6e427-99ce-4873-bacc-697edca3d34e" (UID: "86f6e427-99ce-4873-bacc-697edca3d34e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.628083 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4545581d-5f56-406d-938f-c3b073fdcbce-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4545581d-5f56-406d-938f-c3b073fdcbce" (UID: "4545581d-5f56-406d-938f-c3b073fdcbce"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.628327 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/952d9a1e-efbf-4617-94af-b5ad42cce494-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "952d9a1e-efbf-4617-94af-b5ad42cce494" (UID: "952d9a1e-efbf-4617-94af-b5ad42cce494"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.628493 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81b3d773-720e-42c5-af9e-abddc2180ac7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "81b3d773-720e-42c5-af9e-abddc2180ac7" (UID: "81b3d773-720e-42c5-af9e-abddc2180ac7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.630975 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84c86af5-fd1f-4c53-8978-2b436db59b2a-kube-api-access-pzxl7" (OuterVolumeSpecName: "kube-api-access-pzxl7") pod "84c86af5-fd1f-4c53-8978-2b436db59b2a" (UID: "84c86af5-fd1f-4c53-8978-2b436db59b2a"). InnerVolumeSpecName "kube-api-access-pzxl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.631343 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86f6e427-99ce-4873-bacc-697edca3d34e-kube-api-access-267vz" (OuterVolumeSpecName: "kube-api-access-267vz") pod "86f6e427-99ce-4873-bacc-697edca3d34e" (UID: "86f6e427-99ce-4873-bacc-697edca3d34e"). InnerVolumeSpecName "kube-api-access-267vz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.631409 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/952d9a1e-efbf-4617-94af-b5ad42cce494-kube-api-access-bfb8z" (OuterVolumeSpecName: "kube-api-access-bfb8z") pod "952d9a1e-efbf-4617-94af-b5ad42cce494" (UID: "952d9a1e-efbf-4617-94af-b5ad42cce494"). InnerVolumeSpecName "kube-api-access-bfb8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.631913 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81b3d773-720e-42c5-af9e-abddc2180ac7-kube-api-access-bgrrr" (OuterVolumeSpecName: "kube-api-access-bgrrr") pod "81b3d773-720e-42c5-af9e-abddc2180ac7" (UID: "81b3d773-720e-42c5-af9e-abddc2180ac7"). InnerVolumeSpecName "kube-api-access-bgrrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.633262 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4545581d-5f56-406d-938f-c3b073fdcbce-kube-api-access-2rddx" (OuterVolumeSpecName: "kube-api-access-2rddx") pod "4545581d-5f56-406d-938f-c3b073fdcbce" (UID: "4545581d-5f56-406d-938f-c3b073fdcbce"). InnerVolumeSpecName "kube-api-access-2rddx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.729344 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rddx\" (UniqueName: \"kubernetes.io/projected/4545581d-5f56-406d-938f-c3b073fdcbce-kube-api-access-2rddx\") on node \"crc\" DevicePath \"\"" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.729436 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86f6e427-99ce-4873-bacc-697edca3d34e-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.729460 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-267vz\" (UniqueName: \"kubernetes.io/projected/86f6e427-99ce-4873-bacc-697edca3d34e-kube-api-access-267vz\") on node \"crc\" DevicePath \"\"" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.729478 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84c86af5-fd1f-4c53-8978-2b436db59b2a-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.729488 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzxl7\" (UniqueName: \"kubernetes.io/projected/84c86af5-fd1f-4c53-8978-2b436db59b2a-kube-api-access-pzxl7\") on node \"crc\" DevicePath \"\"" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.729497 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgrrr\" (UniqueName: \"kubernetes.io/projected/81b3d773-720e-42c5-af9e-abddc2180ac7-kube-api-access-bgrrr\") on node \"crc\" DevicePath \"\"" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.729507 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfb8z\" (UniqueName: \"kubernetes.io/projected/952d9a1e-efbf-4617-94af-b5ad42cce494-kube-api-access-bfb8z\") on node \"crc\" DevicePath \"\"" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.729515 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81b3d773-720e-42c5-af9e-abddc2180ac7-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.729524 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4545581d-5f56-406d-938f-c3b073fdcbce-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.729532 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/952d9a1e-efbf-4617-94af-b5ad42cce494-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.887427 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-a831-account-create-update-bq6jl" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.887421 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-a831-account-create-update-bq6jl" event={"ID":"4545581d-5f56-406d-938f-c3b073fdcbce","Type":"ContainerDied","Data":"e185de8351d3e108267a03aa84e999cda5068ec389181389ffbe49bdab27c598"} Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.887998 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e185de8351d3e108267a03aa84e999cda5068ec389181389ffbe49bdab27c598" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.888991 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5g8c9" event={"ID":"952d9a1e-efbf-4617-94af-b5ad42cce494","Type":"ContainerDied","Data":"681ac82fdd84aa4e5d105dd78003928a30e49c90e7fd8b4edd049cfc6b685eff"} Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.889019 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="681ac82fdd84aa4e5d105dd78003928a30e49c90e7fd8b4edd049cfc6b685eff" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.889068 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5g8c9" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.891568 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-e098-account-create-update-s8qnw" event={"ID":"e808813a-e588-4fb9-a15d-588d94a4cd59","Type":"ContainerDied","Data":"2a8731d36142a485e01f80917d9b4cd3eee05d2c848cd6dc5346e696b8c548cb"} Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.891590 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a8731d36142a485e01f80917d9b4cd3eee05d2c848cd6dc5346e696b8c548cb" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.891627 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e098-account-create-update-s8qnw" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.894513 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-fxhzl" event={"ID":"84c86af5-fd1f-4c53-8978-2b436db59b2a","Type":"ContainerDied","Data":"9e71adf9731dfffc02261215475cf37fa6421ef0d218583ccfa458e659a17ae9"} Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.894553 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e71adf9731dfffc02261215475cf37fa6421ef0d218583ccfa458e659a17ae9" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.894579 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fxhzl" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.896513 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wmvjd" event={"ID":"81b3d773-720e-42c5-af9e-abddc2180ac7","Type":"ContainerDied","Data":"74aafe8a75a36a2526a4ba6482968b03120e31e15b066c2c8da6074effe77189"} Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.896535 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wmvjd" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.896549 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74aafe8a75a36a2526a4ba6482968b03120e31e15b066c2c8da6074effe77189" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.899370 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a20c-account-create-update-7pf8l" event={"ID":"86f6e427-99ce-4873-bacc-697edca3d34e","Type":"ContainerDied","Data":"84670f9b7280d1f9e799307995f1eb490a345f26043746f3bece904916c6717d"} Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.899391 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84670f9b7280d1f9e799307995f1eb490a345f26043746f3bece904916c6717d" Jan 27 16:39:24 crc kubenswrapper[4772]: I0127 16:39:24.899512 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a20c-account-create-update-7pf8l" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.141066 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k56t7"] Jan 27 16:39:26 crc kubenswrapper[4772]: E0127 16:39:26.141822 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e808813a-e588-4fb9-a15d-588d94a4cd59" containerName="mariadb-account-create-update" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.141838 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e808813a-e588-4fb9-a15d-588d94a4cd59" containerName="mariadb-account-create-update" Jan 27 16:39:26 crc kubenswrapper[4772]: E0127 16:39:26.141874 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86f6e427-99ce-4873-bacc-697edca3d34e" containerName="mariadb-account-create-update" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.141882 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="86f6e427-99ce-4873-bacc-697edca3d34e" containerName="mariadb-account-create-update" Jan 27 16:39:26 crc kubenswrapper[4772]: E0127 16:39:26.141894 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="952d9a1e-efbf-4617-94af-b5ad42cce494" containerName="mariadb-database-create" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.141901 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="952d9a1e-efbf-4617-94af-b5ad42cce494" containerName="mariadb-database-create" Jan 27 16:39:26 crc kubenswrapper[4772]: E0127 16:39:26.141912 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4545581d-5f56-406d-938f-c3b073fdcbce" containerName="mariadb-account-create-update" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.141919 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4545581d-5f56-406d-938f-c3b073fdcbce" containerName="mariadb-account-create-update" Jan 27 16:39:26 crc kubenswrapper[4772]: E0127 16:39:26.141933 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84c86af5-fd1f-4c53-8978-2b436db59b2a" containerName="mariadb-database-create" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.141941 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="84c86af5-fd1f-4c53-8978-2b436db59b2a" containerName="mariadb-database-create" Jan 27 16:39:26 crc kubenswrapper[4772]: E0127 16:39:26.141956 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81b3d773-720e-42c5-af9e-abddc2180ac7" containerName="mariadb-database-create" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.141965 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="81b3d773-720e-42c5-af9e-abddc2180ac7" containerName="mariadb-database-create" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.142149 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="952d9a1e-efbf-4617-94af-b5ad42cce494" containerName="mariadb-database-create" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.142163 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="86f6e427-99ce-4873-bacc-697edca3d34e" containerName="mariadb-account-create-update" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.142195 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="4545581d-5f56-406d-938f-c3b073fdcbce" containerName="mariadb-account-create-update" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.142220 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="81b3d773-720e-42c5-af9e-abddc2180ac7" containerName="mariadb-database-create" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.142254 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="84c86af5-fd1f-4c53-8978-2b436db59b2a" containerName="mariadb-database-create" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.142264 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="e808813a-e588-4fb9-a15d-588d94a4cd59" containerName="mariadb-account-create-update" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.142935 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k56t7" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.145224 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-6wktq" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.149359 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.155253 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k56t7"] Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.169717 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.254286 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eedd23ad-e532-401a-a991-4bca54fc2711-scripts\") pod \"nova-cell0-conductor-db-sync-k56t7\" (UID: \"eedd23ad-e532-401a-a991-4bca54fc2711\") " pod="openstack/nova-cell0-conductor-db-sync-k56t7" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.254362 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eedd23ad-e532-401a-a991-4bca54fc2711-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-k56t7\" (UID: \"eedd23ad-e532-401a-a991-4bca54fc2711\") " pod="openstack/nova-cell0-conductor-db-sync-k56t7" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.254481 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eedd23ad-e532-401a-a991-4bca54fc2711-config-data\") pod \"nova-cell0-conductor-db-sync-k56t7\" (UID: \"eedd23ad-e532-401a-a991-4bca54fc2711\") " pod="openstack/nova-cell0-conductor-db-sync-k56t7" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.254567 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbf8m\" (UniqueName: \"kubernetes.io/projected/eedd23ad-e532-401a-a991-4bca54fc2711-kube-api-access-hbf8m\") pod \"nova-cell0-conductor-db-sync-k56t7\" (UID: \"eedd23ad-e532-401a-a991-4bca54fc2711\") " pod="openstack/nova-cell0-conductor-db-sync-k56t7" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.356645 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eedd23ad-e532-401a-a991-4bca54fc2711-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-k56t7\" (UID: \"eedd23ad-e532-401a-a991-4bca54fc2711\") " pod="openstack/nova-cell0-conductor-db-sync-k56t7" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.356733 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eedd23ad-e532-401a-a991-4bca54fc2711-config-data\") pod \"nova-cell0-conductor-db-sync-k56t7\" (UID: \"eedd23ad-e532-401a-a991-4bca54fc2711\") " pod="openstack/nova-cell0-conductor-db-sync-k56t7" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.356787 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbf8m\" (UniqueName: \"kubernetes.io/projected/eedd23ad-e532-401a-a991-4bca54fc2711-kube-api-access-hbf8m\") pod \"nova-cell0-conductor-db-sync-k56t7\" (UID: \"eedd23ad-e532-401a-a991-4bca54fc2711\") " pod="openstack/nova-cell0-conductor-db-sync-k56t7" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.356849 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eedd23ad-e532-401a-a991-4bca54fc2711-scripts\") pod \"nova-cell0-conductor-db-sync-k56t7\" (UID: \"eedd23ad-e532-401a-a991-4bca54fc2711\") " pod="openstack/nova-cell0-conductor-db-sync-k56t7" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.368568 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eedd23ad-e532-401a-a991-4bca54fc2711-scripts\") pod \"nova-cell0-conductor-db-sync-k56t7\" (UID: \"eedd23ad-e532-401a-a991-4bca54fc2711\") " pod="openstack/nova-cell0-conductor-db-sync-k56t7" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.369177 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eedd23ad-e532-401a-a991-4bca54fc2711-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-k56t7\" (UID: \"eedd23ad-e532-401a-a991-4bca54fc2711\") " pod="openstack/nova-cell0-conductor-db-sync-k56t7" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.383149 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eedd23ad-e532-401a-a991-4bca54fc2711-config-data\") pod \"nova-cell0-conductor-db-sync-k56t7\" (UID: \"eedd23ad-e532-401a-a991-4bca54fc2711\") " pod="openstack/nova-cell0-conductor-db-sync-k56t7" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.385294 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbf8m\" (UniqueName: \"kubernetes.io/projected/eedd23ad-e532-401a-a991-4bca54fc2711-kube-api-access-hbf8m\") pod \"nova-cell0-conductor-db-sync-k56t7\" (UID: \"eedd23ad-e532-401a-a991-4bca54fc2711\") " pod="openstack/nova-cell0-conductor-db-sync-k56t7" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.484855 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k56t7" Jan 27 16:39:26 crc kubenswrapper[4772]: I0127 16:39:26.984397 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k56t7"] Jan 27 16:39:26 crc kubenswrapper[4772]: W0127 16:39:26.984965 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeedd23ad_e532_401a_a991_4bca54fc2711.slice/crio-9a78b7affa413b3705f382390b02b775a4a3f330d40d06789b404d757989470f WatchSource:0}: Error finding container 9a78b7affa413b3705f382390b02b775a4a3f330d40d06789b404d757989470f: Status 404 returned error can't find the container with id 9a78b7affa413b3705f382390b02b775a4a3f330d40d06789b404d757989470f Jan 27 16:39:27 crc kubenswrapper[4772]: I0127 16:39:27.960358 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k56t7" event={"ID":"eedd23ad-e532-401a-a991-4bca54fc2711","Type":"ContainerStarted","Data":"63117e003669f061743dd5211454454c043cc51173da6608feb5766b651070d3"} Jan 27 16:39:27 crc kubenswrapper[4772]: I0127 16:39:27.960684 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k56t7" event={"ID":"eedd23ad-e532-401a-a991-4bca54fc2711","Type":"ContainerStarted","Data":"9a78b7affa413b3705f382390b02b775a4a3f330d40d06789b404d757989470f"} Jan 27 16:39:37 crc kubenswrapper[4772]: I0127 16:39:37.036903 4772 generic.go:334] "Generic (PLEG): container finished" podID="eedd23ad-e532-401a-a991-4bca54fc2711" containerID="63117e003669f061743dd5211454454c043cc51173da6608feb5766b651070d3" exitCode=0 Jan 27 16:39:37 crc kubenswrapper[4772]: I0127 16:39:37.037031 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k56t7" event={"ID":"eedd23ad-e532-401a-a991-4bca54fc2711","Type":"ContainerDied","Data":"63117e003669f061743dd5211454454c043cc51173da6608feb5766b651070d3"} Jan 27 16:39:38 crc kubenswrapper[4772]: I0127 16:39:38.369363 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k56t7" Jan 27 16:39:38 crc kubenswrapper[4772]: I0127 16:39:38.497533 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eedd23ad-e532-401a-a991-4bca54fc2711-config-data\") pod \"eedd23ad-e532-401a-a991-4bca54fc2711\" (UID: \"eedd23ad-e532-401a-a991-4bca54fc2711\") " Jan 27 16:39:38 crc kubenswrapper[4772]: I0127 16:39:38.497665 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eedd23ad-e532-401a-a991-4bca54fc2711-scripts\") pod \"eedd23ad-e532-401a-a991-4bca54fc2711\" (UID: \"eedd23ad-e532-401a-a991-4bca54fc2711\") " Jan 27 16:39:38 crc kubenswrapper[4772]: I0127 16:39:38.497785 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eedd23ad-e532-401a-a991-4bca54fc2711-combined-ca-bundle\") pod \"eedd23ad-e532-401a-a991-4bca54fc2711\" (UID: \"eedd23ad-e532-401a-a991-4bca54fc2711\") " Jan 27 16:39:38 crc kubenswrapper[4772]: I0127 16:39:38.497846 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbf8m\" (UniqueName: \"kubernetes.io/projected/eedd23ad-e532-401a-a991-4bca54fc2711-kube-api-access-hbf8m\") pod \"eedd23ad-e532-401a-a991-4bca54fc2711\" (UID: \"eedd23ad-e532-401a-a991-4bca54fc2711\") " Jan 27 16:39:38 crc kubenswrapper[4772]: I0127 16:39:38.503736 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eedd23ad-e532-401a-a991-4bca54fc2711-scripts" (OuterVolumeSpecName: "scripts") pod "eedd23ad-e532-401a-a991-4bca54fc2711" (UID: "eedd23ad-e532-401a-a991-4bca54fc2711"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:39:38 crc kubenswrapper[4772]: I0127 16:39:38.503921 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eedd23ad-e532-401a-a991-4bca54fc2711-kube-api-access-hbf8m" (OuterVolumeSpecName: "kube-api-access-hbf8m") pod "eedd23ad-e532-401a-a991-4bca54fc2711" (UID: "eedd23ad-e532-401a-a991-4bca54fc2711"). InnerVolumeSpecName "kube-api-access-hbf8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:39:38 crc kubenswrapper[4772]: I0127 16:39:38.523063 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eedd23ad-e532-401a-a991-4bca54fc2711-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eedd23ad-e532-401a-a991-4bca54fc2711" (UID: "eedd23ad-e532-401a-a991-4bca54fc2711"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:39:38 crc kubenswrapper[4772]: I0127 16:39:38.524487 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eedd23ad-e532-401a-a991-4bca54fc2711-config-data" (OuterVolumeSpecName: "config-data") pod "eedd23ad-e532-401a-a991-4bca54fc2711" (UID: "eedd23ad-e532-401a-a991-4bca54fc2711"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:39:38 crc kubenswrapper[4772]: I0127 16:39:38.600494 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eedd23ad-e532-401a-a991-4bca54fc2711-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:39:38 crc kubenswrapper[4772]: I0127 16:39:38.600731 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbf8m\" (UniqueName: \"kubernetes.io/projected/eedd23ad-e532-401a-a991-4bca54fc2711-kube-api-access-hbf8m\") on node \"crc\" DevicePath \"\"" Jan 27 16:39:38 crc kubenswrapper[4772]: I0127 16:39:38.600791 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eedd23ad-e532-401a-a991-4bca54fc2711-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:39:38 crc kubenswrapper[4772]: I0127 16:39:38.600857 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eedd23ad-e532-401a-a991-4bca54fc2711-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:39:39 crc kubenswrapper[4772]: I0127 16:39:39.058506 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k56t7" event={"ID":"eedd23ad-e532-401a-a991-4bca54fc2711","Type":"ContainerDied","Data":"9a78b7affa413b3705f382390b02b775a4a3f330d40d06789b404d757989470f"} Jan 27 16:39:39 crc kubenswrapper[4772]: I0127 16:39:39.058546 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a78b7affa413b3705f382390b02b775a4a3f330d40d06789b404d757989470f" Jan 27 16:39:39 crc kubenswrapper[4772]: I0127 16:39:39.058592 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k56t7" Jan 27 16:39:39 crc kubenswrapper[4772]: I0127 16:39:39.156809 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 27 16:39:39 crc kubenswrapper[4772]: E0127 16:39:39.157153 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eedd23ad-e532-401a-a991-4bca54fc2711" containerName="nova-cell0-conductor-db-sync" Jan 27 16:39:39 crc kubenswrapper[4772]: I0127 16:39:39.157185 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="eedd23ad-e532-401a-a991-4bca54fc2711" containerName="nova-cell0-conductor-db-sync" Jan 27 16:39:39 crc kubenswrapper[4772]: I0127 16:39:39.157324 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="eedd23ad-e532-401a-a991-4bca54fc2711" containerName="nova-cell0-conductor-db-sync" Jan 27 16:39:39 crc kubenswrapper[4772]: I0127 16:39:39.157922 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 27 16:39:39 crc kubenswrapper[4772]: I0127 16:39:39.159971 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-6wktq" Jan 27 16:39:39 crc kubenswrapper[4772]: I0127 16:39:39.161552 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Jan 27 16:39:39 crc kubenswrapper[4772]: I0127 16:39:39.179897 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 27 16:39:39 crc kubenswrapper[4772]: I0127 16:39:39.313011 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j7l8\" (UniqueName: \"kubernetes.io/projected/4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e-kube-api-access-2j7l8\") pod \"nova-cell0-conductor-0\" (UID: \"4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e\") " pod="openstack/nova-cell0-conductor-0" Jan 27 16:39:39 crc kubenswrapper[4772]: I0127 16:39:39.313145 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e\") " pod="openstack/nova-cell0-conductor-0" Jan 27 16:39:39 crc kubenswrapper[4772]: I0127 16:39:39.313397 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e\") " pod="openstack/nova-cell0-conductor-0" Jan 27 16:39:39 crc kubenswrapper[4772]: I0127 16:39:39.414913 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e\") " pod="openstack/nova-cell0-conductor-0" Jan 27 16:39:39 crc kubenswrapper[4772]: I0127 16:39:39.415007 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e\") " pod="openstack/nova-cell0-conductor-0" Jan 27 16:39:39 crc kubenswrapper[4772]: I0127 16:39:39.415099 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j7l8\" (UniqueName: \"kubernetes.io/projected/4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e-kube-api-access-2j7l8\") pod \"nova-cell0-conductor-0\" (UID: \"4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e\") " pod="openstack/nova-cell0-conductor-0" Jan 27 16:39:39 crc kubenswrapper[4772]: I0127 16:39:39.419868 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e\") " pod="openstack/nova-cell0-conductor-0" Jan 27 16:39:39 crc kubenswrapper[4772]: I0127 16:39:39.422665 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e\") " pod="openstack/nova-cell0-conductor-0" Jan 27 16:39:39 crc kubenswrapper[4772]: I0127 16:39:39.441984 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j7l8\" (UniqueName: \"kubernetes.io/projected/4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e-kube-api-access-2j7l8\") pod \"nova-cell0-conductor-0\" (UID: \"4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e\") " pod="openstack/nova-cell0-conductor-0" Jan 27 16:39:39 crc kubenswrapper[4772]: I0127 16:39:39.512397 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 27 16:39:40 crc kubenswrapper[4772]: I0127 16:39:40.003674 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 27 16:39:40 crc kubenswrapper[4772]: I0127 16:39:40.074006 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e","Type":"ContainerStarted","Data":"30ded49d3176ba704104aa049f568d3100eaac27bd25914a3c0f0f06ce633e73"} Jan 27 16:39:41 crc kubenswrapper[4772]: I0127 16:39:41.084501 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e","Type":"ContainerStarted","Data":"43ecbee0f6dc0baf4a170605cf33f5e90020f6dd199b406980e9515061715e50"} Jan 27 16:39:41 crc kubenswrapper[4772]: I0127 16:39:41.085015 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Jan 27 16:39:41 crc kubenswrapper[4772]: I0127 16:39:41.099349 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.099332574 podStartE2EDuration="2.099332574s" podCreationTimestamp="2026-01-27 16:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:39:41.097772759 +0000 UTC m=+5567.078381867" watchObservedRunningTime="2026-01-27 16:39:41.099332574 +0000 UTC m=+5567.079941672" Jan 27 16:39:49 crc kubenswrapper[4772]: I0127 16:39:49.562904 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.011711 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-9vmft"] Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.013269 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-9vmft" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.015668 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.023216 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.023788 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-9vmft"] Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.114898 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4c79395-a929-4f0d-8aa7-05f24412baed-scripts\") pod \"nova-cell0-cell-mapping-9vmft\" (UID: \"f4c79395-a929-4f0d-8aa7-05f24412baed\") " pod="openstack/nova-cell0-cell-mapping-9vmft" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.114999 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4c79395-a929-4f0d-8aa7-05f24412baed-config-data\") pod \"nova-cell0-cell-mapping-9vmft\" (UID: \"f4c79395-a929-4f0d-8aa7-05f24412baed\") " pod="openstack/nova-cell0-cell-mapping-9vmft" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.115101 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjnmp\" (UniqueName: \"kubernetes.io/projected/f4c79395-a929-4f0d-8aa7-05f24412baed-kube-api-access-mjnmp\") pod \"nova-cell0-cell-mapping-9vmft\" (UID: \"f4c79395-a929-4f0d-8aa7-05f24412baed\") " pod="openstack/nova-cell0-cell-mapping-9vmft" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.115131 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4c79395-a929-4f0d-8aa7-05f24412baed-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-9vmft\" (UID: \"f4c79395-a929-4f0d-8aa7-05f24412baed\") " pod="openstack/nova-cell0-cell-mapping-9vmft" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.164704 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.165795 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.168315 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.180803 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.217606 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjnmp\" (UniqueName: \"kubernetes.io/projected/f4c79395-a929-4f0d-8aa7-05f24412baed-kube-api-access-mjnmp\") pod \"nova-cell0-cell-mapping-9vmft\" (UID: \"f4c79395-a929-4f0d-8aa7-05f24412baed\") " pod="openstack/nova-cell0-cell-mapping-9vmft" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.217652 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4c79395-a929-4f0d-8aa7-05f24412baed-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-9vmft\" (UID: \"f4c79395-a929-4f0d-8aa7-05f24412baed\") " pod="openstack/nova-cell0-cell-mapping-9vmft" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.217713 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4c79395-a929-4f0d-8aa7-05f24412baed-scripts\") pod \"nova-cell0-cell-mapping-9vmft\" (UID: \"f4c79395-a929-4f0d-8aa7-05f24412baed\") " pod="openstack/nova-cell0-cell-mapping-9vmft" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.217769 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4c79395-a929-4f0d-8aa7-05f24412baed-config-data\") pod \"nova-cell0-cell-mapping-9vmft\" (UID: \"f4c79395-a929-4f0d-8aa7-05f24412baed\") " pod="openstack/nova-cell0-cell-mapping-9vmft" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.236968 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4c79395-a929-4f0d-8aa7-05f24412baed-scripts\") pod \"nova-cell0-cell-mapping-9vmft\" (UID: \"f4c79395-a929-4f0d-8aa7-05f24412baed\") " pod="openstack/nova-cell0-cell-mapping-9vmft" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.237689 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4c79395-a929-4f0d-8aa7-05f24412baed-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-9vmft\" (UID: \"f4c79395-a929-4f0d-8aa7-05f24412baed\") " pod="openstack/nova-cell0-cell-mapping-9vmft" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.238607 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4c79395-a929-4f0d-8aa7-05f24412baed-config-data\") pod \"nova-cell0-cell-mapping-9vmft\" (UID: \"f4c79395-a929-4f0d-8aa7-05f24412baed\") " pod="openstack/nova-cell0-cell-mapping-9vmft" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.267864 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.270120 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.280755 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.287798 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjnmp\" (UniqueName: \"kubernetes.io/projected/f4c79395-a929-4f0d-8aa7-05f24412baed-kube-api-access-mjnmp\") pod \"nova-cell0-cell-mapping-9vmft\" (UID: \"f4c79395-a929-4f0d-8aa7-05f24412baed\") " pod="openstack/nova-cell0-cell-mapping-9vmft" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.318815 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26d4226f-e574-498f-a2ce-e0db8f83a8d3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"26d4226f-e574-498f-a2ce-e0db8f83a8d3\") " pod="openstack/nova-scheduler-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.318903 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqgtr\" (UniqueName: \"kubernetes.io/projected/26d4226f-e574-498f-a2ce-e0db8f83a8d3-kube-api-access-tqgtr\") pod \"nova-scheduler-0\" (UID: \"26d4226f-e574-498f-a2ce-e0db8f83a8d3\") " pod="openstack/nova-scheduler-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.318964 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26d4226f-e574-498f-a2ce-e0db8f83a8d3-config-data\") pod \"nova-scheduler-0\" (UID: \"26d4226f-e574-498f-a2ce-e0db8f83a8d3\") " pod="openstack/nova-scheduler-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.325240 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.341601 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-9vmft" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.375649 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.377148 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.382010 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.409224 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.426205 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f993d38-e5bb-4ce4-8a9f-269695614f5e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9f993d38-e5bb-4ce4-8a9f-269695614f5e\") " pod="openstack/nova-metadata-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.426283 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26d4226f-e574-498f-a2ce-e0db8f83a8d3-config-data\") pod \"nova-scheduler-0\" (UID: \"26d4226f-e574-498f-a2ce-e0db8f83a8d3\") " pod="openstack/nova-scheduler-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.426409 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bp727\" (UniqueName: \"kubernetes.io/projected/9f993d38-e5bb-4ce4-8a9f-269695614f5e-kube-api-access-bp727\") pod \"nova-metadata-0\" (UID: \"9f993d38-e5bb-4ce4-8a9f-269695614f5e\") " pod="openstack/nova-metadata-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.426449 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26d4226f-e574-498f-a2ce-e0db8f83a8d3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"26d4226f-e574-498f-a2ce-e0db8f83a8d3\") " pod="openstack/nova-scheduler-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.426474 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f993d38-e5bb-4ce4-8a9f-269695614f5e-config-data\") pod \"nova-metadata-0\" (UID: \"9f993d38-e5bb-4ce4-8a9f-269695614f5e\") " pod="openstack/nova-metadata-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.426527 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqgtr\" (UniqueName: \"kubernetes.io/projected/26d4226f-e574-498f-a2ce-e0db8f83a8d3-kube-api-access-tqgtr\") pod \"nova-scheduler-0\" (UID: \"26d4226f-e574-498f-a2ce-e0db8f83a8d3\") " pod="openstack/nova-scheduler-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.426577 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f993d38-e5bb-4ce4-8a9f-269695614f5e-logs\") pod \"nova-metadata-0\" (UID: \"9f993d38-e5bb-4ce4-8a9f-269695614f5e\") " pod="openstack/nova-metadata-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.435930 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26d4226f-e574-498f-a2ce-e0db8f83a8d3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"26d4226f-e574-498f-a2ce-e0db8f83a8d3\") " pod="openstack/nova-scheduler-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.441995 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26d4226f-e574-498f-a2ce-e0db8f83a8d3-config-data\") pod \"nova-scheduler-0\" (UID: \"26d4226f-e574-498f-a2ce-e0db8f83a8d3\") " pod="openstack/nova-scheduler-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.510399 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqgtr\" (UniqueName: \"kubernetes.io/projected/26d4226f-e574-498f-a2ce-e0db8f83a8d3-kube-api-access-tqgtr\") pod \"nova-scheduler-0\" (UID: \"26d4226f-e574-498f-a2ce-e0db8f83a8d3\") " pod="openstack/nova-scheduler-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.530277 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f993d38-e5bb-4ce4-8a9f-269695614f5e-config-data\") pod \"nova-metadata-0\" (UID: \"9f993d38-e5bb-4ce4-8a9f-269695614f5e\") " pod="openstack/nova-metadata-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.530679 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f993d38-e5bb-4ce4-8a9f-269695614f5e-logs\") pod \"nova-metadata-0\" (UID: \"9f993d38-e5bb-4ce4-8a9f-269695614f5e\") " pod="openstack/nova-metadata-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.530810 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3e14057-d3be-433e-97a7-0b36f18382ee-logs\") pod \"nova-api-0\" (UID: \"e3e14057-d3be-433e-97a7-0b36f18382ee\") " pod="openstack/nova-api-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.530928 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f993d38-e5bb-4ce4-8a9f-269695614f5e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9f993d38-e5bb-4ce4-8a9f-269695614f5e\") " pod="openstack/nova-metadata-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.531065 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfs5m\" (UniqueName: \"kubernetes.io/projected/e3e14057-d3be-433e-97a7-0b36f18382ee-kube-api-access-mfs5m\") pod \"nova-api-0\" (UID: \"e3e14057-d3be-433e-97a7-0b36f18382ee\") " pod="openstack/nova-api-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.531210 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e14057-d3be-433e-97a7-0b36f18382ee-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e3e14057-d3be-433e-97a7-0b36f18382ee\") " pod="openstack/nova-api-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.531361 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3e14057-d3be-433e-97a7-0b36f18382ee-config-data\") pod \"nova-api-0\" (UID: \"e3e14057-d3be-433e-97a7-0b36f18382ee\") " pod="openstack/nova-api-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.531522 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bp727\" (UniqueName: \"kubernetes.io/projected/9f993d38-e5bb-4ce4-8a9f-269695614f5e-kube-api-access-bp727\") pod \"nova-metadata-0\" (UID: \"9f993d38-e5bb-4ce4-8a9f-269695614f5e\") " pod="openstack/nova-metadata-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.534785 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.541153 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f993d38-e5bb-4ce4-8a9f-269695614f5e-logs\") pod \"nova-metadata-0\" (UID: \"9f993d38-e5bb-4ce4-8a9f-269695614f5e\") " pod="openstack/nova-metadata-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.541743 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f993d38-e5bb-4ce4-8a9f-269695614f5e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9f993d38-e5bb-4ce4-8a9f-269695614f5e\") " pod="openstack/nova-metadata-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.542060 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f993d38-e5bb-4ce4-8a9f-269695614f5e-config-data\") pod \"nova-metadata-0\" (UID: \"9f993d38-e5bb-4ce4-8a9f-269695614f5e\") " pod="openstack/nova-metadata-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.569502 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7758766f57-6fk67"] Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.571273 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7758766f57-6fk67" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.604048 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bp727\" (UniqueName: \"kubernetes.io/projected/9f993d38-e5bb-4ce4-8a9f-269695614f5e-kube-api-access-bp727\") pod \"nova-metadata-0\" (UID: \"9f993d38-e5bb-4ce4-8a9f-269695614f5e\") " pod="openstack/nova-metadata-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.625082 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7758766f57-6fk67"] Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.633340 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3e14057-d3be-433e-97a7-0b36f18382ee-logs\") pod \"nova-api-0\" (UID: \"e3e14057-d3be-433e-97a7-0b36f18382ee\") " pod="openstack/nova-api-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.633413 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfs5m\" (UniqueName: \"kubernetes.io/projected/e3e14057-d3be-433e-97a7-0b36f18382ee-kube-api-access-mfs5m\") pod \"nova-api-0\" (UID: \"e3e14057-d3be-433e-97a7-0b36f18382ee\") " pod="openstack/nova-api-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.633444 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e14057-d3be-433e-97a7-0b36f18382ee-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e3e14057-d3be-433e-97a7-0b36f18382ee\") " pod="openstack/nova-api-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.633476 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f90709ae-7118-42ca-aad1-922961a6f858-config\") pod \"dnsmasq-dns-7758766f57-6fk67\" (UID: \"f90709ae-7118-42ca-aad1-922961a6f858\") " pod="openstack/dnsmasq-dns-7758766f57-6fk67" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.633509 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3e14057-d3be-433e-97a7-0b36f18382ee-config-data\") pod \"nova-api-0\" (UID: \"e3e14057-d3be-433e-97a7-0b36f18382ee\") " pod="openstack/nova-api-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.633552 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv5xf\" (UniqueName: \"kubernetes.io/projected/f90709ae-7118-42ca-aad1-922961a6f858-kube-api-access-bv5xf\") pod \"dnsmasq-dns-7758766f57-6fk67\" (UID: \"f90709ae-7118-42ca-aad1-922961a6f858\") " pod="openstack/dnsmasq-dns-7758766f57-6fk67" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.633573 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f90709ae-7118-42ca-aad1-922961a6f858-ovsdbserver-nb\") pod \"dnsmasq-dns-7758766f57-6fk67\" (UID: \"f90709ae-7118-42ca-aad1-922961a6f858\") " pod="openstack/dnsmasq-dns-7758766f57-6fk67" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.633612 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f90709ae-7118-42ca-aad1-922961a6f858-ovsdbserver-sb\") pod \"dnsmasq-dns-7758766f57-6fk67\" (UID: \"f90709ae-7118-42ca-aad1-922961a6f858\") " pod="openstack/dnsmasq-dns-7758766f57-6fk67" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.633633 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f90709ae-7118-42ca-aad1-922961a6f858-dns-svc\") pod \"dnsmasq-dns-7758766f57-6fk67\" (UID: \"f90709ae-7118-42ca-aad1-922961a6f858\") " pod="openstack/dnsmasq-dns-7758766f57-6fk67" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.634017 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3e14057-d3be-433e-97a7-0b36f18382ee-logs\") pod \"nova-api-0\" (UID: \"e3e14057-d3be-433e-97a7-0b36f18382ee\") " pod="openstack/nova-api-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.637982 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e14057-d3be-433e-97a7-0b36f18382ee-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e3e14057-d3be-433e-97a7-0b36f18382ee\") " pod="openstack/nova-api-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.640773 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3e14057-d3be-433e-97a7-0b36f18382ee-config-data\") pod \"nova-api-0\" (UID: \"e3e14057-d3be-433e-97a7-0b36f18382ee\") " pod="openstack/nova-api-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.673972 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfs5m\" (UniqueName: \"kubernetes.io/projected/e3e14057-d3be-433e-97a7-0b36f18382ee-kube-api-access-mfs5m\") pod \"nova-api-0\" (UID: \"e3e14057-d3be-433e-97a7-0b36f18382ee\") " pod="openstack/nova-api-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.732046 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.733195 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.736255 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv5xf\" (UniqueName: \"kubernetes.io/projected/f90709ae-7118-42ca-aad1-922961a6f858-kube-api-access-bv5xf\") pod \"dnsmasq-dns-7758766f57-6fk67\" (UID: \"f90709ae-7118-42ca-aad1-922961a6f858\") " pod="openstack/dnsmasq-dns-7758766f57-6fk67" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.736322 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f90709ae-7118-42ca-aad1-922961a6f858-ovsdbserver-nb\") pod \"dnsmasq-dns-7758766f57-6fk67\" (UID: \"f90709ae-7118-42ca-aad1-922961a6f858\") " pod="openstack/dnsmasq-dns-7758766f57-6fk67" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.736375 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f90709ae-7118-42ca-aad1-922961a6f858-ovsdbserver-sb\") pod \"dnsmasq-dns-7758766f57-6fk67\" (UID: \"f90709ae-7118-42ca-aad1-922961a6f858\") " pod="openstack/dnsmasq-dns-7758766f57-6fk67" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.736401 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f90709ae-7118-42ca-aad1-922961a6f858-dns-svc\") pod \"dnsmasq-dns-7758766f57-6fk67\" (UID: \"f90709ae-7118-42ca-aad1-922961a6f858\") " pod="openstack/dnsmasq-dns-7758766f57-6fk67" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.736534 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f90709ae-7118-42ca-aad1-922961a6f858-config\") pod \"dnsmasq-dns-7758766f57-6fk67\" (UID: \"f90709ae-7118-42ca-aad1-922961a6f858\") " pod="openstack/dnsmasq-dns-7758766f57-6fk67" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.737472 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f90709ae-7118-42ca-aad1-922961a6f858-config\") pod \"dnsmasq-dns-7758766f57-6fk67\" (UID: \"f90709ae-7118-42ca-aad1-922961a6f858\") " pod="openstack/dnsmasq-dns-7758766f57-6fk67" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.737563 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.738426 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f90709ae-7118-42ca-aad1-922961a6f858-ovsdbserver-sb\") pod \"dnsmasq-dns-7758766f57-6fk67\" (UID: \"f90709ae-7118-42ca-aad1-922961a6f858\") " pod="openstack/dnsmasq-dns-7758766f57-6fk67" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.738538 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f90709ae-7118-42ca-aad1-922961a6f858-dns-svc\") pod \"dnsmasq-dns-7758766f57-6fk67\" (UID: \"f90709ae-7118-42ca-aad1-922961a6f858\") " pod="openstack/dnsmasq-dns-7758766f57-6fk67" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.739101 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f90709ae-7118-42ca-aad1-922961a6f858-ovsdbserver-nb\") pod \"dnsmasq-dns-7758766f57-6fk67\" (UID: \"f90709ae-7118-42ca-aad1-922961a6f858\") " pod="openstack/dnsmasq-dns-7758766f57-6fk67" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.773952 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv5xf\" (UniqueName: \"kubernetes.io/projected/f90709ae-7118-42ca-aad1-922961a6f858-kube-api-access-bv5xf\") pod \"dnsmasq-dns-7758766f57-6fk67\" (UID: \"f90709ae-7118-42ca-aad1-922961a6f858\") " pod="openstack/dnsmasq-dns-7758766f57-6fk67" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.775792 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.789255 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.805710 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.839949 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22khv\" (UniqueName: \"kubernetes.io/projected/a30c12eb-4e83-420a-8064-859689d91d2d-kube-api-access-22khv\") pod \"nova-cell1-novncproxy-0\" (UID: \"a30c12eb-4e83-420a-8064-859689d91d2d\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.840490 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a30c12eb-4e83-420a-8064-859689d91d2d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a30c12eb-4e83-420a-8064-859689d91d2d\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.840573 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a30c12eb-4e83-420a-8064-859689d91d2d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a30c12eb-4e83-420a-8064-859689d91d2d\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.945874 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a30c12eb-4e83-420a-8064-859689d91d2d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a30c12eb-4e83-420a-8064-859689d91d2d\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.946094 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22khv\" (UniqueName: \"kubernetes.io/projected/a30c12eb-4e83-420a-8064-859689d91d2d-kube-api-access-22khv\") pod \"nova-cell1-novncproxy-0\" (UID: \"a30c12eb-4e83-420a-8064-859689d91d2d\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.946135 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a30c12eb-4e83-420a-8064-859689d91d2d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a30c12eb-4e83-420a-8064-859689d91d2d\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.956630 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a30c12eb-4e83-420a-8064-859689d91d2d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a30c12eb-4e83-420a-8064-859689d91d2d\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.962592 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a30c12eb-4e83-420a-8064-859689d91d2d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a30c12eb-4e83-420a-8064-859689d91d2d\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.977965 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22khv\" (UniqueName: \"kubernetes.io/projected/a30c12eb-4e83-420a-8064-859689d91d2d-kube-api-access-22khv\") pod \"nova-cell1-novncproxy-0\" (UID: \"a30c12eb-4e83-420a-8064-859689d91d2d\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:39:50 crc kubenswrapper[4772]: I0127 16:39:50.987033 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7758766f57-6fk67" Jan 27 16:39:51 crc kubenswrapper[4772]: I0127 16:39:51.101072 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:39:51 crc kubenswrapper[4772]: I0127 16:39:51.352313 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-9vmft"] Jan 27 16:39:51 crc kubenswrapper[4772]: I0127 16:39:51.399824 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 16:39:51 crc kubenswrapper[4772]: W0127 16:39:51.408141 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26d4226f_e574_498f_a2ce_e0db8f83a8d3.slice/crio-1c139ace8ceb66ebe7302c272fa99b828aba2d79a11db30142d7bb5b26746bbd WatchSource:0}: Error finding container 1c139ace8ceb66ebe7302c272fa99b828aba2d79a11db30142d7bb5b26746bbd: Status 404 returned error can't find the container with id 1c139ace8ceb66ebe7302c272fa99b828aba2d79a11db30142d7bb5b26746bbd Jan 27 16:39:51 crc kubenswrapper[4772]: I0127 16:39:51.526177 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 27 16:39:51 crc kubenswrapper[4772]: I0127 16:39:51.534831 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 16:39:51 crc kubenswrapper[4772]: W0127 16:39:51.542571 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f993d38_e5bb_4ce4_8a9f_269695614f5e.slice/crio-03b13462263b2fb77b677576417743ca4d4b9b8dce8807867be92ad60cefd791 WatchSource:0}: Error finding container 03b13462263b2fb77b677576417743ca4d4b9b8dce8807867be92ad60cefd791: Status 404 returned error can't find the container with id 03b13462263b2fb77b677576417743ca4d4b9b8dce8807867be92ad60cefd791 Jan 27 16:39:51 crc kubenswrapper[4772]: I0127 16:39:51.666812 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7758766f57-6fk67"] Jan 27 16:39:51 crc kubenswrapper[4772]: I0127 16:39:51.699181 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9dk47"] Jan 27 16:39:51 crc kubenswrapper[4772]: I0127 16:39:51.700257 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9dk47" Jan 27 16:39:51 crc kubenswrapper[4772]: I0127 16:39:51.704000 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Jan 27 16:39:51 crc kubenswrapper[4772]: I0127 16:39:51.704393 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Jan 27 16:39:51 crc kubenswrapper[4772]: I0127 16:39:51.712214 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9dk47"] Jan 27 16:39:51 crc kubenswrapper[4772]: I0127 16:39:51.767762 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7zgx\" (UniqueName: \"kubernetes.io/projected/c73df4be-e448-4930-ae5e-d74fde1b4b6d-kube-api-access-q7zgx\") pod \"nova-cell1-conductor-db-sync-9dk47\" (UID: \"c73df4be-e448-4930-ae5e-d74fde1b4b6d\") " pod="openstack/nova-cell1-conductor-db-sync-9dk47" Jan 27 16:39:51 crc kubenswrapper[4772]: I0127 16:39:51.767853 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c73df4be-e448-4930-ae5e-d74fde1b4b6d-config-data\") pod \"nova-cell1-conductor-db-sync-9dk47\" (UID: \"c73df4be-e448-4930-ae5e-d74fde1b4b6d\") " pod="openstack/nova-cell1-conductor-db-sync-9dk47" Jan 27 16:39:51 crc kubenswrapper[4772]: I0127 16:39:51.767881 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c73df4be-e448-4930-ae5e-d74fde1b4b6d-scripts\") pod \"nova-cell1-conductor-db-sync-9dk47\" (UID: \"c73df4be-e448-4930-ae5e-d74fde1b4b6d\") " pod="openstack/nova-cell1-conductor-db-sync-9dk47" Jan 27 16:39:51 crc kubenswrapper[4772]: I0127 16:39:51.767971 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c73df4be-e448-4930-ae5e-d74fde1b4b6d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-9dk47\" (UID: \"c73df4be-e448-4930-ae5e-d74fde1b4b6d\") " pod="openstack/nova-cell1-conductor-db-sync-9dk47" Jan 27 16:39:51 crc kubenswrapper[4772]: I0127 16:39:51.803916 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 27 16:39:51 crc kubenswrapper[4772]: I0127 16:39:51.870251 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c73df4be-e448-4930-ae5e-d74fde1b4b6d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-9dk47\" (UID: \"c73df4be-e448-4930-ae5e-d74fde1b4b6d\") " pod="openstack/nova-cell1-conductor-db-sync-9dk47" Jan 27 16:39:51 crc kubenswrapper[4772]: I0127 16:39:51.870725 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7zgx\" (UniqueName: \"kubernetes.io/projected/c73df4be-e448-4930-ae5e-d74fde1b4b6d-kube-api-access-q7zgx\") pod \"nova-cell1-conductor-db-sync-9dk47\" (UID: \"c73df4be-e448-4930-ae5e-d74fde1b4b6d\") " pod="openstack/nova-cell1-conductor-db-sync-9dk47" Jan 27 16:39:51 crc kubenswrapper[4772]: I0127 16:39:51.870800 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c73df4be-e448-4930-ae5e-d74fde1b4b6d-config-data\") pod \"nova-cell1-conductor-db-sync-9dk47\" (UID: \"c73df4be-e448-4930-ae5e-d74fde1b4b6d\") " pod="openstack/nova-cell1-conductor-db-sync-9dk47" Jan 27 16:39:51 crc kubenswrapper[4772]: I0127 16:39:51.870827 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c73df4be-e448-4930-ae5e-d74fde1b4b6d-scripts\") pod \"nova-cell1-conductor-db-sync-9dk47\" (UID: \"c73df4be-e448-4930-ae5e-d74fde1b4b6d\") " pod="openstack/nova-cell1-conductor-db-sync-9dk47" Jan 27 16:39:51 crc kubenswrapper[4772]: I0127 16:39:51.887913 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c73df4be-e448-4930-ae5e-d74fde1b4b6d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-9dk47\" (UID: \"c73df4be-e448-4930-ae5e-d74fde1b4b6d\") " pod="openstack/nova-cell1-conductor-db-sync-9dk47" Jan 27 16:39:51 crc kubenswrapper[4772]: I0127 16:39:51.888522 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c73df4be-e448-4930-ae5e-d74fde1b4b6d-config-data\") pod \"nova-cell1-conductor-db-sync-9dk47\" (UID: \"c73df4be-e448-4930-ae5e-d74fde1b4b6d\") " pod="openstack/nova-cell1-conductor-db-sync-9dk47" Jan 27 16:39:51 crc kubenswrapper[4772]: I0127 16:39:51.888855 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c73df4be-e448-4930-ae5e-d74fde1b4b6d-scripts\") pod \"nova-cell1-conductor-db-sync-9dk47\" (UID: \"c73df4be-e448-4930-ae5e-d74fde1b4b6d\") " pod="openstack/nova-cell1-conductor-db-sync-9dk47" Jan 27 16:39:51 crc kubenswrapper[4772]: I0127 16:39:51.894375 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7zgx\" (UniqueName: \"kubernetes.io/projected/c73df4be-e448-4930-ae5e-d74fde1b4b6d-kube-api-access-q7zgx\") pod \"nova-cell1-conductor-db-sync-9dk47\" (UID: \"c73df4be-e448-4930-ae5e-d74fde1b4b6d\") " pod="openstack/nova-cell1-conductor-db-sync-9dk47" Jan 27 16:39:52 crc kubenswrapper[4772]: I0127 16:39:52.028812 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9dk47" Jan 27 16:39:52 crc kubenswrapper[4772]: I0127 16:39:52.234816 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-9vmft" event={"ID":"f4c79395-a929-4f0d-8aa7-05f24412baed","Type":"ContainerStarted","Data":"cd101790786da89124bf4fdc7b4bace7b135a40af114a9eeb7dbc7a4372fd732"} Jan 27 16:39:52 crc kubenswrapper[4772]: I0127 16:39:52.234875 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-9vmft" event={"ID":"f4c79395-a929-4f0d-8aa7-05f24412baed","Type":"ContainerStarted","Data":"52cbee2f5d4eb8ed937fe3376681b915bcc858435daf4a58153a4c6bb2cb935d"} Jan 27 16:39:52 crc kubenswrapper[4772]: I0127 16:39:52.256412 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e3e14057-d3be-433e-97a7-0b36f18382ee","Type":"ContainerStarted","Data":"d78776dd87426d6e9e796444fb4d98fcf0b08f9b784ed6f290516aa2ca7cd473"} Jan 27 16:39:52 crc kubenswrapper[4772]: I0127 16:39:52.256472 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e3e14057-d3be-433e-97a7-0b36f18382ee","Type":"ContainerStarted","Data":"0271b9dbb1d2941b26f79c6625f660e41862d12903a306e76d443a4c1adc8635"} Jan 27 16:39:52 crc kubenswrapper[4772]: I0127 16:39:52.256486 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e3e14057-d3be-433e-97a7-0b36f18382ee","Type":"ContainerStarted","Data":"44267cacdffa14217dc9859dca23e969af5b7adbb4750d9c4f117ba9ccd3c77f"} Jan 27 16:39:52 crc kubenswrapper[4772]: I0127 16:39:52.259196 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"26d4226f-e574-498f-a2ce-e0db8f83a8d3","Type":"ContainerStarted","Data":"d6c1d8ec858944c73c32d29d0613f42bb401204fa7425a43dbda27a9aa01ac58"} Jan 27 16:39:52 crc kubenswrapper[4772]: I0127 16:39:52.259226 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"26d4226f-e574-498f-a2ce-e0db8f83a8d3","Type":"ContainerStarted","Data":"1c139ace8ceb66ebe7302c272fa99b828aba2d79a11db30142d7bb5b26746bbd"} Jan 27 16:39:52 crc kubenswrapper[4772]: I0127 16:39:52.270139 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a30c12eb-4e83-420a-8064-859689d91d2d","Type":"ContainerStarted","Data":"fba71bd5b1f04f4ce21ecca46c2028b0aeffaa1f208e5b17978c2d0f906cbf36"} Jan 27 16:39:52 crc kubenswrapper[4772]: I0127 16:39:52.270201 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a30c12eb-4e83-420a-8064-859689d91d2d","Type":"ContainerStarted","Data":"772c077f79e9fa947b145341393fdb1b1ddc0a4d4b6ec55b124c53c8a8bb5534"} Jan 27 16:39:52 crc kubenswrapper[4772]: I0127 16:39:52.281375 4772 generic.go:334] "Generic (PLEG): container finished" podID="f90709ae-7118-42ca-aad1-922961a6f858" containerID="ea368eea29b6534237f47789a13e7bcf3cb4d11f06fe750858483cf68a03ec51" exitCode=0 Jan 27 16:39:52 crc kubenswrapper[4772]: I0127 16:39:52.281766 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7758766f57-6fk67" event={"ID":"f90709ae-7118-42ca-aad1-922961a6f858","Type":"ContainerDied","Data":"ea368eea29b6534237f47789a13e7bcf3cb4d11f06fe750858483cf68a03ec51"} Jan 27 16:39:52 crc kubenswrapper[4772]: I0127 16:39:52.281795 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7758766f57-6fk67" event={"ID":"f90709ae-7118-42ca-aad1-922961a6f858","Type":"ContainerStarted","Data":"b319140f655aa1b23b1e685173361086b6144080f0c58e842877b6cfd23b2922"} Jan 27 16:39:52 crc kubenswrapper[4772]: I0127 16:39:52.305394 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f993d38-e5bb-4ce4-8a9f-269695614f5e","Type":"ContainerStarted","Data":"3a53c1dc2aeb5bf544910214491023eeece1cfbdb9d21ada9d6cf8e1d2675ffb"} Jan 27 16:39:52 crc kubenswrapper[4772]: I0127 16:39:52.305635 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f993d38-e5bb-4ce4-8a9f-269695614f5e","Type":"ContainerStarted","Data":"43c20e864a030473b88bc17f0d3fed51612d4bbdf0ff55280f36b4578127cc8a"} Jan 27 16:39:52 crc kubenswrapper[4772]: I0127 16:39:52.305646 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f993d38-e5bb-4ce4-8a9f-269695614f5e","Type":"ContainerStarted","Data":"03b13462263b2fb77b677576417743ca4d4b9b8dce8807867be92ad60cefd791"} Jan 27 16:39:52 crc kubenswrapper[4772]: I0127 16:39:52.316629 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-9vmft" podStartSLOduration=3.316433564 podStartE2EDuration="3.316433564s" podCreationTimestamp="2026-01-27 16:39:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:39:52.282692362 +0000 UTC m=+5578.263301460" watchObservedRunningTime="2026-01-27 16:39:52.316433564 +0000 UTC m=+5578.297042662" Jan 27 16:39:52 crc kubenswrapper[4772]: I0127 16:39:52.356442 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.356422594 podStartE2EDuration="2.356422594s" podCreationTimestamp="2026-01-27 16:39:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:39:52.330021471 +0000 UTC m=+5578.310630569" watchObservedRunningTime="2026-01-27 16:39:52.356422594 +0000 UTC m=+5578.337031692" Jan 27 16:39:52 crc kubenswrapper[4772]: I0127 16:39:52.359914 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.359903673 podStartE2EDuration="2.359903673s" podCreationTimestamp="2026-01-27 16:39:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:39:52.352325017 +0000 UTC m=+5578.332934125" watchObservedRunningTime="2026-01-27 16:39:52.359903673 +0000 UTC m=+5578.340512771" Jan 27 16:39:52 crc kubenswrapper[4772]: I0127 16:39:52.417696 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.41767385 podStartE2EDuration="2.41767385s" podCreationTimestamp="2026-01-27 16:39:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:39:52.395210519 +0000 UTC m=+5578.375819627" watchObservedRunningTime="2026-01-27 16:39:52.41767385 +0000 UTC m=+5578.398282938" Jan 27 16:39:52 crc kubenswrapper[4772]: I0127 16:39:52.426564 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.426540122 podStartE2EDuration="2.426540122s" podCreationTimestamp="2026-01-27 16:39:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:39:52.41278644 +0000 UTC m=+5578.393395548" watchObservedRunningTime="2026-01-27 16:39:52.426540122 +0000 UTC m=+5578.407149230" Jan 27 16:39:52 crc kubenswrapper[4772]: I0127 16:39:52.582783 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9dk47"] Jan 27 16:39:53 crc kubenswrapper[4772]: I0127 16:39:53.317281 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7758766f57-6fk67" event={"ID":"f90709ae-7118-42ca-aad1-922961a6f858","Type":"ContainerStarted","Data":"97e6d0f1812bfdc92938ed911c0ea776bf32e0bb2e1f04fe8e722bdbc3e2b5e5"} Jan 27 16:39:53 crc kubenswrapper[4772]: I0127 16:39:53.317749 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7758766f57-6fk67" Jan 27 16:39:53 crc kubenswrapper[4772]: I0127 16:39:53.319107 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9dk47" event={"ID":"c73df4be-e448-4930-ae5e-d74fde1b4b6d","Type":"ContainerStarted","Data":"4799caa746e9ac611c5360c3f4fb6d88e8100346e94f7137cac433d747b98815"} Jan 27 16:39:53 crc kubenswrapper[4772]: I0127 16:39:53.319156 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9dk47" event={"ID":"c73df4be-e448-4930-ae5e-d74fde1b4b6d","Type":"ContainerStarted","Data":"916e7f8ec2b785139b190369d0dd78267c4fd066758c7d8bf358fac5d60afeb5"} Jan 27 16:39:53 crc kubenswrapper[4772]: I0127 16:39:53.355871 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-9dk47" podStartSLOduration=2.355854162 podStartE2EDuration="2.355854162s" podCreationTimestamp="2026-01-27 16:39:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:39:53.354893244 +0000 UTC m=+5579.335502342" watchObservedRunningTime="2026-01-27 16:39:53.355854162 +0000 UTC m=+5579.336463260" Jan 27 16:39:53 crc kubenswrapper[4772]: I0127 16:39:53.359658 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7758766f57-6fk67" podStartSLOduration=3.35964491 podStartE2EDuration="3.35964491s" podCreationTimestamp="2026-01-27 16:39:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:39:53.34143194 +0000 UTC m=+5579.322041038" watchObservedRunningTime="2026-01-27 16:39:53.35964491 +0000 UTC m=+5579.340254008" Jan 27 16:39:55 crc kubenswrapper[4772]: I0127 16:39:55.541273 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 27 16:39:55 crc kubenswrapper[4772]: I0127 16:39:55.789748 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 27 16:39:55 crc kubenswrapper[4772]: I0127 16:39:55.789820 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 27 16:39:56 crc kubenswrapper[4772]: I0127 16:39:56.102756 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:39:56 crc kubenswrapper[4772]: I0127 16:39:56.342743 4772 generic.go:334] "Generic (PLEG): container finished" podID="c73df4be-e448-4930-ae5e-d74fde1b4b6d" containerID="4799caa746e9ac611c5360c3f4fb6d88e8100346e94f7137cac433d747b98815" exitCode=0 Jan 27 16:39:56 crc kubenswrapper[4772]: I0127 16:39:56.342788 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9dk47" event={"ID":"c73df4be-e448-4930-ae5e-d74fde1b4b6d","Type":"ContainerDied","Data":"4799caa746e9ac611c5360c3f4fb6d88e8100346e94f7137cac433d747b98815"} Jan 27 16:39:57 crc kubenswrapper[4772]: I0127 16:39:57.354542 4772 generic.go:334] "Generic (PLEG): container finished" podID="f4c79395-a929-4f0d-8aa7-05f24412baed" containerID="cd101790786da89124bf4fdc7b4bace7b135a40af114a9eeb7dbc7a4372fd732" exitCode=0 Jan 27 16:39:57 crc kubenswrapper[4772]: I0127 16:39:57.354605 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-9vmft" event={"ID":"f4c79395-a929-4f0d-8aa7-05f24412baed","Type":"ContainerDied","Data":"cd101790786da89124bf4fdc7b4bace7b135a40af114a9eeb7dbc7a4372fd732"} Jan 27 16:39:57 crc kubenswrapper[4772]: I0127 16:39:57.790246 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9dk47" Jan 27 16:39:57 crc kubenswrapper[4772]: I0127 16:39:57.904460 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c73df4be-e448-4930-ae5e-d74fde1b4b6d-scripts\") pod \"c73df4be-e448-4930-ae5e-d74fde1b4b6d\" (UID: \"c73df4be-e448-4930-ae5e-d74fde1b4b6d\") " Jan 27 16:39:57 crc kubenswrapper[4772]: I0127 16:39:57.904632 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c73df4be-e448-4930-ae5e-d74fde1b4b6d-config-data\") pod \"c73df4be-e448-4930-ae5e-d74fde1b4b6d\" (UID: \"c73df4be-e448-4930-ae5e-d74fde1b4b6d\") " Jan 27 16:39:57 crc kubenswrapper[4772]: I0127 16:39:57.904666 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c73df4be-e448-4930-ae5e-d74fde1b4b6d-combined-ca-bundle\") pod \"c73df4be-e448-4930-ae5e-d74fde1b4b6d\" (UID: \"c73df4be-e448-4930-ae5e-d74fde1b4b6d\") " Jan 27 16:39:57 crc kubenswrapper[4772]: I0127 16:39:57.904713 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7zgx\" (UniqueName: \"kubernetes.io/projected/c73df4be-e448-4930-ae5e-d74fde1b4b6d-kube-api-access-q7zgx\") pod \"c73df4be-e448-4930-ae5e-d74fde1b4b6d\" (UID: \"c73df4be-e448-4930-ae5e-d74fde1b4b6d\") " Jan 27 16:39:57 crc kubenswrapper[4772]: I0127 16:39:57.911615 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c73df4be-e448-4930-ae5e-d74fde1b4b6d-kube-api-access-q7zgx" (OuterVolumeSpecName: "kube-api-access-q7zgx") pod "c73df4be-e448-4930-ae5e-d74fde1b4b6d" (UID: "c73df4be-e448-4930-ae5e-d74fde1b4b6d"). InnerVolumeSpecName "kube-api-access-q7zgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:39:57 crc kubenswrapper[4772]: I0127 16:39:57.911926 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c73df4be-e448-4930-ae5e-d74fde1b4b6d-scripts" (OuterVolumeSpecName: "scripts") pod "c73df4be-e448-4930-ae5e-d74fde1b4b6d" (UID: "c73df4be-e448-4930-ae5e-d74fde1b4b6d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:39:57 crc kubenswrapper[4772]: I0127 16:39:57.932879 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c73df4be-e448-4930-ae5e-d74fde1b4b6d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c73df4be-e448-4930-ae5e-d74fde1b4b6d" (UID: "c73df4be-e448-4930-ae5e-d74fde1b4b6d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:39:57 crc kubenswrapper[4772]: I0127 16:39:57.950592 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c73df4be-e448-4930-ae5e-d74fde1b4b6d-config-data" (OuterVolumeSpecName: "config-data") pod "c73df4be-e448-4930-ae5e-d74fde1b4b6d" (UID: "c73df4be-e448-4930-ae5e-d74fde1b4b6d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:39:58 crc kubenswrapper[4772]: I0127 16:39:58.007028 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c73df4be-e448-4930-ae5e-d74fde1b4b6d-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:39:58 crc kubenswrapper[4772]: I0127 16:39:58.007060 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c73df4be-e448-4930-ae5e-d74fde1b4b6d-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:39:58 crc kubenswrapper[4772]: I0127 16:39:58.007073 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c73df4be-e448-4930-ae5e-d74fde1b4b6d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:39:58 crc kubenswrapper[4772]: I0127 16:39:58.007083 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7zgx\" (UniqueName: \"kubernetes.io/projected/c73df4be-e448-4930-ae5e-d74fde1b4b6d-kube-api-access-q7zgx\") on node \"crc\" DevicePath \"\"" Jan 27 16:39:58 crc kubenswrapper[4772]: I0127 16:39:58.365485 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9dk47" event={"ID":"c73df4be-e448-4930-ae5e-d74fde1b4b6d","Type":"ContainerDied","Data":"916e7f8ec2b785139b190369d0dd78267c4fd066758c7d8bf358fac5d60afeb5"} Jan 27 16:39:58 crc kubenswrapper[4772]: I0127 16:39:58.365529 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="916e7f8ec2b785139b190369d0dd78267c4fd066758c7d8bf358fac5d60afeb5" Jan 27 16:39:58 crc kubenswrapper[4772]: I0127 16:39:58.365582 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9dk47" Jan 27 16:39:58 crc kubenswrapper[4772]: I0127 16:39:58.442100 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 27 16:39:58 crc kubenswrapper[4772]: E0127 16:39:58.442500 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c73df4be-e448-4930-ae5e-d74fde1b4b6d" containerName="nova-cell1-conductor-db-sync" Jan 27 16:39:58 crc kubenswrapper[4772]: I0127 16:39:58.442514 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="c73df4be-e448-4930-ae5e-d74fde1b4b6d" containerName="nova-cell1-conductor-db-sync" Jan 27 16:39:58 crc kubenswrapper[4772]: I0127 16:39:58.442673 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="c73df4be-e448-4930-ae5e-d74fde1b4b6d" containerName="nova-cell1-conductor-db-sync" Jan 27 16:39:58 crc kubenswrapper[4772]: I0127 16:39:58.443226 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Jan 27 16:39:58 crc kubenswrapper[4772]: I0127 16:39:58.444902 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Jan 27 16:39:58 crc kubenswrapper[4772]: I0127 16:39:58.454406 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 27 16:39:58 crc kubenswrapper[4772]: I0127 16:39:58.514983 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdhmd\" (UniqueName: \"kubernetes.io/projected/e1c55305-7d18-44cb-90e5-b6793989abda-kube-api-access-vdhmd\") pod \"nova-cell1-conductor-0\" (UID: \"e1c55305-7d18-44cb-90e5-b6793989abda\") " pod="openstack/nova-cell1-conductor-0" Jan 27 16:39:58 crc kubenswrapper[4772]: I0127 16:39:58.515036 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c55305-7d18-44cb-90e5-b6793989abda-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e1c55305-7d18-44cb-90e5-b6793989abda\") " pod="openstack/nova-cell1-conductor-0" Jan 27 16:39:58 crc kubenswrapper[4772]: I0127 16:39:58.515073 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1c55305-7d18-44cb-90e5-b6793989abda-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e1c55305-7d18-44cb-90e5-b6793989abda\") " pod="openstack/nova-cell1-conductor-0" Jan 27 16:39:58 crc kubenswrapper[4772]: I0127 16:39:58.617222 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdhmd\" (UniqueName: \"kubernetes.io/projected/e1c55305-7d18-44cb-90e5-b6793989abda-kube-api-access-vdhmd\") pod \"nova-cell1-conductor-0\" (UID: \"e1c55305-7d18-44cb-90e5-b6793989abda\") " pod="openstack/nova-cell1-conductor-0" Jan 27 16:39:58 crc kubenswrapper[4772]: I0127 16:39:58.617283 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c55305-7d18-44cb-90e5-b6793989abda-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e1c55305-7d18-44cb-90e5-b6793989abda\") " pod="openstack/nova-cell1-conductor-0" Jan 27 16:39:58 crc kubenswrapper[4772]: I0127 16:39:58.617328 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1c55305-7d18-44cb-90e5-b6793989abda-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e1c55305-7d18-44cb-90e5-b6793989abda\") " pod="openstack/nova-cell1-conductor-0" Jan 27 16:39:58 crc kubenswrapper[4772]: I0127 16:39:58.620913 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1c55305-7d18-44cb-90e5-b6793989abda-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e1c55305-7d18-44cb-90e5-b6793989abda\") " pod="openstack/nova-cell1-conductor-0" Jan 27 16:39:58 crc kubenswrapper[4772]: I0127 16:39:58.622001 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c55305-7d18-44cb-90e5-b6793989abda-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e1c55305-7d18-44cb-90e5-b6793989abda\") " pod="openstack/nova-cell1-conductor-0" Jan 27 16:39:58 crc kubenswrapper[4772]: I0127 16:39:58.640398 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdhmd\" (UniqueName: \"kubernetes.io/projected/e1c55305-7d18-44cb-90e5-b6793989abda-kube-api-access-vdhmd\") pod \"nova-cell1-conductor-0\" (UID: \"e1c55305-7d18-44cb-90e5-b6793989abda\") " pod="openstack/nova-cell1-conductor-0" Jan 27 16:39:58 crc kubenswrapper[4772]: I0127 16:39:58.763826 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Jan 27 16:39:58 crc kubenswrapper[4772]: I0127 16:39:58.882450 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-9vmft" Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.024442 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4c79395-a929-4f0d-8aa7-05f24412baed-config-data\") pod \"f4c79395-a929-4f0d-8aa7-05f24412baed\" (UID: \"f4c79395-a929-4f0d-8aa7-05f24412baed\") " Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.024534 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4c79395-a929-4f0d-8aa7-05f24412baed-scripts\") pod \"f4c79395-a929-4f0d-8aa7-05f24412baed\" (UID: \"f4c79395-a929-4f0d-8aa7-05f24412baed\") " Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.024652 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjnmp\" (UniqueName: \"kubernetes.io/projected/f4c79395-a929-4f0d-8aa7-05f24412baed-kube-api-access-mjnmp\") pod \"f4c79395-a929-4f0d-8aa7-05f24412baed\" (UID: \"f4c79395-a929-4f0d-8aa7-05f24412baed\") " Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.024714 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4c79395-a929-4f0d-8aa7-05f24412baed-combined-ca-bundle\") pod \"f4c79395-a929-4f0d-8aa7-05f24412baed\" (UID: \"f4c79395-a929-4f0d-8aa7-05f24412baed\") " Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.038051 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4c79395-a929-4f0d-8aa7-05f24412baed-scripts" (OuterVolumeSpecName: "scripts") pod "f4c79395-a929-4f0d-8aa7-05f24412baed" (UID: "f4c79395-a929-4f0d-8aa7-05f24412baed"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.038140 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4c79395-a929-4f0d-8aa7-05f24412baed-kube-api-access-mjnmp" (OuterVolumeSpecName: "kube-api-access-mjnmp") pod "f4c79395-a929-4f0d-8aa7-05f24412baed" (UID: "f4c79395-a929-4f0d-8aa7-05f24412baed"). InnerVolumeSpecName "kube-api-access-mjnmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.059562 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4c79395-a929-4f0d-8aa7-05f24412baed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f4c79395-a929-4f0d-8aa7-05f24412baed" (UID: "f4c79395-a929-4f0d-8aa7-05f24412baed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.067091 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4c79395-a929-4f0d-8aa7-05f24412baed-config-data" (OuterVolumeSpecName: "config-data") pod "f4c79395-a929-4f0d-8aa7-05f24412baed" (UID: "f4c79395-a929-4f0d-8aa7-05f24412baed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.126655 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjnmp\" (UniqueName: \"kubernetes.io/projected/f4c79395-a929-4f0d-8aa7-05f24412baed-kube-api-access-mjnmp\") on node \"crc\" DevicePath \"\"" Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.126708 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4c79395-a929-4f0d-8aa7-05f24412baed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.126722 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4c79395-a929-4f0d-8aa7-05f24412baed-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.126734 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4c79395-a929-4f0d-8aa7-05f24412baed-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.253125 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 27 16:39:59 crc kubenswrapper[4772]: W0127 16:39:59.257432 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1c55305_7d18_44cb_90e5_b6793989abda.slice/crio-8f15691a47c277f39b1b0ae53dfcc7f92ebe7de7d545b4f191aa38d15da7fb88 WatchSource:0}: Error finding container 8f15691a47c277f39b1b0ae53dfcc7f92ebe7de7d545b4f191aa38d15da7fb88: Status 404 returned error can't find the container with id 8f15691a47c277f39b1b0ae53dfcc7f92ebe7de7d545b4f191aa38d15da7fb88 Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.389333 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-9vmft" event={"ID":"f4c79395-a929-4f0d-8aa7-05f24412baed","Type":"ContainerDied","Data":"52cbee2f5d4eb8ed937fe3376681b915bcc858435daf4a58153a4c6bb2cb935d"} Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.389827 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52cbee2f5d4eb8ed937fe3376681b915bcc858435daf4a58153a4c6bb2cb935d" Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.389946 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-9vmft" Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.392086 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e1c55305-7d18-44cb-90e5-b6793989abda","Type":"ContainerStarted","Data":"8f15691a47c277f39b1b0ae53dfcc7f92ebe7de7d545b4f191aa38d15da7fb88"} Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.566231 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.566672 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e3e14057-d3be-433e-97a7-0b36f18382ee" containerName="nova-api-log" containerID="cri-o://0271b9dbb1d2941b26f79c6625f660e41862d12903a306e76d443a4c1adc8635" gracePeriod=30 Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.566887 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e3e14057-d3be-433e-97a7-0b36f18382ee" containerName="nova-api-api" containerID="cri-o://d78776dd87426d6e9e796444fb4d98fcf0b08f9b784ed6f290516aa2ca7cd473" gracePeriod=30 Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.575876 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.576114 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="26d4226f-e574-498f-a2ce-e0db8f83a8d3" containerName="nova-scheduler-scheduler" containerID="cri-o://d6c1d8ec858944c73c32d29d0613f42bb401204fa7425a43dbda27a9aa01ac58" gracePeriod=30 Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.636080 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.636451 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9f993d38-e5bb-4ce4-8a9f-269695614f5e" containerName="nova-metadata-metadata" containerID="cri-o://3a53c1dc2aeb5bf544910214491023eeece1cfbdb9d21ada9d6cf8e1d2675ffb" gracePeriod=30 Jan 27 16:39:59 crc kubenswrapper[4772]: I0127 16:39:59.636437 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9f993d38-e5bb-4ce4-8a9f-269695614f5e" containerName="nova-metadata-log" containerID="cri-o://43c20e864a030473b88bc17f0d3fed51612d4bbdf0ff55280f36b4578127cc8a" gracePeriod=30 Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.153308 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.183876 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.248664 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3e14057-d3be-433e-97a7-0b36f18382ee-logs\") pod \"e3e14057-d3be-433e-97a7-0b36f18382ee\" (UID: \"e3e14057-d3be-433e-97a7-0b36f18382ee\") " Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.248779 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfs5m\" (UniqueName: \"kubernetes.io/projected/e3e14057-d3be-433e-97a7-0b36f18382ee-kube-api-access-mfs5m\") pod \"e3e14057-d3be-433e-97a7-0b36f18382ee\" (UID: \"e3e14057-d3be-433e-97a7-0b36f18382ee\") " Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.248853 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e14057-d3be-433e-97a7-0b36f18382ee-combined-ca-bundle\") pod \"e3e14057-d3be-433e-97a7-0b36f18382ee\" (UID: \"e3e14057-d3be-433e-97a7-0b36f18382ee\") " Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.248914 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3e14057-d3be-433e-97a7-0b36f18382ee-config-data\") pod \"e3e14057-d3be-433e-97a7-0b36f18382ee\" (UID: \"e3e14057-d3be-433e-97a7-0b36f18382ee\") " Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.249192 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3e14057-d3be-433e-97a7-0b36f18382ee-logs" (OuterVolumeSpecName: "logs") pod "e3e14057-d3be-433e-97a7-0b36f18382ee" (UID: "e3e14057-d3be-433e-97a7-0b36f18382ee"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.249471 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3e14057-d3be-433e-97a7-0b36f18382ee-logs\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.264431 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3e14057-d3be-433e-97a7-0b36f18382ee-kube-api-access-mfs5m" (OuterVolumeSpecName: "kube-api-access-mfs5m") pod "e3e14057-d3be-433e-97a7-0b36f18382ee" (UID: "e3e14057-d3be-433e-97a7-0b36f18382ee"). InnerVolumeSpecName "kube-api-access-mfs5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.273717 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3e14057-d3be-433e-97a7-0b36f18382ee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e3e14057-d3be-433e-97a7-0b36f18382ee" (UID: "e3e14057-d3be-433e-97a7-0b36f18382ee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.274138 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3e14057-d3be-433e-97a7-0b36f18382ee-config-data" (OuterVolumeSpecName: "config-data") pod "e3e14057-d3be-433e-97a7-0b36f18382ee" (UID: "e3e14057-d3be-433e-97a7-0b36f18382ee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.350412 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f993d38-e5bb-4ce4-8a9f-269695614f5e-config-data\") pod \"9f993d38-e5bb-4ce4-8a9f-269695614f5e\" (UID: \"9f993d38-e5bb-4ce4-8a9f-269695614f5e\") " Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.350540 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f993d38-e5bb-4ce4-8a9f-269695614f5e-logs\") pod \"9f993d38-e5bb-4ce4-8a9f-269695614f5e\" (UID: \"9f993d38-e5bb-4ce4-8a9f-269695614f5e\") " Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.350580 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f993d38-e5bb-4ce4-8a9f-269695614f5e-combined-ca-bundle\") pod \"9f993d38-e5bb-4ce4-8a9f-269695614f5e\" (UID: \"9f993d38-e5bb-4ce4-8a9f-269695614f5e\") " Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.350701 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bp727\" (UniqueName: \"kubernetes.io/projected/9f993d38-e5bb-4ce4-8a9f-269695614f5e-kube-api-access-bp727\") pod \"9f993d38-e5bb-4ce4-8a9f-269695614f5e\" (UID: \"9f993d38-e5bb-4ce4-8a9f-269695614f5e\") " Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.351031 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f993d38-e5bb-4ce4-8a9f-269695614f5e-logs" (OuterVolumeSpecName: "logs") pod "9f993d38-e5bb-4ce4-8a9f-269695614f5e" (UID: "9f993d38-e5bb-4ce4-8a9f-269695614f5e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.351291 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfs5m\" (UniqueName: \"kubernetes.io/projected/e3e14057-d3be-433e-97a7-0b36f18382ee-kube-api-access-mfs5m\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.351318 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f993d38-e5bb-4ce4-8a9f-269695614f5e-logs\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.351336 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3e14057-d3be-433e-97a7-0b36f18382ee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.351348 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3e14057-d3be-433e-97a7-0b36f18382ee-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.353650 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f993d38-e5bb-4ce4-8a9f-269695614f5e-kube-api-access-bp727" (OuterVolumeSpecName: "kube-api-access-bp727") pod "9f993d38-e5bb-4ce4-8a9f-269695614f5e" (UID: "9f993d38-e5bb-4ce4-8a9f-269695614f5e"). InnerVolumeSpecName "kube-api-access-bp727". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.372201 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f993d38-e5bb-4ce4-8a9f-269695614f5e-config-data" (OuterVolumeSpecName: "config-data") pod "9f993d38-e5bb-4ce4-8a9f-269695614f5e" (UID: "9f993d38-e5bb-4ce4-8a9f-269695614f5e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.374290 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f993d38-e5bb-4ce4-8a9f-269695614f5e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f993d38-e5bb-4ce4-8a9f-269695614f5e" (UID: "9f993d38-e5bb-4ce4-8a9f-269695614f5e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.403125 4772 generic.go:334] "Generic (PLEG): container finished" podID="9f993d38-e5bb-4ce4-8a9f-269695614f5e" containerID="3a53c1dc2aeb5bf544910214491023eeece1cfbdb9d21ada9d6cf8e1d2675ffb" exitCode=0 Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.403151 4772 generic.go:334] "Generic (PLEG): container finished" podID="9f993d38-e5bb-4ce4-8a9f-269695614f5e" containerID="43c20e864a030473b88bc17f0d3fed51612d4bbdf0ff55280f36b4578127cc8a" exitCode=143 Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.403198 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f993d38-e5bb-4ce4-8a9f-269695614f5e","Type":"ContainerDied","Data":"3a53c1dc2aeb5bf544910214491023eeece1cfbdb9d21ada9d6cf8e1d2675ffb"} Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.403251 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f993d38-e5bb-4ce4-8a9f-269695614f5e","Type":"ContainerDied","Data":"43c20e864a030473b88bc17f0d3fed51612d4bbdf0ff55280f36b4578127cc8a"} Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.403264 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9f993d38-e5bb-4ce4-8a9f-269695614f5e","Type":"ContainerDied","Data":"03b13462263b2fb77b677576417743ca4d4b9b8dce8807867be92ad60cefd791"} Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.403280 4772 scope.go:117] "RemoveContainer" containerID="3a53c1dc2aeb5bf544910214491023eeece1cfbdb9d21ada9d6cf8e1d2675ffb" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.403386 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.414450 4772 generic.go:334] "Generic (PLEG): container finished" podID="e3e14057-d3be-433e-97a7-0b36f18382ee" containerID="d78776dd87426d6e9e796444fb4d98fcf0b08f9b784ed6f290516aa2ca7cd473" exitCode=0 Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.414478 4772 generic.go:334] "Generic (PLEG): container finished" podID="e3e14057-d3be-433e-97a7-0b36f18382ee" containerID="0271b9dbb1d2941b26f79c6625f660e41862d12903a306e76d443a4c1adc8635" exitCode=143 Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.414533 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e3e14057-d3be-433e-97a7-0b36f18382ee","Type":"ContainerDied","Data":"d78776dd87426d6e9e796444fb4d98fcf0b08f9b784ed6f290516aa2ca7cd473"} Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.414584 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e3e14057-d3be-433e-97a7-0b36f18382ee","Type":"ContainerDied","Data":"0271b9dbb1d2941b26f79c6625f660e41862d12903a306e76d443a4c1adc8635"} Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.414599 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e3e14057-d3be-433e-97a7-0b36f18382ee","Type":"ContainerDied","Data":"44267cacdffa14217dc9859dca23e969af5b7adbb4750d9c4f117ba9ccd3c77f"} Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.414552 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.416240 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e1c55305-7d18-44cb-90e5-b6793989abda","Type":"ContainerStarted","Data":"a2db576fbfc7eed07b697490f956015f0fceaa99ec44f3681ae02d7525bd0c81"} Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.416734 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.453832 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f993d38-e5bb-4ce4-8a9f-269695614f5e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.453874 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bp727\" (UniqueName: \"kubernetes.io/projected/9f993d38-e5bb-4ce4-8a9f-269695614f5e-kube-api-access-bp727\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.453887 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f993d38-e5bb-4ce4-8a9f-269695614f5e-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.466746 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.466723809 podStartE2EDuration="2.466723809s" podCreationTimestamp="2026-01-27 16:39:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:40:00.457801254 +0000 UTC m=+5586.438410362" watchObservedRunningTime="2026-01-27 16:40:00.466723809 +0000 UTC m=+5586.447332907" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.631558 4772 scope.go:117] "RemoveContainer" containerID="43c20e864a030473b88bc17f0d3fed51612d4bbdf0ff55280f36b4578127cc8a" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.632335 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.646295 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.663271 4772 scope.go:117] "RemoveContainer" containerID="3a53c1dc2aeb5bf544910214491023eeece1cfbdb9d21ada9d6cf8e1d2675ffb" Jan 27 16:40:00 crc kubenswrapper[4772]: E0127 16:40:00.663751 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a53c1dc2aeb5bf544910214491023eeece1cfbdb9d21ada9d6cf8e1d2675ffb\": container with ID starting with 3a53c1dc2aeb5bf544910214491023eeece1cfbdb9d21ada9d6cf8e1d2675ffb not found: ID does not exist" containerID="3a53c1dc2aeb5bf544910214491023eeece1cfbdb9d21ada9d6cf8e1d2675ffb" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.663783 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a53c1dc2aeb5bf544910214491023eeece1cfbdb9d21ada9d6cf8e1d2675ffb"} err="failed to get container status \"3a53c1dc2aeb5bf544910214491023eeece1cfbdb9d21ada9d6cf8e1d2675ffb\": rpc error: code = NotFound desc = could not find container \"3a53c1dc2aeb5bf544910214491023eeece1cfbdb9d21ada9d6cf8e1d2675ffb\": container with ID starting with 3a53c1dc2aeb5bf544910214491023eeece1cfbdb9d21ada9d6cf8e1d2675ffb not found: ID does not exist" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.663802 4772 scope.go:117] "RemoveContainer" containerID="43c20e864a030473b88bc17f0d3fed51612d4bbdf0ff55280f36b4578127cc8a" Jan 27 16:40:00 crc kubenswrapper[4772]: E0127 16:40:00.666070 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43c20e864a030473b88bc17f0d3fed51612d4bbdf0ff55280f36b4578127cc8a\": container with ID starting with 43c20e864a030473b88bc17f0d3fed51612d4bbdf0ff55280f36b4578127cc8a not found: ID does not exist" containerID="43c20e864a030473b88bc17f0d3fed51612d4bbdf0ff55280f36b4578127cc8a" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.666099 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43c20e864a030473b88bc17f0d3fed51612d4bbdf0ff55280f36b4578127cc8a"} err="failed to get container status \"43c20e864a030473b88bc17f0d3fed51612d4bbdf0ff55280f36b4578127cc8a\": rpc error: code = NotFound desc = could not find container \"43c20e864a030473b88bc17f0d3fed51612d4bbdf0ff55280f36b4578127cc8a\": container with ID starting with 43c20e864a030473b88bc17f0d3fed51612d4bbdf0ff55280f36b4578127cc8a not found: ID does not exist" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.666114 4772 scope.go:117] "RemoveContainer" containerID="3a53c1dc2aeb5bf544910214491023eeece1cfbdb9d21ada9d6cf8e1d2675ffb" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.667425 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a53c1dc2aeb5bf544910214491023eeece1cfbdb9d21ada9d6cf8e1d2675ffb"} err="failed to get container status \"3a53c1dc2aeb5bf544910214491023eeece1cfbdb9d21ada9d6cf8e1d2675ffb\": rpc error: code = NotFound desc = could not find container \"3a53c1dc2aeb5bf544910214491023eeece1cfbdb9d21ada9d6cf8e1d2675ffb\": container with ID starting with 3a53c1dc2aeb5bf544910214491023eeece1cfbdb9d21ada9d6cf8e1d2675ffb not found: ID does not exist" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.667459 4772 scope.go:117] "RemoveContainer" containerID="43c20e864a030473b88bc17f0d3fed51612d4bbdf0ff55280f36b4578127cc8a" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.667722 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43c20e864a030473b88bc17f0d3fed51612d4bbdf0ff55280f36b4578127cc8a"} err="failed to get container status \"43c20e864a030473b88bc17f0d3fed51612d4bbdf0ff55280f36b4578127cc8a\": rpc error: code = NotFound desc = could not find container \"43c20e864a030473b88bc17f0d3fed51612d4bbdf0ff55280f36b4578127cc8a\": container with ID starting with 43c20e864a030473b88bc17f0d3fed51612d4bbdf0ff55280f36b4578127cc8a not found: ID does not exist" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.667740 4772 scope.go:117] "RemoveContainer" containerID="d78776dd87426d6e9e796444fb4d98fcf0b08f9b784ed6f290516aa2ca7cd473" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.724575 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f993d38-e5bb-4ce4-8a9f-269695614f5e" path="/var/lib/kubelet/pods/9f993d38-e5bb-4ce4-8a9f-269695614f5e/volumes" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.725251 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 27 16:40:00 crc kubenswrapper[4772]: E0127 16:40:00.725536 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4c79395-a929-4f0d-8aa7-05f24412baed" containerName="nova-manage" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.725555 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4c79395-a929-4f0d-8aa7-05f24412baed" containerName="nova-manage" Jan 27 16:40:00 crc kubenswrapper[4772]: E0127 16:40:00.725567 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f993d38-e5bb-4ce4-8a9f-269695614f5e" containerName="nova-metadata-metadata" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.725573 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f993d38-e5bb-4ce4-8a9f-269695614f5e" containerName="nova-metadata-metadata" Jan 27 16:40:00 crc kubenswrapper[4772]: E0127 16:40:00.725590 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f993d38-e5bb-4ce4-8a9f-269695614f5e" containerName="nova-metadata-log" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.725596 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f993d38-e5bb-4ce4-8a9f-269695614f5e" containerName="nova-metadata-log" Jan 27 16:40:00 crc kubenswrapper[4772]: E0127 16:40:00.725607 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3e14057-d3be-433e-97a7-0b36f18382ee" containerName="nova-api-api" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.725613 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3e14057-d3be-433e-97a7-0b36f18382ee" containerName="nova-api-api" Jan 27 16:40:00 crc kubenswrapper[4772]: E0127 16:40:00.725638 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3e14057-d3be-433e-97a7-0b36f18382ee" containerName="nova-api-log" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.725644 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3e14057-d3be-433e-97a7-0b36f18382ee" containerName="nova-api-log" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.725803 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3e14057-d3be-433e-97a7-0b36f18382ee" containerName="nova-api-api" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.725813 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f993d38-e5bb-4ce4-8a9f-269695614f5e" containerName="nova-metadata-log" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.725826 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3e14057-d3be-433e-97a7-0b36f18382ee" containerName="nova-api-log" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.725832 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f993d38-e5bb-4ce4-8a9f-269695614f5e" containerName="nova-metadata-metadata" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.725845 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4c79395-a929-4f0d-8aa7-05f24412baed" containerName="nova-manage" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.726764 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.726783 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.726796 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.726806 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.727225 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.728233 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.728313 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.729875 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.730002 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.765643 4772 scope.go:117] "RemoveContainer" containerID="0271b9dbb1d2941b26f79c6625f660e41862d12903a306e76d443a4c1adc8635" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.792834 4772 scope.go:117] "RemoveContainer" containerID="d78776dd87426d6e9e796444fb4d98fcf0b08f9b784ed6f290516aa2ca7cd473" Jan 27 16:40:00 crc kubenswrapper[4772]: E0127 16:40:00.793281 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d78776dd87426d6e9e796444fb4d98fcf0b08f9b784ed6f290516aa2ca7cd473\": container with ID starting with d78776dd87426d6e9e796444fb4d98fcf0b08f9b784ed6f290516aa2ca7cd473 not found: ID does not exist" containerID="d78776dd87426d6e9e796444fb4d98fcf0b08f9b784ed6f290516aa2ca7cd473" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.793343 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d78776dd87426d6e9e796444fb4d98fcf0b08f9b784ed6f290516aa2ca7cd473"} err="failed to get container status \"d78776dd87426d6e9e796444fb4d98fcf0b08f9b784ed6f290516aa2ca7cd473\": rpc error: code = NotFound desc = could not find container \"d78776dd87426d6e9e796444fb4d98fcf0b08f9b784ed6f290516aa2ca7cd473\": container with ID starting with d78776dd87426d6e9e796444fb4d98fcf0b08f9b784ed6f290516aa2ca7cd473 not found: ID does not exist" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.793368 4772 scope.go:117] "RemoveContainer" containerID="0271b9dbb1d2941b26f79c6625f660e41862d12903a306e76d443a4c1adc8635" Jan 27 16:40:00 crc kubenswrapper[4772]: E0127 16:40:00.793746 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0271b9dbb1d2941b26f79c6625f660e41862d12903a306e76d443a4c1adc8635\": container with ID starting with 0271b9dbb1d2941b26f79c6625f660e41862d12903a306e76d443a4c1adc8635 not found: ID does not exist" containerID="0271b9dbb1d2941b26f79c6625f660e41862d12903a306e76d443a4c1adc8635" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.793776 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0271b9dbb1d2941b26f79c6625f660e41862d12903a306e76d443a4c1adc8635"} err="failed to get container status \"0271b9dbb1d2941b26f79c6625f660e41862d12903a306e76d443a4c1adc8635\": rpc error: code = NotFound desc = could not find container \"0271b9dbb1d2941b26f79c6625f660e41862d12903a306e76d443a4c1adc8635\": container with ID starting with 0271b9dbb1d2941b26f79c6625f660e41862d12903a306e76d443a4c1adc8635 not found: ID does not exist" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.793799 4772 scope.go:117] "RemoveContainer" containerID="d78776dd87426d6e9e796444fb4d98fcf0b08f9b784ed6f290516aa2ca7cd473" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.794027 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d78776dd87426d6e9e796444fb4d98fcf0b08f9b784ed6f290516aa2ca7cd473"} err="failed to get container status \"d78776dd87426d6e9e796444fb4d98fcf0b08f9b784ed6f290516aa2ca7cd473\": rpc error: code = NotFound desc = could not find container \"d78776dd87426d6e9e796444fb4d98fcf0b08f9b784ed6f290516aa2ca7cd473\": container with ID starting with d78776dd87426d6e9e796444fb4d98fcf0b08f9b784ed6f290516aa2ca7cd473 not found: ID does not exist" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.794066 4772 scope.go:117] "RemoveContainer" containerID="0271b9dbb1d2941b26f79c6625f660e41862d12903a306e76d443a4c1adc8635" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.794290 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0271b9dbb1d2941b26f79c6625f660e41862d12903a306e76d443a4c1adc8635"} err="failed to get container status \"0271b9dbb1d2941b26f79c6625f660e41862d12903a306e76d443a4c1adc8635\": rpc error: code = NotFound desc = could not find container \"0271b9dbb1d2941b26f79c6625f660e41862d12903a306e76d443a4c1adc8635\": container with ID starting with 0271b9dbb1d2941b26f79c6625f660e41862d12903a306e76d443a4c1adc8635 not found: ID does not exist" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.865898 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0576a6dc-0ce0-4e70-8be2-179989ed0c03-logs\") pod \"nova-metadata-0\" (UID: \"0576a6dc-0ce0-4e70-8be2-179989ed0c03\") " pod="openstack/nova-metadata-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.865985 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r644m\" (UniqueName: \"kubernetes.io/projected/0576a6dc-0ce0-4e70-8be2-179989ed0c03-kube-api-access-r644m\") pod \"nova-metadata-0\" (UID: \"0576a6dc-0ce0-4e70-8be2-179989ed0c03\") " pod="openstack/nova-metadata-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.866052 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/601c20bf-c7e2-4c26-a2da-7ebce076775a-config-data\") pod \"nova-api-0\" (UID: \"601c20bf-c7e2-4c26-a2da-7ebce076775a\") " pod="openstack/nova-api-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.866094 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0576a6dc-0ce0-4e70-8be2-179989ed0c03-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0576a6dc-0ce0-4e70-8be2-179989ed0c03\") " pod="openstack/nova-metadata-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.866119 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0576a6dc-0ce0-4e70-8be2-179989ed0c03-config-data\") pod \"nova-metadata-0\" (UID: \"0576a6dc-0ce0-4e70-8be2-179989ed0c03\") " pod="openstack/nova-metadata-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.866231 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/601c20bf-c7e2-4c26-a2da-7ebce076775a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"601c20bf-c7e2-4c26-a2da-7ebce076775a\") " pod="openstack/nova-api-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.866251 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/601c20bf-c7e2-4c26-a2da-7ebce076775a-logs\") pod \"nova-api-0\" (UID: \"601c20bf-c7e2-4c26-a2da-7ebce076775a\") " pod="openstack/nova-api-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.866267 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9rxm\" (UniqueName: \"kubernetes.io/projected/601c20bf-c7e2-4c26-a2da-7ebce076775a-kube-api-access-c9rxm\") pod \"nova-api-0\" (UID: \"601c20bf-c7e2-4c26-a2da-7ebce076775a\") " pod="openstack/nova-api-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.967936 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/601c20bf-c7e2-4c26-a2da-7ebce076775a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"601c20bf-c7e2-4c26-a2da-7ebce076775a\") " pod="openstack/nova-api-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.968012 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/601c20bf-c7e2-4c26-a2da-7ebce076775a-logs\") pod \"nova-api-0\" (UID: \"601c20bf-c7e2-4c26-a2da-7ebce076775a\") " pod="openstack/nova-api-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.968039 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9rxm\" (UniqueName: \"kubernetes.io/projected/601c20bf-c7e2-4c26-a2da-7ebce076775a-kube-api-access-c9rxm\") pod \"nova-api-0\" (UID: \"601c20bf-c7e2-4c26-a2da-7ebce076775a\") " pod="openstack/nova-api-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.968090 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0576a6dc-0ce0-4e70-8be2-179989ed0c03-logs\") pod \"nova-metadata-0\" (UID: \"0576a6dc-0ce0-4e70-8be2-179989ed0c03\") " pod="openstack/nova-metadata-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.968139 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r644m\" (UniqueName: \"kubernetes.io/projected/0576a6dc-0ce0-4e70-8be2-179989ed0c03-kube-api-access-r644m\") pod \"nova-metadata-0\" (UID: \"0576a6dc-0ce0-4e70-8be2-179989ed0c03\") " pod="openstack/nova-metadata-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.968199 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/601c20bf-c7e2-4c26-a2da-7ebce076775a-config-data\") pod \"nova-api-0\" (UID: \"601c20bf-c7e2-4c26-a2da-7ebce076775a\") " pod="openstack/nova-api-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.968241 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0576a6dc-0ce0-4e70-8be2-179989ed0c03-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0576a6dc-0ce0-4e70-8be2-179989ed0c03\") " pod="openstack/nova-metadata-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.968261 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0576a6dc-0ce0-4e70-8be2-179989ed0c03-config-data\") pod \"nova-metadata-0\" (UID: \"0576a6dc-0ce0-4e70-8be2-179989ed0c03\") " pod="openstack/nova-metadata-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.968884 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0576a6dc-0ce0-4e70-8be2-179989ed0c03-logs\") pod \"nova-metadata-0\" (UID: \"0576a6dc-0ce0-4e70-8be2-179989ed0c03\") " pod="openstack/nova-metadata-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.969529 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/601c20bf-c7e2-4c26-a2da-7ebce076775a-logs\") pod \"nova-api-0\" (UID: \"601c20bf-c7e2-4c26-a2da-7ebce076775a\") " pod="openstack/nova-api-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.972138 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/601c20bf-c7e2-4c26-a2da-7ebce076775a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"601c20bf-c7e2-4c26-a2da-7ebce076775a\") " pod="openstack/nova-api-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.973207 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0576a6dc-0ce0-4e70-8be2-179989ed0c03-config-data\") pod \"nova-metadata-0\" (UID: \"0576a6dc-0ce0-4e70-8be2-179989ed0c03\") " pod="openstack/nova-metadata-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.974214 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/601c20bf-c7e2-4c26-a2da-7ebce076775a-config-data\") pod \"nova-api-0\" (UID: \"601c20bf-c7e2-4c26-a2da-7ebce076775a\") " pod="openstack/nova-api-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.981089 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0576a6dc-0ce0-4e70-8be2-179989ed0c03-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0576a6dc-0ce0-4e70-8be2-179989ed0c03\") " pod="openstack/nova-metadata-0" Jan 27 16:40:00 crc kubenswrapper[4772]: I0127 16:40:00.998833 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9rxm\" (UniqueName: \"kubernetes.io/projected/601c20bf-c7e2-4c26-a2da-7ebce076775a-kube-api-access-c9rxm\") pod \"nova-api-0\" (UID: \"601c20bf-c7e2-4c26-a2da-7ebce076775a\") " pod="openstack/nova-api-0" Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.000377 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7758766f57-6fk67" Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.008002 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r644m\" (UniqueName: \"kubernetes.io/projected/0576a6dc-0ce0-4e70-8be2-179989ed0c03-kube-api-access-r644m\") pod \"nova-metadata-0\" (UID: \"0576a6dc-0ce0-4e70-8be2-179989ed0c03\") " pod="openstack/nova-metadata-0" Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.061246 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-545c956d45-h25qs"] Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.061482 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-545c956d45-h25qs" podUID="6da65d05-29e3-4d97-869f-d3386a45a38e" containerName="dnsmasq-dns" containerID="cri-o://54f0c375865810e4cb819d2019e6c8ab7827175c7d633ed00deb75c9600fb4b9" gracePeriod=10 Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.082635 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.086123 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.102223 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.131456 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.438887 4772 generic.go:334] "Generic (PLEG): container finished" podID="6da65d05-29e3-4d97-869f-d3386a45a38e" containerID="54f0c375865810e4cb819d2019e6c8ab7827175c7d633ed00deb75c9600fb4b9" exitCode=0 Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.438967 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-545c956d45-h25qs" event={"ID":"6da65d05-29e3-4d97-869f-d3386a45a38e","Type":"ContainerDied","Data":"54f0c375865810e4cb819d2019e6c8ab7827175c7d633ed00deb75c9600fb4b9"} Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.448672 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.514114 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-545c956d45-h25qs" Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.680226 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.683187 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6da65d05-29e3-4d97-869f-d3386a45a38e-ovsdbserver-nb\") pod \"6da65d05-29e3-4d97-869f-d3386a45a38e\" (UID: \"6da65d05-29e3-4d97-869f-d3386a45a38e\") " Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.683244 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6da65d05-29e3-4d97-869f-d3386a45a38e-config\") pod \"6da65d05-29e3-4d97-869f-d3386a45a38e\" (UID: \"6da65d05-29e3-4d97-869f-d3386a45a38e\") " Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.683308 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6da65d05-29e3-4d97-869f-d3386a45a38e-ovsdbserver-sb\") pod \"6da65d05-29e3-4d97-869f-d3386a45a38e\" (UID: \"6da65d05-29e3-4d97-869f-d3386a45a38e\") " Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.683368 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sh8vx\" (UniqueName: \"kubernetes.io/projected/6da65d05-29e3-4d97-869f-d3386a45a38e-kube-api-access-sh8vx\") pod \"6da65d05-29e3-4d97-869f-d3386a45a38e\" (UID: \"6da65d05-29e3-4d97-869f-d3386a45a38e\") " Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.683409 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6da65d05-29e3-4d97-869f-d3386a45a38e-dns-svc\") pod \"6da65d05-29e3-4d97-869f-d3386a45a38e\" (UID: \"6da65d05-29e3-4d97-869f-d3386a45a38e\") " Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.689665 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.693765 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6da65d05-29e3-4d97-869f-d3386a45a38e-kube-api-access-sh8vx" (OuterVolumeSpecName: "kube-api-access-sh8vx") pod "6da65d05-29e3-4d97-869f-d3386a45a38e" (UID: "6da65d05-29e3-4d97-869f-d3386a45a38e"). InnerVolumeSpecName "kube-api-access-sh8vx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.741612 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6da65d05-29e3-4d97-869f-d3386a45a38e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6da65d05-29e3-4d97-869f-d3386a45a38e" (UID: "6da65d05-29e3-4d97-869f-d3386a45a38e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.748409 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6da65d05-29e3-4d97-869f-d3386a45a38e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6da65d05-29e3-4d97-869f-d3386a45a38e" (UID: "6da65d05-29e3-4d97-869f-d3386a45a38e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.770039 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6da65d05-29e3-4d97-869f-d3386a45a38e-config" (OuterVolumeSpecName: "config") pod "6da65d05-29e3-4d97-869f-d3386a45a38e" (UID: "6da65d05-29e3-4d97-869f-d3386a45a38e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.785974 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sh8vx\" (UniqueName: \"kubernetes.io/projected/6da65d05-29e3-4d97-869f-d3386a45a38e-kube-api-access-sh8vx\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.786230 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6da65d05-29e3-4d97-869f-d3386a45a38e-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.786361 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6da65d05-29e3-4d97-869f-d3386a45a38e-config\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.786458 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6da65d05-29e3-4d97-869f-d3386a45a38e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.787580 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6da65d05-29e3-4d97-869f-d3386a45a38e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6da65d05-29e3-4d97-869f-d3386a45a38e" (UID: "6da65d05-29e3-4d97-869f-d3386a45a38e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:40:01 crc kubenswrapper[4772]: I0127 16:40:01.889136 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6da65d05-29e3-4d97-869f-d3386a45a38e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:02 crc kubenswrapper[4772]: I0127 16:40:02.451664 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0576a6dc-0ce0-4e70-8be2-179989ed0c03","Type":"ContainerStarted","Data":"a0a5b8400ff9131c6a732b564464e4682a47bd72327030af67266b99caae0aa5"} Jan 27 16:40:02 crc kubenswrapper[4772]: I0127 16:40:02.452072 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0576a6dc-0ce0-4e70-8be2-179989ed0c03","Type":"ContainerStarted","Data":"b99ac9784926d43b5b9010bb3bf145d936b87ce053aad43267f30ea8ebbbf38a"} Jan 27 16:40:02 crc kubenswrapper[4772]: I0127 16:40:02.452097 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0576a6dc-0ce0-4e70-8be2-179989ed0c03","Type":"ContainerStarted","Data":"272b9d5f7574afede07d814583b3f863aa00e225817941ab1c1713433efa66a6"} Jan 27 16:40:02 crc kubenswrapper[4772]: I0127 16:40:02.455159 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-545c956d45-h25qs" Jan 27 16:40:02 crc kubenswrapper[4772]: I0127 16:40:02.455160 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-545c956d45-h25qs" event={"ID":"6da65d05-29e3-4d97-869f-d3386a45a38e","Type":"ContainerDied","Data":"1b2cc021d3453413ef029602e151c491355c94bca54c6c90c56cd7685dc93518"} Jan 27 16:40:02 crc kubenswrapper[4772]: I0127 16:40:02.455397 4772 scope.go:117] "RemoveContainer" containerID="54f0c375865810e4cb819d2019e6c8ab7827175c7d633ed00deb75c9600fb4b9" Jan 27 16:40:02 crc kubenswrapper[4772]: I0127 16:40:02.459362 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"601c20bf-c7e2-4c26-a2da-7ebce076775a","Type":"ContainerStarted","Data":"6905111d10a31c85326f0556bb66e5eba75c934131e8d3fe500da3f080ca6789"} Jan 27 16:40:02 crc kubenswrapper[4772]: I0127 16:40:02.459399 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"601c20bf-c7e2-4c26-a2da-7ebce076775a","Type":"ContainerStarted","Data":"eae5df57a84c79de36f586452630170afcbfff0b1e0fdb27c896a858c117ff8b"} Jan 27 16:40:02 crc kubenswrapper[4772]: I0127 16:40:02.459421 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"601c20bf-c7e2-4c26-a2da-7ebce076775a","Type":"ContainerStarted","Data":"c71aafaf39ad009fb7eb0ade8a547f4797867d098bc1c0cd4f5e22f563ee2d30"} Jan 27 16:40:02 crc kubenswrapper[4772]: I0127 16:40:02.483954 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.483935716 podStartE2EDuration="2.483935716s" podCreationTimestamp="2026-01-27 16:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:40:02.472333486 +0000 UTC m=+5588.452942604" watchObservedRunningTime="2026-01-27 16:40:02.483935716 +0000 UTC m=+5588.464544814" Jan 27 16:40:02 crc kubenswrapper[4772]: I0127 16:40:02.498027 4772 scope.go:117] "RemoveContainer" containerID="b8551d53299337e281772b00e27dddd2da7a39a65b37574b860847ac8a5c90ba" Jan 27 16:40:02 crc kubenswrapper[4772]: I0127 16:40:02.503720 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.50370346 podStartE2EDuration="2.50370346s" podCreationTimestamp="2026-01-27 16:40:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:40:02.495739433 +0000 UTC m=+5588.476348531" watchObservedRunningTime="2026-01-27 16:40:02.50370346 +0000 UTC m=+5588.484312558" Jan 27 16:40:02 crc kubenswrapper[4772]: I0127 16:40:02.526381 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-545c956d45-h25qs"] Jan 27 16:40:02 crc kubenswrapper[4772]: I0127 16:40:02.537830 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-545c956d45-h25qs"] Jan 27 16:40:02 crc kubenswrapper[4772]: I0127 16:40:02.677476 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6da65d05-29e3-4d97-869f-d3386a45a38e" path="/var/lib/kubelet/pods/6da65d05-29e3-4d97-869f-d3386a45a38e/volumes" Jan 27 16:40:02 crc kubenswrapper[4772]: I0127 16:40:02.678412 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3e14057-d3be-433e-97a7-0b36f18382ee" path="/var/lib/kubelet/pods/e3e14057-d3be-433e-97a7-0b36f18382ee/volumes" Jan 27 16:40:03 crc kubenswrapper[4772]: I0127 16:40:03.871834 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.029381 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26d4226f-e574-498f-a2ce-e0db8f83a8d3-config-data\") pod \"26d4226f-e574-498f-a2ce-e0db8f83a8d3\" (UID: \"26d4226f-e574-498f-a2ce-e0db8f83a8d3\") " Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.029441 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26d4226f-e574-498f-a2ce-e0db8f83a8d3-combined-ca-bundle\") pod \"26d4226f-e574-498f-a2ce-e0db8f83a8d3\" (UID: \"26d4226f-e574-498f-a2ce-e0db8f83a8d3\") " Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.029488 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqgtr\" (UniqueName: \"kubernetes.io/projected/26d4226f-e574-498f-a2ce-e0db8f83a8d3-kube-api-access-tqgtr\") pod \"26d4226f-e574-498f-a2ce-e0db8f83a8d3\" (UID: \"26d4226f-e574-498f-a2ce-e0db8f83a8d3\") " Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.035338 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26d4226f-e574-498f-a2ce-e0db8f83a8d3-kube-api-access-tqgtr" (OuterVolumeSpecName: "kube-api-access-tqgtr") pod "26d4226f-e574-498f-a2ce-e0db8f83a8d3" (UID: "26d4226f-e574-498f-a2ce-e0db8f83a8d3"). InnerVolumeSpecName "kube-api-access-tqgtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.051854 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26d4226f-e574-498f-a2ce-e0db8f83a8d3-config-data" (OuterVolumeSpecName: "config-data") pod "26d4226f-e574-498f-a2ce-e0db8f83a8d3" (UID: "26d4226f-e574-498f-a2ce-e0db8f83a8d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.053891 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26d4226f-e574-498f-a2ce-e0db8f83a8d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "26d4226f-e574-498f-a2ce-e0db8f83a8d3" (UID: "26d4226f-e574-498f-a2ce-e0db8f83a8d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.131831 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26d4226f-e574-498f-a2ce-e0db8f83a8d3-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.131875 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26d4226f-e574-498f-a2ce-e0db8f83a8d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.131894 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqgtr\" (UniqueName: \"kubernetes.io/projected/26d4226f-e574-498f-a2ce-e0db8f83a8d3-kube-api-access-tqgtr\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.478786 4772 generic.go:334] "Generic (PLEG): container finished" podID="26d4226f-e574-498f-a2ce-e0db8f83a8d3" containerID="d6c1d8ec858944c73c32d29d0613f42bb401204fa7425a43dbda27a9aa01ac58" exitCode=0 Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.478847 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.478856 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"26d4226f-e574-498f-a2ce-e0db8f83a8d3","Type":"ContainerDied","Data":"d6c1d8ec858944c73c32d29d0613f42bb401204fa7425a43dbda27a9aa01ac58"} Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.478927 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"26d4226f-e574-498f-a2ce-e0db8f83a8d3","Type":"ContainerDied","Data":"1c139ace8ceb66ebe7302c272fa99b828aba2d79a11db30142d7bb5b26746bbd"} Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.478961 4772 scope.go:117] "RemoveContainer" containerID="d6c1d8ec858944c73c32d29d0613f42bb401204fa7425a43dbda27a9aa01ac58" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.499536 4772 scope.go:117] "RemoveContainer" containerID="d6c1d8ec858944c73c32d29d0613f42bb401204fa7425a43dbda27a9aa01ac58" Jan 27 16:40:04 crc kubenswrapper[4772]: E0127 16:40:04.500065 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6c1d8ec858944c73c32d29d0613f42bb401204fa7425a43dbda27a9aa01ac58\": container with ID starting with d6c1d8ec858944c73c32d29d0613f42bb401204fa7425a43dbda27a9aa01ac58 not found: ID does not exist" containerID="d6c1d8ec858944c73c32d29d0613f42bb401204fa7425a43dbda27a9aa01ac58" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.500113 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6c1d8ec858944c73c32d29d0613f42bb401204fa7425a43dbda27a9aa01ac58"} err="failed to get container status \"d6c1d8ec858944c73c32d29d0613f42bb401204fa7425a43dbda27a9aa01ac58\": rpc error: code = NotFound desc = could not find container \"d6c1d8ec858944c73c32d29d0613f42bb401204fa7425a43dbda27a9aa01ac58\": container with ID starting with d6c1d8ec858944c73c32d29d0613f42bb401204fa7425a43dbda27a9aa01ac58 not found: ID does not exist" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.522344 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.534390 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.546948 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 16:40:04 crc kubenswrapper[4772]: E0127 16:40:04.547376 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26d4226f-e574-498f-a2ce-e0db8f83a8d3" containerName="nova-scheduler-scheduler" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.547393 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="26d4226f-e574-498f-a2ce-e0db8f83a8d3" containerName="nova-scheduler-scheduler" Jan 27 16:40:04 crc kubenswrapper[4772]: E0127 16:40:04.547413 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6da65d05-29e3-4d97-869f-d3386a45a38e" containerName="dnsmasq-dns" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.547421 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="6da65d05-29e3-4d97-869f-d3386a45a38e" containerName="dnsmasq-dns" Jan 27 16:40:04 crc kubenswrapper[4772]: E0127 16:40:04.547436 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6da65d05-29e3-4d97-869f-d3386a45a38e" containerName="init" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.547442 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="6da65d05-29e3-4d97-869f-d3386a45a38e" containerName="init" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.547597 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="26d4226f-e574-498f-a2ce-e0db8f83a8d3" containerName="nova-scheduler-scheduler" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.547610 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="6da65d05-29e3-4d97-869f-d3386a45a38e" containerName="dnsmasq-dns" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.548261 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.550361 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.570678 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.641622 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02c9bfb0-5f95-4692-bb48-ffd47f6682a8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"02c9bfb0-5f95-4692-bb48-ffd47f6682a8\") " pod="openstack/nova-scheduler-0" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.641727 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2khtr\" (UniqueName: \"kubernetes.io/projected/02c9bfb0-5f95-4692-bb48-ffd47f6682a8-kube-api-access-2khtr\") pod \"nova-scheduler-0\" (UID: \"02c9bfb0-5f95-4692-bb48-ffd47f6682a8\") " pod="openstack/nova-scheduler-0" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.641788 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02c9bfb0-5f95-4692-bb48-ffd47f6682a8-config-data\") pod \"nova-scheduler-0\" (UID: \"02c9bfb0-5f95-4692-bb48-ffd47f6682a8\") " pod="openstack/nova-scheduler-0" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.684535 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26d4226f-e574-498f-a2ce-e0db8f83a8d3" path="/var/lib/kubelet/pods/26d4226f-e574-498f-a2ce-e0db8f83a8d3/volumes" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.743945 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2khtr\" (UniqueName: \"kubernetes.io/projected/02c9bfb0-5f95-4692-bb48-ffd47f6682a8-kube-api-access-2khtr\") pod \"nova-scheduler-0\" (UID: \"02c9bfb0-5f95-4692-bb48-ffd47f6682a8\") " pod="openstack/nova-scheduler-0" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.744120 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02c9bfb0-5f95-4692-bb48-ffd47f6682a8-config-data\") pod \"nova-scheduler-0\" (UID: \"02c9bfb0-5f95-4692-bb48-ffd47f6682a8\") " pod="openstack/nova-scheduler-0" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.744354 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02c9bfb0-5f95-4692-bb48-ffd47f6682a8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"02c9bfb0-5f95-4692-bb48-ffd47f6682a8\") " pod="openstack/nova-scheduler-0" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.751753 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02c9bfb0-5f95-4692-bb48-ffd47f6682a8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"02c9bfb0-5f95-4692-bb48-ffd47f6682a8\") " pod="openstack/nova-scheduler-0" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.752349 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02c9bfb0-5f95-4692-bb48-ffd47f6682a8-config-data\") pod \"nova-scheduler-0\" (UID: \"02c9bfb0-5f95-4692-bb48-ffd47f6682a8\") " pod="openstack/nova-scheduler-0" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.761678 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2khtr\" (UniqueName: \"kubernetes.io/projected/02c9bfb0-5f95-4692-bb48-ffd47f6682a8-kube-api-access-2khtr\") pod \"nova-scheduler-0\" (UID: \"02c9bfb0-5f95-4692-bb48-ffd47f6682a8\") " pod="openstack/nova-scheduler-0" Jan 27 16:40:04 crc kubenswrapper[4772]: I0127 16:40:04.877267 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 16:40:05 crc kubenswrapper[4772]: I0127 16:40:05.356675 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 16:40:05 crc kubenswrapper[4772]: W0127 16:40:05.364918 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02c9bfb0_5f95_4692_bb48_ffd47f6682a8.slice/crio-3c6a3e07c72ed8c1bb7f2e42a623b437deaa88e1d9979eefafaa45fb0ff45499 WatchSource:0}: Error finding container 3c6a3e07c72ed8c1bb7f2e42a623b437deaa88e1d9979eefafaa45fb0ff45499: Status 404 returned error can't find the container with id 3c6a3e07c72ed8c1bb7f2e42a623b437deaa88e1d9979eefafaa45fb0ff45499 Jan 27 16:40:05 crc kubenswrapper[4772]: I0127 16:40:05.491523 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"02c9bfb0-5f95-4692-bb48-ffd47f6682a8","Type":"ContainerStarted","Data":"3c6a3e07c72ed8c1bb7f2e42a623b437deaa88e1d9979eefafaa45fb0ff45499"} Jan 27 16:40:06 crc kubenswrapper[4772]: I0127 16:40:06.083330 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 27 16:40:06 crc kubenswrapper[4772]: I0127 16:40:06.083615 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 27 16:40:06 crc kubenswrapper[4772]: I0127 16:40:06.502126 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"02c9bfb0-5f95-4692-bb48-ffd47f6682a8","Type":"ContainerStarted","Data":"6f9e0ed02577ed79c4498daebc32e5d9adfca8bf0094d3a8ddb00a6642f3d336"} Jan 27 16:40:06 crc kubenswrapper[4772]: I0127 16:40:06.529233 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.529209142 podStartE2EDuration="2.529209142s" podCreationTimestamp="2026-01-27 16:40:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:40:06.518220459 +0000 UTC m=+5592.498829577" watchObservedRunningTime="2026-01-27 16:40:06.529209142 +0000 UTC m=+5592.509818260" Jan 27 16:40:07 crc kubenswrapper[4772]: I0127 16:40:07.079527 4772 scope.go:117] "RemoveContainer" containerID="1bbb692ab013dbd6aab71415abc05175b7c980ab72c48771e22a05f7a67573bd" Jan 27 16:40:08 crc kubenswrapper[4772]: I0127 16:40:08.807121 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Jan 27 16:40:09 crc kubenswrapper[4772]: I0127 16:40:09.365628 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-p6mbs"] Jan 27 16:40:09 crc kubenswrapper[4772]: I0127 16:40:09.367262 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-p6mbs" Jan 27 16:40:09 crc kubenswrapper[4772]: I0127 16:40:09.369261 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Jan 27 16:40:09 crc kubenswrapper[4772]: I0127 16:40:09.371095 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Jan 27 16:40:09 crc kubenswrapper[4772]: I0127 16:40:09.406927 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-p6mbs"] Jan 27 16:40:09 crc kubenswrapper[4772]: I0127 16:40:09.429900 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaefb4fd-175a-4431-bd2f-7fc3205684b9-config-data\") pod \"nova-cell1-cell-mapping-p6mbs\" (UID: \"eaefb4fd-175a-4431-bd2f-7fc3205684b9\") " pod="openstack/nova-cell1-cell-mapping-p6mbs" Jan 27 16:40:09 crc kubenswrapper[4772]: I0127 16:40:09.430084 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdxkp\" (UniqueName: \"kubernetes.io/projected/eaefb4fd-175a-4431-bd2f-7fc3205684b9-kube-api-access-vdxkp\") pod \"nova-cell1-cell-mapping-p6mbs\" (UID: \"eaefb4fd-175a-4431-bd2f-7fc3205684b9\") " pod="openstack/nova-cell1-cell-mapping-p6mbs" Jan 27 16:40:09 crc kubenswrapper[4772]: I0127 16:40:09.430451 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaefb4fd-175a-4431-bd2f-7fc3205684b9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-p6mbs\" (UID: \"eaefb4fd-175a-4431-bd2f-7fc3205684b9\") " pod="openstack/nova-cell1-cell-mapping-p6mbs" Jan 27 16:40:09 crc kubenswrapper[4772]: I0127 16:40:09.430506 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaefb4fd-175a-4431-bd2f-7fc3205684b9-scripts\") pod \"nova-cell1-cell-mapping-p6mbs\" (UID: \"eaefb4fd-175a-4431-bd2f-7fc3205684b9\") " pod="openstack/nova-cell1-cell-mapping-p6mbs" Jan 27 16:40:09 crc kubenswrapper[4772]: I0127 16:40:09.532520 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaefb4fd-175a-4431-bd2f-7fc3205684b9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-p6mbs\" (UID: \"eaefb4fd-175a-4431-bd2f-7fc3205684b9\") " pod="openstack/nova-cell1-cell-mapping-p6mbs" Jan 27 16:40:09 crc kubenswrapper[4772]: I0127 16:40:09.532571 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaefb4fd-175a-4431-bd2f-7fc3205684b9-scripts\") pod \"nova-cell1-cell-mapping-p6mbs\" (UID: \"eaefb4fd-175a-4431-bd2f-7fc3205684b9\") " pod="openstack/nova-cell1-cell-mapping-p6mbs" Jan 27 16:40:09 crc kubenswrapper[4772]: I0127 16:40:09.532602 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaefb4fd-175a-4431-bd2f-7fc3205684b9-config-data\") pod \"nova-cell1-cell-mapping-p6mbs\" (UID: \"eaefb4fd-175a-4431-bd2f-7fc3205684b9\") " pod="openstack/nova-cell1-cell-mapping-p6mbs" Jan 27 16:40:09 crc kubenswrapper[4772]: I0127 16:40:09.532693 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdxkp\" (UniqueName: \"kubernetes.io/projected/eaefb4fd-175a-4431-bd2f-7fc3205684b9-kube-api-access-vdxkp\") pod \"nova-cell1-cell-mapping-p6mbs\" (UID: \"eaefb4fd-175a-4431-bd2f-7fc3205684b9\") " pod="openstack/nova-cell1-cell-mapping-p6mbs" Jan 27 16:40:09 crc kubenswrapper[4772]: I0127 16:40:09.539959 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaefb4fd-175a-4431-bd2f-7fc3205684b9-scripts\") pod \"nova-cell1-cell-mapping-p6mbs\" (UID: \"eaefb4fd-175a-4431-bd2f-7fc3205684b9\") " pod="openstack/nova-cell1-cell-mapping-p6mbs" Jan 27 16:40:09 crc kubenswrapper[4772]: I0127 16:40:09.543793 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaefb4fd-175a-4431-bd2f-7fc3205684b9-config-data\") pod \"nova-cell1-cell-mapping-p6mbs\" (UID: \"eaefb4fd-175a-4431-bd2f-7fc3205684b9\") " pod="openstack/nova-cell1-cell-mapping-p6mbs" Jan 27 16:40:09 crc kubenswrapper[4772]: I0127 16:40:09.544383 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaefb4fd-175a-4431-bd2f-7fc3205684b9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-p6mbs\" (UID: \"eaefb4fd-175a-4431-bd2f-7fc3205684b9\") " pod="openstack/nova-cell1-cell-mapping-p6mbs" Jan 27 16:40:09 crc kubenswrapper[4772]: I0127 16:40:09.554799 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdxkp\" (UniqueName: \"kubernetes.io/projected/eaefb4fd-175a-4431-bd2f-7fc3205684b9-kube-api-access-vdxkp\") pod \"nova-cell1-cell-mapping-p6mbs\" (UID: \"eaefb4fd-175a-4431-bd2f-7fc3205684b9\") " pod="openstack/nova-cell1-cell-mapping-p6mbs" Jan 27 16:40:09 crc kubenswrapper[4772]: I0127 16:40:09.704927 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-p6mbs" Jan 27 16:40:09 crc kubenswrapper[4772]: I0127 16:40:09.878427 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 27 16:40:10 crc kubenswrapper[4772]: I0127 16:40:10.138060 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-p6mbs"] Jan 27 16:40:10 crc kubenswrapper[4772]: I0127 16:40:10.537734 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-p6mbs" event={"ID":"eaefb4fd-175a-4431-bd2f-7fc3205684b9","Type":"ContainerStarted","Data":"f57fe908a2ad59651426657b881a5c35f1371cdc115dc93b86dbcb58952ce8a9"} Jan 27 16:40:10 crc kubenswrapper[4772]: I0127 16:40:10.538056 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-p6mbs" event={"ID":"eaefb4fd-175a-4431-bd2f-7fc3205684b9","Type":"ContainerStarted","Data":"abe8b822cc40f21143a3b475b6bbd880c198650912757b0aefe22d25ba054986"} Jan 27 16:40:11 crc kubenswrapper[4772]: I0127 16:40:11.084309 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 27 16:40:11 crc kubenswrapper[4772]: I0127 16:40:11.084357 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 27 16:40:11 crc kubenswrapper[4772]: I0127 16:40:11.086776 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 27 16:40:11 crc kubenswrapper[4772]: I0127 16:40:11.086821 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 27 16:40:12 crc kubenswrapper[4772]: I0127 16:40:12.248416 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="601c20bf-c7e2-4c26-a2da-7ebce076775a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.61:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 27 16:40:12 crc kubenswrapper[4772]: I0127 16:40:12.248882 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0576a6dc-0ce0-4e70-8be2-179989ed0c03" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.60:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 27 16:40:12 crc kubenswrapper[4772]: I0127 16:40:12.249130 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0576a6dc-0ce0-4e70-8be2-179989ed0c03" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.60:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 27 16:40:12 crc kubenswrapper[4772]: I0127 16:40:12.249211 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="601c20bf-c7e2-4c26-a2da-7ebce076775a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.61:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 27 16:40:14 crc kubenswrapper[4772]: I0127 16:40:14.877943 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Jan 27 16:40:14 crc kubenswrapper[4772]: I0127 16:40:14.905815 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Jan 27 16:40:14 crc kubenswrapper[4772]: I0127 16:40:14.925977 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-p6mbs" podStartSLOduration=5.925956122 podStartE2EDuration="5.925956122s" podCreationTimestamp="2026-01-27 16:40:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:40:10.5686014 +0000 UTC m=+5596.549210498" watchObservedRunningTime="2026-01-27 16:40:14.925956122 +0000 UTC m=+5600.906565220" Jan 27 16:40:15 crc kubenswrapper[4772]: I0127 16:40:15.601089 4772 generic.go:334] "Generic (PLEG): container finished" podID="eaefb4fd-175a-4431-bd2f-7fc3205684b9" containerID="f57fe908a2ad59651426657b881a5c35f1371cdc115dc93b86dbcb58952ce8a9" exitCode=0 Jan 27 16:40:15 crc kubenswrapper[4772]: I0127 16:40:15.601193 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-p6mbs" event={"ID":"eaefb4fd-175a-4431-bd2f-7fc3205684b9","Type":"ContainerDied","Data":"f57fe908a2ad59651426657b881a5c35f1371cdc115dc93b86dbcb58952ce8a9"} Jan 27 16:40:15 crc kubenswrapper[4772]: I0127 16:40:15.632608 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Jan 27 16:40:16 crc kubenswrapper[4772]: I0127 16:40:16.915996 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-p6mbs" Jan 27 16:40:16 crc kubenswrapper[4772]: I0127 16:40:16.984508 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdxkp\" (UniqueName: \"kubernetes.io/projected/eaefb4fd-175a-4431-bd2f-7fc3205684b9-kube-api-access-vdxkp\") pod \"eaefb4fd-175a-4431-bd2f-7fc3205684b9\" (UID: \"eaefb4fd-175a-4431-bd2f-7fc3205684b9\") " Jan 27 16:40:16 crc kubenswrapper[4772]: I0127 16:40:16.984583 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaefb4fd-175a-4431-bd2f-7fc3205684b9-combined-ca-bundle\") pod \"eaefb4fd-175a-4431-bd2f-7fc3205684b9\" (UID: \"eaefb4fd-175a-4431-bd2f-7fc3205684b9\") " Jan 27 16:40:16 crc kubenswrapper[4772]: I0127 16:40:16.984631 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaefb4fd-175a-4431-bd2f-7fc3205684b9-config-data\") pod \"eaefb4fd-175a-4431-bd2f-7fc3205684b9\" (UID: \"eaefb4fd-175a-4431-bd2f-7fc3205684b9\") " Jan 27 16:40:16 crc kubenswrapper[4772]: I0127 16:40:16.984663 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaefb4fd-175a-4431-bd2f-7fc3205684b9-scripts\") pod \"eaefb4fd-175a-4431-bd2f-7fc3205684b9\" (UID: \"eaefb4fd-175a-4431-bd2f-7fc3205684b9\") " Jan 27 16:40:16 crc kubenswrapper[4772]: I0127 16:40:16.991464 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaefb4fd-175a-4431-bd2f-7fc3205684b9-kube-api-access-vdxkp" (OuterVolumeSpecName: "kube-api-access-vdxkp") pod "eaefb4fd-175a-4431-bd2f-7fc3205684b9" (UID: "eaefb4fd-175a-4431-bd2f-7fc3205684b9"). InnerVolumeSpecName "kube-api-access-vdxkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:40:16 crc kubenswrapper[4772]: I0127 16:40:16.994024 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaefb4fd-175a-4431-bd2f-7fc3205684b9-scripts" (OuterVolumeSpecName: "scripts") pod "eaefb4fd-175a-4431-bd2f-7fc3205684b9" (UID: "eaefb4fd-175a-4431-bd2f-7fc3205684b9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:40:17 crc kubenswrapper[4772]: I0127 16:40:17.010052 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaefb4fd-175a-4431-bd2f-7fc3205684b9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eaefb4fd-175a-4431-bd2f-7fc3205684b9" (UID: "eaefb4fd-175a-4431-bd2f-7fc3205684b9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:40:17 crc kubenswrapper[4772]: I0127 16:40:17.015159 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaefb4fd-175a-4431-bd2f-7fc3205684b9-config-data" (OuterVolumeSpecName: "config-data") pod "eaefb4fd-175a-4431-bd2f-7fc3205684b9" (UID: "eaefb4fd-175a-4431-bd2f-7fc3205684b9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:40:17 crc kubenswrapper[4772]: I0127 16:40:17.086497 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdxkp\" (UniqueName: \"kubernetes.io/projected/eaefb4fd-175a-4431-bd2f-7fc3205684b9-kube-api-access-vdxkp\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:17 crc kubenswrapper[4772]: I0127 16:40:17.086531 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaefb4fd-175a-4431-bd2f-7fc3205684b9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:17 crc kubenswrapper[4772]: I0127 16:40:17.086541 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaefb4fd-175a-4431-bd2f-7fc3205684b9-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:17 crc kubenswrapper[4772]: I0127 16:40:17.086549 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaefb4fd-175a-4431-bd2f-7fc3205684b9-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:17 crc kubenswrapper[4772]: I0127 16:40:17.619834 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-p6mbs" event={"ID":"eaefb4fd-175a-4431-bd2f-7fc3205684b9","Type":"ContainerDied","Data":"abe8b822cc40f21143a3b475b6bbd880c198650912757b0aefe22d25ba054986"} Jan 27 16:40:17 crc kubenswrapper[4772]: I0127 16:40:17.620193 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abe8b822cc40f21143a3b475b6bbd880c198650912757b0aefe22d25ba054986" Jan 27 16:40:17 crc kubenswrapper[4772]: I0127 16:40:17.619925 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-p6mbs" Jan 27 16:40:17 crc kubenswrapper[4772]: I0127 16:40:17.807585 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 27 16:40:17 crc kubenswrapper[4772]: I0127 16:40:17.807852 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="601c20bf-c7e2-4c26-a2da-7ebce076775a" containerName="nova-api-log" containerID="cri-o://eae5df57a84c79de36f586452630170afcbfff0b1e0fdb27c896a858c117ff8b" gracePeriod=30 Jan 27 16:40:17 crc kubenswrapper[4772]: I0127 16:40:17.807933 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="601c20bf-c7e2-4c26-a2da-7ebce076775a" containerName="nova-api-api" containerID="cri-o://6905111d10a31c85326f0556bb66e5eba75c934131e8d3fe500da3f080ca6789" gracePeriod=30 Jan 27 16:40:17 crc kubenswrapper[4772]: I0127 16:40:17.820046 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 16:40:17 crc kubenswrapper[4772]: I0127 16:40:17.820372 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="02c9bfb0-5f95-4692-bb48-ffd47f6682a8" containerName="nova-scheduler-scheduler" containerID="cri-o://6f9e0ed02577ed79c4498daebc32e5d9adfca8bf0094d3a8ddb00a6642f3d336" gracePeriod=30 Jan 27 16:40:17 crc kubenswrapper[4772]: I0127 16:40:17.843632 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 16:40:17 crc kubenswrapper[4772]: I0127 16:40:17.843857 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0576a6dc-0ce0-4e70-8be2-179989ed0c03" containerName="nova-metadata-log" containerID="cri-o://b99ac9784926d43b5b9010bb3bf145d936b87ce053aad43267f30ea8ebbbf38a" gracePeriod=30 Jan 27 16:40:17 crc kubenswrapper[4772]: I0127 16:40:17.843887 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0576a6dc-0ce0-4e70-8be2-179989ed0c03" containerName="nova-metadata-metadata" containerID="cri-o://a0a5b8400ff9131c6a732b564464e4682a47bd72327030af67266b99caae0aa5" gracePeriod=30 Jan 27 16:40:18 crc kubenswrapper[4772]: I0127 16:40:18.628583 4772 generic.go:334] "Generic (PLEG): container finished" podID="0576a6dc-0ce0-4e70-8be2-179989ed0c03" containerID="b99ac9784926d43b5b9010bb3bf145d936b87ce053aad43267f30ea8ebbbf38a" exitCode=143 Jan 27 16:40:18 crc kubenswrapper[4772]: I0127 16:40:18.628670 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0576a6dc-0ce0-4e70-8be2-179989ed0c03","Type":"ContainerDied","Data":"b99ac9784926d43b5b9010bb3bf145d936b87ce053aad43267f30ea8ebbbf38a"} Jan 27 16:40:18 crc kubenswrapper[4772]: I0127 16:40:18.630241 4772 generic.go:334] "Generic (PLEG): container finished" podID="601c20bf-c7e2-4c26-a2da-7ebce076775a" containerID="eae5df57a84c79de36f586452630170afcbfff0b1e0fdb27c896a858c117ff8b" exitCode=143 Jan 27 16:40:18 crc kubenswrapper[4772]: I0127 16:40:18.630322 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"601c20bf-c7e2-4c26-a2da-7ebce076775a","Type":"ContainerDied","Data":"eae5df57a84c79de36f586452630170afcbfff0b1e0fdb27c896a858c117ff8b"} Jan 27 16:40:19 crc kubenswrapper[4772]: E0127 16:40:19.879501 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6f9e0ed02577ed79c4498daebc32e5d9adfca8bf0094d3a8ddb00a6642f3d336" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 27 16:40:19 crc kubenswrapper[4772]: E0127 16:40:19.880929 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6f9e0ed02577ed79c4498daebc32e5d9adfca8bf0094d3a8ddb00a6642f3d336" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 27 16:40:19 crc kubenswrapper[4772]: E0127 16:40:19.884287 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6f9e0ed02577ed79c4498daebc32e5d9adfca8bf0094d3a8ddb00a6642f3d336" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 27 16:40:19 crc kubenswrapper[4772]: E0127 16:40:19.884358 4772 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="02c9bfb0-5f95-4692-bb48-ffd47f6682a8" containerName="nova-scheduler-scheduler" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.443918 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.452206 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.580864 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9rxm\" (UniqueName: \"kubernetes.io/projected/601c20bf-c7e2-4c26-a2da-7ebce076775a-kube-api-access-c9rxm\") pod \"601c20bf-c7e2-4c26-a2da-7ebce076775a\" (UID: \"601c20bf-c7e2-4c26-a2da-7ebce076775a\") " Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.580940 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/601c20bf-c7e2-4c26-a2da-7ebce076775a-combined-ca-bundle\") pod \"601c20bf-c7e2-4c26-a2da-7ebce076775a\" (UID: \"601c20bf-c7e2-4c26-a2da-7ebce076775a\") " Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.580973 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0576a6dc-0ce0-4e70-8be2-179989ed0c03-config-data\") pod \"0576a6dc-0ce0-4e70-8be2-179989ed0c03\" (UID: \"0576a6dc-0ce0-4e70-8be2-179989ed0c03\") " Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.581015 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0576a6dc-0ce0-4e70-8be2-179989ed0c03-logs\") pod \"0576a6dc-0ce0-4e70-8be2-179989ed0c03\" (UID: \"0576a6dc-0ce0-4e70-8be2-179989ed0c03\") " Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.581036 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/601c20bf-c7e2-4c26-a2da-7ebce076775a-config-data\") pod \"601c20bf-c7e2-4c26-a2da-7ebce076775a\" (UID: \"601c20bf-c7e2-4c26-a2da-7ebce076775a\") " Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.581058 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0576a6dc-0ce0-4e70-8be2-179989ed0c03-combined-ca-bundle\") pod \"0576a6dc-0ce0-4e70-8be2-179989ed0c03\" (UID: \"0576a6dc-0ce0-4e70-8be2-179989ed0c03\") " Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.581113 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/601c20bf-c7e2-4c26-a2da-7ebce076775a-logs\") pod \"601c20bf-c7e2-4c26-a2da-7ebce076775a\" (UID: \"601c20bf-c7e2-4c26-a2da-7ebce076775a\") " Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.581228 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r644m\" (UniqueName: \"kubernetes.io/projected/0576a6dc-0ce0-4e70-8be2-179989ed0c03-kube-api-access-r644m\") pod \"0576a6dc-0ce0-4e70-8be2-179989ed0c03\" (UID: \"0576a6dc-0ce0-4e70-8be2-179989ed0c03\") " Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.582947 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/601c20bf-c7e2-4c26-a2da-7ebce076775a-logs" (OuterVolumeSpecName: "logs") pod "601c20bf-c7e2-4c26-a2da-7ebce076775a" (UID: "601c20bf-c7e2-4c26-a2da-7ebce076775a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.585376 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0576a6dc-0ce0-4e70-8be2-179989ed0c03-logs" (OuterVolumeSpecName: "logs") pod "0576a6dc-0ce0-4e70-8be2-179989ed0c03" (UID: "0576a6dc-0ce0-4e70-8be2-179989ed0c03"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.588388 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/601c20bf-c7e2-4c26-a2da-7ebce076775a-kube-api-access-c9rxm" (OuterVolumeSpecName: "kube-api-access-c9rxm") pod "601c20bf-c7e2-4c26-a2da-7ebce076775a" (UID: "601c20bf-c7e2-4c26-a2da-7ebce076775a"). InnerVolumeSpecName "kube-api-access-c9rxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.588438 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0576a6dc-0ce0-4e70-8be2-179989ed0c03-kube-api-access-r644m" (OuterVolumeSpecName: "kube-api-access-r644m") pod "0576a6dc-0ce0-4e70-8be2-179989ed0c03" (UID: "0576a6dc-0ce0-4e70-8be2-179989ed0c03"). InnerVolumeSpecName "kube-api-access-r644m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.606434 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0576a6dc-0ce0-4e70-8be2-179989ed0c03-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0576a6dc-0ce0-4e70-8be2-179989ed0c03" (UID: "0576a6dc-0ce0-4e70-8be2-179989ed0c03"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.606662 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/601c20bf-c7e2-4c26-a2da-7ebce076775a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "601c20bf-c7e2-4c26-a2da-7ebce076775a" (UID: "601c20bf-c7e2-4c26-a2da-7ebce076775a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.607517 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0576a6dc-0ce0-4e70-8be2-179989ed0c03-config-data" (OuterVolumeSpecName: "config-data") pod "0576a6dc-0ce0-4e70-8be2-179989ed0c03" (UID: "0576a6dc-0ce0-4e70-8be2-179989ed0c03"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.607880 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/601c20bf-c7e2-4c26-a2da-7ebce076775a-config-data" (OuterVolumeSpecName: "config-data") pod "601c20bf-c7e2-4c26-a2da-7ebce076775a" (UID: "601c20bf-c7e2-4c26-a2da-7ebce076775a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.655568 4772 generic.go:334] "Generic (PLEG): container finished" podID="02c9bfb0-5f95-4692-bb48-ffd47f6682a8" containerID="6f9e0ed02577ed79c4498daebc32e5d9adfca8bf0094d3a8ddb00a6642f3d336" exitCode=0 Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.655636 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"02c9bfb0-5f95-4692-bb48-ffd47f6682a8","Type":"ContainerDied","Data":"6f9e0ed02577ed79c4498daebc32e5d9adfca8bf0094d3a8ddb00a6642f3d336"} Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.657441 4772 generic.go:334] "Generic (PLEG): container finished" podID="601c20bf-c7e2-4c26-a2da-7ebce076775a" containerID="6905111d10a31c85326f0556bb66e5eba75c934131e8d3fe500da3f080ca6789" exitCode=0 Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.657501 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"601c20bf-c7e2-4c26-a2da-7ebce076775a","Type":"ContainerDied","Data":"6905111d10a31c85326f0556bb66e5eba75c934131e8d3fe500da3f080ca6789"} Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.657522 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"601c20bf-c7e2-4c26-a2da-7ebce076775a","Type":"ContainerDied","Data":"c71aafaf39ad009fb7eb0ade8a547f4797867d098bc1c0cd4f5e22f563ee2d30"} Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.657520 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.657543 4772 scope.go:117] "RemoveContainer" containerID="6905111d10a31c85326f0556bb66e5eba75c934131e8d3fe500da3f080ca6789" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.659681 4772 generic.go:334] "Generic (PLEG): container finished" podID="0576a6dc-0ce0-4e70-8be2-179989ed0c03" containerID="a0a5b8400ff9131c6a732b564464e4682a47bd72327030af67266b99caae0aa5" exitCode=0 Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.659731 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0576a6dc-0ce0-4e70-8be2-179989ed0c03","Type":"ContainerDied","Data":"a0a5b8400ff9131c6a732b564464e4682a47bd72327030af67266b99caae0aa5"} Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.660387 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0576a6dc-0ce0-4e70-8be2-179989ed0c03","Type":"ContainerDied","Data":"272b9d5f7574afede07d814583b3f863aa00e225817941ab1c1713433efa66a6"} Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.659758 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.684213 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/601c20bf-c7e2-4c26-a2da-7ebce076775a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.684271 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0576a6dc-0ce0-4e70-8be2-179989ed0c03-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.684281 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0576a6dc-0ce0-4e70-8be2-179989ed0c03-logs\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.684289 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/601c20bf-c7e2-4c26-a2da-7ebce076775a-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.684298 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0576a6dc-0ce0-4e70-8be2-179989ed0c03-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.684305 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/601c20bf-c7e2-4c26-a2da-7ebce076775a-logs\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.684333 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r644m\" (UniqueName: \"kubernetes.io/projected/0576a6dc-0ce0-4e70-8be2-179989ed0c03-kube-api-access-r644m\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.684346 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9rxm\" (UniqueName: \"kubernetes.io/projected/601c20bf-c7e2-4c26-a2da-7ebce076775a-kube-api-access-c9rxm\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.712283 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.716624 4772 scope.go:117] "RemoveContainer" containerID="eae5df57a84c79de36f586452630170afcbfff0b1e0fdb27c896a858c117ff8b" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.727865 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.741354 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.749361 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.758245 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 27 16:40:21 crc kubenswrapper[4772]: E0127 16:40:21.758779 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="601c20bf-c7e2-4c26-a2da-7ebce076775a" containerName="nova-api-log" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.758795 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="601c20bf-c7e2-4c26-a2da-7ebce076775a" containerName="nova-api-log" Jan 27 16:40:21 crc kubenswrapper[4772]: E0127 16:40:21.758821 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaefb4fd-175a-4431-bd2f-7fc3205684b9" containerName="nova-manage" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.758829 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaefb4fd-175a-4431-bd2f-7fc3205684b9" containerName="nova-manage" Jan 27 16:40:21 crc kubenswrapper[4772]: E0127 16:40:21.758845 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0576a6dc-0ce0-4e70-8be2-179989ed0c03" containerName="nova-metadata-metadata" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.758856 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0576a6dc-0ce0-4e70-8be2-179989ed0c03" containerName="nova-metadata-metadata" Jan 27 16:40:21 crc kubenswrapper[4772]: E0127 16:40:21.758875 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0576a6dc-0ce0-4e70-8be2-179989ed0c03" containerName="nova-metadata-log" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.758882 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0576a6dc-0ce0-4e70-8be2-179989ed0c03" containerName="nova-metadata-log" Jan 27 16:40:21 crc kubenswrapper[4772]: E0127 16:40:21.758904 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="601c20bf-c7e2-4c26-a2da-7ebce076775a" containerName="nova-api-api" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.758911 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="601c20bf-c7e2-4c26-a2da-7ebce076775a" containerName="nova-api-api" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.759160 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="601c20bf-c7e2-4c26-a2da-7ebce076775a" containerName="nova-api-api" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.759194 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaefb4fd-175a-4431-bd2f-7fc3205684b9" containerName="nova-manage" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.759208 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="601c20bf-c7e2-4c26-a2da-7ebce076775a" containerName="nova-api-log" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.759222 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="0576a6dc-0ce0-4e70-8be2-179989ed0c03" containerName="nova-metadata-log" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.759236 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="0576a6dc-0ce0-4e70-8be2-179989ed0c03" containerName="nova-metadata-metadata" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.760379 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.765960 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.771184 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.773635 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.777830 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.778017 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.791103 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.794715 4772 scope.go:117] "RemoveContainer" containerID="6905111d10a31c85326f0556bb66e5eba75c934131e8d3fe500da3f080ca6789" Jan 27 16:40:21 crc kubenswrapper[4772]: E0127 16:40:21.802262 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6905111d10a31c85326f0556bb66e5eba75c934131e8d3fe500da3f080ca6789\": container with ID starting with 6905111d10a31c85326f0556bb66e5eba75c934131e8d3fe500da3f080ca6789 not found: ID does not exist" containerID="6905111d10a31c85326f0556bb66e5eba75c934131e8d3fe500da3f080ca6789" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.802307 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6905111d10a31c85326f0556bb66e5eba75c934131e8d3fe500da3f080ca6789"} err="failed to get container status \"6905111d10a31c85326f0556bb66e5eba75c934131e8d3fe500da3f080ca6789\": rpc error: code = NotFound desc = could not find container \"6905111d10a31c85326f0556bb66e5eba75c934131e8d3fe500da3f080ca6789\": container with ID starting with 6905111d10a31c85326f0556bb66e5eba75c934131e8d3fe500da3f080ca6789 not found: ID does not exist" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.802330 4772 scope.go:117] "RemoveContainer" containerID="eae5df57a84c79de36f586452630170afcbfff0b1e0fdb27c896a858c117ff8b" Jan 27 16:40:21 crc kubenswrapper[4772]: E0127 16:40:21.805512 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eae5df57a84c79de36f586452630170afcbfff0b1e0fdb27c896a858c117ff8b\": container with ID starting with eae5df57a84c79de36f586452630170afcbfff0b1e0fdb27c896a858c117ff8b not found: ID does not exist" containerID="eae5df57a84c79de36f586452630170afcbfff0b1e0fdb27c896a858c117ff8b" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.805556 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eae5df57a84c79de36f586452630170afcbfff0b1e0fdb27c896a858c117ff8b"} err="failed to get container status \"eae5df57a84c79de36f586452630170afcbfff0b1e0fdb27c896a858c117ff8b\": rpc error: code = NotFound desc = could not find container \"eae5df57a84c79de36f586452630170afcbfff0b1e0fdb27c896a858c117ff8b\": container with ID starting with eae5df57a84c79de36f586452630170afcbfff0b1e0fdb27c896a858c117ff8b not found: ID does not exist" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.805579 4772 scope.go:117] "RemoveContainer" containerID="a0a5b8400ff9131c6a732b564464e4682a47bd72327030af67266b99caae0aa5" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.832098 4772 scope.go:117] "RemoveContainer" containerID="b99ac9784926d43b5b9010bb3bf145d936b87ce053aad43267f30ea8ebbbf38a" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.854482 4772 scope.go:117] "RemoveContainer" containerID="a0a5b8400ff9131c6a732b564464e4682a47bd72327030af67266b99caae0aa5" Jan 27 16:40:21 crc kubenswrapper[4772]: E0127 16:40:21.855108 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0a5b8400ff9131c6a732b564464e4682a47bd72327030af67266b99caae0aa5\": container with ID starting with a0a5b8400ff9131c6a732b564464e4682a47bd72327030af67266b99caae0aa5 not found: ID does not exist" containerID="a0a5b8400ff9131c6a732b564464e4682a47bd72327030af67266b99caae0aa5" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.855154 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0a5b8400ff9131c6a732b564464e4682a47bd72327030af67266b99caae0aa5"} err="failed to get container status \"a0a5b8400ff9131c6a732b564464e4682a47bd72327030af67266b99caae0aa5\": rpc error: code = NotFound desc = could not find container \"a0a5b8400ff9131c6a732b564464e4682a47bd72327030af67266b99caae0aa5\": container with ID starting with a0a5b8400ff9131c6a732b564464e4682a47bd72327030af67266b99caae0aa5 not found: ID does not exist" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.855200 4772 scope.go:117] "RemoveContainer" containerID="b99ac9784926d43b5b9010bb3bf145d936b87ce053aad43267f30ea8ebbbf38a" Jan 27 16:40:21 crc kubenswrapper[4772]: E0127 16:40:21.855480 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b99ac9784926d43b5b9010bb3bf145d936b87ce053aad43267f30ea8ebbbf38a\": container with ID starting with b99ac9784926d43b5b9010bb3bf145d936b87ce053aad43267f30ea8ebbbf38a not found: ID does not exist" containerID="b99ac9784926d43b5b9010bb3bf145d936b87ce053aad43267f30ea8ebbbf38a" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.855519 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b99ac9784926d43b5b9010bb3bf145d936b87ce053aad43267f30ea8ebbbf38a"} err="failed to get container status \"b99ac9784926d43b5b9010bb3bf145d936b87ce053aad43267f30ea8ebbbf38a\": rpc error: code = NotFound desc = could not find container \"b99ac9784926d43b5b9010bb3bf145d936b87ce053aad43267f30ea8ebbbf38a\": container with ID starting with b99ac9784926d43b5b9010bb3bf145d936b87ce053aad43267f30ea8ebbbf38a not found: ID does not exist" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.886972 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a518103b-f26e-4f91-9ca9-93f1f8d5e113-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a518103b-f26e-4f91-9ca9-93f1f8d5e113\") " pod="openstack/nova-api-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.887040 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vg99\" (UniqueName: \"kubernetes.io/projected/a518103b-f26e-4f91-9ca9-93f1f8d5e113-kube-api-access-8vg99\") pod \"nova-api-0\" (UID: \"a518103b-f26e-4f91-9ca9-93f1f8d5e113\") " pod="openstack/nova-api-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.887071 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c652921-712d-46ee-9683-fd6312e33d1e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8c652921-712d-46ee-9683-fd6312e33d1e\") " pod="openstack/nova-metadata-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.887127 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a518103b-f26e-4f91-9ca9-93f1f8d5e113-logs\") pod \"nova-api-0\" (UID: \"a518103b-f26e-4f91-9ca9-93f1f8d5e113\") " pod="openstack/nova-api-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.887154 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c652921-712d-46ee-9683-fd6312e33d1e-config-data\") pod \"nova-metadata-0\" (UID: \"8c652921-712d-46ee-9683-fd6312e33d1e\") " pod="openstack/nova-metadata-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.887189 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a518103b-f26e-4f91-9ca9-93f1f8d5e113-config-data\") pod \"nova-api-0\" (UID: \"a518103b-f26e-4f91-9ca9-93f1f8d5e113\") " pod="openstack/nova-api-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.887215 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cfjb\" (UniqueName: \"kubernetes.io/projected/8c652921-712d-46ee-9683-fd6312e33d1e-kube-api-access-7cfjb\") pod \"nova-metadata-0\" (UID: \"8c652921-712d-46ee-9683-fd6312e33d1e\") " pod="openstack/nova-metadata-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.887257 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c652921-712d-46ee-9683-fd6312e33d1e-logs\") pod \"nova-metadata-0\" (UID: \"8c652921-712d-46ee-9683-fd6312e33d1e\") " pod="openstack/nova-metadata-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.933657 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.988266 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2khtr\" (UniqueName: \"kubernetes.io/projected/02c9bfb0-5f95-4692-bb48-ffd47f6682a8-kube-api-access-2khtr\") pod \"02c9bfb0-5f95-4692-bb48-ffd47f6682a8\" (UID: \"02c9bfb0-5f95-4692-bb48-ffd47f6682a8\") " Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.988397 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02c9bfb0-5f95-4692-bb48-ffd47f6682a8-combined-ca-bundle\") pod \"02c9bfb0-5f95-4692-bb48-ffd47f6682a8\" (UID: \"02c9bfb0-5f95-4692-bb48-ffd47f6682a8\") " Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.988450 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02c9bfb0-5f95-4692-bb48-ffd47f6682a8-config-data\") pod \"02c9bfb0-5f95-4692-bb48-ffd47f6682a8\" (UID: \"02c9bfb0-5f95-4692-bb48-ffd47f6682a8\") " Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.988701 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c652921-712d-46ee-9683-fd6312e33d1e-config-data\") pod \"nova-metadata-0\" (UID: \"8c652921-712d-46ee-9683-fd6312e33d1e\") " pod="openstack/nova-metadata-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.988720 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a518103b-f26e-4f91-9ca9-93f1f8d5e113-config-data\") pod \"nova-api-0\" (UID: \"a518103b-f26e-4f91-9ca9-93f1f8d5e113\") " pod="openstack/nova-api-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.988738 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cfjb\" (UniqueName: \"kubernetes.io/projected/8c652921-712d-46ee-9683-fd6312e33d1e-kube-api-access-7cfjb\") pod \"nova-metadata-0\" (UID: \"8c652921-712d-46ee-9683-fd6312e33d1e\") " pod="openstack/nova-metadata-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.988777 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c652921-712d-46ee-9683-fd6312e33d1e-logs\") pod \"nova-metadata-0\" (UID: \"8c652921-712d-46ee-9683-fd6312e33d1e\") " pod="openstack/nova-metadata-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.988821 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a518103b-f26e-4f91-9ca9-93f1f8d5e113-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a518103b-f26e-4f91-9ca9-93f1f8d5e113\") " pod="openstack/nova-api-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.988846 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vg99\" (UniqueName: \"kubernetes.io/projected/a518103b-f26e-4f91-9ca9-93f1f8d5e113-kube-api-access-8vg99\") pod \"nova-api-0\" (UID: \"a518103b-f26e-4f91-9ca9-93f1f8d5e113\") " pod="openstack/nova-api-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.988866 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c652921-712d-46ee-9683-fd6312e33d1e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8c652921-712d-46ee-9683-fd6312e33d1e\") " pod="openstack/nova-metadata-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.988935 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a518103b-f26e-4f91-9ca9-93f1f8d5e113-logs\") pod \"nova-api-0\" (UID: \"a518103b-f26e-4f91-9ca9-93f1f8d5e113\") " pod="openstack/nova-api-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.989406 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a518103b-f26e-4f91-9ca9-93f1f8d5e113-logs\") pod \"nova-api-0\" (UID: \"a518103b-f26e-4f91-9ca9-93f1f8d5e113\") " pod="openstack/nova-api-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.989850 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c652921-712d-46ee-9683-fd6312e33d1e-logs\") pod \"nova-metadata-0\" (UID: \"8c652921-712d-46ee-9683-fd6312e33d1e\") " pod="openstack/nova-metadata-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.994972 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a518103b-f26e-4f91-9ca9-93f1f8d5e113-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a518103b-f26e-4f91-9ca9-93f1f8d5e113\") " pod="openstack/nova-api-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.994972 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c652921-712d-46ee-9683-fd6312e33d1e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8c652921-712d-46ee-9683-fd6312e33d1e\") " pod="openstack/nova-metadata-0" Jan 27 16:40:21 crc kubenswrapper[4772]: I0127 16:40:21.995473 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a518103b-f26e-4f91-9ca9-93f1f8d5e113-config-data\") pod \"nova-api-0\" (UID: \"a518103b-f26e-4f91-9ca9-93f1f8d5e113\") " pod="openstack/nova-api-0" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.002366 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02c9bfb0-5f95-4692-bb48-ffd47f6682a8-kube-api-access-2khtr" (OuterVolumeSpecName: "kube-api-access-2khtr") pod "02c9bfb0-5f95-4692-bb48-ffd47f6682a8" (UID: "02c9bfb0-5f95-4692-bb48-ffd47f6682a8"). InnerVolumeSpecName "kube-api-access-2khtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.003632 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c652921-712d-46ee-9683-fd6312e33d1e-config-data\") pod \"nova-metadata-0\" (UID: \"8c652921-712d-46ee-9683-fd6312e33d1e\") " pod="openstack/nova-metadata-0" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.014341 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cfjb\" (UniqueName: \"kubernetes.io/projected/8c652921-712d-46ee-9683-fd6312e33d1e-kube-api-access-7cfjb\") pod \"nova-metadata-0\" (UID: \"8c652921-712d-46ee-9683-fd6312e33d1e\") " pod="openstack/nova-metadata-0" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.014774 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vg99\" (UniqueName: \"kubernetes.io/projected/a518103b-f26e-4f91-9ca9-93f1f8d5e113-kube-api-access-8vg99\") pod \"nova-api-0\" (UID: \"a518103b-f26e-4f91-9ca9-93f1f8d5e113\") " pod="openstack/nova-api-0" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.022569 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02c9bfb0-5f95-4692-bb48-ffd47f6682a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02c9bfb0-5f95-4692-bb48-ffd47f6682a8" (UID: "02c9bfb0-5f95-4692-bb48-ffd47f6682a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.027404 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02c9bfb0-5f95-4692-bb48-ffd47f6682a8-config-data" (OuterVolumeSpecName: "config-data") pod "02c9bfb0-5f95-4692-bb48-ffd47f6682a8" (UID: "02c9bfb0-5f95-4692-bb48-ffd47f6682a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.090306 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2khtr\" (UniqueName: \"kubernetes.io/projected/02c9bfb0-5f95-4692-bb48-ffd47f6682a8-kube-api-access-2khtr\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.090336 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02c9bfb0-5f95-4692-bb48-ffd47f6682a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.090344 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02c9bfb0-5f95-4692-bb48-ffd47f6682a8-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.091967 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.099990 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 16:40:22 crc kubenswrapper[4772]: W0127 16:40:22.571103 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c652921_712d_46ee_9683_fd6312e33d1e.slice/crio-8059f8ca79e1f853876a311e130550fa4d5e0bfbe87838df84c9d7f7b757e1d9 WatchSource:0}: Error finding container 8059f8ca79e1f853876a311e130550fa4d5e0bfbe87838df84c9d7f7b757e1d9: Status 404 returned error can't find the container with id 8059f8ca79e1f853876a311e130550fa4d5e0bfbe87838df84c9d7f7b757e1d9 Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.572620 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.613646 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.678794 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0576a6dc-0ce0-4e70-8be2-179989ed0c03" path="/var/lib/kubelet/pods/0576a6dc-0ce0-4e70-8be2-179989ed0c03/volumes" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.679538 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="601c20bf-c7e2-4c26-a2da-7ebce076775a" path="/var/lib/kubelet/pods/601c20bf-c7e2-4c26-a2da-7ebce076775a/volumes" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.680203 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.680555 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c652921-712d-46ee-9683-fd6312e33d1e","Type":"ContainerStarted","Data":"8059f8ca79e1f853876a311e130550fa4d5e0bfbe87838df84c9d7f7b757e1d9"} Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.680589 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a518103b-f26e-4f91-9ca9-93f1f8d5e113","Type":"ContainerStarted","Data":"ac2ae0550771f81842cdb596165399767b9e27c96df8f6a9cc1d4cc43f6c15e1"} Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.680605 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"02c9bfb0-5f95-4692-bb48-ffd47f6682a8","Type":"ContainerDied","Data":"3c6a3e07c72ed8c1bb7f2e42a623b437deaa88e1d9979eefafaa45fb0ff45499"} Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.680625 4772 scope.go:117] "RemoveContainer" containerID="6f9e0ed02577ed79c4498daebc32e5d9adfca8bf0094d3a8ddb00a6642f3d336" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.738449 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.746129 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.755042 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 16:40:22 crc kubenswrapper[4772]: E0127 16:40:22.755667 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02c9bfb0-5f95-4692-bb48-ffd47f6682a8" containerName="nova-scheduler-scheduler" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.755685 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="02c9bfb0-5f95-4692-bb48-ffd47f6682a8" containerName="nova-scheduler-scheduler" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.755848 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="02c9bfb0-5f95-4692-bb48-ffd47f6682a8" containerName="nova-scheduler-scheduler" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.756477 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.762726 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.770382 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.802552 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce725b6a-06bb-4339-819f-fee8819078f0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ce725b6a-06bb-4339-819f-fee8819078f0\") " pod="openstack/nova-scheduler-0" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.802709 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ckmf\" (UniqueName: \"kubernetes.io/projected/ce725b6a-06bb-4339-819f-fee8819078f0-kube-api-access-7ckmf\") pod \"nova-scheduler-0\" (UID: \"ce725b6a-06bb-4339-819f-fee8819078f0\") " pod="openstack/nova-scheduler-0" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.802759 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce725b6a-06bb-4339-819f-fee8819078f0-config-data\") pod \"nova-scheduler-0\" (UID: \"ce725b6a-06bb-4339-819f-fee8819078f0\") " pod="openstack/nova-scheduler-0" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.904586 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ckmf\" (UniqueName: \"kubernetes.io/projected/ce725b6a-06bb-4339-819f-fee8819078f0-kube-api-access-7ckmf\") pod \"nova-scheduler-0\" (UID: \"ce725b6a-06bb-4339-819f-fee8819078f0\") " pod="openstack/nova-scheduler-0" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.904670 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce725b6a-06bb-4339-819f-fee8819078f0-config-data\") pod \"nova-scheduler-0\" (UID: \"ce725b6a-06bb-4339-819f-fee8819078f0\") " pod="openstack/nova-scheduler-0" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.904704 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce725b6a-06bb-4339-819f-fee8819078f0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ce725b6a-06bb-4339-819f-fee8819078f0\") " pod="openstack/nova-scheduler-0" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.910125 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce725b6a-06bb-4339-819f-fee8819078f0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ce725b6a-06bb-4339-819f-fee8819078f0\") " pod="openstack/nova-scheduler-0" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.910319 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce725b6a-06bb-4339-819f-fee8819078f0-config-data\") pod \"nova-scheduler-0\" (UID: \"ce725b6a-06bb-4339-819f-fee8819078f0\") " pod="openstack/nova-scheduler-0" Jan 27 16:40:22 crc kubenswrapper[4772]: I0127 16:40:22.921974 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ckmf\" (UniqueName: \"kubernetes.io/projected/ce725b6a-06bb-4339-819f-fee8819078f0-kube-api-access-7ckmf\") pod \"nova-scheduler-0\" (UID: \"ce725b6a-06bb-4339-819f-fee8819078f0\") " pod="openstack/nova-scheduler-0" Jan 27 16:40:23 crc kubenswrapper[4772]: I0127 16:40:23.077945 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 16:40:23 crc kubenswrapper[4772]: I0127 16:40:23.524815 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 16:40:23 crc kubenswrapper[4772]: I0127 16:40:23.690446 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a518103b-f26e-4f91-9ca9-93f1f8d5e113","Type":"ContainerStarted","Data":"67369ccebf4181eb0be2f61ead1ad9a24b9c3e7147fed849be8e7d8b2afab075"} Jan 27 16:40:23 crc kubenswrapper[4772]: I0127 16:40:23.690490 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a518103b-f26e-4f91-9ca9-93f1f8d5e113","Type":"ContainerStarted","Data":"8f94df1da4ef644e3b2171358f7c0dcee05549e2b5ef22dd8252e4943b46f5c0"} Jan 27 16:40:23 crc kubenswrapper[4772]: I0127 16:40:23.693881 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ce725b6a-06bb-4339-819f-fee8819078f0","Type":"ContainerStarted","Data":"33aa342145a07c52aa025581718207e0ed3316f1c48915a16075320a2dbeed5c"} Jan 27 16:40:23 crc kubenswrapper[4772]: I0127 16:40:23.693918 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ce725b6a-06bb-4339-819f-fee8819078f0","Type":"ContainerStarted","Data":"c5ffffe294d87ab27553633999a0d84408edaf82429449f00a94658621086fd0"} Jan 27 16:40:23 crc kubenswrapper[4772]: I0127 16:40:23.696916 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c652921-712d-46ee-9683-fd6312e33d1e","Type":"ContainerStarted","Data":"b7973b4ea385a2bd5b179204a29b45f97b045488c54fbb72800c5104d45dc193"} Jan 27 16:40:23 crc kubenswrapper[4772]: I0127 16:40:23.696955 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c652921-712d-46ee-9683-fd6312e33d1e","Type":"ContainerStarted","Data":"193d54411bb1309c9287f87812da125aae0f9213050e8e77849e515da35b0fc4"} Jan 27 16:40:23 crc kubenswrapper[4772]: I0127 16:40:23.710125 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.7101066940000003 podStartE2EDuration="2.710106694s" podCreationTimestamp="2026-01-27 16:40:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:40:23.707543661 +0000 UTC m=+5609.688152779" watchObservedRunningTime="2026-01-27 16:40:23.710106694 +0000 UTC m=+5609.690715792" Jan 27 16:40:23 crc kubenswrapper[4772]: I0127 16:40:23.728138 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.728117408 podStartE2EDuration="2.728117408s" podCreationTimestamp="2026-01-27 16:40:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:40:23.724613108 +0000 UTC m=+5609.705222236" watchObservedRunningTime="2026-01-27 16:40:23.728117408 +0000 UTC m=+5609.708726506" Jan 27 16:40:23 crc kubenswrapper[4772]: I0127 16:40:23.749583 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.749558399 podStartE2EDuration="1.749558399s" podCreationTimestamp="2026-01-27 16:40:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:40:23.740930343 +0000 UTC m=+5609.721539451" watchObservedRunningTime="2026-01-27 16:40:23.749558399 +0000 UTC m=+5609.730167507" Jan 27 16:40:24 crc kubenswrapper[4772]: I0127 16:40:24.705131 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02c9bfb0-5f95-4692-bb48-ffd47f6682a8" path="/var/lib/kubelet/pods/02c9bfb0-5f95-4692-bb48-ffd47f6682a8/volumes" Jan 27 16:40:27 crc kubenswrapper[4772]: I0127 16:40:27.092694 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 27 16:40:27 crc kubenswrapper[4772]: I0127 16:40:27.093008 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 27 16:40:28 crc kubenswrapper[4772]: I0127 16:40:28.078546 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 27 16:40:32 crc kubenswrapper[4772]: I0127 16:40:32.093416 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 27 16:40:32 crc kubenswrapper[4772]: I0127 16:40:32.093955 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 27 16:40:32 crc kubenswrapper[4772]: I0127 16:40:32.100941 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 27 16:40:32 crc kubenswrapper[4772]: I0127 16:40:32.100985 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 27 16:40:33 crc kubenswrapper[4772]: I0127 16:40:33.078825 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Jan 27 16:40:33 crc kubenswrapper[4772]: I0127 16:40:33.109544 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Jan 27 16:40:33 crc kubenswrapper[4772]: I0127 16:40:33.217451 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a518103b-f26e-4f91-9ca9-93f1f8d5e113" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.65:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 27 16:40:33 crc kubenswrapper[4772]: I0127 16:40:33.258432 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a518103b-f26e-4f91-9ca9-93f1f8d5e113" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.65:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 27 16:40:33 crc kubenswrapper[4772]: I0127 16:40:33.258719 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8c652921-712d-46ee-9683-fd6312e33d1e" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.64:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 27 16:40:33 crc kubenswrapper[4772]: I0127 16:40:33.258884 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8c652921-712d-46ee-9683-fd6312e33d1e" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.64:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 27 16:40:33 crc kubenswrapper[4772]: I0127 16:40:33.831868 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Jan 27 16:40:42 crc kubenswrapper[4772]: I0127 16:40:42.058666 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:40:42 crc kubenswrapper[4772]: I0127 16:40:42.059397 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:40:42 crc kubenswrapper[4772]: I0127 16:40:42.094820 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 27 16:40:42 crc kubenswrapper[4772]: I0127 16:40:42.095275 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 27 16:40:42 crc kubenswrapper[4772]: I0127 16:40:42.096761 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 27 16:40:42 crc kubenswrapper[4772]: I0127 16:40:42.104205 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 27 16:40:42 crc kubenswrapper[4772]: I0127 16:40:42.104563 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 27 16:40:42 crc kubenswrapper[4772]: I0127 16:40:42.107838 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 27 16:40:42 crc kubenswrapper[4772]: I0127 16:40:42.111743 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 27 16:40:42 crc kubenswrapper[4772]: I0127 16:40:42.898497 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 27 16:40:42 crc kubenswrapper[4772]: I0127 16:40:42.921259 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 27 16:40:42 crc kubenswrapper[4772]: I0127 16:40:42.961785 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 27 16:40:43 crc kubenswrapper[4772]: I0127 16:40:43.250317 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c56cfbf-9f62r"] Jan 27 16:40:43 crc kubenswrapper[4772]: I0127 16:40:43.251989 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" Jan 27 16:40:43 crc kubenswrapper[4772]: I0127 16:40:43.262782 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c56cfbf-9f62r"] Jan 27 16:40:43 crc kubenswrapper[4772]: I0127 16:40:43.308831 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/487e140f-d3fb-4ece-a41c-7a1c55a37534-ovsdbserver-sb\") pod \"dnsmasq-dns-89c56cfbf-9f62r\" (UID: \"487e140f-d3fb-4ece-a41c-7a1c55a37534\") " pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" Jan 27 16:40:43 crc kubenswrapper[4772]: I0127 16:40:43.308907 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/487e140f-d3fb-4ece-a41c-7a1c55a37534-dns-svc\") pod \"dnsmasq-dns-89c56cfbf-9f62r\" (UID: \"487e140f-d3fb-4ece-a41c-7a1c55a37534\") " pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" Jan 27 16:40:43 crc kubenswrapper[4772]: I0127 16:40:43.308935 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/487e140f-d3fb-4ece-a41c-7a1c55a37534-config\") pod \"dnsmasq-dns-89c56cfbf-9f62r\" (UID: \"487e140f-d3fb-4ece-a41c-7a1c55a37534\") " pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" Jan 27 16:40:43 crc kubenswrapper[4772]: I0127 16:40:43.309004 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dkhk\" (UniqueName: \"kubernetes.io/projected/487e140f-d3fb-4ece-a41c-7a1c55a37534-kube-api-access-2dkhk\") pod \"dnsmasq-dns-89c56cfbf-9f62r\" (UID: \"487e140f-d3fb-4ece-a41c-7a1c55a37534\") " pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" Jan 27 16:40:43 crc kubenswrapper[4772]: I0127 16:40:43.309057 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/487e140f-d3fb-4ece-a41c-7a1c55a37534-ovsdbserver-nb\") pod \"dnsmasq-dns-89c56cfbf-9f62r\" (UID: \"487e140f-d3fb-4ece-a41c-7a1c55a37534\") " pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" Jan 27 16:40:43 crc kubenswrapper[4772]: I0127 16:40:43.410933 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/487e140f-d3fb-4ece-a41c-7a1c55a37534-dns-svc\") pod \"dnsmasq-dns-89c56cfbf-9f62r\" (UID: \"487e140f-d3fb-4ece-a41c-7a1c55a37534\") " pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" Jan 27 16:40:43 crc kubenswrapper[4772]: I0127 16:40:43.410972 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/487e140f-d3fb-4ece-a41c-7a1c55a37534-config\") pod \"dnsmasq-dns-89c56cfbf-9f62r\" (UID: \"487e140f-d3fb-4ece-a41c-7a1c55a37534\") " pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" Jan 27 16:40:43 crc kubenswrapper[4772]: I0127 16:40:43.411038 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dkhk\" (UniqueName: \"kubernetes.io/projected/487e140f-d3fb-4ece-a41c-7a1c55a37534-kube-api-access-2dkhk\") pod \"dnsmasq-dns-89c56cfbf-9f62r\" (UID: \"487e140f-d3fb-4ece-a41c-7a1c55a37534\") " pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" Jan 27 16:40:43 crc kubenswrapper[4772]: I0127 16:40:43.411073 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/487e140f-d3fb-4ece-a41c-7a1c55a37534-ovsdbserver-nb\") pod \"dnsmasq-dns-89c56cfbf-9f62r\" (UID: \"487e140f-d3fb-4ece-a41c-7a1c55a37534\") " pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" Jan 27 16:40:43 crc kubenswrapper[4772]: I0127 16:40:43.411166 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/487e140f-d3fb-4ece-a41c-7a1c55a37534-ovsdbserver-sb\") pod \"dnsmasq-dns-89c56cfbf-9f62r\" (UID: \"487e140f-d3fb-4ece-a41c-7a1c55a37534\") " pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" Jan 27 16:40:43 crc kubenswrapper[4772]: I0127 16:40:43.411930 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/487e140f-d3fb-4ece-a41c-7a1c55a37534-config\") pod \"dnsmasq-dns-89c56cfbf-9f62r\" (UID: \"487e140f-d3fb-4ece-a41c-7a1c55a37534\") " pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" Jan 27 16:40:43 crc kubenswrapper[4772]: I0127 16:40:43.412076 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/487e140f-d3fb-4ece-a41c-7a1c55a37534-dns-svc\") pod \"dnsmasq-dns-89c56cfbf-9f62r\" (UID: \"487e140f-d3fb-4ece-a41c-7a1c55a37534\") " pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" Jan 27 16:40:43 crc kubenswrapper[4772]: I0127 16:40:43.412237 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/487e140f-d3fb-4ece-a41c-7a1c55a37534-ovsdbserver-sb\") pod \"dnsmasq-dns-89c56cfbf-9f62r\" (UID: \"487e140f-d3fb-4ece-a41c-7a1c55a37534\") " pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" Jan 27 16:40:43 crc kubenswrapper[4772]: I0127 16:40:43.412777 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/487e140f-d3fb-4ece-a41c-7a1c55a37534-ovsdbserver-nb\") pod \"dnsmasq-dns-89c56cfbf-9f62r\" (UID: \"487e140f-d3fb-4ece-a41c-7a1c55a37534\") " pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" Jan 27 16:40:43 crc kubenswrapper[4772]: I0127 16:40:43.435477 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dkhk\" (UniqueName: \"kubernetes.io/projected/487e140f-d3fb-4ece-a41c-7a1c55a37534-kube-api-access-2dkhk\") pod \"dnsmasq-dns-89c56cfbf-9f62r\" (UID: \"487e140f-d3fb-4ece-a41c-7a1c55a37534\") " pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" Jan 27 16:40:43 crc kubenswrapper[4772]: I0127 16:40:43.623316 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" Jan 27 16:40:44 crc kubenswrapper[4772]: I0127 16:40:44.095324 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c56cfbf-9f62r"] Jan 27 16:40:44 crc kubenswrapper[4772]: I0127 16:40:44.927157 4772 generic.go:334] "Generic (PLEG): container finished" podID="487e140f-d3fb-4ece-a41c-7a1c55a37534" containerID="3b284b2b9feaaf7731a5ba79953b14bcd6f7aec90b9b76040821cf8f6a1cb2f9" exitCode=0 Jan 27 16:40:44 crc kubenswrapper[4772]: I0127 16:40:44.927288 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" event={"ID":"487e140f-d3fb-4ece-a41c-7a1c55a37534","Type":"ContainerDied","Data":"3b284b2b9feaaf7731a5ba79953b14bcd6f7aec90b9b76040821cf8f6a1cb2f9"} Jan 27 16:40:44 crc kubenswrapper[4772]: I0127 16:40:44.928318 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" event={"ID":"487e140f-d3fb-4ece-a41c-7a1c55a37534","Type":"ContainerStarted","Data":"e6f437da52e8342f357b6a59971f74fc9403dcd77f6fa7a76fcf902ca6dd800f"} Jan 27 16:40:45 crc kubenswrapper[4772]: I0127 16:40:45.942721 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" event={"ID":"487e140f-d3fb-4ece-a41c-7a1c55a37534","Type":"ContainerStarted","Data":"b740d63c9d1013074313725144a776e8a3b5f1fd4cbe0771f67139ace29cab8f"} Jan 27 16:40:45 crc kubenswrapper[4772]: I0127 16:40:45.944314 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" Jan 27 16:40:45 crc kubenswrapper[4772]: I0127 16:40:45.994000 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" podStartSLOduration=2.993977669 podStartE2EDuration="2.993977669s" podCreationTimestamp="2026-01-27 16:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:40:45.985442756 +0000 UTC m=+5631.966051874" watchObservedRunningTime="2026-01-27 16:40:45.993977669 +0000 UTC m=+5631.974586767" Jan 27 16:40:53 crc kubenswrapper[4772]: I0127 16:40:53.625489 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" Jan 27 16:40:53 crc kubenswrapper[4772]: I0127 16:40:53.708941 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7758766f57-6fk67"] Jan 27 16:40:53 crc kubenswrapper[4772]: I0127 16:40:53.712525 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7758766f57-6fk67" podUID="f90709ae-7118-42ca-aad1-922961a6f858" containerName="dnsmasq-dns" containerID="cri-o://97e6d0f1812bfdc92938ed911c0ea776bf32e0bb2e1f04fe8e722bdbc3e2b5e5" gracePeriod=10 Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.193765 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7758766f57-6fk67" Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.369309 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f90709ae-7118-42ca-aad1-922961a6f858-ovsdbserver-nb\") pod \"f90709ae-7118-42ca-aad1-922961a6f858\" (UID: \"f90709ae-7118-42ca-aad1-922961a6f858\") " Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.369361 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f90709ae-7118-42ca-aad1-922961a6f858-dns-svc\") pod \"f90709ae-7118-42ca-aad1-922961a6f858\" (UID: \"f90709ae-7118-42ca-aad1-922961a6f858\") " Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.369400 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f90709ae-7118-42ca-aad1-922961a6f858-ovsdbserver-sb\") pod \"f90709ae-7118-42ca-aad1-922961a6f858\" (UID: \"f90709ae-7118-42ca-aad1-922961a6f858\") " Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.369419 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f90709ae-7118-42ca-aad1-922961a6f858-config\") pod \"f90709ae-7118-42ca-aad1-922961a6f858\" (UID: \"f90709ae-7118-42ca-aad1-922961a6f858\") " Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.369527 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bv5xf\" (UniqueName: \"kubernetes.io/projected/f90709ae-7118-42ca-aad1-922961a6f858-kube-api-access-bv5xf\") pod \"f90709ae-7118-42ca-aad1-922961a6f858\" (UID: \"f90709ae-7118-42ca-aad1-922961a6f858\") " Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.374504 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f90709ae-7118-42ca-aad1-922961a6f858-kube-api-access-bv5xf" (OuterVolumeSpecName: "kube-api-access-bv5xf") pod "f90709ae-7118-42ca-aad1-922961a6f858" (UID: "f90709ae-7118-42ca-aad1-922961a6f858"). InnerVolumeSpecName "kube-api-access-bv5xf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.425337 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f90709ae-7118-42ca-aad1-922961a6f858-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f90709ae-7118-42ca-aad1-922961a6f858" (UID: "f90709ae-7118-42ca-aad1-922961a6f858"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.427790 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f90709ae-7118-42ca-aad1-922961a6f858-config" (OuterVolumeSpecName: "config") pod "f90709ae-7118-42ca-aad1-922961a6f858" (UID: "f90709ae-7118-42ca-aad1-922961a6f858"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.439667 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f90709ae-7118-42ca-aad1-922961a6f858-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f90709ae-7118-42ca-aad1-922961a6f858" (UID: "f90709ae-7118-42ca-aad1-922961a6f858"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.448861 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f90709ae-7118-42ca-aad1-922961a6f858-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f90709ae-7118-42ca-aad1-922961a6f858" (UID: "f90709ae-7118-42ca-aad1-922961a6f858"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.463483 4772 generic.go:334] "Generic (PLEG): container finished" podID="f90709ae-7118-42ca-aad1-922961a6f858" containerID="97e6d0f1812bfdc92938ed911c0ea776bf32e0bb2e1f04fe8e722bdbc3e2b5e5" exitCode=0 Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.463526 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7758766f57-6fk67" event={"ID":"f90709ae-7118-42ca-aad1-922961a6f858","Type":"ContainerDied","Data":"97e6d0f1812bfdc92938ed911c0ea776bf32e0bb2e1f04fe8e722bdbc3e2b5e5"} Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.463551 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7758766f57-6fk67" event={"ID":"f90709ae-7118-42ca-aad1-922961a6f858","Type":"ContainerDied","Data":"b319140f655aa1b23b1e685173361086b6144080f0c58e842877b6cfd23b2922"} Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.463570 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7758766f57-6fk67" Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.463624 4772 scope.go:117] "RemoveContainer" containerID="97e6d0f1812bfdc92938ed911c0ea776bf32e0bb2e1f04fe8e722bdbc3e2b5e5" Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.471252 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f90709ae-7118-42ca-aad1-922961a6f858-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.471276 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f90709ae-7118-42ca-aad1-922961a6f858-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.471286 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f90709ae-7118-42ca-aad1-922961a6f858-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.471294 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f90709ae-7118-42ca-aad1-922961a6f858-config\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.471303 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bv5xf\" (UniqueName: \"kubernetes.io/projected/f90709ae-7118-42ca-aad1-922961a6f858-kube-api-access-bv5xf\") on node \"crc\" DevicePath \"\"" Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.499779 4772 scope.go:117] "RemoveContainer" containerID="ea368eea29b6534237f47789a13e7bcf3cb4d11f06fe750858483cf68a03ec51" Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.506261 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7758766f57-6fk67"] Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.514305 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7758766f57-6fk67"] Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.519920 4772 scope.go:117] "RemoveContainer" containerID="97e6d0f1812bfdc92938ed911c0ea776bf32e0bb2e1f04fe8e722bdbc3e2b5e5" Jan 27 16:40:54 crc kubenswrapper[4772]: E0127 16:40:54.520448 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97e6d0f1812bfdc92938ed911c0ea776bf32e0bb2e1f04fe8e722bdbc3e2b5e5\": container with ID starting with 97e6d0f1812bfdc92938ed911c0ea776bf32e0bb2e1f04fe8e722bdbc3e2b5e5 not found: ID does not exist" containerID="97e6d0f1812bfdc92938ed911c0ea776bf32e0bb2e1f04fe8e722bdbc3e2b5e5" Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.520483 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97e6d0f1812bfdc92938ed911c0ea776bf32e0bb2e1f04fe8e722bdbc3e2b5e5"} err="failed to get container status \"97e6d0f1812bfdc92938ed911c0ea776bf32e0bb2e1f04fe8e722bdbc3e2b5e5\": rpc error: code = NotFound desc = could not find container \"97e6d0f1812bfdc92938ed911c0ea776bf32e0bb2e1f04fe8e722bdbc3e2b5e5\": container with ID starting with 97e6d0f1812bfdc92938ed911c0ea776bf32e0bb2e1f04fe8e722bdbc3e2b5e5 not found: ID does not exist" Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.520507 4772 scope.go:117] "RemoveContainer" containerID="ea368eea29b6534237f47789a13e7bcf3cb4d11f06fe750858483cf68a03ec51" Jan 27 16:40:54 crc kubenswrapper[4772]: E0127 16:40:54.520814 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea368eea29b6534237f47789a13e7bcf3cb4d11f06fe750858483cf68a03ec51\": container with ID starting with ea368eea29b6534237f47789a13e7bcf3cb4d11f06fe750858483cf68a03ec51 not found: ID does not exist" containerID="ea368eea29b6534237f47789a13e7bcf3cb4d11f06fe750858483cf68a03ec51" Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.520836 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea368eea29b6534237f47789a13e7bcf3cb4d11f06fe750858483cf68a03ec51"} err="failed to get container status \"ea368eea29b6534237f47789a13e7bcf3cb4d11f06fe750858483cf68a03ec51\": rpc error: code = NotFound desc = could not find container \"ea368eea29b6534237f47789a13e7bcf3cb4d11f06fe750858483cf68a03ec51\": container with ID starting with ea368eea29b6534237f47789a13e7bcf3cb4d11f06fe750858483cf68a03ec51 not found: ID does not exist" Jan 27 16:40:54 crc kubenswrapper[4772]: I0127 16:40:54.676513 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f90709ae-7118-42ca-aad1-922961a6f858" path="/var/lib/kubelet/pods/f90709ae-7118-42ca-aad1-922961a6f858/volumes" Jan 27 16:40:56 crc kubenswrapper[4772]: I0127 16:40:56.007359 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-hk6d6"] Jan 27 16:40:56 crc kubenswrapper[4772]: E0127 16:40:56.008009 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f90709ae-7118-42ca-aad1-922961a6f858" containerName="init" Jan 27 16:40:56 crc kubenswrapper[4772]: I0127 16:40:56.008023 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f90709ae-7118-42ca-aad1-922961a6f858" containerName="init" Jan 27 16:40:56 crc kubenswrapper[4772]: E0127 16:40:56.008048 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f90709ae-7118-42ca-aad1-922961a6f858" containerName="dnsmasq-dns" Jan 27 16:40:56 crc kubenswrapper[4772]: I0127 16:40:56.008054 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f90709ae-7118-42ca-aad1-922961a6f858" containerName="dnsmasq-dns" Jan 27 16:40:56 crc kubenswrapper[4772]: I0127 16:40:56.008234 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="f90709ae-7118-42ca-aad1-922961a6f858" containerName="dnsmasq-dns" Jan 27 16:40:56 crc kubenswrapper[4772]: I0127 16:40:56.008875 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-hk6d6" Jan 27 16:40:56 crc kubenswrapper[4772]: I0127 16:40:56.023856 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-hk6d6"] Jan 27 16:40:56 crc kubenswrapper[4772]: I0127 16:40:56.099375 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-e17c-account-create-update-4ccmc"] Jan 27 16:40:56 crc kubenswrapper[4772]: I0127 16:40:56.100600 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e17c-account-create-update-4ccmc" Jan 27 16:40:56 crc kubenswrapper[4772]: I0127 16:40:56.102866 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Jan 27 16:40:56 crc kubenswrapper[4772]: I0127 16:40:56.113598 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-e17c-account-create-update-4ccmc"] Jan 27 16:40:56 crc kubenswrapper[4772]: I0127 16:40:56.201565 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/357c50eb-1246-4dd8-975c-b10d09439cbd-operator-scripts\") pod \"cinder-db-create-hk6d6\" (UID: \"357c50eb-1246-4dd8-975c-b10d09439cbd\") " pod="openstack/cinder-db-create-hk6d6" Jan 27 16:40:56 crc kubenswrapper[4772]: I0127 16:40:56.201633 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlzz8\" (UniqueName: \"kubernetes.io/projected/357c50eb-1246-4dd8-975c-b10d09439cbd-kube-api-access-wlzz8\") pod \"cinder-db-create-hk6d6\" (UID: \"357c50eb-1246-4dd8-975c-b10d09439cbd\") " pod="openstack/cinder-db-create-hk6d6" Jan 27 16:40:56 crc kubenswrapper[4772]: I0127 16:40:56.304104 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsjgt\" (UniqueName: \"kubernetes.io/projected/d67b98cc-0659-4ebd-a96a-025044731558-kube-api-access-wsjgt\") pod \"cinder-e17c-account-create-update-4ccmc\" (UID: \"d67b98cc-0659-4ebd-a96a-025044731558\") " pod="openstack/cinder-e17c-account-create-update-4ccmc" Jan 27 16:40:56 crc kubenswrapper[4772]: I0127 16:40:56.304211 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d67b98cc-0659-4ebd-a96a-025044731558-operator-scripts\") pod \"cinder-e17c-account-create-update-4ccmc\" (UID: \"d67b98cc-0659-4ebd-a96a-025044731558\") " pod="openstack/cinder-e17c-account-create-update-4ccmc" Jan 27 16:40:56 crc kubenswrapper[4772]: I0127 16:40:56.304287 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/357c50eb-1246-4dd8-975c-b10d09439cbd-operator-scripts\") pod \"cinder-db-create-hk6d6\" (UID: \"357c50eb-1246-4dd8-975c-b10d09439cbd\") " pod="openstack/cinder-db-create-hk6d6" Jan 27 16:40:56 crc kubenswrapper[4772]: I0127 16:40:56.304315 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlzz8\" (UniqueName: \"kubernetes.io/projected/357c50eb-1246-4dd8-975c-b10d09439cbd-kube-api-access-wlzz8\") pod \"cinder-db-create-hk6d6\" (UID: \"357c50eb-1246-4dd8-975c-b10d09439cbd\") " pod="openstack/cinder-db-create-hk6d6" Jan 27 16:40:56 crc kubenswrapper[4772]: I0127 16:40:56.305373 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/357c50eb-1246-4dd8-975c-b10d09439cbd-operator-scripts\") pod \"cinder-db-create-hk6d6\" (UID: \"357c50eb-1246-4dd8-975c-b10d09439cbd\") " pod="openstack/cinder-db-create-hk6d6" Jan 27 16:40:56 crc kubenswrapper[4772]: I0127 16:40:56.321187 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlzz8\" (UniqueName: \"kubernetes.io/projected/357c50eb-1246-4dd8-975c-b10d09439cbd-kube-api-access-wlzz8\") pod \"cinder-db-create-hk6d6\" (UID: \"357c50eb-1246-4dd8-975c-b10d09439cbd\") " pod="openstack/cinder-db-create-hk6d6" Jan 27 16:40:56 crc kubenswrapper[4772]: I0127 16:40:56.333472 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-hk6d6" Jan 27 16:40:56 crc kubenswrapper[4772]: I0127 16:40:56.406347 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d67b98cc-0659-4ebd-a96a-025044731558-operator-scripts\") pod \"cinder-e17c-account-create-update-4ccmc\" (UID: \"d67b98cc-0659-4ebd-a96a-025044731558\") " pod="openstack/cinder-e17c-account-create-update-4ccmc" Jan 27 16:40:56 crc kubenswrapper[4772]: I0127 16:40:56.407117 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsjgt\" (UniqueName: \"kubernetes.io/projected/d67b98cc-0659-4ebd-a96a-025044731558-kube-api-access-wsjgt\") pod \"cinder-e17c-account-create-update-4ccmc\" (UID: \"d67b98cc-0659-4ebd-a96a-025044731558\") " pod="openstack/cinder-e17c-account-create-update-4ccmc" Jan 27 16:40:56 crc kubenswrapper[4772]: I0127 16:40:56.407236 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d67b98cc-0659-4ebd-a96a-025044731558-operator-scripts\") pod \"cinder-e17c-account-create-update-4ccmc\" (UID: \"d67b98cc-0659-4ebd-a96a-025044731558\") " pod="openstack/cinder-e17c-account-create-update-4ccmc" Jan 27 16:40:56 crc kubenswrapper[4772]: I0127 16:40:56.430044 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsjgt\" (UniqueName: \"kubernetes.io/projected/d67b98cc-0659-4ebd-a96a-025044731558-kube-api-access-wsjgt\") pod \"cinder-e17c-account-create-update-4ccmc\" (UID: \"d67b98cc-0659-4ebd-a96a-025044731558\") " pod="openstack/cinder-e17c-account-create-update-4ccmc" Jan 27 16:40:57 crc kubenswrapper[4772]: I0127 16:40:56.718284 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e17c-account-create-update-4ccmc" Jan 27 16:40:57 crc kubenswrapper[4772]: I0127 16:40:57.398691 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-hk6d6"] Jan 27 16:40:57 crc kubenswrapper[4772]: I0127 16:40:57.492667 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-hk6d6" event={"ID":"357c50eb-1246-4dd8-975c-b10d09439cbd","Type":"ContainerStarted","Data":"376bf0f983c24165f1737a8abc049bd61ef568f18bb3826a169d2c905cf7cb55"} Jan 27 16:40:57 crc kubenswrapper[4772]: I0127 16:40:57.505979 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-e17c-account-create-update-4ccmc"] Jan 27 16:40:58 crc kubenswrapper[4772]: I0127 16:40:58.503075 4772 generic.go:334] "Generic (PLEG): container finished" podID="357c50eb-1246-4dd8-975c-b10d09439cbd" containerID="2746702aa9709faf5d661e469190b51cee57d01ba7b40596d6751ad98441dd4d" exitCode=0 Jan 27 16:40:58 crc kubenswrapper[4772]: I0127 16:40:58.503133 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-hk6d6" event={"ID":"357c50eb-1246-4dd8-975c-b10d09439cbd","Type":"ContainerDied","Data":"2746702aa9709faf5d661e469190b51cee57d01ba7b40596d6751ad98441dd4d"} Jan 27 16:40:58 crc kubenswrapper[4772]: I0127 16:40:58.505264 4772 generic.go:334] "Generic (PLEG): container finished" podID="d67b98cc-0659-4ebd-a96a-025044731558" containerID="dab4307441a9ab5d1178a5547c03edae0ead3f783db4ef2bf29e8414026bb08f" exitCode=0 Jan 27 16:40:58 crc kubenswrapper[4772]: I0127 16:40:58.505310 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-e17c-account-create-update-4ccmc" event={"ID":"d67b98cc-0659-4ebd-a96a-025044731558","Type":"ContainerDied","Data":"dab4307441a9ab5d1178a5547c03edae0ead3f783db4ef2bf29e8414026bb08f"} Jan 27 16:40:58 crc kubenswrapper[4772]: I0127 16:40:58.505339 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-e17c-account-create-update-4ccmc" event={"ID":"d67b98cc-0659-4ebd-a96a-025044731558","Type":"ContainerStarted","Data":"e8c6958939d862ea28c0f21a1d8ba7e761ac273dda2bdefeb7b657d0d56e4fa0"} Jan 27 16:41:00 crc kubenswrapper[4772]: I0127 16:41:00.008568 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e17c-account-create-update-4ccmc" Jan 27 16:41:00 crc kubenswrapper[4772]: I0127 16:41:00.017999 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-hk6d6" Jan 27 16:41:00 crc kubenswrapper[4772]: I0127 16:41:00.175245 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsjgt\" (UniqueName: \"kubernetes.io/projected/d67b98cc-0659-4ebd-a96a-025044731558-kube-api-access-wsjgt\") pod \"d67b98cc-0659-4ebd-a96a-025044731558\" (UID: \"d67b98cc-0659-4ebd-a96a-025044731558\") " Jan 27 16:41:00 crc kubenswrapper[4772]: I0127 16:41:00.175308 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/357c50eb-1246-4dd8-975c-b10d09439cbd-operator-scripts\") pod \"357c50eb-1246-4dd8-975c-b10d09439cbd\" (UID: \"357c50eb-1246-4dd8-975c-b10d09439cbd\") " Jan 27 16:41:00 crc kubenswrapper[4772]: I0127 16:41:00.175357 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlzz8\" (UniqueName: \"kubernetes.io/projected/357c50eb-1246-4dd8-975c-b10d09439cbd-kube-api-access-wlzz8\") pod \"357c50eb-1246-4dd8-975c-b10d09439cbd\" (UID: \"357c50eb-1246-4dd8-975c-b10d09439cbd\") " Jan 27 16:41:00 crc kubenswrapper[4772]: I0127 16:41:00.175382 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d67b98cc-0659-4ebd-a96a-025044731558-operator-scripts\") pod \"d67b98cc-0659-4ebd-a96a-025044731558\" (UID: \"d67b98cc-0659-4ebd-a96a-025044731558\") " Jan 27 16:41:00 crc kubenswrapper[4772]: I0127 16:41:00.175820 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/357c50eb-1246-4dd8-975c-b10d09439cbd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "357c50eb-1246-4dd8-975c-b10d09439cbd" (UID: "357c50eb-1246-4dd8-975c-b10d09439cbd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:41:00 crc kubenswrapper[4772]: I0127 16:41:00.175954 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/357c50eb-1246-4dd8-975c-b10d09439cbd-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:00 crc kubenswrapper[4772]: I0127 16:41:00.176360 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d67b98cc-0659-4ebd-a96a-025044731558-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d67b98cc-0659-4ebd-a96a-025044731558" (UID: "d67b98cc-0659-4ebd-a96a-025044731558"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:41:00 crc kubenswrapper[4772]: I0127 16:41:00.185410 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/357c50eb-1246-4dd8-975c-b10d09439cbd-kube-api-access-wlzz8" (OuterVolumeSpecName: "kube-api-access-wlzz8") pod "357c50eb-1246-4dd8-975c-b10d09439cbd" (UID: "357c50eb-1246-4dd8-975c-b10d09439cbd"). InnerVolumeSpecName "kube-api-access-wlzz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:41:00 crc kubenswrapper[4772]: I0127 16:41:00.185525 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d67b98cc-0659-4ebd-a96a-025044731558-kube-api-access-wsjgt" (OuterVolumeSpecName: "kube-api-access-wsjgt") pod "d67b98cc-0659-4ebd-a96a-025044731558" (UID: "d67b98cc-0659-4ebd-a96a-025044731558"). InnerVolumeSpecName "kube-api-access-wsjgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:41:00 crc kubenswrapper[4772]: I0127 16:41:00.276911 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsjgt\" (UniqueName: \"kubernetes.io/projected/d67b98cc-0659-4ebd-a96a-025044731558-kube-api-access-wsjgt\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:00 crc kubenswrapper[4772]: I0127 16:41:00.277075 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlzz8\" (UniqueName: \"kubernetes.io/projected/357c50eb-1246-4dd8-975c-b10d09439cbd-kube-api-access-wlzz8\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:00 crc kubenswrapper[4772]: I0127 16:41:00.277127 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d67b98cc-0659-4ebd-a96a-025044731558-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:00 crc kubenswrapper[4772]: I0127 16:41:00.536218 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-hk6d6" event={"ID":"357c50eb-1246-4dd8-975c-b10d09439cbd","Type":"ContainerDied","Data":"376bf0f983c24165f1737a8abc049bd61ef568f18bb3826a169d2c905cf7cb55"} Jan 27 16:41:00 crc kubenswrapper[4772]: I0127 16:41:00.536549 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="376bf0f983c24165f1737a8abc049bd61ef568f18bb3826a169d2c905cf7cb55" Jan 27 16:41:00 crc kubenswrapper[4772]: I0127 16:41:00.536706 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-hk6d6" Jan 27 16:41:00 crc kubenswrapper[4772]: I0127 16:41:00.539021 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-e17c-account-create-update-4ccmc" event={"ID":"d67b98cc-0659-4ebd-a96a-025044731558","Type":"ContainerDied","Data":"e8c6958939d862ea28c0f21a1d8ba7e761ac273dda2bdefeb7b657d0d56e4fa0"} Jan 27 16:41:00 crc kubenswrapper[4772]: I0127 16:41:00.539075 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8c6958939d862ea28c0f21a1d8ba7e761ac273dda2bdefeb7b657d0d56e4fa0" Jan 27 16:41:00 crc kubenswrapper[4772]: I0127 16:41:00.539158 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e17c-account-create-update-4ccmc" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.187686 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-66rdf"] Jan 27 16:41:06 crc kubenswrapper[4772]: E0127 16:41:06.188889 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="357c50eb-1246-4dd8-975c-b10d09439cbd" containerName="mariadb-database-create" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.188912 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="357c50eb-1246-4dd8-975c-b10d09439cbd" containerName="mariadb-database-create" Jan 27 16:41:06 crc kubenswrapper[4772]: E0127 16:41:06.188948 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d67b98cc-0659-4ebd-a96a-025044731558" containerName="mariadb-account-create-update" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.188960 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="d67b98cc-0659-4ebd-a96a-025044731558" containerName="mariadb-account-create-update" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.189287 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="357c50eb-1246-4dd8-975c-b10d09439cbd" containerName="mariadb-database-create" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.189320 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="d67b98cc-0659-4ebd-a96a-025044731558" containerName="mariadb-account-create-update" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.190320 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-66rdf" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.193132 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.193133 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-p7qxg" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.195726 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.200978 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-66rdf"] Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.320127 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e9bc1548-ca21-4230-a5db-a9321ab69a37-etc-machine-id\") pod \"cinder-db-sync-66rdf\" (UID: \"e9bc1548-ca21-4230-a5db-a9321ab69a37\") " pod="openstack/cinder-db-sync-66rdf" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.320251 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e9bc1548-ca21-4230-a5db-a9321ab69a37-db-sync-config-data\") pod \"cinder-db-sync-66rdf\" (UID: \"e9bc1548-ca21-4230-a5db-a9321ab69a37\") " pod="openstack/cinder-db-sync-66rdf" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.320307 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9bc1548-ca21-4230-a5db-a9321ab69a37-scripts\") pod \"cinder-db-sync-66rdf\" (UID: \"e9bc1548-ca21-4230-a5db-a9321ab69a37\") " pod="openstack/cinder-db-sync-66rdf" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.320335 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9bc1548-ca21-4230-a5db-a9321ab69a37-combined-ca-bundle\") pod \"cinder-db-sync-66rdf\" (UID: \"e9bc1548-ca21-4230-a5db-a9321ab69a37\") " pod="openstack/cinder-db-sync-66rdf" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.320363 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhjk7\" (UniqueName: \"kubernetes.io/projected/e9bc1548-ca21-4230-a5db-a9321ab69a37-kube-api-access-dhjk7\") pod \"cinder-db-sync-66rdf\" (UID: \"e9bc1548-ca21-4230-a5db-a9321ab69a37\") " pod="openstack/cinder-db-sync-66rdf" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.320386 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9bc1548-ca21-4230-a5db-a9321ab69a37-config-data\") pod \"cinder-db-sync-66rdf\" (UID: \"e9bc1548-ca21-4230-a5db-a9321ab69a37\") " pod="openstack/cinder-db-sync-66rdf" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.421434 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9bc1548-ca21-4230-a5db-a9321ab69a37-combined-ca-bundle\") pod \"cinder-db-sync-66rdf\" (UID: \"e9bc1548-ca21-4230-a5db-a9321ab69a37\") " pod="openstack/cinder-db-sync-66rdf" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.421492 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhjk7\" (UniqueName: \"kubernetes.io/projected/e9bc1548-ca21-4230-a5db-a9321ab69a37-kube-api-access-dhjk7\") pod \"cinder-db-sync-66rdf\" (UID: \"e9bc1548-ca21-4230-a5db-a9321ab69a37\") " pod="openstack/cinder-db-sync-66rdf" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.421515 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9bc1548-ca21-4230-a5db-a9321ab69a37-config-data\") pod \"cinder-db-sync-66rdf\" (UID: \"e9bc1548-ca21-4230-a5db-a9321ab69a37\") " pod="openstack/cinder-db-sync-66rdf" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.421569 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e9bc1548-ca21-4230-a5db-a9321ab69a37-etc-machine-id\") pod \"cinder-db-sync-66rdf\" (UID: \"e9bc1548-ca21-4230-a5db-a9321ab69a37\") " pod="openstack/cinder-db-sync-66rdf" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.421629 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e9bc1548-ca21-4230-a5db-a9321ab69a37-db-sync-config-data\") pod \"cinder-db-sync-66rdf\" (UID: \"e9bc1548-ca21-4230-a5db-a9321ab69a37\") " pod="openstack/cinder-db-sync-66rdf" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.421661 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9bc1548-ca21-4230-a5db-a9321ab69a37-scripts\") pod \"cinder-db-sync-66rdf\" (UID: \"e9bc1548-ca21-4230-a5db-a9321ab69a37\") " pod="openstack/cinder-db-sync-66rdf" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.421960 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e9bc1548-ca21-4230-a5db-a9321ab69a37-etc-machine-id\") pod \"cinder-db-sync-66rdf\" (UID: \"e9bc1548-ca21-4230-a5db-a9321ab69a37\") " pod="openstack/cinder-db-sync-66rdf" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.427330 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9bc1548-ca21-4230-a5db-a9321ab69a37-scripts\") pod \"cinder-db-sync-66rdf\" (UID: \"e9bc1548-ca21-4230-a5db-a9321ab69a37\") " pod="openstack/cinder-db-sync-66rdf" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.427964 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9bc1548-ca21-4230-a5db-a9321ab69a37-config-data\") pod \"cinder-db-sync-66rdf\" (UID: \"e9bc1548-ca21-4230-a5db-a9321ab69a37\") " pod="openstack/cinder-db-sync-66rdf" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.431641 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e9bc1548-ca21-4230-a5db-a9321ab69a37-db-sync-config-data\") pod \"cinder-db-sync-66rdf\" (UID: \"e9bc1548-ca21-4230-a5db-a9321ab69a37\") " pod="openstack/cinder-db-sync-66rdf" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.431913 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9bc1548-ca21-4230-a5db-a9321ab69a37-combined-ca-bundle\") pod \"cinder-db-sync-66rdf\" (UID: \"e9bc1548-ca21-4230-a5db-a9321ab69a37\") " pod="openstack/cinder-db-sync-66rdf" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.443937 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhjk7\" (UniqueName: \"kubernetes.io/projected/e9bc1548-ca21-4230-a5db-a9321ab69a37-kube-api-access-dhjk7\") pod \"cinder-db-sync-66rdf\" (UID: \"e9bc1548-ca21-4230-a5db-a9321ab69a37\") " pod="openstack/cinder-db-sync-66rdf" Jan 27 16:41:06 crc kubenswrapper[4772]: I0127 16:41:06.519181 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-66rdf" Jan 27 16:41:07 crc kubenswrapper[4772]: I0127 16:41:07.100398 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-66rdf"] Jan 27 16:41:07 crc kubenswrapper[4772]: I0127 16:41:07.639615 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-66rdf" event={"ID":"e9bc1548-ca21-4230-a5db-a9321ab69a37","Type":"ContainerStarted","Data":"1da868c60af8c377653c0df4d049fe41db3740b753fe0048ef2e596877f8308c"} Jan 27 16:41:08 crc kubenswrapper[4772]: I0127 16:41:08.651138 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-66rdf" event={"ID":"e9bc1548-ca21-4230-a5db-a9321ab69a37","Type":"ContainerStarted","Data":"d0051faf5f33fa9d044b3023d9e4654d63902fac62af135831e6d6e9a248c7b6"} Jan 27 16:41:08 crc kubenswrapper[4772]: I0127 16:41:08.674358 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-66rdf" podStartSLOduration=2.674341098 podStartE2EDuration="2.674341098s" podCreationTimestamp="2026-01-27 16:41:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:41:08.667306317 +0000 UTC m=+5654.647915415" watchObservedRunningTime="2026-01-27 16:41:08.674341098 +0000 UTC m=+5654.654950196" Jan 27 16:41:10 crc kubenswrapper[4772]: I0127 16:41:10.675205 4772 generic.go:334] "Generic (PLEG): container finished" podID="e9bc1548-ca21-4230-a5db-a9321ab69a37" containerID="d0051faf5f33fa9d044b3023d9e4654d63902fac62af135831e6d6e9a248c7b6" exitCode=0 Jan 27 16:41:10 crc kubenswrapper[4772]: I0127 16:41:10.686935 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-66rdf" event={"ID":"e9bc1548-ca21-4230-a5db-a9321ab69a37","Type":"ContainerDied","Data":"d0051faf5f33fa9d044b3023d9e4654d63902fac62af135831e6d6e9a248c7b6"} Jan 27 16:41:12 crc kubenswrapper[4772]: I0127 16:41:12.058657 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:41:12 crc kubenswrapper[4772]: I0127 16:41:12.059233 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:41:12 crc kubenswrapper[4772]: I0127 16:41:12.075487 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-66rdf" Jan 27 16:41:12 crc kubenswrapper[4772]: I0127 16:41:12.147475 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e9bc1548-ca21-4230-a5db-a9321ab69a37-db-sync-config-data\") pod \"e9bc1548-ca21-4230-a5db-a9321ab69a37\" (UID: \"e9bc1548-ca21-4230-a5db-a9321ab69a37\") " Jan 27 16:41:12 crc kubenswrapper[4772]: I0127 16:41:12.147547 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e9bc1548-ca21-4230-a5db-a9321ab69a37-etc-machine-id\") pod \"e9bc1548-ca21-4230-a5db-a9321ab69a37\" (UID: \"e9bc1548-ca21-4230-a5db-a9321ab69a37\") " Jan 27 16:41:12 crc kubenswrapper[4772]: I0127 16:41:12.147598 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9bc1548-ca21-4230-a5db-a9321ab69a37-combined-ca-bundle\") pod \"e9bc1548-ca21-4230-a5db-a9321ab69a37\" (UID: \"e9bc1548-ca21-4230-a5db-a9321ab69a37\") " Jan 27 16:41:12 crc kubenswrapper[4772]: I0127 16:41:12.147738 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhjk7\" (UniqueName: \"kubernetes.io/projected/e9bc1548-ca21-4230-a5db-a9321ab69a37-kube-api-access-dhjk7\") pod \"e9bc1548-ca21-4230-a5db-a9321ab69a37\" (UID: \"e9bc1548-ca21-4230-a5db-a9321ab69a37\") " Jan 27 16:41:12 crc kubenswrapper[4772]: I0127 16:41:12.147780 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9bc1548-ca21-4230-a5db-a9321ab69a37-config-data\") pod \"e9bc1548-ca21-4230-a5db-a9321ab69a37\" (UID: \"e9bc1548-ca21-4230-a5db-a9321ab69a37\") " Jan 27 16:41:12 crc kubenswrapper[4772]: I0127 16:41:12.147909 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9bc1548-ca21-4230-a5db-a9321ab69a37-scripts\") pod \"e9bc1548-ca21-4230-a5db-a9321ab69a37\" (UID: \"e9bc1548-ca21-4230-a5db-a9321ab69a37\") " Jan 27 16:41:12 crc kubenswrapper[4772]: I0127 16:41:12.148411 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e9bc1548-ca21-4230-a5db-a9321ab69a37-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e9bc1548-ca21-4230-a5db-a9321ab69a37" (UID: "e9bc1548-ca21-4230-a5db-a9321ab69a37"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 16:41:12 crc kubenswrapper[4772]: I0127 16:41:12.153658 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9bc1548-ca21-4230-a5db-a9321ab69a37-kube-api-access-dhjk7" (OuterVolumeSpecName: "kube-api-access-dhjk7") pod "e9bc1548-ca21-4230-a5db-a9321ab69a37" (UID: "e9bc1548-ca21-4230-a5db-a9321ab69a37"). InnerVolumeSpecName "kube-api-access-dhjk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:41:12 crc kubenswrapper[4772]: I0127 16:41:12.154256 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9bc1548-ca21-4230-a5db-a9321ab69a37-scripts" (OuterVolumeSpecName: "scripts") pod "e9bc1548-ca21-4230-a5db-a9321ab69a37" (UID: "e9bc1548-ca21-4230-a5db-a9321ab69a37"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:41:12 crc kubenswrapper[4772]: I0127 16:41:12.156195 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9bc1548-ca21-4230-a5db-a9321ab69a37-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "e9bc1548-ca21-4230-a5db-a9321ab69a37" (UID: "e9bc1548-ca21-4230-a5db-a9321ab69a37"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:41:12 crc kubenswrapper[4772]: I0127 16:41:12.179460 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9bc1548-ca21-4230-a5db-a9321ab69a37-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e9bc1548-ca21-4230-a5db-a9321ab69a37" (UID: "e9bc1548-ca21-4230-a5db-a9321ab69a37"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:41:12 crc kubenswrapper[4772]: I0127 16:41:12.194729 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9bc1548-ca21-4230-a5db-a9321ab69a37-config-data" (OuterVolumeSpecName: "config-data") pod "e9bc1548-ca21-4230-a5db-a9321ab69a37" (UID: "e9bc1548-ca21-4230-a5db-a9321ab69a37"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:41:12 crc kubenswrapper[4772]: I0127 16:41:12.248917 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9bc1548-ca21-4230-a5db-a9321ab69a37-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:12 crc kubenswrapper[4772]: I0127 16:41:12.248947 4772 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e9bc1548-ca21-4230-a5db-a9321ab69a37-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:12 crc kubenswrapper[4772]: I0127 16:41:12.248958 4772 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e9bc1548-ca21-4230-a5db-a9321ab69a37-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:12 crc kubenswrapper[4772]: I0127 16:41:12.248966 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9bc1548-ca21-4230-a5db-a9321ab69a37-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:12 crc kubenswrapper[4772]: I0127 16:41:12.248975 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhjk7\" (UniqueName: \"kubernetes.io/projected/e9bc1548-ca21-4230-a5db-a9321ab69a37-kube-api-access-dhjk7\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:12 crc kubenswrapper[4772]: I0127 16:41:12.248984 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9bc1548-ca21-4230-a5db-a9321ab69a37-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:12 crc kubenswrapper[4772]: I0127 16:41:12.702150 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-66rdf" event={"ID":"e9bc1548-ca21-4230-a5db-a9321ab69a37","Type":"ContainerDied","Data":"1da868c60af8c377653c0df4d049fe41db3740b753fe0048ef2e596877f8308c"} Jan 27 16:41:12 crc kubenswrapper[4772]: I0127 16:41:12.702222 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1da868c60af8c377653c0df4d049fe41db3740b753fe0048ef2e596877f8308c" Jan 27 16:41:12 crc kubenswrapper[4772]: I0127 16:41:12.702253 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-66rdf" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.019566 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8896c5c8c-s6z7x"] Jan 27 16:41:13 crc kubenswrapper[4772]: E0127 16:41:13.020027 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9bc1548-ca21-4230-a5db-a9321ab69a37" containerName="cinder-db-sync" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.020042 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9bc1548-ca21-4230-a5db-a9321ab69a37" containerName="cinder-db-sync" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.023864 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9bc1548-ca21-4230-a5db-a9321ab69a37" containerName="cinder-db-sync" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.025428 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8896c5c8c-s6z7x" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.052638 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8896c5c8c-s6z7x"] Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.137869 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.139817 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.146426 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-p7qxg" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.146609 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.146805 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.147284 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.148846 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.164702 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22e08251-8371-4470-bc3e-d88d673d56f3-dns-svc\") pod \"dnsmasq-dns-8896c5c8c-s6z7x\" (UID: \"22e08251-8371-4470-bc3e-d88d673d56f3\") " pod="openstack/dnsmasq-dns-8896c5c8c-s6z7x" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.164819 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22e08251-8371-4470-bc3e-d88d673d56f3-ovsdbserver-sb\") pod \"dnsmasq-dns-8896c5c8c-s6z7x\" (UID: \"22e08251-8371-4470-bc3e-d88d673d56f3\") " pod="openstack/dnsmasq-dns-8896c5c8c-s6z7x" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.164859 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22e08251-8371-4470-bc3e-d88d673d56f3-config\") pod \"dnsmasq-dns-8896c5c8c-s6z7x\" (UID: \"22e08251-8371-4470-bc3e-d88d673d56f3\") " pod="openstack/dnsmasq-dns-8896c5c8c-s6z7x" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.164922 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24z7g\" (UniqueName: \"kubernetes.io/projected/22e08251-8371-4470-bc3e-d88d673d56f3-kube-api-access-24z7g\") pod \"dnsmasq-dns-8896c5c8c-s6z7x\" (UID: \"22e08251-8371-4470-bc3e-d88d673d56f3\") " pod="openstack/dnsmasq-dns-8896c5c8c-s6z7x" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.164999 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/22e08251-8371-4470-bc3e-d88d673d56f3-ovsdbserver-nb\") pod \"dnsmasq-dns-8896c5c8c-s6z7x\" (UID: \"22e08251-8371-4470-bc3e-d88d673d56f3\") " pod="openstack/dnsmasq-dns-8896c5c8c-s6z7x" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.266892 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " pod="openstack/cinder-api-0" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.266957 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-config-data-custom\") pod \"cinder-api-0\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " pod="openstack/cinder-api-0" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.267009 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/22e08251-8371-4470-bc3e-d88d673d56f3-ovsdbserver-nb\") pod \"dnsmasq-dns-8896c5c8c-s6z7x\" (UID: \"22e08251-8371-4470-bc3e-d88d673d56f3\") " pod="openstack/dnsmasq-dns-8896c5c8c-s6z7x" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.267119 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " pod="openstack/cinder-api-0" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.267145 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22e08251-8371-4470-bc3e-d88d673d56f3-dns-svc\") pod \"dnsmasq-dns-8896c5c8c-s6z7x\" (UID: \"22e08251-8371-4470-bc3e-d88d673d56f3\") " pod="openstack/dnsmasq-dns-8896c5c8c-s6z7x" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.267194 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-logs\") pod \"cinder-api-0\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " pod="openstack/cinder-api-0" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.267231 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-scripts\") pod \"cinder-api-0\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " pod="openstack/cinder-api-0" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.267473 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22e08251-8371-4470-bc3e-d88d673d56f3-ovsdbserver-sb\") pod \"dnsmasq-dns-8896c5c8c-s6z7x\" (UID: \"22e08251-8371-4470-bc3e-d88d673d56f3\") " pod="openstack/dnsmasq-dns-8896c5c8c-s6z7x" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.267499 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf7qv\" (UniqueName: \"kubernetes.io/projected/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-kube-api-access-rf7qv\") pod \"cinder-api-0\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " pod="openstack/cinder-api-0" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.267521 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-config-data\") pod \"cinder-api-0\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " pod="openstack/cinder-api-0" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.267543 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22e08251-8371-4470-bc3e-d88d673d56f3-config\") pod \"dnsmasq-dns-8896c5c8c-s6z7x\" (UID: \"22e08251-8371-4470-bc3e-d88d673d56f3\") " pod="openstack/dnsmasq-dns-8896c5c8c-s6z7x" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.267565 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24z7g\" (UniqueName: \"kubernetes.io/projected/22e08251-8371-4470-bc3e-d88d673d56f3-kube-api-access-24z7g\") pod \"dnsmasq-dns-8896c5c8c-s6z7x\" (UID: \"22e08251-8371-4470-bc3e-d88d673d56f3\") " pod="openstack/dnsmasq-dns-8896c5c8c-s6z7x" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.268948 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/22e08251-8371-4470-bc3e-d88d673d56f3-ovsdbserver-nb\") pod \"dnsmasq-dns-8896c5c8c-s6z7x\" (UID: \"22e08251-8371-4470-bc3e-d88d673d56f3\") " pod="openstack/dnsmasq-dns-8896c5c8c-s6z7x" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.269069 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/22e08251-8371-4470-bc3e-d88d673d56f3-dns-svc\") pod \"dnsmasq-dns-8896c5c8c-s6z7x\" (UID: \"22e08251-8371-4470-bc3e-d88d673d56f3\") " pod="openstack/dnsmasq-dns-8896c5c8c-s6z7x" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.278783 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22e08251-8371-4470-bc3e-d88d673d56f3-config\") pod \"dnsmasq-dns-8896c5c8c-s6z7x\" (UID: \"22e08251-8371-4470-bc3e-d88d673d56f3\") " pod="openstack/dnsmasq-dns-8896c5c8c-s6z7x" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.282800 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/22e08251-8371-4470-bc3e-d88d673d56f3-ovsdbserver-sb\") pod \"dnsmasq-dns-8896c5c8c-s6z7x\" (UID: \"22e08251-8371-4470-bc3e-d88d673d56f3\") " pod="openstack/dnsmasq-dns-8896c5c8c-s6z7x" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.324461 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24z7g\" (UniqueName: \"kubernetes.io/projected/22e08251-8371-4470-bc3e-d88d673d56f3-kube-api-access-24z7g\") pod \"dnsmasq-dns-8896c5c8c-s6z7x\" (UID: \"22e08251-8371-4470-bc3e-d88d673d56f3\") " pod="openstack/dnsmasq-dns-8896c5c8c-s6z7x" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.370200 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " pod="openstack/cinder-api-0" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.370267 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-logs\") pod \"cinder-api-0\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " pod="openstack/cinder-api-0" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.370304 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-scripts\") pod \"cinder-api-0\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " pod="openstack/cinder-api-0" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.370331 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf7qv\" (UniqueName: \"kubernetes.io/projected/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-kube-api-access-rf7qv\") pod \"cinder-api-0\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " pod="openstack/cinder-api-0" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.370356 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-config-data\") pod \"cinder-api-0\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " pod="openstack/cinder-api-0" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.370402 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " pod="openstack/cinder-api-0" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.370429 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-config-data-custom\") pod \"cinder-api-0\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " pod="openstack/cinder-api-0" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.371475 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-logs\") pod \"cinder-api-0\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " pod="openstack/cinder-api-0" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.372305 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " pod="openstack/cinder-api-0" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.374737 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-config-data-custom\") pod \"cinder-api-0\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " pod="openstack/cinder-api-0" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.375090 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8896c5c8c-s6z7x" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.375901 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-scripts\") pod \"cinder-api-0\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " pod="openstack/cinder-api-0" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.376176 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " pod="openstack/cinder-api-0" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.392448 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-config-data\") pod \"cinder-api-0\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " pod="openstack/cinder-api-0" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.410155 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf7qv\" (UniqueName: \"kubernetes.io/projected/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-kube-api-access-rf7qv\") pod \"cinder-api-0\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " pod="openstack/cinder-api-0" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.464070 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 27 16:41:13 crc kubenswrapper[4772]: I0127 16:41:13.931659 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8896c5c8c-s6z7x"] Jan 27 16:41:14 crc kubenswrapper[4772]: I0127 16:41:14.097096 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 27 16:41:14 crc kubenswrapper[4772]: I0127 16:41:14.758624 4772 generic.go:334] "Generic (PLEG): container finished" podID="22e08251-8371-4470-bc3e-d88d673d56f3" containerID="f8de1200ab762c5af38758e595e2a2a98ba09782a2b206aac9f09f13a5fccfdb" exitCode=0 Jan 27 16:41:14 crc kubenswrapper[4772]: I0127 16:41:14.759111 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8896c5c8c-s6z7x" event={"ID":"22e08251-8371-4470-bc3e-d88d673d56f3","Type":"ContainerDied","Data":"f8de1200ab762c5af38758e595e2a2a98ba09782a2b206aac9f09f13a5fccfdb"} Jan 27 16:41:14 crc kubenswrapper[4772]: I0127 16:41:14.759145 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8896c5c8c-s6z7x" event={"ID":"22e08251-8371-4470-bc3e-d88d673d56f3","Type":"ContainerStarted","Data":"f53bbe315294f89e556d416ad0831cf8d150079c678fca10d9cd19f4263b92e5"} Jan 27 16:41:14 crc kubenswrapper[4772]: I0127 16:41:14.767657 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f","Type":"ContainerStarted","Data":"714b9b6c5c6438e8d52919984e494ff614fd56b03d77fad32ca28a5a24726e46"} Jan 27 16:41:14 crc kubenswrapper[4772]: I0127 16:41:14.767697 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f","Type":"ContainerStarted","Data":"1985a43f384e22d29818a239c3a060f96131fc47f07c59ec7571d217b8454dc4"} Jan 27 16:41:15 crc kubenswrapper[4772]: I0127 16:41:15.776433 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8896c5c8c-s6z7x" event={"ID":"22e08251-8371-4470-bc3e-d88d673d56f3","Type":"ContainerStarted","Data":"5a92c845e63b80a27d43fbe97b07377bb4993d6fc150d43118a8c3a2e9f82858"} Jan 27 16:41:15 crc kubenswrapper[4772]: I0127 16:41:15.776657 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8896c5c8c-s6z7x" Jan 27 16:41:15 crc kubenswrapper[4772]: I0127 16:41:15.780081 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f","Type":"ContainerStarted","Data":"31e3d55d1d113b922a3f84db51ffbe33b9ce1af8297930f881b9d6d32892374f"} Jan 27 16:41:15 crc kubenswrapper[4772]: I0127 16:41:15.780273 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Jan 27 16:41:15 crc kubenswrapper[4772]: I0127 16:41:15.798947 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8896c5c8c-s6z7x" podStartSLOduration=3.7989276050000003 podStartE2EDuration="3.798927605s" podCreationTimestamp="2026-01-27 16:41:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:41:15.794821838 +0000 UTC m=+5661.775430936" watchObservedRunningTime="2026-01-27 16:41:15.798927605 +0000 UTC m=+5661.779536703" Jan 27 16:41:15 crc kubenswrapper[4772]: I0127 16:41:15.823917 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.823897317 podStartE2EDuration="2.823897317s" podCreationTimestamp="2026-01-27 16:41:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:41:15.815207689 +0000 UTC m=+5661.795816787" watchObservedRunningTime="2026-01-27 16:41:15.823897317 +0000 UTC m=+5661.804506415" Jan 27 16:41:17 crc kubenswrapper[4772]: I0127 16:41:17.530920 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qzhn5"] Jan 27 16:41:17 crc kubenswrapper[4772]: I0127 16:41:17.533990 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qzhn5" Jan 27 16:41:17 crc kubenswrapper[4772]: I0127 16:41:17.545772 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzhn5"] Jan 27 16:41:17 crc kubenswrapper[4772]: I0127 16:41:17.663344 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e18a7da1-2037-446c-8646-76917bb9544b-catalog-content\") pod \"redhat-marketplace-qzhn5\" (UID: \"e18a7da1-2037-446c-8646-76917bb9544b\") " pod="openshift-marketplace/redhat-marketplace-qzhn5" Jan 27 16:41:17 crc kubenswrapper[4772]: I0127 16:41:17.663562 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e18a7da1-2037-446c-8646-76917bb9544b-utilities\") pod \"redhat-marketplace-qzhn5\" (UID: \"e18a7da1-2037-446c-8646-76917bb9544b\") " pod="openshift-marketplace/redhat-marketplace-qzhn5" Jan 27 16:41:17 crc kubenswrapper[4772]: I0127 16:41:17.663637 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcwvs\" (UniqueName: \"kubernetes.io/projected/e18a7da1-2037-446c-8646-76917bb9544b-kube-api-access-gcwvs\") pod \"redhat-marketplace-qzhn5\" (UID: \"e18a7da1-2037-446c-8646-76917bb9544b\") " pod="openshift-marketplace/redhat-marketplace-qzhn5" Jan 27 16:41:17 crc kubenswrapper[4772]: I0127 16:41:17.765264 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e18a7da1-2037-446c-8646-76917bb9544b-utilities\") pod \"redhat-marketplace-qzhn5\" (UID: \"e18a7da1-2037-446c-8646-76917bb9544b\") " pod="openshift-marketplace/redhat-marketplace-qzhn5" Jan 27 16:41:17 crc kubenswrapper[4772]: I0127 16:41:17.765354 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcwvs\" (UniqueName: \"kubernetes.io/projected/e18a7da1-2037-446c-8646-76917bb9544b-kube-api-access-gcwvs\") pod \"redhat-marketplace-qzhn5\" (UID: \"e18a7da1-2037-446c-8646-76917bb9544b\") " pod="openshift-marketplace/redhat-marketplace-qzhn5" Jan 27 16:41:17 crc kubenswrapper[4772]: I0127 16:41:17.765389 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e18a7da1-2037-446c-8646-76917bb9544b-catalog-content\") pod \"redhat-marketplace-qzhn5\" (UID: \"e18a7da1-2037-446c-8646-76917bb9544b\") " pod="openshift-marketplace/redhat-marketplace-qzhn5" Jan 27 16:41:17 crc kubenswrapper[4772]: I0127 16:41:17.765947 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e18a7da1-2037-446c-8646-76917bb9544b-utilities\") pod \"redhat-marketplace-qzhn5\" (UID: \"e18a7da1-2037-446c-8646-76917bb9544b\") " pod="openshift-marketplace/redhat-marketplace-qzhn5" Jan 27 16:41:17 crc kubenswrapper[4772]: I0127 16:41:17.765979 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e18a7da1-2037-446c-8646-76917bb9544b-catalog-content\") pod \"redhat-marketplace-qzhn5\" (UID: \"e18a7da1-2037-446c-8646-76917bb9544b\") " pod="openshift-marketplace/redhat-marketplace-qzhn5" Jan 27 16:41:17 crc kubenswrapper[4772]: I0127 16:41:17.785138 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcwvs\" (UniqueName: \"kubernetes.io/projected/e18a7da1-2037-446c-8646-76917bb9544b-kube-api-access-gcwvs\") pod \"redhat-marketplace-qzhn5\" (UID: \"e18a7da1-2037-446c-8646-76917bb9544b\") " pod="openshift-marketplace/redhat-marketplace-qzhn5" Jan 27 16:41:17 crc kubenswrapper[4772]: I0127 16:41:17.853691 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qzhn5" Jan 27 16:41:18 crc kubenswrapper[4772]: I0127 16:41:18.335389 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzhn5"] Jan 27 16:41:18 crc kubenswrapper[4772]: W0127 16:41:18.346296 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode18a7da1_2037_446c_8646_76917bb9544b.slice/crio-54fcdad16f86f21c30f016c1e9b9c8d1af9bb4af36e051b32118c18fdeb26dbf WatchSource:0}: Error finding container 54fcdad16f86f21c30f016c1e9b9c8d1af9bb4af36e051b32118c18fdeb26dbf: Status 404 returned error can't find the container with id 54fcdad16f86f21c30f016c1e9b9c8d1af9bb4af36e051b32118c18fdeb26dbf Jan 27 16:41:18 crc kubenswrapper[4772]: I0127 16:41:18.806103 4772 generic.go:334] "Generic (PLEG): container finished" podID="e18a7da1-2037-446c-8646-76917bb9544b" containerID="5ba8daff8c2db9de85548929fc9125eac49769ff0aed125f736c35420b1b3caf" exitCode=0 Jan 27 16:41:18 crc kubenswrapper[4772]: I0127 16:41:18.806157 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzhn5" event={"ID":"e18a7da1-2037-446c-8646-76917bb9544b","Type":"ContainerDied","Data":"5ba8daff8c2db9de85548929fc9125eac49769ff0aed125f736c35420b1b3caf"} Jan 27 16:41:18 crc kubenswrapper[4772]: I0127 16:41:18.806502 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzhn5" event={"ID":"e18a7da1-2037-446c-8646-76917bb9544b","Type":"ContainerStarted","Data":"54fcdad16f86f21c30f016c1e9b9c8d1af9bb4af36e051b32118c18fdeb26dbf"} Jan 27 16:41:21 crc kubenswrapper[4772]: I0127 16:41:21.840932 4772 generic.go:334] "Generic (PLEG): container finished" podID="e18a7da1-2037-446c-8646-76917bb9544b" containerID="93df129a7d9ca49629bdaa9f5049664a8fdbaa8353cdb2fc27b598bded3cdb48" exitCode=0 Jan 27 16:41:21 crc kubenswrapper[4772]: I0127 16:41:21.840983 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzhn5" event={"ID":"e18a7da1-2037-446c-8646-76917bb9544b","Type":"ContainerDied","Data":"93df129a7d9ca49629bdaa9f5049664a8fdbaa8353cdb2fc27b598bded3cdb48"} Jan 27 16:41:22 crc kubenswrapper[4772]: I0127 16:41:22.851547 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzhn5" event={"ID":"e18a7da1-2037-446c-8646-76917bb9544b","Type":"ContainerStarted","Data":"2438505e8dbbe6ceb8683042eba5b21297e5463c63d380b7ca6a75115f1e523f"} Jan 27 16:41:22 crc kubenswrapper[4772]: I0127 16:41:22.871836 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qzhn5" podStartSLOduration=2.403662734 podStartE2EDuration="5.87181893s" podCreationTimestamp="2026-01-27 16:41:17 +0000 UTC" firstStartedPulling="2026-01-27 16:41:18.807745778 +0000 UTC m=+5664.788354876" lastFinishedPulling="2026-01-27 16:41:22.275901974 +0000 UTC m=+5668.256511072" observedRunningTime="2026-01-27 16:41:22.867880888 +0000 UTC m=+5668.848489986" watchObservedRunningTime="2026-01-27 16:41:22.87181893 +0000 UTC m=+5668.852428028" Jan 27 16:41:23 crc kubenswrapper[4772]: I0127 16:41:23.376696 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8896c5c8c-s6z7x" Jan 27 16:41:23 crc kubenswrapper[4772]: I0127 16:41:23.437569 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c56cfbf-9f62r"] Jan 27 16:41:23 crc kubenswrapper[4772]: I0127 16:41:23.437871 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" podUID="487e140f-d3fb-4ece-a41c-7a1c55a37534" containerName="dnsmasq-dns" containerID="cri-o://b740d63c9d1013074313725144a776e8a3b5f1fd4cbe0771f67139ace29cab8f" gracePeriod=10 Jan 27 16:41:23 crc kubenswrapper[4772]: I0127 16:41:23.624439 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" podUID="487e140f-d3fb-4ece-a41c-7a1c55a37534" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.67:5353: connect: connection refused" Jan 27 16:41:23 crc kubenswrapper[4772]: I0127 16:41:23.865123 4772 generic.go:334] "Generic (PLEG): container finished" podID="487e140f-d3fb-4ece-a41c-7a1c55a37534" containerID="b740d63c9d1013074313725144a776e8a3b5f1fd4cbe0771f67139ace29cab8f" exitCode=0 Jan 27 16:41:23 crc kubenswrapper[4772]: I0127 16:41:23.865556 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" event={"ID":"487e140f-d3fb-4ece-a41c-7a1c55a37534","Type":"ContainerDied","Data":"b740d63c9d1013074313725144a776e8a3b5f1fd4cbe0771f67139ace29cab8f"} Jan 27 16:41:24 crc kubenswrapper[4772]: I0127 16:41:24.087363 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" Jan 27 16:41:24 crc kubenswrapper[4772]: I0127 16:41:24.180639 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/487e140f-d3fb-4ece-a41c-7a1c55a37534-dns-svc\") pod \"487e140f-d3fb-4ece-a41c-7a1c55a37534\" (UID: \"487e140f-d3fb-4ece-a41c-7a1c55a37534\") " Jan 27 16:41:24 crc kubenswrapper[4772]: I0127 16:41:24.180850 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dkhk\" (UniqueName: \"kubernetes.io/projected/487e140f-d3fb-4ece-a41c-7a1c55a37534-kube-api-access-2dkhk\") pod \"487e140f-d3fb-4ece-a41c-7a1c55a37534\" (UID: \"487e140f-d3fb-4ece-a41c-7a1c55a37534\") " Jan 27 16:41:24 crc kubenswrapper[4772]: I0127 16:41:24.180939 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/487e140f-d3fb-4ece-a41c-7a1c55a37534-ovsdbserver-sb\") pod \"487e140f-d3fb-4ece-a41c-7a1c55a37534\" (UID: \"487e140f-d3fb-4ece-a41c-7a1c55a37534\") " Jan 27 16:41:24 crc kubenswrapper[4772]: I0127 16:41:24.180988 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/487e140f-d3fb-4ece-a41c-7a1c55a37534-ovsdbserver-nb\") pod \"487e140f-d3fb-4ece-a41c-7a1c55a37534\" (UID: \"487e140f-d3fb-4ece-a41c-7a1c55a37534\") " Jan 27 16:41:24 crc kubenswrapper[4772]: I0127 16:41:24.181025 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/487e140f-d3fb-4ece-a41c-7a1c55a37534-config\") pod \"487e140f-d3fb-4ece-a41c-7a1c55a37534\" (UID: \"487e140f-d3fb-4ece-a41c-7a1c55a37534\") " Jan 27 16:41:24 crc kubenswrapper[4772]: I0127 16:41:24.204574 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/487e140f-d3fb-4ece-a41c-7a1c55a37534-kube-api-access-2dkhk" (OuterVolumeSpecName: "kube-api-access-2dkhk") pod "487e140f-d3fb-4ece-a41c-7a1c55a37534" (UID: "487e140f-d3fb-4ece-a41c-7a1c55a37534"). InnerVolumeSpecName "kube-api-access-2dkhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:41:24 crc kubenswrapper[4772]: I0127 16:41:24.242131 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/487e140f-d3fb-4ece-a41c-7a1c55a37534-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "487e140f-d3fb-4ece-a41c-7a1c55a37534" (UID: "487e140f-d3fb-4ece-a41c-7a1c55a37534"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:41:24 crc kubenswrapper[4772]: I0127 16:41:24.242576 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/487e140f-d3fb-4ece-a41c-7a1c55a37534-config" (OuterVolumeSpecName: "config") pod "487e140f-d3fb-4ece-a41c-7a1c55a37534" (UID: "487e140f-d3fb-4ece-a41c-7a1c55a37534"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:41:24 crc kubenswrapper[4772]: I0127 16:41:24.247277 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/487e140f-d3fb-4ece-a41c-7a1c55a37534-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "487e140f-d3fb-4ece-a41c-7a1c55a37534" (UID: "487e140f-d3fb-4ece-a41c-7a1c55a37534"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:41:24 crc kubenswrapper[4772]: I0127 16:41:24.254602 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/487e140f-d3fb-4ece-a41c-7a1c55a37534-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "487e140f-d3fb-4ece-a41c-7a1c55a37534" (UID: "487e140f-d3fb-4ece-a41c-7a1c55a37534"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:41:24 crc kubenswrapper[4772]: I0127 16:41:24.283711 4772 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/487e140f-d3fb-4ece-a41c-7a1c55a37534-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:24 crc kubenswrapper[4772]: I0127 16:41:24.283763 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dkhk\" (UniqueName: \"kubernetes.io/projected/487e140f-d3fb-4ece-a41c-7a1c55a37534-kube-api-access-2dkhk\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:24 crc kubenswrapper[4772]: I0127 16:41:24.283778 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/487e140f-d3fb-4ece-a41c-7a1c55a37534-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:24 crc kubenswrapper[4772]: I0127 16:41:24.283791 4772 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/487e140f-d3fb-4ece-a41c-7a1c55a37534-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:24 crc kubenswrapper[4772]: I0127 16:41:24.283806 4772 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/487e140f-d3fb-4ece-a41c-7a1c55a37534-config\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:24 crc kubenswrapper[4772]: I0127 16:41:24.876121 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" event={"ID":"487e140f-d3fb-4ece-a41c-7a1c55a37534","Type":"ContainerDied","Data":"e6f437da52e8342f357b6a59971f74fc9403dcd77f6fa7a76fcf902ca6dd800f"} Jan 27 16:41:24 crc kubenswrapper[4772]: I0127 16:41:24.876522 4772 scope.go:117] "RemoveContainer" containerID="b740d63c9d1013074313725144a776e8a3b5f1fd4cbe0771f67139ace29cab8f" Jan 27 16:41:24 crc kubenswrapper[4772]: I0127 16:41:24.876214 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c56cfbf-9f62r" Jan 27 16:41:24 crc kubenswrapper[4772]: I0127 16:41:24.904570 4772 scope.go:117] "RemoveContainer" containerID="3b284b2b9feaaf7731a5ba79953b14bcd6f7aec90b9b76040821cf8f6a1cb2f9" Jan 27 16:41:24 crc kubenswrapper[4772]: I0127 16:41:24.909421 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c56cfbf-9f62r"] Jan 27 16:41:24 crc kubenswrapper[4772]: I0127 16:41:24.922305 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c56cfbf-9f62r"] Jan 27 16:41:24 crc kubenswrapper[4772]: I0127 16:41:24.952181 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 27 16:41:24 crc kubenswrapper[4772]: I0127 16:41:24.952448 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e" containerName="nova-cell0-conductor-conductor" containerID="cri-o://43ecbee0f6dc0baf4a170605cf33f5e90020f6dd199b406980e9515061715e50" gracePeriod=30 Jan 27 16:41:25 crc kubenswrapper[4772]: I0127 16:41:25.021549 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 16:41:25 crc kubenswrapper[4772]: I0127 16:41:25.021835 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8c652921-712d-46ee-9683-fd6312e33d1e" containerName="nova-metadata-log" containerID="cri-o://193d54411bb1309c9287f87812da125aae0f9213050e8e77849e515da35b0fc4" gracePeriod=30 Jan 27 16:41:25 crc kubenswrapper[4772]: I0127 16:41:25.022388 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8c652921-712d-46ee-9683-fd6312e33d1e" containerName="nova-metadata-metadata" containerID="cri-o://b7973b4ea385a2bd5b179204a29b45f97b045488c54fbb72800c5104d45dc193" gracePeriod=30 Jan 27 16:41:25 crc kubenswrapper[4772]: I0127 16:41:25.049266 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 27 16:41:25 crc kubenswrapper[4772]: I0127 16:41:25.049535 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a518103b-f26e-4f91-9ca9-93f1f8d5e113" containerName="nova-api-log" containerID="cri-o://8f94df1da4ef644e3b2171358f7c0dcee05549e2b5ef22dd8252e4943b46f5c0" gracePeriod=30 Jan 27 16:41:25 crc kubenswrapper[4772]: I0127 16:41:25.049943 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a518103b-f26e-4f91-9ca9-93f1f8d5e113" containerName="nova-api-api" containerID="cri-o://67369ccebf4181eb0be2f61ead1ad9a24b9c3e7147fed849be8e7d8b2afab075" gracePeriod=30 Jan 27 16:41:25 crc kubenswrapper[4772]: I0127 16:41:25.071505 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 16:41:25 crc kubenswrapper[4772]: I0127 16:41:25.071750 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="ce725b6a-06bb-4339-819f-fee8819078f0" containerName="nova-scheduler-scheduler" containerID="cri-o://33aa342145a07c52aa025581718207e0ed3316f1c48915a16075320a2dbeed5c" gracePeriod=30 Jan 27 16:41:25 crc kubenswrapper[4772]: I0127 16:41:25.085729 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 27 16:41:25 crc kubenswrapper[4772]: I0127 16:41:25.086638 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="a30c12eb-4e83-420a-8064-859689d91d2d" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://fba71bd5b1f04f4ce21ecca46c2028b0aeffaa1f208e5b17978c2d0f906cbf36" gracePeriod=30 Jan 27 16:41:25 crc kubenswrapper[4772]: I0127 16:41:25.094305 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 27 16:41:25 crc kubenswrapper[4772]: I0127 16:41:25.094690 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="e1c55305-7d18-44cb-90e5-b6793989abda" containerName="nova-cell1-conductor-conductor" containerID="cri-o://a2db576fbfc7eed07b697490f956015f0fceaa99ec44f3681ae02d7525bd0c81" gracePeriod=30 Jan 27 16:41:25 crc kubenswrapper[4772]: I0127 16:41:25.567897 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Jan 27 16:41:25 crc kubenswrapper[4772]: I0127 16:41:25.889018 4772 generic.go:334] "Generic (PLEG): container finished" podID="a518103b-f26e-4f91-9ca9-93f1f8d5e113" containerID="8f94df1da4ef644e3b2171358f7c0dcee05549e2b5ef22dd8252e4943b46f5c0" exitCode=143 Jan 27 16:41:25 crc kubenswrapper[4772]: I0127 16:41:25.889199 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a518103b-f26e-4f91-9ca9-93f1f8d5e113","Type":"ContainerDied","Data":"8f94df1da4ef644e3b2171358f7c0dcee05549e2b5ef22dd8252e4943b46f5c0"} Jan 27 16:41:25 crc kubenswrapper[4772]: I0127 16:41:25.891981 4772 generic.go:334] "Generic (PLEG): container finished" podID="a30c12eb-4e83-420a-8064-859689d91d2d" containerID="fba71bd5b1f04f4ce21ecca46c2028b0aeffaa1f208e5b17978c2d0f906cbf36" exitCode=0 Jan 27 16:41:25 crc kubenswrapper[4772]: I0127 16:41:25.892039 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a30c12eb-4e83-420a-8064-859689d91d2d","Type":"ContainerDied","Data":"fba71bd5b1f04f4ce21ecca46c2028b0aeffaa1f208e5b17978c2d0f906cbf36"} Jan 27 16:41:25 crc kubenswrapper[4772]: I0127 16:41:25.892061 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a30c12eb-4e83-420a-8064-859689d91d2d","Type":"ContainerDied","Data":"772c077f79e9fa947b145341393fdb1b1ddc0a4d4b6ec55b124c53c8a8bb5534"} Jan 27 16:41:25 crc kubenswrapper[4772]: I0127 16:41:25.892076 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="772c077f79e9fa947b145341393fdb1b1ddc0a4d4b6ec55b124c53c8a8bb5534" Jan 27 16:41:25 crc kubenswrapper[4772]: I0127 16:41:25.898406 4772 generic.go:334] "Generic (PLEG): container finished" podID="8c652921-712d-46ee-9683-fd6312e33d1e" containerID="193d54411bb1309c9287f87812da125aae0f9213050e8e77849e515da35b0fc4" exitCode=143 Jan 27 16:41:25 crc kubenswrapper[4772]: I0127 16:41:25.898442 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c652921-712d-46ee-9683-fd6312e33d1e","Type":"ContainerDied","Data":"193d54411bb1309c9287f87812da125aae0f9213050e8e77849e515da35b0fc4"} Jan 27 16:41:26 crc kubenswrapper[4772]: I0127 16:41:26.012492 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:41:26 crc kubenswrapper[4772]: I0127 16:41:26.143672 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22khv\" (UniqueName: \"kubernetes.io/projected/a30c12eb-4e83-420a-8064-859689d91d2d-kube-api-access-22khv\") pod \"a30c12eb-4e83-420a-8064-859689d91d2d\" (UID: \"a30c12eb-4e83-420a-8064-859689d91d2d\") " Jan 27 16:41:26 crc kubenswrapper[4772]: I0127 16:41:26.143785 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a30c12eb-4e83-420a-8064-859689d91d2d-config-data\") pod \"a30c12eb-4e83-420a-8064-859689d91d2d\" (UID: \"a30c12eb-4e83-420a-8064-859689d91d2d\") " Jan 27 16:41:26 crc kubenswrapper[4772]: I0127 16:41:26.143878 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a30c12eb-4e83-420a-8064-859689d91d2d-combined-ca-bundle\") pod \"a30c12eb-4e83-420a-8064-859689d91d2d\" (UID: \"a30c12eb-4e83-420a-8064-859689d91d2d\") " Jan 27 16:41:26 crc kubenswrapper[4772]: I0127 16:41:26.149629 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a30c12eb-4e83-420a-8064-859689d91d2d-kube-api-access-22khv" (OuterVolumeSpecName: "kube-api-access-22khv") pod "a30c12eb-4e83-420a-8064-859689d91d2d" (UID: "a30c12eb-4e83-420a-8064-859689d91d2d"). InnerVolumeSpecName "kube-api-access-22khv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:41:26 crc kubenswrapper[4772]: I0127 16:41:26.173232 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a30c12eb-4e83-420a-8064-859689d91d2d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a30c12eb-4e83-420a-8064-859689d91d2d" (UID: "a30c12eb-4e83-420a-8064-859689d91d2d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:41:26 crc kubenswrapper[4772]: I0127 16:41:26.194568 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a30c12eb-4e83-420a-8064-859689d91d2d-config-data" (OuterVolumeSpecName: "config-data") pod "a30c12eb-4e83-420a-8064-859689d91d2d" (UID: "a30c12eb-4e83-420a-8064-859689d91d2d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:41:26 crc kubenswrapper[4772]: I0127 16:41:26.245555 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a30c12eb-4e83-420a-8064-859689d91d2d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:26 crc kubenswrapper[4772]: I0127 16:41:26.245762 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22khv\" (UniqueName: \"kubernetes.io/projected/a30c12eb-4e83-420a-8064-859689d91d2d-kube-api-access-22khv\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:26 crc kubenswrapper[4772]: I0127 16:41:26.245831 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a30c12eb-4e83-420a-8064-859689d91d2d-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:26 crc kubenswrapper[4772]: I0127 16:41:26.674180 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="487e140f-d3fb-4ece-a41c-7a1c55a37534" path="/var/lib/kubelet/pods/487e140f-d3fb-4ece-a41c-7a1c55a37534/volumes" Jan 27 16:41:26 crc kubenswrapper[4772]: I0127 16:41:26.907699 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:41:26 crc kubenswrapper[4772]: I0127 16:41:26.932546 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 27 16:41:26 crc kubenswrapper[4772]: I0127 16:41:26.957236 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 27 16:41:26 crc kubenswrapper[4772]: I0127 16:41:26.967492 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 27 16:41:26 crc kubenswrapper[4772]: E0127 16:41:26.968107 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="487e140f-d3fb-4ece-a41c-7a1c55a37534" containerName="dnsmasq-dns" Jan 27 16:41:26 crc kubenswrapper[4772]: I0127 16:41:26.968209 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="487e140f-d3fb-4ece-a41c-7a1c55a37534" containerName="dnsmasq-dns" Jan 27 16:41:26 crc kubenswrapper[4772]: E0127 16:41:26.968284 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="487e140f-d3fb-4ece-a41c-7a1c55a37534" containerName="init" Jan 27 16:41:26 crc kubenswrapper[4772]: I0127 16:41:26.968334 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="487e140f-d3fb-4ece-a41c-7a1c55a37534" containerName="init" Jan 27 16:41:26 crc kubenswrapper[4772]: E0127 16:41:26.968423 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a30c12eb-4e83-420a-8064-859689d91d2d" containerName="nova-cell1-novncproxy-novncproxy" Jan 27 16:41:26 crc kubenswrapper[4772]: I0127 16:41:26.968473 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a30c12eb-4e83-420a-8064-859689d91d2d" containerName="nova-cell1-novncproxy-novncproxy" Jan 27 16:41:26 crc kubenswrapper[4772]: I0127 16:41:26.968695 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="a30c12eb-4e83-420a-8064-859689d91d2d" containerName="nova-cell1-novncproxy-novncproxy" Jan 27 16:41:26 crc kubenswrapper[4772]: I0127 16:41:26.968788 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="487e140f-d3fb-4ece-a41c-7a1c55a37534" containerName="dnsmasq-dns" Jan 27 16:41:26 crc kubenswrapper[4772]: I0127 16:41:26.969449 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:41:26 crc kubenswrapper[4772]: I0127 16:41:26.974351 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Jan 27 16:41:26 crc kubenswrapper[4772]: I0127 16:41:26.989975 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 27 16:41:27 crc kubenswrapper[4772]: I0127 16:41:27.160874 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db0a7ce2-c175-4632-abe5-f35a6b5ce680-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"db0a7ce2-c175-4632-abe5-f35a6b5ce680\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:41:27 crc kubenswrapper[4772]: I0127 16:41:27.161119 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db0a7ce2-c175-4632-abe5-f35a6b5ce680-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"db0a7ce2-c175-4632-abe5-f35a6b5ce680\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:41:27 crc kubenswrapper[4772]: I0127 16:41:27.161243 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl8m4\" (UniqueName: \"kubernetes.io/projected/db0a7ce2-c175-4632-abe5-f35a6b5ce680-kube-api-access-bl8m4\") pod \"nova-cell1-novncproxy-0\" (UID: \"db0a7ce2-c175-4632-abe5-f35a6b5ce680\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:41:27 crc kubenswrapper[4772]: I0127 16:41:27.263098 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db0a7ce2-c175-4632-abe5-f35a6b5ce680-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"db0a7ce2-c175-4632-abe5-f35a6b5ce680\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:41:27 crc kubenswrapper[4772]: I0127 16:41:27.263211 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl8m4\" (UniqueName: \"kubernetes.io/projected/db0a7ce2-c175-4632-abe5-f35a6b5ce680-kube-api-access-bl8m4\") pod \"nova-cell1-novncproxy-0\" (UID: \"db0a7ce2-c175-4632-abe5-f35a6b5ce680\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:41:27 crc kubenswrapper[4772]: I0127 16:41:27.263305 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db0a7ce2-c175-4632-abe5-f35a6b5ce680-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"db0a7ce2-c175-4632-abe5-f35a6b5ce680\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:41:27 crc kubenswrapper[4772]: I0127 16:41:27.276375 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db0a7ce2-c175-4632-abe5-f35a6b5ce680-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"db0a7ce2-c175-4632-abe5-f35a6b5ce680\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:41:27 crc kubenswrapper[4772]: I0127 16:41:27.280020 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db0a7ce2-c175-4632-abe5-f35a6b5ce680-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"db0a7ce2-c175-4632-abe5-f35a6b5ce680\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:41:27 crc kubenswrapper[4772]: I0127 16:41:27.280468 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl8m4\" (UniqueName: \"kubernetes.io/projected/db0a7ce2-c175-4632-abe5-f35a6b5ce680-kube-api-access-bl8m4\") pod \"nova-cell1-novncproxy-0\" (UID: \"db0a7ce2-c175-4632-abe5-f35a6b5ce680\") " pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:41:27 crc kubenswrapper[4772]: I0127 16:41:27.299640 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:41:27 crc kubenswrapper[4772]: I0127 16:41:27.766941 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 27 16:41:27 crc kubenswrapper[4772]: I0127 16:41:27.854942 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qzhn5" Jan 27 16:41:27 crc kubenswrapper[4772]: I0127 16:41:27.854985 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qzhn5" Jan 27 16:41:27 crc kubenswrapper[4772]: I0127 16:41:27.915880 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qzhn5" Jan 27 16:41:27 crc kubenswrapper[4772]: I0127 16:41:27.922590 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"db0a7ce2-c175-4632-abe5-f35a6b5ce680","Type":"ContainerStarted","Data":"cc983685cb7f18229c2e45eee47ef7546e8b63e99119756d8a92f563e3b11e6e"} Jan 27 16:41:27 crc kubenswrapper[4772]: I0127 16:41:27.971220 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qzhn5" Jan 27 16:41:28 crc kubenswrapper[4772]: E0127 16:41:28.080235 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="33aa342145a07c52aa025581718207e0ed3316f1c48915a16075320a2dbeed5c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 27 16:41:28 crc kubenswrapper[4772]: E0127 16:41:28.082354 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="33aa342145a07c52aa025581718207e0ed3316f1c48915a16075320a2dbeed5c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 27 16:41:28 crc kubenswrapper[4772]: E0127 16:41:28.084992 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="33aa342145a07c52aa025581718207e0ed3316f1c48915a16075320a2dbeed5c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 27 16:41:28 crc kubenswrapper[4772]: E0127 16:41:28.085048 4772 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="ce725b6a-06bb-4339-819f-fee8819078f0" containerName="nova-scheduler-scheduler" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.153808 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="8c652921-712d-46ee-9683-fd6312e33d1e" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.64:8775/\": read tcp 10.217.0.2:34972->10.217.1.64:8775: read: connection reset by peer" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.154201 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzhn5"] Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.155229 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="8c652921-712d-46ee-9683-fd6312e33d1e" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.64:8775/\": read tcp 10.217.0.2:34970->10.217.1.64:8775: read: connection reset by peer" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.494204 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.586915 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c55305-7d18-44cb-90e5-b6793989abda-combined-ca-bundle\") pod \"e1c55305-7d18-44cb-90e5-b6793989abda\" (UID: \"e1c55305-7d18-44cb-90e5-b6793989abda\") " Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.587142 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1c55305-7d18-44cb-90e5-b6793989abda-config-data\") pod \"e1c55305-7d18-44cb-90e5-b6793989abda\" (UID: \"e1c55305-7d18-44cb-90e5-b6793989abda\") " Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.587186 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdhmd\" (UniqueName: \"kubernetes.io/projected/e1c55305-7d18-44cb-90e5-b6793989abda-kube-api-access-vdhmd\") pod \"e1c55305-7d18-44cb-90e5-b6793989abda\" (UID: \"e1c55305-7d18-44cb-90e5-b6793989abda\") " Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.596388 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1c55305-7d18-44cb-90e5-b6793989abda-kube-api-access-vdhmd" (OuterVolumeSpecName: "kube-api-access-vdhmd") pod "e1c55305-7d18-44cb-90e5-b6793989abda" (UID: "e1c55305-7d18-44cb-90e5-b6793989abda"). InnerVolumeSpecName "kube-api-access-vdhmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.641973 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c55305-7d18-44cb-90e5-b6793989abda-config-data" (OuterVolumeSpecName: "config-data") pod "e1c55305-7d18-44cb-90e5-b6793989abda" (UID: "e1c55305-7d18-44cb-90e5-b6793989abda"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.668371 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c55305-7d18-44cb-90e5-b6793989abda-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e1c55305-7d18-44cb-90e5-b6793989abda" (UID: "e1c55305-7d18-44cb-90e5-b6793989abda"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.691346 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1c55305-7d18-44cb-90e5-b6793989abda-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.691371 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdhmd\" (UniqueName: \"kubernetes.io/projected/e1c55305-7d18-44cb-90e5-b6793989abda-kube-api-access-vdhmd\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.691380 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c55305-7d18-44cb-90e5-b6793989abda-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.696934 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a30c12eb-4e83-420a-8064-859689d91d2d" path="/var/lib/kubelet/pods/a30c12eb-4e83-420a-8064-859689d91d2d/volumes" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.713121 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.787285 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.793040 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cfjb\" (UniqueName: \"kubernetes.io/projected/8c652921-712d-46ee-9683-fd6312e33d1e-kube-api-access-7cfjb\") pod \"8c652921-712d-46ee-9683-fd6312e33d1e\" (UID: \"8c652921-712d-46ee-9683-fd6312e33d1e\") " Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.793602 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c652921-712d-46ee-9683-fd6312e33d1e-config-data\") pod \"8c652921-712d-46ee-9683-fd6312e33d1e\" (UID: \"8c652921-712d-46ee-9683-fd6312e33d1e\") " Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.793798 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c652921-712d-46ee-9683-fd6312e33d1e-logs\") pod \"8c652921-712d-46ee-9683-fd6312e33d1e\" (UID: \"8c652921-712d-46ee-9683-fd6312e33d1e\") " Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.793834 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c652921-712d-46ee-9683-fd6312e33d1e-combined-ca-bundle\") pod \"8c652921-712d-46ee-9683-fd6312e33d1e\" (UID: \"8c652921-712d-46ee-9683-fd6312e33d1e\") " Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.796060 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c652921-712d-46ee-9683-fd6312e33d1e-logs" (OuterVolumeSpecName: "logs") pod "8c652921-712d-46ee-9683-fd6312e33d1e" (UID: "8c652921-712d-46ee-9683-fd6312e33d1e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.799362 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c652921-712d-46ee-9683-fd6312e33d1e-kube-api-access-7cfjb" (OuterVolumeSpecName: "kube-api-access-7cfjb") pod "8c652921-712d-46ee-9683-fd6312e33d1e" (UID: "8c652921-712d-46ee-9683-fd6312e33d1e"). InnerVolumeSpecName "kube-api-access-7cfjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.824688 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c652921-712d-46ee-9683-fd6312e33d1e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c652921-712d-46ee-9683-fd6312e33d1e" (UID: "8c652921-712d-46ee-9683-fd6312e33d1e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.836940 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c652921-712d-46ee-9683-fd6312e33d1e-config-data" (OuterVolumeSpecName: "config-data") pod "8c652921-712d-46ee-9683-fd6312e33d1e" (UID: "8c652921-712d-46ee-9683-fd6312e33d1e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.870050 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.897148 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a518103b-f26e-4f91-9ca9-93f1f8d5e113-combined-ca-bundle\") pod \"a518103b-f26e-4f91-9ca9-93f1f8d5e113\" (UID: \"a518103b-f26e-4f91-9ca9-93f1f8d5e113\") " Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.897224 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a518103b-f26e-4f91-9ca9-93f1f8d5e113-config-data\") pod \"a518103b-f26e-4f91-9ca9-93f1f8d5e113\" (UID: \"a518103b-f26e-4f91-9ca9-93f1f8d5e113\") " Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.897267 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a518103b-f26e-4f91-9ca9-93f1f8d5e113-logs\") pod \"a518103b-f26e-4f91-9ca9-93f1f8d5e113\" (UID: \"a518103b-f26e-4f91-9ca9-93f1f8d5e113\") " Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.897291 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vg99\" (UniqueName: \"kubernetes.io/projected/a518103b-f26e-4f91-9ca9-93f1f8d5e113-kube-api-access-8vg99\") pod \"a518103b-f26e-4f91-9ca9-93f1f8d5e113\" (UID: \"a518103b-f26e-4f91-9ca9-93f1f8d5e113\") " Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.897660 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cfjb\" (UniqueName: \"kubernetes.io/projected/8c652921-712d-46ee-9683-fd6312e33d1e-kube-api-access-7cfjb\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.897657 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a518103b-f26e-4f91-9ca9-93f1f8d5e113-logs" (OuterVolumeSpecName: "logs") pod "a518103b-f26e-4f91-9ca9-93f1f8d5e113" (UID: "a518103b-f26e-4f91-9ca9-93f1f8d5e113"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.897676 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c652921-712d-46ee-9683-fd6312e33d1e-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.897724 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c652921-712d-46ee-9683-fd6312e33d1e-logs\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.897738 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c652921-712d-46ee-9683-fd6312e33d1e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.904495 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a518103b-f26e-4f91-9ca9-93f1f8d5e113-kube-api-access-8vg99" (OuterVolumeSpecName: "kube-api-access-8vg99") pod "a518103b-f26e-4f91-9ca9-93f1f8d5e113" (UID: "a518103b-f26e-4f91-9ca9-93f1f8d5e113"). InnerVolumeSpecName "kube-api-access-8vg99". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.930341 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a518103b-f26e-4f91-9ca9-93f1f8d5e113-config-data" (OuterVolumeSpecName: "config-data") pod "a518103b-f26e-4f91-9ca9-93f1f8d5e113" (UID: "a518103b-f26e-4f91-9ca9-93f1f8d5e113"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.935063 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a518103b-f26e-4f91-9ca9-93f1f8d5e113-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a518103b-f26e-4f91-9ca9-93f1f8d5e113" (UID: "a518103b-f26e-4f91-9ca9-93f1f8d5e113"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.936282 4772 generic.go:334] "Generic (PLEG): container finished" podID="a518103b-f26e-4f91-9ca9-93f1f8d5e113" containerID="67369ccebf4181eb0be2f61ead1ad9a24b9c3e7147fed849be8e7d8b2afab075" exitCode=0 Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.936346 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a518103b-f26e-4f91-9ca9-93f1f8d5e113","Type":"ContainerDied","Data":"67369ccebf4181eb0be2f61ead1ad9a24b9c3e7147fed849be8e7d8b2afab075"} Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.936371 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a518103b-f26e-4f91-9ca9-93f1f8d5e113","Type":"ContainerDied","Data":"ac2ae0550771f81842cdb596165399767b9e27c96df8f6a9cc1d4cc43f6c15e1"} Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.936389 4772 scope.go:117] "RemoveContainer" containerID="67369ccebf4181eb0be2f61ead1ad9a24b9c3e7147fed849be8e7d8b2afab075" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.936432 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.938693 4772 generic.go:334] "Generic (PLEG): container finished" podID="4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e" containerID="43ecbee0f6dc0baf4a170605cf33f5e90020f6dd199b406980e9515061715e50" exitCode=0 Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.938772 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e","Type":"ContainerDied","Data":"43ecbee0f6dc0baf4a170605cf33f5e90020f6dd199b406980e9515061715e50"} Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.938820 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e","Type":"ContainerDied","Data":"30ded49d3176ba704104aa049f568d3100eaac27bd25914a3c0f0f06ce633e73"} Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.938894 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.950733 4772 generic.go:334] "Generic (PLEG): container finished" podID="8c652921-712d-46ee-9683-fd6312e33d1e" containerID="b7973b4ea385a2bd5b179204a29b45f97b045488c54fbb72800c5104d45dc193" exitCode=0 Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.950836 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.950833 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c652921-712d-46ee-9683-fd6312e33d1e","Type":"ContainerDied","Data":"b7973b4ea385a2bd5b179204a29b45f97b045488c54fbb72800c5104d45dc193"} Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.950972 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c652921-712d-46ee-9683-fd6312e33d1e","Type":"ContainerDied","Data":"8059f8ca79e1f853876a311e130550fa4d5e0bfbe87838df84c9d7f7b757e1d9"} Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.954959 4772 generic.go:334] "Generic (PLEG): container finished" podID="e1c55305-7d18-44cb-90e5-b6793989abda" containerID="a2db576fbfc7eed07b697490f956015f0fceaa99ec44f3681ae02d7525bd0c81" exitCode=0 Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.955126 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.956001 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e1c55305-7d18-44cb-90e5-b6793989abda","Type":"ContainerDied","Data":"a2db576fbfc7eed07b697490f956015f0fceaa99ec44f3681ae02d7525bd0c81"} Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.956031 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e1c55305-7d18-44cb-90e5-b6793989abda","Type":"ContainerDied","Data":"8f15691a47c277f39b1b0ae53dfcc7f92ebe7de7d545b4f191aa38d15da7fb88"} Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.962798 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"db0a7ce2-c175-4632-abe5-f35a6b5ce680","Type":"ContainerStarted","Data":"7311b0a3ed84e943d3cdbe87e7d103f916c68b9c89bff3b2198ec268ff4f96f3"} Jan 27 16:41:28 crc kubenswrapper[4772]: I0127 16:41:28.995054 4772 scope.go:117] "RemoveContainer" containerID="8f94df1da4ef644e3b2171358f7c0dcee05549e2b5ef22dd8252e4943b46f5c0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.014662 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e-combined-ca-bundle\") pod \"4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e\" (UID: \"4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e\") " Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.015063 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2j7l8\" (UniqueName: \"kubernetes.io/projected/4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e-kube-api-access-2j7l8\") pod \"4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e\" (UID: \"4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e\") " Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.015217 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e-config-data\") pod \"4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e\" (UID: \"4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e\") " Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.015826 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a518103b-f26e-4f91-9ca9-93f1f8d5e113-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.015844 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a518103b-f26e-4f91-9ca9-93f1f8d5e113-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.015855 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a518103b-f26e-4f91-9ca9-93f1f8d5e113-logs\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.015866 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vg99\" (UniqueName: \"kubernetes.io/projected/a518103b-f26e-4f91-9ca9-93f1f8d5e113-kube-api-access-8vg99\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.026099 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e-kube-api-access-2j7l8" (OuterVolumeSpecName: "kube-api-access-2j7l8") pod "4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e" (UID: "4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e"). InnerVolumeSpecName "kube-api-access-2j7l8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.058354 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.063159 4772 scope.go:117] "RemoveContainer" containerID="67369ccebf4181eb0be2f61ead1ad9a24b9c3e7147fed849be8e7d8b2afab075" Jan 27 16:41:29 crc kubenswrapper[4772]: E0127 16:41:29.064055 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67369ccebf4181eb0be2f61ead1ad9a24b9c3e7147fed849be8e7d8b2afab075\": container with ID starting with 67369ccebf4181eb0be2f61ead1ad9a24b9c3e7147fed849be8e7d8b2afab075 not found: ID does not exist" containerID="67369ccebf4181eb0be2f61ead1ad9a24b9c3e7147fed849be8e7d8b2afab075" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.064084 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67369ccebf4181eb0be2f61ead1ad9a24b9c3e7147fed849be8e7d8b2afab075"} err="failed to get container status \"67369ccebf4181eb0be2f61ead1ad9a24b9c3e7147fed849be8e7d8b2afab075\": rpc error: code = NotFound desc = could not find container \"67369ccebf4181eb0be2f61ead1ad9a24b9c3e7147fed849be8e7d8b2afab075\": container with ID starting with 67369ccebf4181eb0be2f61ead1ad9a24b9c3e7147fed849be8e7d8b2afab075 not found: ID does not exist" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.064108 4772 scope.go:117] "RemoveContainer" containerID="8f94df1da4ef644e3b2171358f7c0dcee05549e2b5ef22dd8252e4943b46f5c0" Jan 27 16:41:29 crc kubenswrapper[4772]: E0127 16:41:29.064341 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f94df1da4ef644e3b2171358f7c0dcee05549e2b5ef22dd8252e4943b46f5c0\": container with ID starting with 8f94df1da4ef644e3b2171358f7c0dcee05549e2b5ef22dd8252e4943b46f5c0 not found: ID does not exist" containerID="8f94df1da4ef644e3b2171358f7c0dcee05549e2b5ef22dd8252e4943b46f5c0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.064356 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f94df1da4ef644e3b2171358f7c0dcee05549e2b5ef22dd8252e4943b46f5c0"} err="failed to get container status \"8f94df1da4ef644e3b2171358f7c0dcee05549e2b5ef22dd8252e4943b46f5c0\": rpc error: code = NotFound desc = could not find container \"8f94df1da4ef644e3b2171358f7c0dcee05549e2b5ef22dd8252e4943b46f5c0\": container with ID starting with 8f94df1da4ef644e3b2171358f7c0dcee05549e2b5ef22dd8252e4943b46f5c0 not found: ID does not exist" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.064367 4772 scope.go:117] "RemoveContainer" containerID="43ecbee0f6dc0baf4a170605cf33f5e90020f6dd199b406980e9515061715e50" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.066088 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e" (UID: "4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.097291 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.102275 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e-config-data" (OuterVolumeSpecName: "config-data") pod "4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e" (UID: "4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.119385 4772 scope.go:117] "RemoveContainer" containerID="43ecbee0f6dc0baf4a170605cf33f5e90020f6dd199b406980e9515061715e50" Jan 27 16:41:29 crc kubenswrapper[4772]: E0127 16:41:29.120391 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43ecbee0f6dc0baf4a170605cf33f5e90020f6dd199b406980e9515061715e50\": container with ID starting with 43ecbee0f6dc0baf4a170605cf33f5e90020f6dd199b406980e9515061715e50 not found: ID does not exist" containerID="43ecbee0f6dc0baf4a170605cf33f5e90020f6dd199b406980e9515061715e50" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.120445 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43ecbee0f6dc0baf4a170605cf33f5e90020f6dd199b406980e9515061715e50"} err="failed to get container status \"43ecbee0f6dc0baf4a170605cf33f5e90020f6dd199b406980e9515061715e50\": rpc error: code = NotFound desc = could not find container \"43ecbee0f6dc0baf4a170605cf33f5e90020f6dd199b406980e9515061715e50\": container with ID starting with 43ecbee0f6dc0baf4a170605cf33f5e90020f6dd199b406980e9515061715e50 not found: ID does not exist" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.120476 4772 scope.go:117] "RemoveContainer" containerID="b7973b4ea385a2bd5b179204a29b45f97b045488c54fbb72800c5104d45dc193" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.120539 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2j7l8\" (UniqueName: \"kubernetes.io/projected/4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e-kube-api-access-2j7l8\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.120556 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.120566 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.124236 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 27 16:41:29 crc kubenswrapper[4772]: E0127 16:41:29.124682 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1c55305-7d18-44cb-90e5-b6793989abda" containerName="nova-cell1-conductor-conductor" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.124700 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1c55305-7d18-44cb-90e5-b6793989abda" containerName="nova-cell1-conductor-conductor" Jan 27 16:41:29 crc kubenswrapper[4772]: E0127 16:41:29.124715 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e" containerName="nova-cell0-conductor-conductor" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.124721 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e" containerName="nova-cell0-conductor-conductor" Jan 27 16:41:29 crc kubenswrapper[4772]: E0127 16:41:29.124730 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a518103b-f26e-4f91-9ca9-93f1f8d5e113" containerName="nova-api-api" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.124739 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a518103b-f26e-4f91-9ca9-93f1f8d5e113" containerName="nova-api-api" Jan 27 16:41:29 crc kubenswrapper[4772]: E0127 16:41:29.124756 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c652921-712d-46ee-9683-fd6312e33d1e" containerName="nova-metadata-log" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.124762 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c652921-712d-46ee-9683-fd6312e33d1e" containerName="nova-metadata-log" Jan 27 16:41:29 crc kubenswrapper[4772]: E0127 16:41:29.124778 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c652921-712d-46ee-9683-fd6312e33d1e" containerName="nova-metadata-metadata" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.124784 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c652921-712d-46ee-9683-fd6312e33d1e" containerName="nova-metadata-metadata" Jan 27 16:41:29 crc kubenswrapper[4772]: E0127 16:41:29.124795 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a518103b-f26e-4f91-9ca9-93f1f8d5e113" containerName="nova-api-log" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.124801 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a518103b-f26e-4f91-9ca9-93f1f8d5e113" containerName="nova-api-log" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.124953 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c652921-712d-46ee-9683-fd6312e33d1e" containerName="nova-metadata-log" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.124970 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="a518103b-f26e-4f91-9ca9-93f1f8d5e113" containerName="nova-api-log" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.124993 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="a518103b-f26e-4f91-9ca9-93f1f8d5e113" containerName="nova-api-api" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.125006 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1c55305-7d18-44cb-90e5-b6793989abda" containerName="nova-cell1-conductor-conductor" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.125016 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e" containerName="nova-cell0-conductor-conductor" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.125024 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c652921-712d-46ee-9683-fd6312e33d1e" containerName="nova-metadata-metadata" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.126003 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.129161 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.158656 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.166280 4772 scope.go:117] "RemoveContainer" containerID="193d54411bb1309c9287f87812da125aae0f9213050e8e77849e515da35b0fc4" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.170609 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.170587188 podStartE2EDuration="3.170587188s" podCreationTimestamp="2026-01-27 16:41:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:41:29.010932598 +0000 UTC m=+5674.991541706" watchObservedRunningTime="2026-01-27 16:41:29.170587188 +0000 UTC m=+5675.151196286" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.194574 4772 scope.go:117] "RemoveContainer" containerID="b7973b4ea385a2bd5b179204a29b45f97b045488c54fbb72800c5104d45dc193" Jan 27 16:41:29 crc kubenswrapper[4772]: E0127 16:41:29.195234 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7973b4ea385a2bd5b179204a29b45f97b045488c54fbb72800c5104d45dc193\": container with ID starting with b7973b4ea385a2bd5b179204a29b45f97b045488c54fbb72800c5104d45dc193 not found: ID does not exist" containerID="b7973b4ea385a2bd5b179204a29b45f97b045488c54fbb72800c5104d45dc193" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.195324 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7973b4ea385a2bd5b179204a29b45f97b045488c54fbb72800c5104d45dc193"} err="failed to get container status \"b7973b4ea385a2bd5b179204a29b45f97b045488c54fbb72800c5104d45dc193\": rpc error: code = NotFound desc = could not find container \"b7973b4ea385a2bd5b179204a29b45f97b045488c54fbb72800c5104d45dc193\": container with ID starting with b7973b4ea385a2bd5b179204a29b45f97b045488c54fbb72800c5104d45dc193 not found: ID does not exist" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.195348 4772 scope.go:117] "RemoveContainer" containerID="193d54411bb1309c9287f87812da125aae0f9213050e8e77849e515da35b0fc4" Jan 27 16:41:29 crc kubenswrapper[4772]: E0127 16:41:29.196206 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"193d54411bb1309c9287f87812da125aae0f9213050e8e77849e515da35b0fc4\": container with ID starting with 193d54411bb1309c9287f87812da125aae0f9213050e8e77849e515da35b0fc4 not found: ID does not exist" containerID="193d54411bb1309c9287f87812da125aae0f9213050e8e77849e515da35b0fc4" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.196241 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"193d54411bb1309c9287f87812da125aae0f9213050e8e77849e515da35b0fc4"} err="failed to get container status \"193d54411bb1309c9287f87812da125aae0f9213050e8e77849e515da35b0fc4\": rpc error: code = NotFound desc = could not find container \"193d54411bb1309c9287f87812da125aae0f9213050e8e77849e515da35b0fc4\": container with ID starting with 193d54411bb1309c9287f87812da125aae0f9213050e8e77849e515da35b0fc4 not found: ID does not exist" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.196262 4772 scope.go:117] "RemoveContainer" containerID="a2db576fbfc7eed07b697490f956015f0fceaa99ec44f3681ae02d7525bd0c81" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.196584 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.208854 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.218544 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.219923 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.221720 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f178d7e3-af69-4014-8209-5e766a130997-logs\") pod \"nova-api-0\" (UID: \"f178d7e3-af69-4014-8209-5e766a130997\") " pod="openstack/nova-api-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.221779 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f178d7e3-af69-4014-8209-5e766a130997-config-data\") pod \"nova-api-0\" (UID: \"f178d7e3-af69-4014-8209-5e766a130997\") " pod="openstack/nova-api-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.221807 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qptf4\" (UniqueName: \"kubernetes.io/projected/f178d7e3-af69-4014-8209-5e766a130997-kube-api-access-qptf4\") pod \"nova-api-0\" (UID: \"f178d7e3-af69-4014-8209-5e766a130997\") " pod="openstack/nova-api-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.221834 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f178d7e3-af69-4014-8209-5e766a130997-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f178d7e3-af69-4014-8209-5e766a130997\") " pod="openstack/nova-api-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.222414 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.228654 4772 scope.go:117] "RemoveContainer" containerID="a2db576fbfc7eed07b697490f956015f0fceaa99ec44f3681ae02d7525bd0c81" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.229713 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 16:41:29 crc kubenswrapper[4772]: E0127 16:41:29.230991 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2db576fbfc7eed07b697490f956015f0fceaa99ec44f3681ae02d7525bd0c81\": container with ID starting with a2db576fbfc7eed07b697490f956015f0fceaa99ec44f3681ae02d7525bd0c81 not found: ID does not exist" containerID="a2db576fbfc7eed07b697490f956015f0fceaa99ec44f3681ae02d7525bd0c81" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.231021 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2db576fbfc7eed07b697490f956015f0fceaa99ec44f3681ae02d7525bd0c81"} err="failed to get container status \"a2db576fbfc7eed07b697490f956015f0fceaa99ec44f3681ae02d7525bd0c81\": rpc error: code = NotFound desc = could not find container \"a2db576fbfc7eed07b697490f956015f0fceaa99ec44f3681ae02d7525bd0c81\": container with ID starting with a2db576fbfc7eed07b697490f956015f0fceaa99ec44f3681ae02d7525bd0c81 not found: ID does not exist" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.242669 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.251920 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 27 16:41:29 crc kubenswrapper[4772]: E0127 16:41:29.253434 4772 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c652921_712d_46ee_9683_fd6312e33d1e.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c652921_712d_46ee_9683_fd6312e33d1e.slice/crio-8059f8ca79e1f853876a311e130550fa4d5e0bfbe87838df84c9d7f7b757e1d9\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda518103b_f26e_4f91_9ca9_93f1f8d5e113.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda518103b_f26e_4f91_9ca9_93f1f8d5e113.slice/crio-ac2ae0550771f81842cdb596165399767b9e27c96df8f6a9cc1d4cc43f6c15e1\": RecentStats: unable to find data in memory cache]" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.291107 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.315380 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.315509 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.318538 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.324555 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qptf4\" (UniqueName: \"kubernetes.io/projected/f178d7e3-af69-4014-8209-5e766a130997-kube-api-access-qptf4\") pod \"nova-api-0\" (UID: \"f178d7e3-af69-4014-8209-5e766a130997\") " pod="openstack/nova-api-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.324618 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f178d7e3-af69-4014-8209-5e766a130997-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f178d7e3-af69-4014-8209-5e766a130997\") " pod="openstack/nova-api-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.324685 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9549c89c-f55f-484d-80b2-ca1ad19bf758-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"9549c89c-f55f-484d-80b2-ca1ad19bf758\") " pod="openstack/nova-cell1-conductor-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.324707 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9549c89c-f55f-484d-80b2-ca1ad19bf758-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"9549c89c-f55f-484d-80b2-ca1ad19bf758\") " pod="openstack/nova-cell1-conductor-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.324762 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f178d7e3-af69-4014-8209-5e766a130997-logs\") pod \"nova-api-0\" (UID: \"f178d7e3-af69-4014-8209-5e766a130997\") " pod="openstack/nova-api-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.324816 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f178d7e3-af69-4014-8209-5e766a130997-config-data\") pod \"nova-api-0\" (UID: \"f178d7e3-af69-4014-8209-5e766a130997\") " pod="openstack/nova-api-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.324841 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgqfs\" (UniqueName: \"kubernetes.io/projected/9549c89c-f55f-484d-80b2-ca1ad19bf758-kube-api-access-pgqfs\") pod \"nova-cell1-conductor-0\" (UID: \"9549c89c-f55f-484d-80b2-ca1ad19bf758\") " pod="openstack/nova-cell1-conductor-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.326146 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f178d7e3-af69-4014-8209-5e766a130997-logs\") pod \"nova-api-0\" (UID: \"f178d7e3-af69-4014-8209-5e766a130997\") " pod="openstack/nova-api-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.341843 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f178d7e3-af69-4014-8209-5e766a130997-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f178d7e3-af69-4014-8209-5e766a130997\") " pod="openstack/nova-api-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.343655 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f178d7e3-af69-4014-8209-5e766a130997-config-data\") pod \"nova-api-0\" (UID: \"f178d7e3-af69-4014-8209-5e766a130997\") " pod="openstack/nova-api-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.349823 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qptf4\" (UniqueName: \"kubernetes.io/projected/f178d7e3-af69-4014-8209-5e766a130997-kube-api-access-qptf4\") pod \"nova-api-0\" (UID: \"f178d7e3-af69-4014-8209-5e766a130997\") " pod="openstack/nova-api-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.374359 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.388791 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.404580 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.406008 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.408664 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.414559 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.426290 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a983cf0-2c51-4d6a-af53-f115f3a57360-config-data\") pod \"nova-metadata-0\" (UID: \"3a983cf0-2c51-4d6a-af53-f115f3a57360\") " pod="openstack/nova-metadata-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.426423 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-427mh\" (UniqueName: \"kubernetes.io/projected/3a983cf0-2c51-4d6a-af53-f115f3a57360-kube-api-access-427mh\") pod \"nova-metadata-0\" (UID: \"3a983cf0-2c51-4d6a-af53-f115f3a57360\") " pod="openstack/nova-metadata-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.426496 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgqfs\" (UniqueName: \"kubernetes.io/projected/9549c89c-f55f-484d-80b2-ca1ad19bf758-kube-api-access-pgqfs\") pod \"nova-cell1-conductor-0\" (UID: \"9549c89c-f55f-484d-80b2-ca1ad19bf758\") " pod="openstack/nova-cell1-conductor-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.426562 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a983cf0-2c51-4d6a-af53-f115f3a57360-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3a983cf0-2c51-4d6a-af53-f115f3a57360\") " pod="openstack/nova-metadata-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.426621 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9549c89c-f55f-484d-80b2-ca1ad19bf758-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"9549c89c-f55f-484d-80b2-ca1ad19bf758\") " pod="openstack/nova-cell1-conductor-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.426664 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9549c89c-f55f-484d-80b2-ca1ad19bf758-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"9549c89c-f55f-484d-80b2-ca1ad19bf758\") " pod="openstack/nova-cell1-conductor-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.426711 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a983cf0-2c51-4d6a-af53-f115f3a57360-logs\") pod \"nova-metadata-0\" (UID: \"3a983cf0-2c51-4d6a-af53-f115f3a57360\") " pod="openstack/nova-metadata-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.431104 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9549c89c-f55f-484d-80b2-ca1ad19bf758-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"9549c89c-f55f-484d-80b2-ca1ad19bf758\") " pod="openstack/nova-cell1-conductor-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.437972 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9549c89c-f55f-484d-80b2-ca1ad19bf758-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"9549c89c-f55f-484d-80b2-ca1ad19bf758\") " pod="openstack/nova-cell1-conductor-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.441954 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgqfs\" (UniqueName: \"kubernetes.io/projected/9549c89c-f55f-484d-80b2-ca1ad19bf758-kube-api-access-pgqfs\") pod \"nova-cell1-conductor-0\" (UID: \"9549c89c-f55f-484d-80b2-ca1ad19bf758\") " pod="openstack/nova-cell1-conductor-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.469282 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.528713 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a983cf0-2c51-4d6a-af53-f115f3a57360-config-data\") pod \"nova-metadata-0\" (UID: \"3a983cf0-2c51-4d6a-af53-f115f3a57360\") " pod="openstack/nova-metadata-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.528775 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-427mh\" (UniqueName: \"kubernetes.io/projected/3a983cf0-2c51-4d6a-af53-f115f3a57360-kube-api-access-427mh\") pod \"nova-metadata-0\" (UID: \"3a983cf0-2c51-4d6a-af53-f115f3a57360\") " pod="openstack/nova-metadata-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.528818 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp92q\" (UniqueName: \"kubernetes.io/projected/3331c1dd-ff2d-4a41-9cb3-731297ae0dc3-kube-api-access-fp92q\") pod \"nova-cell0-conductor-0\" (UID: \"3331c1dd-ff2d-4a41-9cb3-731297ae0dc3\") " pod="openstack/nova-cell0-conductor-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.528852 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3331c1dd-ff2d-4a41-9cb3-731297ae0dc3-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3331c1dd-ff2d-4a41-9cb3-731297ae0dc3\") " pod="openstack/nova-cell0-conductor-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.528905 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a983cf0-2c51-4d6a-af53-f115f3a57360-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3a983cf0-2c51-4d6a-af53-f115f3a57360\") " pod="openstack/nova-metadata-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.528948 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3331c1dd-ff2d-4a41-9cb3-731297ae0dc3-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3331c1dd-ff2d-4a41-9cb3-731297ae0dc3\") " pod="openstack/nova-cell0-conductor-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.528981 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a983cf0-2c51-4d6a-af53-f115f3a57360-logs\") pod \"nova-metadata-0\" (UID: \"3a983cf0-2c51-4d6a-af53-f115f3a57360\") " pod="openstack/nova-metadata-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.531001 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a983cf0-2c51-4d6a-af53-f115f3a57360-logs\") pod \"nova-metadata-0\" (UID: \"3a983cf0-2c51-4d6a-af53-f115f3a57360\") " pod="openstack/nova-metadata-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.532506 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a983cf0-2c51-4d6a-af53-f115f3a57360-config-data\") pod \"nova-metadata-0\" (UID: \"3a983cf0-2c51-4d6a-af53-f115f3a57360\") " pod="openstack/nova-metadata-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.539804 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a983cf0-2c51-4d6a-af53-f115f3a57360-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3a983cf0-2c51-4d6a-af53-f115f3a57360\") " pod="openstack/nova-metadata-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.548108 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.554410 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-427mh\" (UniqueName: \"kubernetes.io/projected/3a983cf0-2c51-4d6a-af53-f115f3a57360-kube-api-access-427mh\") pod \"nova-metadata-0\" (UID: \"3a983cf0-2c51-4d6a-af53-f115f3a57360\") " pod="openstack/nova-metadata-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.631220 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3331c1dd-ff2d-4a41-9cb3-731297ae0dc3-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3331c1dd-ff2d-4a41-9cb3-731297ae0dc3\") " pod="openstack/nova-cell0-conductor-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.631381 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp92q\" (UniqueName: \"kubernetes.io/projected/3331c1dd-ff2d-4a41-9cb3-731297ae0dc3-kube-api-access-fp92q\") pod \"nova-cell0-conductor-0\" (UID: \"3331c1dd-ff2d-4a41-9cb3-731297ae0dc3\") " pod="openstack/nova-cell0-conductor-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.631430 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3331c1dd-ff2d-4a41-9cb3-731297ae0dc3-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3331c1dd-ff2d-4a41-9cb3-731297ae0dc3\") " pod="openstack/nova-cell0-conductor-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.637375 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3331c1dd-ff2d-4a41-9cb3-731297ae0dc3-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3331c1dd-ff2d-4a41-9cb3-731297ae0dc3\") " pod="openstack/nova-cell0-conductor-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.641845 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3331c1dd-ff2d-4a41-9cb3-731297ae0dc3-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3331c1dd-ff2d-4a41-9cb3-731297ae0dc3\") " pod="openstack/nova-cell0-conductor-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.658252 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp92q\" (UniqueName: \"kubernetes.io/projected/3331c1dd-ff2d-4a41-9cb3-731297ae0dc3-kube-api-access-fp92q\") pod \"nova-cell0-conductor-0\" (UID: \"3331c1dd-ff2d-4a41-9cb3-731297ae0dc3\") " pod="openstack/nova-cell0-conductor-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.727674 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.849327 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 27 16:41:29 crc kubenswrapper[4772]: I0127 16:41:29.945964 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 27 16:41:30 crc kubenswrapper[4772]: I0127 16:41:30.009799 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qzhn5" podUID="e18a7da1-2037-446c-8646-76917bb9544b" containerName="registry-server" containerID="cri-o://2438505e8dbbe6ceb8683042eba5b21297e5463c63d380b7ca6a75115f1e523f" gracePeriod=2 Jan 27 16:41:30 crc kubenswrapper[4772]: I0127 16:41:30.090614 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 27 16:41:30 crc kubenswrapper[4772]: I0127 16:41:30.244144 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 27 16:41:30 crc kubenswrapper[4772]: W0127 16:41:30.253053 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a983cf0_2c51_4d6a_af53_f115f3a57360.slice/crio-d4be5456d48858af909844028917bca2da1295787038c08194bdc6f50d2b0ce3 WatchSource:0}: Error finding container d4be5456d48858af909844028917bca2da1295787038c08194bdc6f50d2b0ce3: Status 404 returned error can't find the container with id d4be5456d48858af909844028917bca2da1295787038c08194bdc6f50d2b0ce3 Jan 27 16:41:30 crc kubenswrapper[4772]: W0127 16:41:30.406262 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3331c1dd_ff2d_4a41_9cb3_731297ae0dc3.slice/crio-4487a1101aa85dfdee48a3a3947a7279c9954c304d0cf3715f872609665629fc WatchSource:0}: Error finding container 4487a1101aa85dfdee48a3a3947a7279c9954c304d0cf3715f872609665629fc: Status 404 returned error can't find the container with id 4487a1101aa85dfdee48a3a3947a7279c9954c304d0cf3715f872609665629fc Jan 27 16:41:30 crc kubenswrapper[4772]: I0127 16:41:30.409782 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 27 16:41:30 crc kubenswrapper[4772]: I0127 16:41:30.675504 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e" path="/var/lib/kubelet/pods/4b851a7f-5a0c-41fd-9c4a-a5e30c8d389e/volumes" Jan 27 16:41:30 crc kubenswrapper[4772]: I0127 16:41:30.676307 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c652921-712d-46ee-9683-fd6312e33d1e" path="/var/lib/kubelet/pods/8c652921-712d-46ee-9683-fd6312e33d1e/volumes" Jan 27 16:41:30 crc kubenswrapper[4772]: I0127 16:41:30.677065 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a518103b-f26e-4f91-9ca9-93f1f8d5e113" path="/var/lib/kubelet/pods/a518103b-f26e-4f91-9ca9-93f1f8d5e113/volumes" Jan 27 16:41:30 crc kubenswrapper[4772]: I0127 16:41:30.678434 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1c55305-7d18-44cb-90e5-b6793989abda" path="/var/lib/kubelet/pods/e1c55305-7d18-44cb-90e5-b6793989abda/volumes" Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.018804 4772 generic.go:334] "Generic (PLEG): container finished" podID="e18a7da1-2037-446c-8646-76917bb9544b" containerID="2438505e8dbbe6ceb8683042eba5b21297e5463c63d380b7ca6a75115f1e523f" exitCode=0 Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.018887 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzhn5" event={"ID":"e18a7da1-2037-446c-8646-76917bb9544b","Type":"ContainerDied","Data":"2438505e8dbbe6ceb8683042eba5b21297e5463c63d380b7ca6a75115f1e523f"} Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.022027 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3a983cf0-2c51-4d6a-af53-f115f3a57360","Type":"ContainerStarted","Data":"64d6d17b19ec020b3dfb6e4355e68bf1830f515f1e0ec8a1112c49d0c4efca66"} Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.022068 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3a983cf0-2c51-4d6a-af53-f115f3a57360","Type":"ContainerStarted","Data":"d4be5456d48858af909844028917bca2da1295787038c08194bdc6f50d2b0ce3"} Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.023151 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"9549c89c-f55f-484d-80b2-ca1ad19bf758","Type":"ContainerStarted","Data":"b3bbbec2af58eb256af98ed83442ff28dc180942162c5cd5afdf54e6d460aa0e"} Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.023287 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"9549c89c-f55f-484d-80b2-ca1ad19bf758","Type":"ContainerStarted","Data":"4b21d0ea7a82793879133c00776797657ee4dd5c87a48f04228ecfaaf39dbb27"} Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.024745 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.029199 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"3331c1dd-ff2d-4a41-9cb3-731297ae0dc3","Type":"ContainerStarted","Data":"82f4585169c63c6ceec89dc5a4784bbd9d685d5832c552d6bd7983d785cc412c"} Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.029448 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"3331c1dd-ff2d-4a41-9cb3-731297ae0dc3","Type":"ContainerStarted","Data":"4487a1101aa85dfdee48a3a3947a7279c9954c304d0cf3715f872609665629fc"} Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.030398 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.034917 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f178d7e3-af69-4014-8209-5e766a130997","Type":"ContainerStarted","Data":"7a03bd9d50fe00835e95690a3cd780a1c18122b457f26c56197834d3528434c9"} Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.035099 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f178d7e3-af69-4014-8209-5e766a130997","Type":"ContainerStarted","Data":"6434c4d76d62b8e8f47a32fd939db880888344c2fda956f49ede17e91fe6bfd0"} Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.035191 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f178d7e3-af69-4014-8209-5e766a130997","Type":"ContainerStarted","Data":"e41ed01429c113075baaa97bbdf723baf1b7aa1af73c8d2579134878b9e4b420"} Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.039761 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qzhn5" Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.085590 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.085557303 podStartE2EDuration="2.085557303s" podCreationTimestamp="2026-01-27 16:41:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:41:31.050047821 +0000 UTC m=+5677.030656919" watchObservedRunningTime="2026-01-27 16:41:31.085557303 +0000 UTC m=+5677.066166391" Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.136636 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.136613818 podStartE2EDuration="2.136613818s" podCreationTimestamp="2026-01-27 16:41:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:41:31.125425089 +0000 UTC m=+5677.106034197" watchObservedRunningTime="2026-01-27 16:41:31.136613818 +0000 UTC m=+5677.117222916" Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.149946 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.149921668 podStartE2EDuration="3.149921668s" podCreationTimestamp="2026-01-27 16:41:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:41:31.143180035 +0000 UTC m=+5677.123789153" watchObservedRunningTime="2026-01-27 16:41:31.149921668 +0000 UTC m=+5677.130530766" Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.171118 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e18a7da1-2037-446c-8646-76917bb9544b-catalog-content\") pod \"e18a7da1-2037-446c-8646-76917bb9544b\" (UID: \"e18a7da1-2037-446c-8646-76917bb9544b\") " Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.171554 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcwvs\" (UniqueName: \"kubernetes.io/projected/e18a7da1-2037-446c-8646-76917bb9544b-kube-api-access-gcwvs\") pod \"e18a7da1-2037-446c-8646-76917bb9544b\" (UID: \"e18a7da1-2037-446c-8646-76917bb9544b\") " Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.171681 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e18a7da1-2037-446c-8646-76917bb9544b-utilities\") pod \"e18a7da1-2037-446c-8646-76917bb9544b\" (UID: \"e18a7da1-2037-446c-8646-76917bb9544b\") " Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.172444 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e18a7da1-2037-446c-8646-76917bb9544b-utilities" (OuterVolumeSpecName: "utilities") pod "e18a7da1-2037-446c-8646-76917bb9544b" (UID: "e18a7da1-2037-446c-8646-76917bb9544b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.173316 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e18a7da1-2037-446c-8646-76917bb9544b-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.176265 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e18a7da1-2037-446c-8646-76917bb9544b-kube-api-access-gcwvs" (OuterVolumeSpecName: "kube-api-access-gcwvs") pod "e18a7da1-2037-446c-8646-76917bb9544b" (UID: "e18a7da1-2037-446c-8646-76917bb9544b"). InnerVolumeSpecName "kube-api-access-gcwvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.206153 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e18a7da1-2037-446c-8646-76917bb9544b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e18a7da1-2037-446c-8646-76917bb9544b" (UID: "e18a7da1-2037-446c-8646-76917bb9544b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.274711 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e18a7da1-2037-446c-8646-76917bb9544b-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:31 crc kubenswrapper[4772]: I0127 16:41:31.274755 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcwvs\" (UniqueName: \"kubernetes.io/projected/e18a7da1-2037-446c-8646-76917bb9544b-kube-api-access-gcwvs\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:32 crc kubenswrapper[4772]: I0127 16:41:32.056338 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzhn5" event={"ID":"e18a7da1-2037-446c-8646-76917bb9544b","Type":"ContainerDied","Data":"54fcdad16f86f21c30f016c1e9b9c8d1af9bb4af36e051b32118c18fdeb26dbf"} Jan 27 16:41:32 crc kubenswrapper[4772]: I0127 16:41:32.057872 4772 scope.go:117] "RemoveContainer" containerID="2438505e8dbbe6ceb8683042eba5b21297e5463c63d380b7ca6a75115f1e523f" Jan 27 16:41:32 crc kubenswrapper[4772]: I0127 16:41:32.058146 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qzhn5" Jan 27 16:41:32 crc kubenswrapper[4772]: I0127 16:41:32.082778 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3a983cf0-2c51-4d6a-af53-f115f3a57360","Type":"ContainerStarted","Data":"c609b64efb90cd41df1661bc2a9ccff091d7aba93a74339f5dc29950cc106710"} Jan 27 16:41:32 crc kubenswrapper[4772]: I0127 16:41:32.085272 4772 generic.go:334] "Generic (PLEG): container finished" podID="ce725b6a-06bb-4339-819f-fee8819078f0" containerID="33aa342145a07c52aa025581718207e0ed3316f1c48915a16075320a2dbeed5c" exitCode=0 Jan 27 16:41:32 crc kubenswrapper[4772]: I0127 16:41:32.085390 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ce725b6a-06bb-4339-819f-fee8819078f0","Type":"ContainerDied","Data":"33aa342145a07c52aa025581718207e0ed3316f1c48915a16075320a2dbeed5c"} Jan 27 16:41:32 crc kubenswrapper[4772]: I0127 16:41:32.120147 4772 scope.go:117] "RemoveContainer" containerID="93df129a7d9ca49629bdaa9f5049664a8fdbaa8353cdb2fc27b598bded3cdb48" Jan 27 16:41:32 crc kubenswrapper[4772]: I0127 16:41:32.130034 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.129998193 podStartE2EDuration="3.129998193s" podCreationTimestamp="2026-01-27 16:41:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:41:32.117932529 +0000 UTC m=+5678.098541617" watchObservedRunningTime="2026-01-27 16:41:32.129998193 +0000 UTC m=+5678.110607291" Jan 27 16:41:32 crc kubenswrapper[4772]: I0127 16:41:32.141109 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzhn5"] Jan 27 16:41:32 crc kubenswrapper[4772]: I0127 16:41:32.149144 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzhn5"] Jan 27 16:41:32 crc kubenswrapper[4772]: I0127 16:41:32.161442 4772 scope.go:117] "RemoveContainer" containerID="5ba8daff8c2db9de85548929fc9125eac49769ff0aed125f736c35420b1b3caf" Jan 27 16:41:32 crc kubenswrapper[4772]: I0127 16:41:32.300255 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:41:32 crc kubenswrapper[4772]: I0127 16:41:32.564932 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 16:41:32 crc kubenswrapper[4772]: I0127 16:41:32.674665 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e18a7da1-2037-446c-8646-76917bb9544b" path="/var/lib/kubelet/pods/e18a7da1-2037-446c-8646-76917bb9544b/volumes" Jan 27 16:41:32 crc kubenswrapper[4772]: I0127 16:41:32.710328 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ckmf\" (UniqueName: \"kubernetes.io/projected/ce725b6a-06bb-4339-819f-fee8819078f0-kube-api-access-7ckmf\") pod \"ce725b6a-06bb-4339-819f-fee8819078f0\" (UID: \"ce725b6a-06bb-4339-819f-fee8819078f0\") " Jan 27 16:41:32 crc kubenswrapper[4772]: I0127 16:41:32.710552 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce725b6a-06bb-4339-819f-fee8819078f0-combined-ca-bundle\") pod \"ce725b6a-06bb-4339-819f-fee8819078f0\" (UID: \"ce725b6a-06bb-4339-819f-fee8819078f0\") " Jan 27 16:41:32 crc kubenswrapper[4772]: I0127 16:41:32.710604 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce725b6a-06bb-4339-819f-fee8819078f0-config-data\") pod \"ce725b6a-06bb-4339-819f-fee8819078f0\" (UID: \"ce725b6a-06bb-4339-819f-fee8819078f0\") " Jan 27 16:41:32 crc kubenswrapper[4772]: I0127 16:41:32.728929 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce725b6a-06bb-4339-819f-fee8819078f0-kube-api-access-7ckmf" (OuterVolumeSpecName: "kube-api-access-7ckmf") pod "ce725b6a-06bb-4339-819f-fee8819078f0" (UID: "ce725b6a-06bb-4339-819f-fee8819078f0"). InnerVolumeSpecName "kube-api-access-7ckmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:41:32 crc kubenswrapper[4772]: I0127 16:41:32.735005 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce725b6a-06bb-4339-819f-fee8819078f0-config-data" (OuterVolumeSpecName: "config-data") pod "ce725b6a-06bb-4339-819f-fee8819078f0" (UID: "ce725b6a-06bb-4339-819f-fee8819078f0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:41:32 crc kubenswrapper[4772]: I0127 16:41:32.736825 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce725b6a-06bb-4339-819f-fee8819078f0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce725b6a-06bb-4339-819f-fee8819078f0" (UID: "ce725b6a-06bb-4339-819f-fee8819078f0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:41:32 crc kubenswrapper[4772]: I0127 16:41:32.813690 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce725b6a-06bb-4339-819f-fee8819078f0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:32 crc kubenswrapper[4772]: I0127 16:41:32.813737 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce725b6a-06bb-4339-819f-fee8819078f0-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:32 crc kubenswrapper[4772]: I0127 16:41:32.813751 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ckmf\" (UniqueName: \"kubernetes.io/projected/ce725b6a-06bb-4339-819f-fee8819078f0-kube-api-access-7ckmf\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.097031 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.103030 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ce725b6a-06bb-4339-819f-fee8819078f0","Type":"ContainerDied","Data":"c5ffffe294d87ab27553633999a0d84408edaf82429449f00a94658621086fd0"} Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.103109 4772 scope.go:117] "RemoveContainer" containerID="33aa342145a07c52aa025581718207e0ed3316f1c48915a16075320a2dbeed5c" Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.151698 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.175138 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.184705 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 16:41:33 crc kubenswrapper[4772]: E0127 16:41:33.185576 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e18a7da1-2037-446c-8646-76917bb9544b" containerName="registry-server" Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.185599 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e18a7da1-2037-446c-8646-76917bb9544b" containerName="registry-server" Jan 27 16:41:33 crc kubenswrapper[4772]: E0127 16:41:33.185644 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e18a7da1-2037-446c-8646-76917bb9544b" containerName="extract-content" Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.185652 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e18a7da1-2037-446c-8646-76917bb9544b" containerName="extract-content" Jan 27 16:41:33 crc kubenswrapper[4772]: E0127 16:41:33.185678 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce725b6a-06bb-4339-819f-fee8819078f0" containerName="nova-scheduler-scheduler" Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.185688 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce725b6a-06bb-4339-819f-fee8819078f0" containerName="nova-scheduler-scheduler" Jan 27 16:41:33 crc kubenswrapper[4772]: E0127 16:41:33.185702 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e18a7da1-2037-446c-8646-76917bb9544b" containerName="extract-utilities" Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.185710 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e18a7da1-2037-446c-8646-76917bb9544b" containerName="extract-utilities" Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.186007 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="e18a7da1-2037-446c-8646-76917bb9544b" containerName="registry-server" Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.186030 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce725b6a-06bb-4339-819f-fee8819078f0" containerName="nova-scheduler-scheduler" Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.188055 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.193692 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.195968 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.324239 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85lz2\" (UniqueName: \"kubernetes.io/projected/c9832ca2-4d35-4533-bdb3-7ac3773e5242-kube-api-access-85lz2\") pod \"nova-scheduler-0\" (UID: \"c9832ca2-4d35-4533-bdb3-7ac3773e5242\") " pod="openstack/nova-scheduler-0" Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.324302 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9832ca2-4d35-4533-bdb3-7ac3773e5242-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c9832ca2-4d35-4533-bdb3-7ac3773e5242\") " pod="openstack/nova-scheduler-0" Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.324324 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9832ca2-4d35-4533-bdb3-7ac3773e5242-config-data\") pod \"nova-scheduler-0\" (UID: \"c9832ca2-4d35-4533-bdb3-7ac3773e5242\") " pod="openstack/nova-scheduler-0" Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.425779 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85lz2\" (UniqueName: \"kubernetes.io/projected/c9832ca2-4d35-4533-bdb3-7ac3773e5242-kube-api-access-85lz2\") pod \"nova-scheduler-0\" (UID: \"c9832ca2-4d35-4533-bdb3-7ac3773e5242\") " pod="openstack/nova-scheduler-0" Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.426470 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9832ca2-4d35-4533-bdb3-7ac3773e5242-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c9832ca2-4d35-4533-bdb3-7ac3773e5242\") " pod="openstack/nova-scheduler-0" Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.427425 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9832ca2-4d35-4533-bdb3-7ac3773e5242-config-data\") pod \"nova-scheduler-0\" (UID: \"c9832ca2-4d35-4533-bdb3-7ac3773e5242\") " pod="openstack/nova-scheduler-0" Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.431005 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9832ca2-4d35-4533-bdb3-7ac3773e5242-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c9832ca2-4d35-4533-bdb3-7ac3773e5242\") " pod="openstack/nova-scheduler-0" Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.431373 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9832ca2-4d35-4533-bdb3-7ac3773e5242-config-data\") pod \"nova-scheduler-0\" (UID: \"c9832ca2-4d35-4533-bdb3-7ac3773e5242\") " pod="openstack/nova-scheduler-0" Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.448209 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85lz2\" (UniqueName: \"kubernetes.io/projected/c9832ca2-4d35-4533-bdb3-7ac3773e5242-kube-api-access-85lz2\") pod \"nova-scheduler-0\" (UID: \"c9832ca2-4d35-4533-bdb3-7ac3773e5242\") " pod="openstack/nova-scheduler-0" Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.509836 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 27 16:41:33 crc kubenswrapper[4772]: I0127 16:41:33.836927 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 27 16:41:34 crc kubenswrapper[4772]: I0127 16:41:34.109390 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c9832ca2-4d35-4533-bdb3-7ac3773e5242","Type":"ContainerStarted","Data":"65a9e9b4d2896ed39525d206b510fca660ca26b4570c062a20320a6ceaa82ab3"} Jan 27 16:41:34 crc kubenswrapper[4772]: I0127 16:41:34.109673 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c9832ca2-4d35-4533-bdb3-7ac3773e5242","Type":"ContainerStarted","Data":"8c64ae7d9f465a3928274a6bfabff3ff3e174b69628594eef3cbb9e8a9068613"} Jan 27 16:41:34 crc kubenswrapper[4772]: I0127 16:41:34.673599 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce725b6a-06bb-4339-819f-fee8819078f0" path="/var/lib/kubelet/pods/ce725b6a-06bb-4339-819f-fee8819078f0/volumes" Jan 27 16:41:34 crc kubenswrapper[4772]: I0127 16:41:34.728660 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 27 16:41:34 crc kubenswrapper[4772]: I0127 16:41:34.728818 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 27 16:41:35 crc kubenswrapper[4772]: I0127 16:41:35.143373 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.143352506 podStartE2EDuration="2.143352506s" podCreationTimestamp="2026-01-27 16:41:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:41:35.138715513 +0000 UTC m=+5681.119324641" watchObservedRunningTime="2026-01-27 16:41:35.143352506 +0000 UTC m=+5681.123961604" Jan 27 16:41:37 crc kubenswrapper[4772]: I0127 16:41:37.302464 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:41:37 crc kubenswrapper[4772]: I0127 16:41:37.314133 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:41:38 crc kubenswrapper[4772]: I0127 16:41:38.161368 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Jan 27 16:41:38 crc kubenswrapper[4772]: I0127 16:41:38.511283 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 27 16:41:39 crc kubenswrapper[4772]: I0127 16:41:39.469779 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 27 16:41:39 crc kubenswrapper[4772]: I0127 16:41:39.470375 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 27 16:41:39 crc kubenswrapper[4772]: I0127 16:41:39.588595 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Jan 27 16:41:39 crc kubenswrapper[4772]: I0127 16:41:39.727768 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 27 16:41:39 crc kubenswrapper[4772]: I0127 16:41:39.728002 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 27 16:41:39 crc kubenswrapper[4772]: I0127 16:41:39.878697 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Jan 27 16:41:40 crc kubenswrapper[4772]: I0127 16:41:40.551426 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f178d7e3-af69-4014-8209-5e766a130997" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.75:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 27 16:41:40 crc kubenswrapper[4772]: I0127 16:41:40.551715 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f178d7e3-af69-4014-8209-5e766a130997" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.75:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 27 16:41:40 crc kubenswrapper[4772]: I0127 16:41:40.811393 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3a983cf0-2c51-4d6a-af53-f115f3a57360" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.77:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 27 16:41:40 crc kubenswrapper[4772]: I0127 16:41:40.811421 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3a983cf0-2c51-4d6a-af53-f115f3a57360" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.77:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 27 16:41:42 crc kubenswrapper[4772]: I0127 16:41:42.058448 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:41:42 crc kubenswrapper[4772]: I0127 16:41:42.058748 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:41:42 crc kubenswrapper[4772]: I0127 16:41:42.058791 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 16:41:42 crc kubenswrapper[4772]: I0127 16:41:42.059529 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 16:41:42 crc kubenswrapper[4772]: I0127 16:41:42.059588 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" gracePeriod=600 Jan 27 16:41:42 crc kubenswrapper[4772]: E0127 16:41:42.189422 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.206600 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" exitCode=0 Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.207266 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69"} Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.207354 4772 scope.go:117] "RemoveContainer" containerID="90e27c06727cf113f54cd7c0344565bfa447b15cc343fc7033a04f41dddb22f9" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.212621 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:41:43 crc kubenswrapper[4772]: E0127 16:41:43.213715 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.231314 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.232846 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.238640 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.267284 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.312393 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.312467 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-config-data\") pod \"cinder-scheduler-0\" (UID: \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.312570 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zw7z\" (UniqueName: \"kubernetes.io/projected/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-kube-api-access-8zw7z\") pod \"cinder-scheduler-0\" (UID: \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.312596 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-scripts\") pod \"cinder-scheduler-0\" (UID: \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.312612 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.312649 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.414034 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zw7z\" (UniqueName: \"kubernetes.io/projected/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-kube-api-access-8zw7z\") pod \"cinder-scheduler-0\" (UID: \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.414094 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-scripts\") pod \"cinder-scheduler-0\" (UID: \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.414112 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.414146 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.414214 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.414250 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-config-data\") pod \"cinder-scheduler-0\" (UID: \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.414289 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.420956 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.421407 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-scripts\") pod \"cinder-scheduler-0\" (UID: \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.421421 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.422947 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-config-data\") pod \"cinder-scheduler-0\" (UID: \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.435850 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zw7z\" (UniqueName: \"kubernetes.io/projected/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-kube-api-access-8zw7z\") pod \"cinder-scheduler-0\" (UID: \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.511339 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.540653 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Jan 27 16:41:43 crc kubenswrapper[4772]: I0127 16:41:43.552897 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 27 16:41:44 crc kubenswrapper[4772]: W0127 16:41:44.200370 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb6ca922_71e5_4fa3_ae7b_5137b0e58397.slice/crio-9c8e4b7251239d087f479e7cdf47b4e5ff69599faa90d12e66a25ebf0083f84b WatchSource:0}: Error finding container 9c8e4b7251239d087f479e7cdf47b4e5ff69599faa90d12e66a25ebf0083f84b: Status 404 returned error can't find the container with id 9c8e4b7251239d087f479e7cdf47b4e5ff69599faa90d12e66a25ebf0083f84b Jan 27 16:41:44 crc kubenswrapper[4772]: I0127 16:41:44.200506 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 27 16:41:44 crc kubenswrapper[4772]: I0127 16:41:44.219576 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb6ca922-71e5-4fa3-ae7b-5137b0e58397","Type":"ContainerStarted","Data":"9c8e4b7251239d087f479e7cdf47b4e5ff69599faa90d12e66a25ebf0083f84b"} Jan 27 16:41:44 crc kubenswrapper[4772]: I0127 16:41:44.252527 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Jan 27 16:41:44 crc kubenswrapper[4772]: I0127 16:41:44.725155 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Jan 27 16:41:44 crc kubenswrapper[4772]: I0127 16:41:44.725797 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="1d692e3b-b4e1-4af1-8cb1-a64a6e51916f" containerName="cinder-api-log" containerID="cri-o://714b9b6c5c6438e8d52919984e494ff614fd56b03d77fad32ca28a5a24726e46" gracePeriod=30 Jan 27 16:41:44 crc kubenswrapper[4772]: I0127 16:41:44.725881 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="1d692e3b-b4e1-4af1-8cb1-a64a6e51916f" containerName="cinder-api" containerID="cri-o://31e3d55d1d113b922a3f84db51ffbe33b9ce1af8297930f881b9d6d32892374f" gracePeriod=30 Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.236159 4772 generic.go:334] "Generic (PLEG): container finished" podID="1d692e3b-b4e1-4af1-8cb1-a64a6e51916f" containerID="714b9b6c5c6438e8d52919984e494ff614fd56b03d77fad32ca28a5a24726e46" exitCode=143 Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.236489 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f","Type":"ContainerDied","Data":"714b9b6c5c6438e8d52919984e494ff614fd56b03d77fad32ca28a5a24726e46"} Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.240969 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb6ca922-71e5-4fa3-ae7b-5137b0e58397","Type":"ContainerStarted","Data":"be1b9f8d0c257a0e1d0419cf2c1278df0e274331eb3adeef663b50de9e49f2f3"} Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.262238 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.264239 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.268757 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.305662 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.354564 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-dev\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.354616 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.354652 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.354695 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.354749 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.354787 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.354818 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.354848 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.354870 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.354896 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-sys\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.354925 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-run\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.355036 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.355065 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.355087 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55dzs\" (UniqueName: \"kubernetes.io/projected/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-kube-api-access-55dzs\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.355111 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.355134 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457201 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457267 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457303 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457330 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457337 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457397 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457438 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-sys\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457473 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-run\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457474 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457499 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457518 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-run\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457536 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-sys\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457580 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457611 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457614 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457631 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55dzs\" (UniqueName: \"kubernetes.io/projected/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-kube-api-access-55dzs\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457660 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457681 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457708 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-dev\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457710 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457731 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457764 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457789 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457788 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457917 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-dev\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.457931 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.463525 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.465749 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.466148 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.470871 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.478717 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.481723 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55dzs\" (UniqueName: \"kubernetes.io/projected/7b6810b2-bc50-486d-9a87-cf4cd50d33c5-kube-api-access-55dzs\") pod \"cinder-volume-volume1-0\" (UID: \"7b6810b2-bc50-486d-9a87-cf4cd50d33c5\") " pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.589874 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.792006 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.795416 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.797958 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.811723 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.869514 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.869576 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/37170132-cd9f-44e7-827d-b98486cefb39-ceph\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.869635 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86z5c\" (UniqueName: \"kubernetes.io/projected/37170132-cd9f-44e7-827d-b98486cefb39-kube-api-access-86z5c\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.869678 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-dev\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.869710 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37170132-cd9f-44e7-827d-b98486cefb39-scripts\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.869741 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-lib-modules\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.869768 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-etc-nvme\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.869802 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.869851 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-sys\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.869898 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-run\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.869925 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37170132-cd9f-44e7-827d-b98486cefb39-config-data-custom\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.869960 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.870006 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.870030 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37170132-cd9f-44e7-827d-b98486cefb39-config-data\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.870074 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.870114 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37170132-cd9f-44e7-827d-b98486cefb39-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.971653 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.971712 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/37170132-cd9f-44e7-827d-b98486cefb39-ceph\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.971739 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86z5c\" (UniqueName: \"kubernetes.io/projected/37170132-cd9f-44e7-827d-b98486cefb39-kube-api-access-86z5c\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.971769 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-dev\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.971795 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37170132-cd9f-44e7-827d-b98486cefb39-scripts\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.971834 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-lib-modules\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.971854 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-etc-nvme\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.971886 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.971932 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-sys\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.971968 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-run\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.971990 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37170132-cd9f-44e7-827d-b98486cefb39-config-data-custom\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.972022 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.972063 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.972084 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37170132-cd9f-44e7-827d-b98486cefb39-config-data\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.972104 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.972130 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37170132-cd9f-44e7-827d-b98486cefb39-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.972269 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.972314 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-sys\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.972316 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.972341 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-run\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.972379 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-dev\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.972403 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.972679 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.972731 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-lib-modules\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.972764 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.972879 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/37170132-cd9f-44e7-827d-b98486cefb39-etc-nvme\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.976754 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/37170132-cd9f-44e7-827d-b98486cefb39-ceph\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.977621 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37170132-cd9f-44e7-827d-b98486cefb39-scripts\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.977818 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37170132-cd9f-44e7-827d-b98486cefb39-config-data-custom\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.992485 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37170132-cd9f-44e7-827d-b98486cefb39-config-data\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.994386 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37170132-cd9f-44e7-827d-b98486cefb39-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:45 crc kubenswrapper[4772]: I0127 16:41:45.994632 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86z5c\" (UniqueName: \"kubernetes.io/projected/37170132-cd9f-44e7-827d-b98486cefb39-kube-api-access-86z5c\") pod \"cinder-backup-0\" (UID: \"37170132-cd9f-44e7-827d-b98486cefb39\") " pod="openstack/cinder-backup-0" Jan 27 16:41:46 crc kubenswrapper[4772]: I0127 16:41:46.128232 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Jan 27 16:41:46 crc kubenswrapper[4772]: I0127 16:41:46.204501 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Jan 27 16:41:46 crc kubenswrapper[4772]: W0127 16:41:46.213304 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b6810b2_bc50_486d_9a87_cf4cd50d33c5.slice/crio-f6537ff34af5a48afbc027cf852a989774ed686a8c36aef611b39b70b4d39635 WatchSource:0}: Error finding container f6537ff34af5a48afbc027cf852a989774ed686a8c36aef611b39b70b4d39635: Status 404 returned error can't find the container with id f6537ff34af5a48afbc027cf852a989774ed686a8c36aef611b39b70b4d39635 Jan 27 16:41:46 crc kubenswrapper[4772]: I0127 16:41:46.266008 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb6ca922-71e5-4fa3-ae7b-5137b0e58397","Type":"ContainerStarted","Data":"ece33e666a3e58ee419da4cf74156d864bda223101f1f560e453f6c6c89635ef"} Jan 27 16:41:46 crc kubenswrapper[4772]: I0127 16:41:46.277433 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"7b6810b2-bc50-486d-9a87-cf4cd50d33c5","Type":"ContainerStarted","Data":"f6537ff34af5a48afbc027cf852a989774ed686a8c36aef611b39b70b4d39635"} Jan 27 16:41:46 crc kubenswrapper[4772]: I0127 16:41:46.294541 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.2944987550000002 podStartE2EDuration="3.294498755s" podCreationTimestamp="2026-01-27 16:41:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:41:46.287590048 +0000 UTC m=+5692.268199146" watchObservedRunningTime="2026-01-27 16:41:46.294498755 +0000 UTC m=+5692.275107853" Jan 27 16:41:46 crc kubenswrapper[4772]: I0127 16:41:46.691388 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Jan 27 16:41:47 crc kubenswrapper[4772]: I0127 16:41:47.299357 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"37170132-cd9f-44e7-827d-b98486cefb39","Type":"ContainerStarted","Data":"22a8fa900207d0302c8abf2e457cf9376b2ee0a0d9600c94dbfb5249680a1692"} Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.316691 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"37170132-cd9f-44e7-827d-b98486cefb39","Type":"ContainerStarted","Data":"dc6e30a448451b97f1266ec3d9ad50ad6ae4ebb2ac72cb94a8a9cb432bf1f514"} Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.317621 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"37170132-cd9f-44e7-827d-b98486cefb39","Type":"ContainerStarted","Data":"a2c049f67758489a77e066df2431ee96e2428bbd04db38037a834c967aeda936"} Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.319408 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"7b6810b2-bc50-486d-9a87-cf4cd50d33c5","Type":"ContainerStarted","Data":"07fb537d1a94745b9dd392726b4097426196d8f9517eb850bd68fd5c08583620"} Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.319451 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"7b6810b2-bc50-486d-9a87-cf4cd50d33c5","Type":"ContainerStarted","Data":"356a72ff887fc4a895e6fe05bc640c45244dd9e2d55a5f7750e045dc0184530f"} Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.324233 4772 generic.go:334] "Generic (PLEG): container finished" podID="1d692e3b-b4e1-4af1-8cb1-a64a6e51916f" containerID="31e3d55d1d113b922a3f84db51ffbe33b9ce1af8297930f881b9d6d32892374f" exitCode=0 Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.324281 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f","Type":"ContainerDied","Data":"31e3d55d1d113b922a3f84db51ffbe33b9ce1af8297930f881b9d6d32892374f"} Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.349782 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.54014595 podStartE2EDuration="3.349737807s" podCreationTimestamp="2026-01-27 16:41:45 +0000 UTC" firstStartedPulling="2026-01-27 16:41:46.714420344 +0000 UTC m=+5692.695029432" lastFinishedPulling="2026-01-27 16:41:47.524012191 +0000 UTC m=+5693.504621289" observedRunningTime="2026-01-27 16:41:48.344451337 +0000 UTC m=+5694.325060425" watchObservedRunningTime="2026-01-27 16:41:48.349737807 +0000 UTC m=+5694.330346905" Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.389912 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=2.443403814 podStartE2EDuration="3.389882692s" podCreationTimestamp="2026-01-27 16:41:45 +0000 UTC" firstStartedPulling="2026-01-27 16:41:46.215672518 +0000 UTC m=+5692.196281616" lastFinishedPulling="2026-01-27 16:41:47.162151396 +0000 UTC m=+5693.142760494" observedRunningTime="2026-01-27 16:41:48.382955414 +0000 UTC m=+5694.363564522" watchObservedRunningTime="2026-01-27 16:41:48.389882692 +0000 UTC m=+5694.370491790" Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.430742 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.539778 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-config-data\") pod \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.540155 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-etc-machine-id\") pod \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.540234 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-logs\") pod \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.541063 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-combined-ca-bundle\") pod \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.541312 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1d692e3b-b4e1-4af1-8cb1-a64a6e51916f" (UID: "1d692e3b-b4e1-4af1-8cb1-a64a6e51916f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.541665 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-config-data-custom\") pod \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.541772 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rf7qv\" (UniqueName: \"kubernetes.io/projected/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-kube-api-access-rf7qv\") pod \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.541938 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-scripts\") pod \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\" (UID: \"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f\") " Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.542203 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-logs" (OuterVolumeSpecName: "logs") pod "1d692e3b-b4e1-4af1-8cb1-a64a6e51916f" (UID: "1d692e3b-b4e1-4af1-8cb1-a64a6e51916f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.543019 4772 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.543035 4772 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-logs\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.546511 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-scripts" (OuterVolumeSpecName: "scripts") pod "1d692e3b-b4e1-4af1-8cb1-a64a6e51916f" (UID: "1d692e3b-b4e1-4af1-8cb1-a64a6e51916f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.546967 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-kube-api-access-rf7qv" (OuterVolumeSpecName: "kube-api-access-rf7qv") pod "1d692e3b-b4e1-4af1-8cb1-a64a6e51916f" (UID: "1d692e3b-b4e1-4af1-8cb1-a64a6e51916f"). InnerVolumeSpecName "kube-api-access-rf7qv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.547598 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1d692e3b-b4e1-4af1-8cb1-a64a6e51916f" (UID: "1d692e3b-b4e1-4af1-8cb1-a64a6e51916f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.554735 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.615888 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d692e3b-b4e1-4af1-8cb1-a64a6e51916f" (UID: "1d692e3b-b4e1-4af1-8cb1-a64a6e51916f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.643256 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-config-data" (OuterVolumeSpecName: "config-data") pod "1d692e3b-b4e1-4af1-8cb1-a64a6e51916f" (UID: "1d692e3b-b4e1-4af1-8cb1-a64a6e51916f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.645456 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rf7qv\" (UniqueName: \"kubernetes.io/projected/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-kube-api-access-rf7qv\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.645493 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.645508 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.645522 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:48 crc kubenswrapper[4772]: I0127 16:41:48.645535 4772 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.339760 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.339763 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1d692e3b-b4e1-4af1-8cb1-a64a6e51916f","Type":"ContainerDied","Data":"1985a43f384e22d29818a239c3a060f96131fc47f07c59ec7571d217b8454dc4"} Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.340727 4772 scope.go:117] "RemoveContainer" containerID="31e3d55d1d113b922a3f84db51ffbe33b9ce1af8297930f881b9d6d32892374f" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.399087 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.404541 4772 scope.go:117] "RemoveContainer" containerID="714b9b6c5c6438e8d52919984e494ff614fd56b03d77fad32ca28a5a24726e46" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.409794 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.420015 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Jan 27 16:41:49 crc kubenswrapper[4772]: E0127 16:41:49.420584 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d692e3b-b4e1-4af1-8cb1-a64a6e51916f" containerName="cinder-api" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.420614 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d692e3b-b4e1-4af1-8cb1-a64a6e51916f" containerName="cinder-api" Jan 27 16:41:49 crc kubenswrapper[4772]: E0127 16:41:49.420635 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d692e3b-b4e1-4af1-8cb1-a64a6e51916f" containerName="cinder-api-log" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.420646 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d692e3b-b4e1-4af1-8cb1-a64a6e51916f" containerName="cinder-api-log" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.420865 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d692e3b-b4e1-4af1-8cb1-a64a6e51916f" containerName="cinder-api-log" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.420884 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d692e3b-b4e1-4af1-8cb1-a64a6e51916f" containerName="cinder-api" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.422104 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.428480 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.428521 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.473653 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.473978 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.476043 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.480332 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.560475 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a46febaf-97b6-4ed3-8958-316e2a542a5f-scripts\") pod \"cinder-api-0\" (UID: \"a46febaf-97b6-4ed3-8958-316e2a542a5f\") " pod="openstack/cinder-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.561579 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlbg8\" (UniqueName: \"kubernetes.io/projected/a46febaf-97b6-4ed3-8958-316e2a542a5f-kube-api-access-wlbg8\") pod \"cinder-api-0\" (UID: \"a46febaf-97b6-4ed3-8958-316e2a542a5f\") " pod="openstack/cinder-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.561827 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a46febaf-97b6-4ed3-8958-316e2a542a5f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a46febaf-97b6-4ed3-8958-316e2a542a5f\") " pod="openstack/cinder-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.561935 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a46febaf-97b6-4ed3-8958-316e2a542a5f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a46febaf-97b6-4ed3-8958-316e2a542a5f\") " pod="openstack/cinder-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.562016 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a46febaf-97b6-4ed3-8958-316e2a542a5f-logs\") pod \"cinder-api-0\" (UID: \"a46febaf-97b6-4ed3-8958-316e2a542a5f\") " pod="openstack/cinder-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.562130 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a46febaf-97b6-4ed3-8958-316e2a542a5f-config-data-custom\") pod \"cinder-api-0\" (UID: \"a46febaf-97b6-4ed3-8958-316e2a542a5f\") " pod="openstack/cinder-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.562255 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a46febaf-97b6-4ed3-8958-316e2a542a5f-config-data\") pod \"cinder-api-0\" (UID: \"a46febaf-97b6-4ed3-8958-316e2a542a5f\") " pod="openstack/cinder-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.663248 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlbg8\" (UniqueName: \"kubernetes.io/projected/a46febaf-97b6-4ed3-8958-316e2a542a5f-kube-api-access-wlbg8\") pod \"cinder-api-0\" (UID: \"a46febaf-97b6-4ed3-8958-316e2a542a5f\") " pod="openstack/cinder-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.663637 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a46febaf-97b6-4ed3-8958-316e2a542a5f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a46febaf-97b6-4ed3-8958-316e2a542a5f\") " pod="openstack/cinder-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.663664 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a46febaf-97b6-4ed3-8958-316e2a542a5f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a46febaf-97b6-4ed3-8958-316e2a542a5f\") " pod="openstack/cinder-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.663683 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a46febaf-97b6-4ed3-8958-316e2a542a5f-logs\") pod \"cinder-api-0\" (UID: \"a46febaf-97b6-4ed3-8958-316e2a542a5f\") " pod="openstack/cinder-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.663714 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a46febaf-97b6-4ed3-8958-316e2a542a5f-config-data-custom\") pod \"cinder-api-0\" (UID: \"a46febaf-97b6-4ed3-8958-316e2a542a5f\") " pod="openstack/cinder-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.663740 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a46febaf-97b6-4ed3-8958-316e2a542a5f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a46febaf-97b6-4ed3-8958-316e2a542a5f\") " pod="openstack/cinder-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.663747 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a46febaf-97b6-4ed3-8958-316e2a542a5f-config-data\") pod \"cinder-api-0\" (UID: \"a46febaf-97b6-4ed3-8958-316e2a542a5f\") " pod="openstack/cinder-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.663910 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a46febaf-97b6-4ed3-8958-316e2a542a5f-scripts\") pod \"cinder-api-0\" (UID: \"a46febaf-97b6-4ed3-8958-316e2a542a5f\") " pod="openstack/cinder-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.664807 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a46febaf-97b6-4ed3-8958-316e2a542a5f-logs\") pod \"cinder-api-0\" (UID: \"a46febaf-97b6-4ed3-8958-316e2a542a5f\") " pod="openstack/cinder-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.669602 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a46febaf-97b6-4ed3-8958-316e2a542a5f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a46febaf-97b6-4ed3-8958-316e2a542a5f\") " pod="openstack/cinder-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.671010 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a46febaf-97b6-4ed3-8958-316e2a542a5f-config-data-custom\") pod \"cinder-api-0\" (UID: \"a46febaf-97b6-4ed3-8958-316e2a542a5f\") " pod="openstack/cinder-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.671144 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a46febaf-97b6-4ed3-8958-316e2a542a5f-config-data\") pod \"cinder-api-0\" (UID: \"a46febaf-97b6-4ed3-8958-316e2a542a5f\") " pod="openstack/cinder-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.680933 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a46febaf-97b6-4ed3-8958-316e2a542a5f-scripts\") pod \"cinder-api-0\" (UID: \"a46febaf-97b6-4ed3-8958-316e2a542a5f\") " pod="openstack/cinder-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.683814 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlbg8\" (UniqueName: \"kubernetes.io/projected/a46febaf-97b6-4ed3-8958-316e2a542a5f-kube-api-access-wlbg8\") pod \"cinder-api-0\" (UID: \"a46febaf-97b6-4ed3-8958-316e2a542a5f\") " pod="openstack/cinder-api-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.730663 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.731474 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.732530 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 27 16:41:49 crc kubenswrapper[4772]: I0127 16:41:49.752559 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 27 16:41:50 crc kubenswrapper[4772]: I0127 16:41:50.207766 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 27 16:41:50 crc kubenswrapper[4772]: I0127 16:41:50.352733 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a46febaf-97b6-4ed3-8958-316e2a542a5f","Type":"ContainerStarted","Data":"287b252ce68976793728fff56a71a7cca1bfb0340ced44587ebaf86607e9529a"} Jan 27 16:41:50 crc kubenswrapper[4772]: I0127 16:41:50.353653 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 27 16:41:50 crc kubenswrapper[4772]: I0127 16:41:50.355589 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 27 16:41:50 crc kubenswrapper[4772]: I0127 16:41:50.357433 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 27 16:41:50 crc kubenswrapper[4772]: I0127 16:41:50.593389 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:50 crc kubenswrapper[4772]: I0127 16:41:50.681011 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d692e3b-b4e1-4af1-8cb1-a64a6e51916f" path="/var/lib/kubelet/pods/1d692e3b-b4e1-4af1-8cb1-a64a6e51916f/volumes" Jan 27 16:41:51 crc kubenswrapper[4772]: I0127 16:41:51.129553 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Jan 27 16:41:51 crc kubenswrapper[4772]: I0127 16:41:51.367801 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a46febaf-97b6-4ed3-8958-316e2a542a5f","Type":"ContainerStarted","Data":"d7165bee75b99f5d5e7f1a0cfe3be5a235bea586658dcfb51148e8422a49f2ba"} Jan 27 16:41:52 crc kubenswrapper[4772]: I0127 16:41:52.378123 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a46febaf-97b6-4ed3-8958-316e2a542a5f","Type":"ContainerStarted","Data":"bebfac65418099316deacbc1c27d9c9d4cefb5cd8c9460548100c4cee37449e7"} Jan 27 16:41:52 crc kubenswrapper[4772]: I0127 16:41:52.402002 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.4019864220000002 podStartE2EDuration="3.401986422s" podCreationTimestamp="2026-01-27 16:41:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:41:52.39911032 +0000 UTC m=+5698.379719438" watchObservedRunningTime="2026-01-27 16:41:52.401986422 +0000 UTC m=+5698.382595520" Jan 27 16:41:53 crc kubenswrapper[4772]: I0127 16:41:53.385642 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Jan 27 16:41:53 crc kubenswrapper[4772]: I0127 16:41:53.783740 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Jan 27 16:41:53 crc kubenswrapper[4772]: I0127 16:41:53.841398 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 27 16:41:54 crc kubenswrapper[4772]: I0127 16:41:54.399380 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="cb6ca922-71e5-4fa3-ae7b-5137b0e58397" containerName="cinder-scheduler" containerID="cri-o://be1b9f8d0c257a0e1d0419cf2c1278df0e274331eb3adeef663b50de9e49f2f3" gracePeriod=30 Jan 27 16:41:54 crc kubenswrapper[4772]: I0127 16:41:54.400485 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="cb6ca922-71e5-4fa3-ae7b-5137b0e58397" containerName="probe" containerID="cri-o://ece33e666a3e58ee419da4cf74156d864bda223101f1f560e453f6c6c89635ef" gracePeriod=30 Jan 27 16:41:54 crc kubenswrapper[4772]: I0127 16:41:54.682976 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:41:54 crc kubenswrapper[4772]: E0127 16:41:54.683440 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:41:55 crc kubenswrapper[4772]: I0127 16:41:55.415337 4772 generic.go:334] "Generic (PLEG): container finished" podID="cb6ca922-71e5-4fa3-ae7b-5137b0e58397" containerID="ece33e666a3e58ee419da4cf74156d864bda223101f1f560e453f6c6c89635ef" exitCode=0 Jan 27 16:41:55 crc kubenswrapper[4772]: I0127 16:41:55.415458 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb6ca922-71e5-4fa3-ae7b-5137b0e58397","Type":"ContainerDied","Data":"ece33e666a3e58ee419da4cf74156d864bda223101f1f560e453f6c6c89635ef"} Jan 27 16:41:55 crc kubenswrapper[4772]: I0127 16:41:55.785241 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Jan 27 16:41:56 crc kubenswrapper[4772]: I0127 16:41:56.349334 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Jan 27 16:41:56 crc kubenswrapper[4772]: I0127 16:41:56.971248 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.114219 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-config-data\") pod \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\" (UID: \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\") " Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.114305 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-config-data-custom\") pod \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\" (UID: \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\") " Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.114362 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zw7z\" (UniqueName: \"kubernetes.io/projected/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-kube-api-access-8zw7z\") pod \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\" (UID: \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\") " Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.114407 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-combined-ca-bundle\") pod \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\" (UID: \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\") " Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.114430 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-etc-machine-id\") pod \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\" (UID: \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\") " Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.114545 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-scripts\") pod \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\" (UID: \"cb6ca922-71e5-4fa3-ae7b-5137b0e58397\") " Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.114665 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "cb6ca922-71e5-4fa3-ae7b-5137b0e58397" (UID: "cb6ca922-71e5-4fa3-ae7b-5137b0e58397"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.115103 4772 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.123430 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-kube-api-access-8zw7z" (OuterVolumeSpecName: "kube-api-access-8zw7z") pod "cb6ca922-71e5-4fa3-ae7b-5137b0e58397" (UID: "cb6ca922-71e5-4fa3-ae7b-5137b0e58397"). InnerVolumeSpecName "kube-api-access-8zw7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.123465 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-scripts" (OuterVolumeSpecName: "scripts") pod "cb6ca922-71e5-4fa3-ae7b-5137b0e58397" (UID: "cb6ca922-71e5-4fa3-ae7b-5137b0e58397"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.123430 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cb6ca922-71e5-4fa3-ae7b-5137b0e58397" (UID: "cb6ca922-71e5-4fa3-ae7b-5137b0e58397"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.165900 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb6ca922-71e5-4fa3-ae7b-5137b0e58397" (UID: "cb6ca922-71e5-4fa3-ae7b-5137b0e58397"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.216604 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.216635 4772 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.216646 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zw7z\" (UniqueName: \"kubernetes.io/projected/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-kube-api-access-8zw7z\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.216659 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.228813 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-config-data" (OuterVolumeSpecName: "config-data") pod "cb6ca922-71e5-4fa3-ae7b-5137b0e58397" (UID: "cb6ca922-71e5-4fa3-ae7b-5137b0e58397"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.318059 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb6ca922-71e5-4fa3-ae7b-5137b0e58397-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.437493 4772 generic.go:334] "Generic (PLEG): container finished" podID="cb6ca922-71e5-4fa3-ae7b-5137b0e58397" containerID="be1b9f8d0c257a0e1d0419cf2c1278df0e274331eb3adeef663b50de9e49f2f3" exitCode=0 Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.437540 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb6ca922-71e5-4fa3-ae7b-5137b0e58397","Type":"ContainerDied","Data":"be1b9f8d0c257a0e1d0419cf2c1278df0e274331eb3adeef663b50de9e49f2f3"} Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.437568 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb6ca922-71e5-4fa3-ae7b-5137b0e58397","Type":"ContainerDied","Data":"9c8e4b7251239d087f479e7cdf47b4e5ff69599faa90d12e66a25ebf0083f84b"} Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.437590 4772 scope.go:117] "RemoveContainer" containerID="ece33e666a3e58ee419da4cf74156d864bda223101f1f560e453f6c6c89635ef" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.437591 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.473087 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.474901 4772 scope.go:117] "RemoveContainer" containerID="be1b9f8d0c257a0e1d0419cf2c1278df0e274331eb3adeef663b50de9e49f2f3" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.483987 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.495619 4772 scope.go:117] "RemoveContainer" containerID="ece33e666a3e58ee419da4cf74156d864bda223101f1f560e453f6c6c89635ef" Jan 27 16:41:57 crc kubenswrapper[4772]: E0127 16:41:57.496438 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ece33e666a3e58ee419da4cf74156d864bda223101f1f560e453f6c6c89635ef\": container with ID starting with ece33e666a3e58ee419da4cf74156d864bda223101f1f560e453f6c6c89635ef not found: ID does not exist" containerID="ece33e666a3e58ee419da4cf74156d864bda223101f1f560e453f6c6c89635ef" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.496595 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ece33e666a3e58ee419da4cf74156d864bda223101f1f560e453f6c6c89635ef"} err="failed to get container status \"ece33e666a3e58ee419da4cf74156d864bda223101f1f560e453f6c6c89635ef\": rpc error: code = NotFound desc = could not find container \"ece33e666a3e58ee419da4cf74156d864bda223101f1f560e453f6c6c89635ef\": container with ID starting with ece33e666a3e58ee419da4cf74156d864bda223101f1f560e453f6c6c89635ef not found: ID does not exist" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.496733 4772 scope.go:117] "RemoveContainer" containerID="be1b9f8d0c257a0e1d0419cf2c1278df0e274331eb3adeef663b50de9e49f2f3" Jan 27 16:41:57 crc kubenswrapper[4772]: E0127 16:41:57.497289 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be1b9f8d0c257a0e1d0419cf2c1278df0e274331eb3adeef663b50de9e49f2f3\": container with ID starting with be1b9f8d0c257a0e1d0419cf2c1278df0e274331eb3adeef663b50de9e49f2f3 not found: ID does not exist" containerID="be1b9f8d0c257a0e1d0419cf2c1278df0e274331eb3adeef663b50de9e49f2f3" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.497349 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be1b9f8d0c257a0e1d0419cf2c1278df0e274331eb3adeef663b50de9e49f2f3"} err="failed to get container status \"be1b9f8d0c257a0e1d0419cf2c1278df0e274331eb3adeef663b50de9e49f2f3\": rpc error: code = NotFound desc = could not find container \"be1b9f8d0c257a0e1d0419cf2c1278df0e274331eb3adeef663b50de9e49f2f3\": container with ID starting with be1b9f8d0c257a0e1d0419cf2c1278df0e274331eb3adeef663b50de9e49f2f3 not found: ID does not exist" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.507631 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Jan 27 16:41:57 crc kubenswrapper[4772]: E0127 16:41:57.508066 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb6ca922-71e5-4fa3-ae7b-5137b0e58397" containerName="cinder-scheduler" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.508090 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb6ca922-71e5-4fa3-ae7b-5137b0e58397" containerName="cinder-scheduler" Jan 27 16:41:57 crc kubenswrapper[4772]: E0127 16:41:57.508120 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb6ca922-71e5-4fa3-ae7b-5137b0e58397" containerName="probe" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.508128 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb6ca922-71e5-4fa3-ae7b-5137b0e58397" containerName="probe" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.508381 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb6ca922-71e5-4fa3-ae7b-5137b0e58397" containerName="probe" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.508408 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb6ca922-71e5-4fa3-ae7b-5137b0e58397" containerName="cinder-scheduler" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.511493 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.513560 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.519855 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.623012 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/66512484-80ba-4887-b9a9-9cc87a65ad18-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"66512484-80ba-4887-b9a9-9cc87a65ad18\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.623112 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/66512484-80ba-4887-b9a9-9cc87a65ad18-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"66512484-80ba-4887-b9a9-9cc87a65ad18\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.623130 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66512484-80ba-4887-b9a9-9cc87a65ad18-scripts\") pod \"cinder-scheduler-0\" (UID: \"66512484-80ba-4887-b9a9-9cc87a65ad18\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.623153 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66512484-80ba-4887-b9a9-9cc87a65ad18-config-data\") pod \"cinder-scheduler-0\" (UID: \"66512484-80ba-4887-b9a9-9cc87a65ad18\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.623259 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6vtw\" (UniqueName: \"kubernetes.io/projected/66512484-80ba-4887-b9a9-9cc87a65ad18-kube-api-access-m6vtw\") pod \"cinder-scheduler-0\" (UID: \"66512484-80ba-4887-b9a9-9cc87a65ad18\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.623281 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66512484-80ba-4887-b9a9-9cc87a65ad18-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"66512484-80ba-4887-b9a9-9cc87a65ad18\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.726400 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6vtw\" (UniqueName: \"kubernetes.io/projected/66512484-80ba-4887-b9a9-9cc87a65ad18-kube-api-access-m6vtw\") pod \"cinder-scheduler-0\" (UID: \"66512484-80ba-4887-b9a9-9cc87a65ad18\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.726479 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66512484-80ba-4887-b9a9-9cc87a65ad18-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"66512484-80ba-4887-b9a9-9cc87a65ad18\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.726527 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/66512484-80ba-4887-b9a9-9cc87a65ad18-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"66512484-80ba-4887-b9a9-9cc87a65ad18\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.726741 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/66512484-80ba-4887-b9a9-9cc87a65ad18-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"66512484-80ba-4887-b9a9-9cc87a65ad18\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.726786 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66512484-80ba-4887-b9a9-9cc87a65ad18-scripts\") pod \"cinder-scheduler-0\" (UID: \"66512484-80ba-4887-b9a9-9cc87a65ad18\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.726836 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66512484-80ba-4887-b9a9-9cc87a65ad18-config-data\") pod \"cinder-scheduler-0\" (UID: \"66512484-80ba-4887-b9a9-9cc87a65ad18\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.728343 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/66512484-80ba-4887-b9a9-9cc87a65ad18-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"66512484-80ba-4887-b9a9-9cc87a65ad18\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.733144 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66512484-80ba-4887-b9a9-9cc87a65ad18-scripts\") pod \"cinder-scheduler-0\" (UID: \"66512484-80ba-4887-b9a9-9cc87a65ad18\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.733243 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66512484-80ba-4887-b9a9-9cc87a65ad18-config-data\") pod \"cinder-scheduler-0\" (UID: \"66512484-80ba-4887-b9a9-9cc87a65ad18\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.733821 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66512484-80ba-4887-b9a9-9cc87a65ad18-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"66512484-80ba-4887-b9a9-9cc87a65ad18\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.737352 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/66512484-80ba-4887-b9a9-9cc87a65ad18-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"66512484-80ba-4887-b9a9-9cc87a65ad18\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.760363 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6vtw\" (UniqueName: \"kubernetes.io/projected/66512484-80ba-4887-b9a9-9cc87a65ad18-kube-api-access-m6vtw\") pod \"cinder-scheduler-0\" (UID: \"66512484-80ba-4887-b9a9-9cc87a65ad18\") " pod="openstack/cinder-scheduler-0" Jan 27 16:41:57 crc kubenswrapper[4772]: I0127 16:41:57.847209 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 27 16:41:58 crc kubenswrapper[4772]: I0127 16:41:58.299630 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 27 16:41:58 crc kubenswrapper[4772]: W0127 16:41:58.306186 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66512484_80ba_4887_b9a9_9cc87a65ad18.slice/crio-985270c438f483f088dfdeaaea8bfdcc4d8df2c83364557f3669e0b747fb7c38 WatchSource:0}: Error finding container 985270c438f483f088dfdeaaea8bfdcc4d8df2c83364557f3669e0b747fb7c38: Status 404 returned error can't find the container with id 985270c438f483f088dfdeaaea8bfdcc4d8df2c83364557f3669e0b747fb7c38 Jan 27 16:41:58 crc kubenswrapper[4772]: I0127 16:41:58.448904 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"66512484-80ba-4887-b9a9-9cc87a65ad18","Type":"ContainerStarted","Data":"985270c438f483f088dfdeaaea8bfdcc4d8df2c83364557f3669e0b747fb7c38"} Jan 27 16:41:58 crc kubenswrapper[4772]: I0127 16:41:58.674124 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb6ca922-71e5-4fa3-ae7b-5137b0e58397" path="/var/lib/kubelet/pods/cb6ca922-71e5-4fa3-ae7b-5137b0e58397/volumes" Jan 27 16:41:59 crc kubenswrapper[4772]: I0127 16:41:59.476638 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"66512484-80ba-4887-b9a9-9cc87a65ad18","Type":"ContainerStarted","Data":"4190615e9df1af6789d9abb9ce07935a57894feb9b066d1087504e28e29c1ae6"} Jan 27 16:41:59 crc kubenswrapper[4772]: I0127 16:41:59.477030 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"66512484-80ba-4887-b9a9-9cc87a65ad18","Type":"ContainerStarted","Data":"8998b51dedfa1e5b10014c825ed5e1f50ffed9f00543215a444700336d946133"} Jan 27 16:41:59 crc kubenswrapper[4772]: I0127 16:41:59.506966 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.506947911 podStartE2EDuration="2.506947911s" podCreationTimestamp="2026-01-27 16:41:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:41:59.504581053 +0000 UTC m=+5705.485190161" watchObservedRunningTime="2026-01-27 16:41:59.506947911 +0000 UTC m=+5705.487557009" Jan 27 16:42:01 crc kubenswrapper[4772]: I0127 16:42:01.553594 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Jan 27 16:42:02 crc kubenswrapper[4772]: I0127 16:42:02.847918 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Jan 27 16:42:08 crc kubenswrapper[4772]: I0127 16:42:08.054581 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Jan 27 16:42:09 crc kubenswrapper[4772]: I0127 16:42:09.664483 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:42:09 crc kubenswrapper[4772]: E0127 16:42:09.664786 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:42:22 crc kubenswrapper[4772]: I0127 16:42:22.664447 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:42:22 crc kubenswrapper[4772]: E0127 16:42:22.666017 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:42:35 crc kubenswrapper[4772]: I0127 16:42:35.663193 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:42:35 crc kubenswrapper[4772]: E0127 16:42:35.664043 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:42:48 crc kubenswrapper[4772]: I0127 16:42:48.662913 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:42:48 crc kubenswrapper[4772]: E0127 16:42:48.663808 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:42:59 crc kubenswrapper[4772]: I0127 16:42:59.221201 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2zjm2"] Jan 27 16:42:59 crc kubenswrapper[4772]: I0127 16:42:59.223956 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2zjm2" Jan 27 16:42:59 crc kubenswrapper[4772]: I0127 16:42:59.229857 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2zjm2"] Jan 27 16:42:59 crc kubenswrapper[4772]: I0127 16:42:59.302625 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/459295b8-ff6b-4e68-86ef-2db5ccb48ebd-utilities\") pod \"certified-operators-2zjm2\" (UID: \"459295b8-ff6b-4e68-86ef-2db5ccb48ebd\") " pod="openshift-marketplace/certified-operators-2zjm2" Jan 27 16:42:59 crc kubenswrapper[4772]: I0127 16:42:59.302952 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/459295b8-ff6b-4e68-86ef-2db5ccb48ebd-catalog-content\") pod \"certified-operators-2zjm2\" (UID: \"459295b8-ff6b-4e68-86ef-2db5ccb48ebd\") " pod="openshift-marketplace/certified-operators-2zjm2" Jan 27 16:42:59 crc kubenswrapper[4772]: I0127 16:42:59.303051 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6phf\" (UniqueName: \"kubernetes.io/projected/459295b8-ff6b-4e68-86ef-2db5ccb48ebd-kube-api-access-w6phf\") pod \"certified-operators-2zjm2\" (UID: \"459295b8-ff6b-4e68-86ef-2db5ccb48ebd\") " pod="openshift-marketplace/certified-operators-2zjm2" Jan 27 16:42:59 crc kubenswrapper[4772]: I0127 16:42:59.404798 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6phf\" (UniqueName: \"kubernetes.io/projected/459295b8-ff6b-4e68-86ef-2db5ccb48ebd-kube-api-access-w6phf\") pod \"certified-operators-2zjm2\" (UID: \"459295b8-ff6b-4e68-86ef-2db5ccb48ebd\") " pod="openshift-marketplace/certified-operators-2zjm2" Jan 27 16:42:59 crc kubenswrapper[4772]: I0127 16:42:59.404912 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/459295b8-ff6b-4e68-86ef-2db5ccb48ebd-utilities\") pod \"certified-operators-2zjm2\" (UID: \"459295b8-ff6b-4e68-86ef-2db5ccb48ebd\") " pod="openshift-marketplace/certified-operators-2zjm2" Jan 27 16:42:59 crc kubenswrapper[4772]: I0127 16:42:59.404942 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/459295b8-ff6b-4e68-86ef-2db5ccb48ebd-catalog-content\") pod \"certified-operators-2zjm2\" (UID: \"459295b8-ff6b-4e68-86ef-2db5ccb48ebd\") " pod="openshift-marketplace/certified-operators-2zjm2" Jan 27 16:42:59 crc kubenswrapper[4772]: I0127 16:42:59.405636 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/459295b8-ff6b-4e68-86ef-2db5ccb48ebd-catalog-content\") pod \"certified-operators-2zjm2\" (UID: \"459295b8-ff6b-4e68-86ef-2db5ccb48ebd\") " pod="openshift-marketplace/certified-operators-2zjm2" Jan 27 16:42:59 crc kubenswrapper[4772]: I0127 16:42:59.405725 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/459295b8-ff6b-4e68-86ef-2db5ccb48ebd-utilities\") pod \"certified-operators-2zjm2\" (UID: \"459295b8-ff6b-4e68-86ef-2db5ccb48ebd\") " pod="openshift-marketplace/certified-operators-2zjm2" Jan 27 16:42:59 crc kubenswrapper[4772]: I0127 16:42:59.428059 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6phf\" (UniqueName: \"kubernetes.io/projected/459295b8-ff6b-4e68-86ef-2db5ccb48ebd-kube-api-access-w6phf\") pod \"certified-operators-2zjm2\" (UID: \"459295b8-ff6b-4e68-86ef-2db5ccb48ebd\") " pod="openshift-marketplace/certified-operators-2zjm2" Jan 27 16:42:59 crc kubenswrapper[4772]: I0127 16:42:59.559999 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2zjm2" Jan 27 16:43:00 crc kubenswrapper[4772]: I0127 16:43:00.175427 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2zjm2"] Jan 27 16:43:01 crc kubenswrapper[4772]: I0127 16:43:01.019032 4772 generic.go:334] "Generic (PLEG): container finished" podID="459295b8-ff6b-4e68-86ef-2db5ccb48ebd" containerID="3f844d16bdc9ff2e0498bc2f66feb122a928a2a7de272c003efb7d14c3fb197a" exitCode=0 Jan 27 16:43:01 crc kubenswrapper[4772]: I0127 16:43:01.019101 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2zjm2" event={"ID":"459295b8-ff6b-4e68-86ef-2db5ccb48ebd","Type":"ContainerDied","Data":"3f844d16bdc9ff2e0498bc2f66feb122a928a2a7de272c003efb7d14c3fb197a"} Jan 27 16:43:01 crc kubenswrapper[4772]: I0127 16:43:01.019493 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2zjm2" event={"ID":"459295b8-ff6b-4e68-86ef-2db5ccb48ebd","Type":"ContainerStarted","Data":"ebcd2a276f29c5a3831867dbda33105f239b6023bcb91f062252b087f3f0823f"} Jan 27 16:43:01 crc kubenswrapper[4772]: I0127 16:43:01.664152 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:43:01 crc kubenswrapper[4772]: E0127 16:43:01.664681 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:43:03 crc kubenswrapper[4772]: I0127 16:43:03.038428 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2zjm2" event={"ID":"459295b8-ff6b-4e68-86ef-2db5ccb48ebd","Type":"ContainerStarted","Data":"603104fcf1485a95a1485cbf6e7d7243393b9a93d0ca768f32888f4b677dac6b"} Jan 27 16:43:04 crc kubenswrapper[4772]: I0127 16:43:04.048325 4772 generic.go:334] "Generic (PLEG): container finished" podID="459295b8-ff6b-4e68-86ef-2db5ccb48ebd" containerID="603104fcf1485a95a1485cbf6e7d7243393b9a93d0ca768f32888f4b677dac6b" exitCode=0 Jan 27 16:43:04 crc kubenswrapper[4772]: I0127 16:43:04.048429 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2zjm2" event={"ID":"459295b8-ff6b-4e68-86ef-2db5ccb48ebd","Type":"ContainerDied","Data":"603104fcf1485a95a1485cbf6e7d7243393b9a93d0ca768f32888f4b677dac6b"} Jan 27 16:43:05 crc kubenswrapper[4772]: I0127 16:43:05.060712 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2zjm2" event={"ID":"459295b8-ff6b-4e68-86ef-2db5ccb48ebd","Type":"ContainerStarted","Data":"d32818b18c7ff11f1c88b751ea1897cc762fa68917a75c87760f093e1d49259d"} Jan 27 16:43:05 crc kubenswrapper[4772]: I0127 16:43:05.088361 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2zjm2" podStartSLOduration=2.5869534119999997 podStartE2EDuration="6.088337615s" podCreationTimestamp="2026-01-27 16:42:59 +0000 UTC" firstStartedPulling="2026-01-27 16:43:01.022244837 +0000 UTC m=+5767.002853935" lastFinishedPulling="2026-01-27 16:43:04.52362904 +0000 UTC m=+5770.504238138" observedRunningTime="2026-01-27 16:43:05.077387143 +0000 UTC m=+5771.057996241" watchObservedRunningTime="2026-01-27 16:43:05.088337615 +0000 UTC m=+5771.068946713" Jan 27 16:43:07 crc kubenswrapper[4772]: I0127 16:43:07.483415 4772 scope.go:117] "RemoveContainer" containerID="f3387679ac491bfa95cee52f7df561b5ae6817e29d93d8687efd8b7b43af3938" Jan 27 16:43:07 crc kubenswrapper[4772]: I0127 16:43:07.504004 4772 scope.go:117] "RemoveContainer" containerID="71552776ddecc62c8f2d51f59b6ec4b6ba66501bb9fcfb2256ecea66da9231d2" Jan 27 16:43:09 crc kubenswrapper[4772]: I0127 16:43:09.561096 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2zjm2" Jan 27 16:43:09 crc kubenswrapper[4772]: I0127 16:43:09.561414 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2zjm2" Jan 27 16:43:09 crc kubenswrapper[4772]: I0127 16:43:09.609062 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2zjm2" Jan 27 16:43:10 crc kubenswrapper[4772]: I0127 16:43:10.150854 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2zjm2" Jan 27 16:43:10 crc kubenswrapper[4772]: I0127 16:43:10.201910 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2zjm2"] Jan 27 16:43:12 crc kubenswrapper[4772]: I0127 16:43:12.121155 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2zjm2" podUID="459295b8-ff6b-4e68-86ef-2db5ccb48ebd" containerName="registry-server" containerID="cri-o://d32818b18c7ff11f1c88b751ea1897cc762fa68917a75c87760f093e1d49259d" gracePeriod=2 Jan 27 16:43:12 crc kubenswrapper[4772]: I0127 16:43:12.596453 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2zjm2" Jan 27 16:43:12 crc kubenswrapper[4772]: I0127 16:43:12.660443 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/459295b8-ff6b-4e68-86ef-2db5ccb48ebd-utilities\") pod \"459295b8-ff6b-4e68-86ef-2db5ccb48ebd\" (UID: \"459295b8-ff6b-4e68-86ef-2db5ccb48ebd\") " Jan 27 16:43:12 crc kubenswrapper[4772]: I0127 16:43:12.660564 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6phf\" (UniqueName: \"kubernetes.io/projected/459295b8-ff6b-4e68-86ef-2db5ccb48ebd-kube-api-access-w6phf\") pod \"459295b8-ff6b-4e68-86ef-2db5ccb48ebd\" (UID: \"459295b8-ff6b-4e68-86ef-2db5ccb48ebd\") " Jan 27 16:43:12 crc kubenswrapper[4772]: I0127 16:43:12.660667 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/459295b8-ff6b-4e68-86ef-2db5ccb48ebd-catalog-content\") pod \"459295b8-ff6b-4e68-86ef-2db5ccb48ebd\" (UID: \"459295b8-ff6b-4e68-86ef-2db5ccb48ebd\") " Jan 27 16:43:12 crc kubenswrapper[4772]: I0127 16:43:12.661368 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/459295b8-ff6b-4e68-86ef-2db5ccb48ebd-utilities" (OuterVolumeSpecName: "utilities") pod "459295b8-ff6b-4e68-86ef-2db5ccb48ebd" (UID: "459295b8-ff6b-4e68-86ef-2db5ccb48ebd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:43:12 crc kubenswrapper[4772]: I0127 16:43:12.665523 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/459295b8-ff6b-4e68-86ef-2db5ccb48ebd-kube-api-access-w6phf" (OuterVolumeSpecName: "kube-api-access-w6phf") pod "459295b8-ff6b-4e68-86ef-2db5ccb48ebd" (UID: "459295b8-ff6b-4e68-86ef-2db5ccb48ebd"). InnerVolumeSpecName "kube-api-access-w6phf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:43:12 crc kubenswrapper[4772]: I0127 16:43:12.709419 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/459295b8-ff6b-4e68-86ef-2db5ccb48ebd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "459295b8-ff6b-4e68-86ef-2db5ccb48ebd" (UID: "459295b8-ff6b-4e68-86ef-2db5ccb48ebd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:43:12 crc kubenswrapper[4772]: I0127 16:43:12.762937 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/459295b8-ff6b-4e68-86ef-2db5ccb48ebd-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 16:43:12 crc kubenswrapper[4772]: I0127 16:43:12.762967 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/459295b8-ff6b-4e68-86ef-2db5ccb48ebd-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 16:43:12 crc kubenswrapper[4772]: I0127 16:43:12.762978 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6phf\" (UniqueName: \"kubernetes.io/projected/459295b8-ff6b-4e68-86ef-2db5ccb48ebd-kube-api-access-w6phf\") on node \"crc\" DevicePath \"\"" Jan 27 16:43:13 crc kubenswrapper[4772]: I0127 16:43:13.131695 4772 generic.go:334] "Generic (PLEG): container finished" podID="459295b8-ff6b-4e68-86ef-2db5ccb48ebd" containerID="d32818b18c7ff11f1c88b751ea1897cc762fa68917a75c87760f093e1d49259d" exitCode=0 Jan 27 16:43:13 crc kubenswrapper[4772]: I0127 16:43:13.131740 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2zjm2" event={"ID":"459295b8-ff6b-4e68-86ef-2db5ccb48ebd","Type":"ContainerDied","Data":"d32818b18c7ff11f1c88b751ea1897cc762fa68917a75c87760f093e1d49259d"} Jan 27 16:43:13 crc kubenswrapper[4772]: I0127 16:43:13.131772 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2zjm2" Jan 27 16:43:13 crc kubenswrapper[4772]: I0127 16:43:13.131782 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2zjm2" event={"ID":"459295b8-ff6b-4e68-86ef-2db5ccb48ebd","Type":"ContainerDied","Data":"ebcd2a276f29c5a3831867dbda33105f239b6023bcb91f062252b087f3f0823f"} Jan 27 16:43:13 crc kubenswrapper[4772]: I0127 16:43:13.131807 4772 scope.go:117] "RemoveContainer" containerID="d32818b18c7ff11f1c88b751ea1897cc762fa68917a75c87760f093e1d49259d" Jan 27 16:43:13 crc kubenswrapper[4772]: I0127 16:43:13.154477 4772 scope.go:117] "RemoveContainer" containerID="603104fcf1485a95a1485cbf6e7d7243393b9a93d0ca768f32888f4b677dac6b" Jan 27 16:43:13 crc kubenswrapper[4772]: I0127 16:43:13.181820 4772 scope.go:117] "RemoveContainer" containerID="3f844d16bdc9ff2e0498bc2f66feb122a928a2a7de272c003efb7d14c3fb197a" Jan 27 16:43:13 crc kubenswrapper[4772]: I0127 16:43:13.185224 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2zjm2"] Jan 27 16:43:13 crc kubenswrapper[4772]: I0127 16:43:13.195245 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2zjm2"] Jan 27 16:43:13 crc kubenswrapper[4772]: I0127 16:43:13.218810 4772 scope.go:117] "RemoveContainer" containerID="d32818b18c7ff11f1c88b751ea1897cc762fa68917a75c87760f093e1d49259d" Jan 27 16:43:13 crc kubenswrapper[4772]: E0127 16:43:13.219454 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d32818b18c7ff11f1c88b751ea1897cc762fa68917a75c87760f093e1d49259d\": container with ID starting with d32818b18c7ff11f1c88b751ea1897cc762fa68917a75c87760f093e1d49259d not found: ID does not exist" containerID="d32818b18c7ff11f1c88b751ea1897cc762fa68917a75c87760f093e1d49259d" Jan 27 16:43:13 crc kubenswrapper[4772]: I0127 16:43:13.219516 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d32818b18c7ff11f1c88b751ea1897cc762fa68917a75c87760f093e1d49259d"} err="failed to get container status \"d32818b18c7ff11f1c88b751ea1897cc762fa68917a75c87760f093e1d49259d\": rpc error: code = NotFound desc = could not find container \"d32818b18c7ff11f1c88b751ea1897cc762fa68917a75c87760f093e1d49259d\": container with ID starting with d32818b18c7ff11f1c88b751ea1897cc762fa68917a75c87760f093e1d49259d not found: ID does not exist" Jan 27 16:43:13 crc kubenswrapper[4772]: I0127 16:43:13.219559 4772 scope.go:117] "RemoveContainer" containerID="603104fcf1485a95a1485cbf6e7d7243393b9a93d0ca768f32888f4b677dac6b" Jan 27 16:43:13 crc kubenswrapper[4772]: E0127 16:43:13.220034 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"603104fcf1485a95a1485cbf6e7d7243393b9a93d0ca768f32888f4b677dac6b\": container with ID starting with 603104fcf1485a95a1485cbf6e7d7243393b9a93d0ca768f32888f4b677dac6b not found: ID does not exist" containerID="603104fcf1485a95a1485cbf6e7d7243393b9a93d0ca768f32888f4b677dac6b" Jan 27 16:43:13 crc kubenswrapper[4772]: I0127 16:43:13.220075 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"603104fcf1485a95a1485cbf6e7d7243393b9a93d0ca768f32888f4b677dac6b"} err="failed to get container status \"603104fcf1485a95a1485cbf6e7d7243393b9a93d0ca768f32888f4b677dac6b\": rpc error: code = NotFound desc = could not find container \"603104fcf1485a95a1485cbf6e7d7243393b9a93d0ca768f32888f4b677dac6b\": container with ID starting with 603104fcf1485a95a1485cbf6e7d7243393b9a93d0ca768f32888f4b677dac6b not found: ID does not exist" Jan 27 16:43:13 crc kubenswrapper[4772]: I0127 16:43:13.220117 4772 scope.go:117] "RemoveContainer" containerID="3f844d16bdc9ff2e0498bc2f66feb122a928a2a7de272c003efb7d14c3fb197a" Jan 27 16:43:13 crc kubenswrapper[4772]: E0127 16:43:13.220580 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f844d16bdc9ff2e0498bc2f66feb122a928a2a7de272c003efb7d14c3fb197a\": container with ID starting with 3f844d16bdc9ff2e0498bc2f66feb122a928a2a7de272c003efb7d14c3fb197a not found: ID does not exist" containerID="3f844d16bdc9ff2e0498bc2f66feb122a928a2a7de272c003efb7d14c3fb197a" Jan 27 16:43:13 crc kubenswrapper[4772]: I0127 16:43:13.220636 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f844d16bdc9ff2e0498bc2f66feb122a928a2a7de272c003efb7d14c3fb197a"} err="failed to get container status \"3f844d16bdc9ff2e0498bc2f66feb122a928a2a7de272c003efb7d14c3fb197a\": rpc error: code = NotFound desc = could not find container \"3f844d16bdc9ff2e0498bc2f66feb122a928a2a7de272c003efb7d14c3fb197a\": container with ID starting with 3f844d16bdc9ff2e0498bc2f66feb122a928a2a7de272c003efb7d14c3fb197a not found: ID does not exist" Jan 27 16:43:14 crc kubenswrapper[4772]: I0127 16:43:14.670088 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:43:14 crc kubenswrapper[4772]: E0127 16:43:14.670366 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:43:14 crc kubenswrapper[4772]: I0127 16:43:14.674402 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="459295b8-ff6b-4e68-86ef-2db5ccb48ebd" path="/var/lib/kubelet/pods/459295b8-ff6b-4e68-86ef-2db5ccb48ebd/volumes" Jan 27 16:43:26 crc kubenswrapper[4772]: I0127 16:43:26.663386 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:43:26 crc kubenswrapper[4772]: E0127 16:43:26.664564 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:43:31 crc kubenswrapper[4772]: I0127 16:43:31.041869 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-jb5ch"] Jan 27 16:43:31 crc kubenswrapper[4772]: I0127 16:43:31.052663 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-jb5ch"] Jan 27 16:43:32 crc kubenswrapper[4772]: I0127 16:43:32.028206 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-0c73-account-create-update-nz44z"] Jan 27 16:43:32 crc kubenswrapper[4772]: I0127 16:43:32.036311 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-0c73-account-create-update-nz44z"] Jan 27 16:43:32 crc kubenswrapper[4772]: I0127 16:43:32.675265 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5292a043-9ee5-4d14-a991-c50dbf4d136e" path="/var/lib/kubelet/pods/5292a043-9ee5-4d14-a991-c50dbf4d136e/volumes" Jan 27 16:43:32 crc kubenswrapper[4772]: I0127 16:43:32.676052 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7e8906f-73a7-4580-81c8-ec81439faea5" path="/var/lib/kubelet/pods/d7e8906f-73a7-4580-81c8-ec81439faea5/volumes" Jan 27 16:43:37 crc kubenswrapper[4772]: I0127 16:43:37.057637 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-ls87v"] Jan 27 16:43:37 crc kubenswrapper[4772]: I0127 16:43:37.066257 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-ls87v"] Jan 27 16:43:38 crc kubenswrapper[4772]: I0127 16:43:38.677924 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e179bb8-f9e4-434d-9636-84cc97d632fb" path="/var/lib/kubelet/pods/9e179bb8-f9e4-434d-9636-84cc97d632fb/volumes" Jan 27 16:43:41 crc kubenswrapper[4772]: I0127 16:43:41.663717 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:43:41 crc kubenswrapper[4772]: E0127 16:43:41.664299 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:43:47 crc kubenswrapper[4772]: I0127 16:43:47.937353 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-s5xbr"] Jan 27 16:43:47 crc kubenswrapper[4772]: E0127 16:43:47.970627 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="459295b8-ff6b-4e68-86ef-2db5ccb48ebd" containerName="extract-utilities" Jan 27 16:43:47 crc kubenswrapper[4772]: I0127 16:43:47.970674 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="459295b8-ff6b-4e68-86ef-2db5ccb48ebd" containerName="extract-utilities" Jan 27 16:43:47 crc kubenswrapper[4772]: E0127 16:43:47.970725 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="459295b8-ff6b-4e68-86ef-2db5ccb48ebd" containerName="extract-content" Jan 27 16:43:47 crc kubenswrapper[4772]: I0127 16:43:47.970734 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="459295b8-ff6b-4e68-86ef-2db5ccb48ebd" containerName="extract-content" Jan 27 16:43:47 crc kubenswrapper[4772]: E0127 16:43:47.970753 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="459295b8-ff6b-4e68-86ef-2db5ccb48ebd" containerName="registry-server" Jan 27 16:43:47 crc kubenswrapper[4772]: I0127 16:43:47.970761 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="459295b8-ff6b-4e68-86ef-2db5ccb48ebd" containerName="registry-server" Jan 27 16:43:47 crc kubenswrapper[4772]: I0127 16:43:47.971275 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="459295b8-ff6b-4e68-86ef-2db5ccb48ebd" containerName="registry-server" Jan 27 16:43:47 crc kubenswrapper[4772]: I0127 16:43:47.975943 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-s5xbr" Jan 27 16:43:47 crc kubenswrapper[4772]: I0127 16:43:47.979230 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Jan 27 16:43:47 crc kubenswrapper[4772]: I0127 16:43:47.979479 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-zmzrv" Jan 27 16:43:47 crc kubenswrapper[4772]: I0127 16:43:47.990116 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jv694"] Jan 27 16:43:47 crc kubenswrapper[4772]: I0127 16:43:47.992824 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jv694" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.011755 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-s5xbr"] Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.025980 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jv694"] Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.051820 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/febb140e-d26e-43db-9924-0f06739b9a4a-var-run\") pod \"ovn-controller-jv694\" (UID: \"febb140e-d26e-43db-9924-0f06739b9a4a\") " pod="openstack/ovn-controller-jv694" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.051875 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/7e5eabb2-229a-4d75-b62d-65be688f753a-var-log\") pod \"ovn-controller-ovs-s5xbr\" (UID: \"7e5eabb2-229a-4d75-b62d-65be688f753a\") " pod="openstack/ovn-controller-ovs-s5xbr" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.051928 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7e5eabb2-229a-4d75-b62d-65be688f753a-var-run\") pod \"ovn-controller-ovs-s5xbr\" (UID: \"7e5eabb2-229a-4d75-b62d-65be688f753a\") " pod="openstack/ovn-controller-ovs-s5xbr" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.052021 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvxsl\" (UniqueName: \"kubernetes.io/projected/7e5eabb2-229a-4d75-b62d-65be688f753a-kube-api-access-xvxsl\") pod \"ovn-controller-ovs-s5xbr\" (UID: \"7e5eabb2-229a-4d75-b62d-65be688f753a\") " pod="openstack/ovn-controller-ovs-s5xbr" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.052063 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/7e5eabb2-229a-4d75-b62d-65be688f753a-var-lib\") pod \"ovn-controller-ovs-s5xbr\" (UID: \"7e5eabb2-229a-4d75-b62d-65be688f753a\") " pod="openstack/ovn-controller-ovs-s5xbr" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.052092 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/7e5eabb2-229a-4d75-b62d-65be688f753a-etc-ovs\") pod \"ovn-controller-ovs-s5xbr\" (UID: \"7e5eabb2-229a-4d75-b62d-65be688f753a\") " pod="openstack/ovn-controller-ovs-s5xbr" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.052117 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/febb140e-d26e-43db-9924-0f06739b9a4a-var-run-ovn\") pod \"ovn-controller-jv694\" (UID: \"febb140e-d26e-43db-9924-0f06739b9a4a\") " pod="openstack/ovn-controller-jv694" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.052451 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/febb140e-d26e-43db-9924-0f06739b9a4a-scripts\") pod \"ovn-controller-jv694\" (UID: \"febb140e-d26e-43db-9924-0f06739b9a4a\") " pod="openstack/ovn-controller-jv694" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.052489 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/febb140e-d26e-43db-9924-0f06739b9a4a-var-log-ovn\") pod \"ovn-controller-jv694\" (UID: \"febb140e-d26e-43db-9924-0f06739b9a4a\") " pod="openstack/ovn-controller-jv694" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.052534 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e5eabb2-229a-4d75-b62d-65be688f753a-scripts\") pod \"ovn-controller-ovs-s5xbr\" (UID: \"7e5eabb2-229a-4d75-b62d-65be688f753a\") " pod="openstack/ovn-controller-ovs-s5xbr" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.052562 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5bh9\" (UniqueName: \"kubernetes.io/projected/febb140e-d26e-43db-9924-0f06739b9a4a-kube-api-access-p5bh9\") pod \"ovn-controller-jv694\" (UID: \"febb140e-d26e-43db-9924-0f06739b9a4a\") " pod="openstack/ovn-controller-jv694" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.153986 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/febb140e-d26e-43db-9924-0f06739b9a4a-scripts\") pod \"ovn-controller-jv694\" (UID: \"febb140e-d26e-43db-9924-0f06739b9a4a\") " pod="openstack/ovn-controller-jv694" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.154124 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/febb140e-d26e-43db-9924-0f06739b9a4a-var-log-ovn\") pod \"ovn-controller-jv694\" (UID: \"febb140e-d26e-43db-9924-0f06739b9a4a\") " pod="openstack/ovn-controller-jv694" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.154189 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e5eabb2-229a-4d75-b62d-65be688f753a-scripts\") pod \"ovn-controller-ovs-s5xbr\" (UID: \"7e5eabb2-229a-4d75-b62d-65be688f753a\") " pod="openstack/ovn-controller-ovs-s5xbr" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.154211 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5bh9\" (UniqueName: \"kubernetes.io/projected/febb140e-d26e-43db-9924-0f06739b9a4a-kube-api-access-p5bh9\") pod \"ovn-controller-jv694\" (UID: \"febb140e-d26e-43db-9924-0f06739b9a4a\") " pod="openstack/ovn-controller-jv694" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.154227 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/febb140e-d26e-43db-9924-0f06739b9a4a-var-run\") pod \"ovn-controller-jv694\" (UID: \"febb140e-d26e-43db-9924-0f06739b9a4a\") " pod="openstack/ovn-controller-jv694" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.154264 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/7e5eabb2-229a-4d75-b62d-65be688f753a-var-log\") pod \"ovn-controller-ovs-s5xbr\" (UID: \"7e5eabb2-229a-4d75-b62d-65be688f753a\") " pod="openstack/ovn-controller-ovs-s5xbr" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.154297 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7e5eabb2-229a-4d75-b62d-65be688f753a-var-run\") pod \"ovn-controller-ovs-s5xbr\" (UID: \"7e5eabb2-229a-4d75-b62d-65be688f753a\") " pod="openstack/ovn-controller-ovs-s5xbr" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.154397 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvxsl\" (UniqueName: \"kubernetes.io/projected/7e5eabb2-229a-4d75-b62d-65be688f753a-kube-api-access-xvxsl\") pod \"ovn-controller-ovs-s5xbr\" (UID: \"7e5eabb2-229a-4d75-b62d-65be688f753a\") " pod="openstack/ovn-controller-ovs-s5xbr" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.154436 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/7e5eabb2-229a-4d75-b62d-65be688f753a-var-lib\") pod \"ovn-controller-ovs-s5xbr\" (UID: \"7e5eabb2-229a-4d75-b62d-65be688f753a\") " pod="openstack/ovn-controller-ovs-s5xbr" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.154459 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/7e5eabb2-229a-4d75-b62d-65be688f753a-etc-ovs\") pod \"ovn-controller-ovs-s5xbr\" (UID: \"7e5eabb2-229a-4d75-b62d-65be688f753a\") " pod="openstack/ovn-controller-ovs-s5xbr" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.154505 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/febb140e-d26e-43db-9924-0f06739b9a4a-var-run-ovn\") pod \"ovn-controller-jv694\" (UID: \"febb140e-d26e-43db-9924-0f06739b9a4a\") " pod="openstack/ovn-controller-jv694" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.154930 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/febb140e-d26e-43db-9924-0f06739b9a4a-var-run\") pod \"ovn-controller-jv694\" (UID: \"febb140e-d26e-43db-9924-0f06739b9a4a\") " pod="openstack/ovn-controller-jv694" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.155015 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/7e5eabb2-229a-4d75-b62d-65be688f753a-etc-ovs\") pod \"ovn-controller-ovs-s5xbr\" (UID: \"7e5eabb2-229a-4d75-b62d-65be688f753a\") " pod="openstack/ovn-controller-ovs-s5xbr" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.155034 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7e5eabb2-229a-4d75-b62d-65be688f753a-var-run\") pod \"ovn-controller-ovs-s5xbr\" (UID: \"7e5eabb2-229a-4d75-b62d-65be688f753a\") " pod="openstack/ovn-controller-ovs-s5xbr" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.154978 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/7e5eabb2-229a-4d75-b62d-65be688f753a-var-log\") pod \"ovn-controller-ovs-s5xbr\" (UID: \"7e5eabb2-229a-4d75-b62d-65be688f753a\") " pod="openstack/ovn-controller-ovs-s5xbr" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.154978 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/febb140e-d26e-43db-9924-0f06739b9a4a-var-log-ovn\") pod \"ovn-controller-jv694\" (UID: \"febb140e-d26e-43db-9924-0f06739b9a4a\") " pod="openstack/ovn-controller-jv694" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.154933 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/febb140e-d26e-43db-9924-0f06739b9a4a-var-run-ovn\") pod \"ovn-controller-jv694\" (UID: \"febb140e-d26e-43db-9924-0f06739b9a4a\") " pod="openstack/ovn-controller-jv694" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.155040 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/7e5eabb2-229a-4d75-b62d-65be688f753a-var-lib\") pod \"ovn-controller-ovs-s5xbr\" (UID: \"7e5eabb2-229a-4d75-b62d-65be688f753a\") " pod="openstack/ovn-controller-ovs-s5xbr" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.156869 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/febb140e-d26e-43db-9924-0f06739b9a4a-scripts\") pod \"ovn-controller-jv694\" (UID: \"febb140e-d26e-43db-9924-0f06739b9a4a\") " pod="openstack/ovn-controller-jv694" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.157248 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e5eabb2-229a-4d75-b62d-65be688f753a-scripts\") pod \"ovn-controller-ovs-s5xbr\" (UID: \"7e5eabb2-229a-4d75-b62d-65be688f753a\") " pod="openstack/ovn-controller-ovs-s5xbr" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.177261 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5bh9\" (UniqueName: \"kubernetes.io/projected/febb140e-d26e-43db-9924-0f06739b9a4a-kube-api-access-p5bh9\") pod \"ovn-controller-jv694\" (UID: \"febb140e-d26e-43db-9924-0f06739b9a4a\") " pod="openstack/ovn-controller-jv694" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.177398 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvxsl\" (UniqueName: \"kubernetes.io/projected/7e5eabb2-229a-4d75-b62d-65be688f753a-kube-api-access-xvxsl\") pod \"ovn-controller-ovs-s5xbr\" (UID: \"7e5eabb2-229a-4d75-b62d-65be688f753a\") " pod="openstack/ovn-controller-ovs-s5xbr" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.326185 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-s5xbr" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.336978 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jv694" Jan 27 16:43:48 crc kubenswrapper[4772]: I0127 16:43:48.978619 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jv694"] Jan 27 16:43:49 crc kubenswrapper[4772]: I0127 16:43:49.227344 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-s5xbr"] Jan 27 16:43:49 crc kubenswrapper[4772]: W0127 16:43:49.242107 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e5eabb2_229a_4d75_b62d_65be688f753a.slice/crio-5a385b760c7c2a71eac14dab5e06934c0c974193b8985265d741d828f09cd81e WatchSource:0}: Error finding container 5a385b760c7c2a71eac14dab5e06934c0c974193b8985265d741d828f09cd81e: Status 404 returned error can't find the container with id 5a385b760c7c2a71eac14dab5e06934c0c974193b8985265d741d828f09cd81e Jan 27 16:43:49 crc kubenswrapper[4772]: I0127 16:43:49.525238 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-s5xbr" event={"ID":"7e5eabb2-229a-4d75-b62d-65be688f753a","Type":"ContainerStarted","Data":"5a385b760c7c2a71eac14dab5e06934c0c974193b8985265d741d828f09cd81e"} Jan 27 16:43:49 crc kubenswrapper[4772]: I0127 16:43:49.527255 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jv694" event={"ID":"febb140e-d26e-43db-9924-0f06739b9a4a","Type":"ContainerStarted","Data":"dd533541b3646484c7bf20aa3cbde14d63fba54329547e4d5b3fb8effd1c14d4"} Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.491145 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-sx2ff"] Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.492597 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-sx2ff" Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.494864 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.499423 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d12bbd4-3d0b-444b-a462-b620a7a5d73d-config\") pod \"ovn-controller-metrics-sx2ff\" (UID: \"3d12bbd4-3d0b-444b-a462-b620a7a5d73d\") " pod="openstack/ovn-controller-metrics-sx2ff" Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.499473 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3d12bbd4-3d0b-444b-a462-b620a7a5d73d-ovn-rundir\") pod \"ovn-controller-metrics-sx2ff\" (UID: \"3d12bbd4-3d0b-444b-a462-b620a7a5d73d\") " pod="openstack/ovn-controller-metrics-sx2ff" Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.499507 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3d12bbd4-3d0b-444b-a462-b620a7a5d73d-ovs-rundir\") pod \"ovn-controller-metrics-sx2ff\" (UID: \"3d12bbd4-3d0b-444b-a462-b620a7a5d73d\") " pod="openstack/ovn-controller-metrics-sx2ff" Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.499546 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj7wr\" (UniqueName: \"kubernetes.io/projected/3d12bbd4-3d0b-444b-a462-b620a7a5d73d-kube-api-access-mj7wr\") pod \"ovn-controller-metrics-sx2ff\" (UID: \"3d12bbd4-3d0b-444b-a462-b620a7a5d73d\") " pod="openstack/ovn-controller-metrics-sx2ff" Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.512864 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-sx2ff"] Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.541828 4772 generic.go:334] "Generic (PLEG): container finished" podID="7e5eabb2-229a-4d75-b62d-65be688f753a" containerID="e9cc226b04b7fdd0ec9bbe67c4798430f89dc91b6835ece11ecc4a31026a5e90" exitCode=0 Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.542152 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-s5xbr" event={"ID":"7e5eabb2-229a-4d75-b62d-65be688f753a","Type":"ContainerDied","Data":"e9cc226b04b7fdd0ec9bbe67c4798430f89dc91b6835ece11ecc4a31026a5e90"} Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.544982 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jv694" event={"ID":"febb140e-d26e-43db-9924-0f06739b9a4a","Type":"ContainerStarted","Data":"95a19528aa0ed75e0d341a0f15c898fc3ac16cf14ea038307b7611297762c3e3"} Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.545293 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-jv694" Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.601367 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d12bbd4-3d0b-444b-a462-b620a7a5d73d-config\") pod \"ovn-controller-metrics-sx2ff\" (UID: \"3d12bbd4-3d0b-444b-a462-b620a7a5d73d\") " pod="openstack/ovn-controller-metrics-sx2ff" Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.601433 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3d12bbd4-3d0b-444b-a462-b620a7a5d73d-ovn-rundir\") pod \"ovn-controller-metrics-sx2ff\" (UID: \"3d12bbd4-3d0b-444b-a462-b620a7a5d73d\") " pod="openstack/ovn-controller-metrics-sx2ff" Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.601471 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3d12bbd4-3d0b-444b-a462-b620a7a5d73d-ovs-rundir\") pod \"ovn-controller-metrics-sx2ff\" (UID: \"3d12bbd4-3d0b-444b-a462-b620a7a5d73d\") " pod="openstack/ovn-controller-metrics-sx2ff" Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.601538 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj7wr\" (UniqueName: \"kubernetes.io/projected/3d12bbd4-3d0b-444b-a462-b620a7a5d73d-kube-api-access-mj7wr\") pod \"ovn-controller-metrics-sx2ff\" (UID: \"3d12bbd4-3d0b-444b-a462-b620a7a5d73d\") " pod="openstack/ovn-controller-metrics-sx2ff" Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.603453 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3d12bbd4-3d0b-444b-a462-b620a7a5d73d-ovn-rundir\") pod \"ovn-controller-metrics-sx2ff\" (UID: \"3d12bbd4-3d0b-444b-a462-b620a7a5d73d\") " pod="openstack/ovn-controller-metrics-sx2ff" Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.603516 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3d12bbd4-3d0b-444b-a462-b620a7a5d73d-ovs-rundir\") pod \"ovn-controller-metrics-sx2ff\" (UID: \"3d12bbd4-3d0b-444b-a462-b620a7a5d73d\") " pod="openstack/ovn-controller-metrics-sx2ff" Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.605983 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d12bbd4-3d0b-444b-a462-b620a7a5d73d-config\") pod \"ovn-controller-metrics-sx2ff\" (UID: \"3d12bbd4-3d0b-444b-a462-b620a7a5d73d\") " pod="openstack/ovn-controller-metrics-sx2ff" Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.607726 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-jv694" podStartSLOduration=3.607707312 podStartE2EDuration="3.607707312s" podCreationTimestamp="2026-01-27 16:43:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:43:50.600101355 +0000 UTC m=+5816.580710473" watchObservedRunningTime="2026-01-27 16:43:50.607707312 +0000 UTC m=+5816.588316410" Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.638290 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj7wr\" (UniqueName: \"kubernetes.io/projected/3d12bbd4-3d0b-444b-a462-b620a7a5d73d-kube-api-access-mj7wr\") pod \"ovn-controller-metrics-sx2ff\" (UID: \"3d12bbd4-3d0b-444b-a462-b620a7a5d73d\") " pod="openstack/ovn-controller-metrics-sx2ff" Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.690230 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-create-fw49r"] Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.691831 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-fw49r" Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.711924 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-fw49r"] Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.805001 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ccfn\" (UniqueName: \"kubernetes.io/projected/6985c580-efad-46fe-8e20-9f932ce3af7d-kube-api-access-9ccfn\") pod \"octavia-db-create-fw49r\" (UID: \"6985c580-efad-46fe-8e20-9f932ce3af7d\") " pod="openstack/octavia-db-create-fw49r" Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.805400 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6985c580-efad-46fe-8e20-9f932ce3af7d-operator-scripts\") pod \"octavia-db-create-fw49r\" (UID: \"6985c580-efad-46fe-8e20-9f932ce3af7d\") " pod="openstack/octavia-db-create-fw49r" Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.826536 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-sx2ff" Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.907239 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6985c580-efad-46fe-8e20-9f932ce3af7d-operator-scripts\") pod \"octavia-db-create-fw49r\" (UID: \"6985c580-efad-46fe-8e20-9f932ce3af7d\") " pod="openstack/octavia-db-create-fw49r" Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.907423 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ccfn\" (UniqueName: \"kubernetes.io/projected/6985c580-efad-46fe-8e20-9f932ce3af7d-kube-api-access-9ccfn\") pod \"octavia-db-create-fw49r\" (UID: \"6985c580-efad-46fe-8e20-9f932ce3af7d\") " pod="openstack/octavia-db-create-fw49r" Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.909344 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6985c580-efad-46fe-8e20-9f932ce3af7d-operator-scripts\") pod \"octavia-db-create-fw49r\" (UID: \"6985c580-efad-46fe-8e20-9f932ce3af7d\") " pod="openstack/octavia-db-create-fw49r" Jan 27 16:43:50 crc kubenswrapper[4772]: I0127 16:43:50.928988 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ccfn\" (UniqueName: \"kubernetes.io/projected/6985c580-efad-46fe-8e20-9f932ce3af7d-kube-api-access-9ccfn\") pod \"octavia-db-create-fw49r\" (UID: \"6985c580-efad-46fe-8e20-9f932ce3af7d\") " pod="openstack/octavia-db-create-fw49r" Jan 27 16:43:51 crc kubenswrapper[4772]: I0127 16:43:51.017523 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-fw49r" Jan 27 16:43:51 crc kubenswrapper[4772]: I0127 16:43:51.286161 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-sx2ff"] Jan 27 16:43:51 crc kubenswrapper[4772]: I0127 16:43:51.483771 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-fw49r"] Jan 27 16:43:51 crc kubenswrapper[4772]: W0127 16:43:51.486953 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6985c580_efad_46fe_8e20_9f932ce3af7d.slice/crio-c36e42e52f193c7a6be7631f052325519c9df5bcaa5b03bea3f71227fc45caeb WatchSource:0}: Error finding container c36e42e52f193c7a6be7631f052325519c9df5bcaa5b03bea3f71227fc45caeb: Status 404 returned error can't find the container with id c36e42e52f193c7a6be7631f052325519c9df5bcaa5b03bea3f71227fc45caeb Jan 27 16:43:51 crc kubenswrapper[4772]: I0127 16:43:51.559746 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-s5xbr" event={"ID":"7e5eabb2-229a-4d75-b62d-65be688f753a","Type":"ContainerStarted","Data":"ddaa40cdf12e5368b170788f23af626c70d6c92a1e1f8bea4c264cf7a5e59c60"} Jan 27 16:43:51 crc kubenswrapper[4772]: I0127 16:43:51.560074 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-s5xbr" Jan 27 16:43:51 crc kubenswrapper[4772]: I0127 16:43:51.560091 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-s5xbr" event={"ID":"7e5eabb2-229a-4d75-b62d-65be688f753a","Type":"ContainerStarted","Data":"9a5a9cf84dee4ca22e579aea9eb7f9bdecc66933783caa6f5694692162c4c67b"} Jan 27 16:43:51 crc kubenswrapper[4772]: I0127 16:43:51.560104 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-s5xbr" Jan 27 16:43:51 crc kubenswrapper[4772]: I0127 16:43:51.563000 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-fw49r" event={"ID":"6985c580-efad-46fe-8e20-9f932ce3af7d","Type":"ContainerStarted","Data":"c36e42e52f193c7a6be7631f052325519c9df5bcaa5b03bea3f71227fc45caeb"} Jan 27 16:43:51 crc kubenswrapper[4772]: I0127 16:43:51.564598 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-sx2ff" event={"ID":"3d12bbd4-3d0b-444b-a462-b620a7a5d73d","Type":"ContainerStarted","Data":"56265457e5f04d0ed907eafc6fe108ad439fc0cc44816cbf12164bb2fecad158"} Jan 27 16:43:51 crc kubenswrapper[4772]: I0127 16:43:51.564623 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-sx2ff" event={"ID":"3d12bbd4-3d0b-444b-a462-b620a7a5d73d","Type":"ContainerStarted","Data":"e8d8314c18e87a69cb5868bf30d42a1cdbe11175a97f8adcf90b32a51c767f89"} Jan 27 16:43:51 crc kubenswrapper[4772]: I0127 16:43:51.585293 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-s5xbr" podStartSLOduration=4.585271596 podStartE2EDuration="4.585271596s" podCreationTimestamp="2026-01-27 16:43:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:43:51.581885049 +0000 UTC m=+5817.562494157" watchObservedRunningTime="2026-01-27 16:43:51.585271596 +0000 UTC m=+5817.565880704" Jan 27 16:43:51 crc kubenswrapper[4772]: I0127 16:43:51.611017 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-sx2ff" podStartSLOduration=1.6110012889999998 podStartE2EDuration="1.611001289s" podCreationTimestamp="2026-01-27 16:43:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:43:51.605314457 +0000 UTC m=+5817.585923555" watchObservedRunningTime="2026-01-27 16:43:51.611001289 +0000 UTC m=+5817.591610377" Jan 27 16:43:52 crc kubenswrapper[4772]: I0127 16:43:52.029386 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-8lxch"] Jan 27 16:43:52 crc kubenswrapper[4772]: I0127 16:43:52.039614 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-8lxch"] Jan 27 16:43:52 crc kubenswrapper[4772]: I0127 16:43:52.406603 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-c028-account-create-update-x9k2v"] Jan 27 16:43:52 crc kubenswrapper[4772]: I0127 16:43:52.408100 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-c028-account-create-update-x9k2v" Jan 27 16:43:52 crc kubenswrapper[4772]: I0127 16:43:52.411080 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-db-secret" Jan 27 16:43:52 crc kubenswrapper[4772]: I0127 16:43:52.421246 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-c028-account-create-update-x9k2v"] Jan 27 16:43:52 crc kubenswrapper[4772]: I0127 16:43:52.545002 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1cd343bc-4a97-4ad2-aa82-12ea527398d8-operator-scripts\") pod \"octavia-c028-account-create-update-x9k2v\" (UID: \"1cd343bc-4a97-4ad2-aa82-12ea527398d8\") " pod="openstack/octavia-c028-account-create-update-x9k2v" Jan 27 16:43:52 crc kubenswrapper[4772]: I0127 16:43:52.545074 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt2lh\" (UniqueName: \"kubernetes.io/projected/1cd343bc-4a97-4ad2-aa82-12ea527398d8-kube-api-access-mt2lh\") pod \"octavia-c028-account-create-update-x9k2v\" (UID: \"1cd343bc-4a97-4ad2-aa82-12ea527398d8\") " pod="openstack/octavia-c028-account-create-update-x9k2v" Jan 27 16:43:52 crc kubenswrapper[4772]: I0127 16:43:52.578598 4772 generic.go:334] "Generic (PLEG): container finished" podID="6985c580-efad-46fe-8e20-9f932ce3af7d" containerID="742607c6bfa7c1d2a5f8ceee5a81852f3b14a71713e837fcc9210fdc6dccc556" exitCode=0 Jan 27 16:43:52 crc kubenswrapper[4772]: I0127 16:43:52.578660 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-fw49r" event={"ID":"6985c580-efad-46fe-8e20-9f932ce3af7d","Type":"ContainerDied","Data":"742607c6bfa7c1d2a5f8ceee5a81852f3b14a71713e837fcc9210fdc6dccc556"} Jan 27 16:43:52 crc kubenswrapper[4772]: I0127 16:43:52.646832 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt2lh\" (UniqueName: \"kubernetes.io/projected/1cd343bc-4a97-4ad2-aa82-12ea527398d8-kube-api-access-mt2lh\") pod \"octavia-c028-account-create-update-x9k2v\" (UID: \"1cd343bc-4a97-4ad2-aa82-12ea527398d8\") " pod="openstack/octavia-c028-account-create-update-x9k2v" Jan 27 16:43:52 crc kubenswrapper[4772]: I0127 16:43:52.646884 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1cd343bc-4a97-4ad2-aa82-12ea527398d8-operator-scripts\") pod \"octavia-c028-account-create-update-x9k2v\" (UID: \"1cd343bc-4a97-4ad2-aa82-12ea527398d8\") " pod="openstack/octavia-c028-account-create-update-x9k2v" Jan 27 16:43:52 crc kubenswrapper[4772]: I0127 16:43:52.647673 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1cd343bc-4a97-4ad2-aa82-12ea527398d8-operator-scripts\") pod \"octavia-c028-account-create-update-x9k2v\" (UID: \"1cd343bc-4a97-4ad2-aa82-12ea527398d8\") " pod="openstack/octavia-c028-account-create-update-x9k2v" Jan 27 16:43:52 crc kubenswrapper[4772]: I0127 16:43:52.673088 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt2lh\" (UniqueName: \"kubernetes.io/projected/1cd343bc-4a97-4ad2-aa82-12ea527398d8-kube-api-access-mt2lh\") pod \"octavia-c028-account-create-update-x9k2v\" (UID: \"1cd343bc-4a97-4ad2-aa82-12ea527398d8\") " pod="openstack/octavia-c028-account-create-update-x9k2v" Jan 27 16:43:52 crc kubenswrapper[4772]: I0127 16:43:52.677282 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4b64159-de94-4b31-85ce-b845fdb391b3" path="/var/lib/kubelet/pods/e4b64159-de94-4b31-85ce-b845fdb391b3/volumes" Jan 27 16:43:52 crc kubenswrapper[4772]: I0127 16:43:52.726048 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-c028-account-create-update-x9k2v" Jan 27 16:43:53 crc kubenswrapper[4772]: W0127 16:43:53.225019 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1cd343bc_4a97_4ad2_aa82_12ea527398d8.slice/crio-b8a814b0fd5cc8c4f7f3f644db99ccfb0e5e7d7ce60566fdb27aab472cf71cb1 WatchSource:0}: Error finding container b8a814b0fd5cc8c4f7f3f644db99ccfb0e5e7d7ce60566fdb27aab472cf71cb1: Status 404 returned error can't find the container with id b8a814b0fd5cc8c4f7f3f644db99ccfb0e5e7d7ce60566fdb27aab472cf71cb1 Jan 27 16:43:53 crc kubenswrapper[4772]: I0127 16:43:53.225111 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-c028-account-create-update-x9k2v"] Jan 27 16:43:53 crc kubenswrapper[4772]: I0127 16:43:53.587765 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-c028-account-create-update-x9k2v" event={"ID":"1cd343bc-4a97-4ad2-aa82-12ea527398d8","Type":"ContainerStarted","Data":"5ed8b1fb179ec96a663871b1030549de3060884fefe55bd4b363bd01934e9e74"} Jan 27 16:43:53 crc kubenswrapper[4772]: I0127 16:43:53.588043 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-c028-account-create-update-x9k2v" event={"ID":"1cd343bc-4a97-4ad2-aa82-12ea527398d8","Type":"ContainerStarted","Data":"b8a814b0fd5cc8c4f7f3f644db99ccfb0e5e7d7ce60566fdb27aab472cf71cb1"} Jan 27 16:43:53 crc kubenswrapper[4772]: I0127 16:43:53.604391 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-c028-account-create-update-x9k2v" podStartSLOduration=1.604367918 podStartE2EDuration="1.604367918s" podCreationTimestamp="2026-01-27 16:43:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:43:53.600711224 +0000 UTC m=+5819.581320342" watchObservedRunningTime="2026-01-27 16:43:53.604367918 +0000 UTC m=+5819.584977016" Jan 27 16:43:53 crc kubenswrapper[4772]: I0127 16:43:53.664925 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:43:53 crc kubenswrapper[4772]: E0127 16:43:53.667741 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:43:53 crc kubenswrapper[4772]: I0127 16:43:53.932835 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-fw49r" Jan 27 16:43:54 crc kubenswrapper[4772]: I0127 16:43:54.077356 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6985c580-efad-46fe-8e20-9f932ce3af7d-operator-scripts\") pod \"6985c580-efad-46fe-8e20-9f932ce3af7d\" (UID: \"6985c580-efad-46fe-8e20-9f932ce3af7d\") " Jan 27 16:43:54 crc kubenswrapper[4772]: I0127 16:43:54.077608 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ccfn\" (UniqueName: \"kubernetes.io/projected/6985c580-efad-46fe-8e20-9f932ce3af7d-kube-api-access-9ccfn\") pod \"6985c580-efad-46fe-8e20-9f932ce3af7d\" (UID: \"6985c580-efad-46fe-8e20-9f932ce3af7d\") " Jan 27 16:43:54 crc kubenswrapper[4772]: I0127 16:43:54.078678 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6985c580-efad-46fe-8e20-9f932ce3af7d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6985c580-efad-46fe-8e20-9f932ce3af7d" (UID: "6985c580-efad-46fe-8e20-9f932ce3af7d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:43:54 crc kubenswrapper[4772]: I0127 16:43:54.082505 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6985c580-efad-46fe-8e20-9f932ce3af7d-kube-api-access-9ccfn" (OuterVolumeSpecName: "kube-api-access-9ccfn") pod "6985c580-efad-46fe-8e20-9f932ce3af7d" (UID: "6985c580-efad-46fe-8e20-9f932ce3af7d"). InnerVolumeSpecName "kube-api-access-9ccfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:43:54 crc kubenswrapper[4772]: I0127 16:43:54.180573 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6985c580-efad-46fe-8e20-9f932ce3af7d-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:43:54 crc kubenswrapper[4772]: I0127 16:43:54.180667 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ccfn\" (UniqueName: \"kubernetes.io/projected/6985c580-efad-46fe-8e20-9f932ce3af7d-kube-api-access-9ccfn\") on node \"crc\" DevicePath \"\"" Jan 27 16:43:54 crc kubenswrapper[4772]: I0127 16:43:54.606649 4772 generic.go:334] "Generic (PLEG): container finished" podID="1cd343bc-4a97-4ad2-aa82-12ea527398d8" containerID="5ed8b1fb179ec96a663871b1030549de3060884fefe55bd4b363bd01934e9e74" exitCode=0 Jan 27 16:43:54 crc kubenswrapper[4772]: I0127 16:43:54.607108 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-c028-account-create-update-x9k2v" event={"ID":"1cd343bc-4a97-4ad2-aa82-12ea527398d8","Type":"ContainerDied","Data":"5ed8b1fb179ec96a663871b1030549de3060884fefe55bd4b363bd01934e9e74"} Jan 27 16:43:54 crc kubenswrapper[4772]: I0127 16:43:54.613661 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-fw49r" event={"ID":"6985c580-efad-46fe-8e20-9f932ce3af7d","Type":"ContainerDied","Data":"c36e42e52f193c7a6be7631f052325519c9df5bcaa5b03bea3f71227fc45caeb"} Jan 27 16:43:54 crc kubenswrapper[4772]: I0127 16:43:54.613708 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c36e42e52f193c7a6be7631f052325519c9df5bcaa5b03bea3f71227fc45caeb" Jan 27 16:43:54 crc kubenswrapper[4772]: I0127 16:43:54.613727 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-fw49r" Jan 27 16:43:55 crc kubenswrapper[4772]: I0127 16:43:55.989208 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-c028-account-create-update-x9k2v" Jan 27 16:43:56 crc kubenswrapper[4772]: I0127 16:43:56.115571 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1cd343bc-4a97-4ad2-aa82-12ea527398d8-operator-scripts\") pod \"1cd343bc-4a97-4ad2-aa82-12ea527398d8\" (UID: \"1cd343bc-4a97-4ad2-aa82-12ea527398d8\") " Jan 27 16:43:56 crc kubenswrapper[4772]: I0127 16:43:56.115754 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mt2lh\" (UniqueName: \"kubernetes.io/projected/1cd343bc-4a97-4ad2-aa82-12ea527398d8-kube-api-access-mt2lh\") pod \"1cd343bc-4a97-4ad2-aa82-12ea527398d8\" (UID: \"1cd343bc-4a97-4ad2-aa82-12ea527398d8\") " Jan 27 16:43:56 crc kubenswrapper[4772]: I0127 16:43:56.116027 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cd343bc-4a97-4ad2-aa82-12ea527398d8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1cd343bc-4a97-4ad2-aa82-12ea527398d8" (UID: "1cd343bc-4a97-4ad2-aa82-12ea527398d8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:43:56 crc kubenswrapper[4772]: I0127 16:43:56.116160 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1cd343bc-4a97-4ad2-aa82-12ea527398d8-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:43:56 crc kubenswrapper[4772]: I0127 16:43:56.121553 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cd343bc-4a97-4ad2-aa82-12ea527398d8-kube-api-access-mt2lh" (OuterVolumeSpecName: "kube-api-access-mt2lh") pod "1cd343bc-4a97-4ad2-aa82-12ea527398d8" (UID: "1cd343bc-4a97-4ad2-aa82-12ea527398d8"). InnerVolumeSpecName "kube-api-access-mt2lh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:43:56 crc kubenswrapper[4772]: I0127 16:43:56.219473 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mt2lh\" (UniqueName: \"kubernetes.io/projected/1cd343bc-4a97-4ad2-aa82-12ea527398d8-kube-api-access-mt2lh\") on node \"crc\" DevicePath \"\"" Jan 27 16:43:56 crc kubenswrapper[4772]: I0127 16:43:56.634101 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-c028-account-create-update-x9k2v" event={"ID":"1cd343bc-4a97-4ad2-aa82-12ea527398d8","Type":"ContainerDied","Data":"b8a814b0fd5cc8c4f7f3f644db99ccfb0e5e7d7ce60566fdb27aab472cf71cb1"} Jan 27 16:43:56 crc kubenswrapper[4772]: I0127 16:43:56.634440 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8a814b0fd5cc8c4f7f3f644db99ccfb0e5e7d7ce60566fdb27aab472cf71cb1" Jan 27 16:43:56 crc kubenswrapper[4772]: I0127 16:43:56.634189 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-c028-account-create-update-x9k2v" Jan 27 16:43:58 crc kubenswrapper[4772]: I0127 16:43:58.451820 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-persistence-db-create-4mfnz"] Jan 27 16:43:58 crc kubenswrapper[4772]: E0127 16:43:58.452213 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6985c580-efad-46fe-8e20-9f932ce3af7d" containerName="mariadb-database-create" Jan 27 16:43:58 crc kubenswrapper[4772]: I0127 16:43:58.452225 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="6985c580-efad-46fe-8e20-9f932ce3af7d" containerName="mariadb-database-create" Jan 27 16:43:58 crc kubenswrapper[4772]: E0127 16:43:58.452257 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cd343bc-4a97-4ad2-aa82-12ea527398d8" containerName="mariadb-account-create-update" Jan 27 16:43:58 crc kubenswrapper[4772]: I0127 16:43:58.452263 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cd343bc-4a97-4ad2-aa82-12ea527398d8" containerName="mariadb-account-create-update" Jan 27 16:43:58 crc kubenswrapper[4772]: I0127 16:43:58.452440 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="6985c580-efad-46fe-8e20-9f932ce3af7d" containerName="mariadb-database-create" Jan 27 16:43:58 crc kubenswrapper[4772]: I0127 16:43:58.452459 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cd343bc-4a97-4ad2-aa82-12ea527398d8" containerName="mariadb-account-create-update" Jan 27 16:43:58 crc kubenswrapper[4772]: I0127 16:43:58.453046 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-4mfnz" Jan 27 16:43:58 crc kubenswrapper[4772]: I0127 16:43:58.470834 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-4mfnz"] Jan 27 16:43:58 crc kubenswrapper[4772]: I0127 16:43:58.566324 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xv4r4\" (UniqueName: \"kubernetes.io/projected/c203bc37-6216-4aa6-8a51-1e0a2f01bb43-kube-api-access-xv4r4\") pod \"octavia-persistence-db-create-4mfnz\" (UID: \"c203bc37-6216-4aa6-8a51-1e0a2f01bb43\") " pod="openstack/octavia-persistence-db-create-4mfnz" Jan 27 16:43:58 crc kubenswrapper[4772]: I0127 16:43:58.566386 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c203bc37-6216-4aa6-8a51-1e0a2f01bb43-operator-scripts\") pod \"octavia-persistence-db-create-4mfnz\" (UID: \"c203bc37-6216-4aa6-8a51-1e0a2f01bb43\") " pod="openstack/octavia-persistence-db-create-4mfnz" Jan 27 16:43:58 crc kubenswrapper[4772]: I0127 16:43:58.668541 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xv4r4\" (UniqueName: \"kubernetes.io/projected/c203bc37-6216-4aa6-8a51-1e0a2f01bb43-kube-api-access-xv4r4\") pod \"octavia-persistence-db-create-4mfnz\" (UID: \"c203bc37-6216-4aa6-8a51-1e0a2f01bb43\") " pod="openstack/octavia-persistence-db-create-4mfnz" Jan 27 16:43:58 crc kubenswrapper[4772]: I0127 16:43:58.668595 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c203bc37-6216-4aa6-8a51-1e0a2f01bb43-operator-scripts\") pod \"octavia-persistence-db-create-4mfnz\" (UID: \"c203bc37-6216-4aa6-8a51-1e0a2f01bb43\") " pod="openstack/octavia-persistence-db-create-4mfnz" Jan 27 16:43:58 crc kubenswrapper[4772]: I0127 16:43:58.670400 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c203bc37-6216-4aa6-8a51-1e0a2f01bb43-operator-scripts\") pod \"octavia-persistence-db-create-4mfnz\" (UID: \"c203bc37-6216-4aa6-8a51-1e0a2f01bb43\") " pod="openstack/octavia-persistence-db-create-4mfnz" Jan 27 16:43:58 crc kubenswrapper[4772]: I0127 16:43:58.699560 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xv4r4\" (UniqueName: \"kubernetes.io/projected/c203bc37-6216-4aa6-8a51-1e0a2f01bb43-kube-api-access-xv4r4\") pod \"octavia-persistence-db-create-4mfnz\" (UID: \"c203bc37-6216-4aa6-8a51-1e0a2f01bb43\") " pod="openstack/octavia-persistence-db-create-4mfnz" Jan 27 16:43:58 crc kubenswrapper[4772]: I0127 16:43:58.772602 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-4mfnz" Jan 27 16:43:58 crc kubenswrapper[4772]: I0127 16:43:58.934993 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-c1c0-account-create-update-cjqxk"] Jan 27 16:43:58 crc kubenswrapper[4772]: I0127 16:43:58.936867 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-c1c0-account-create-update-cjqxk" Jan 27 16:43:58 crc kubenswrapper[4772]: I0127 16:43:58.938714 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-persistence-db-secret" Jan 27 16:43:58 crc kubenswrapper[4772]: I0127 16:43:58.946626 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-c1c0-account-create-update-cjqxk"] Jan 27 16:43:58 crc kubenswrapper[4772]: I0127 16:43:58.974976 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnn2n\" (UniqueName: \"kubernetes.io/projected/7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4-kube-api-access-qnn2n\") pod \"octavia-c1c0-account-create-update-cjqxk\" (UID: \"7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4\") " pod="openstack/octavia-c1c0-account-create-update-cjqxk" Jan 27 16:43:58 crc kubenswrapper[4772]: I0127 16:43:58.975316 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4-operator-scripts\") pod \"octavia-c1c0-account-create-update-cjqxk\" (UID: \"7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4\") " pod="openstack/octavia-c1c0-account-create-update-cjqxk" Jan 27 16:43:59 crc kubenswrapper[4772]: I0127 16:43:59.076988 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnn2n\" (UniqueName: \"kubernetes.io/projected/7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4-kube-api-access-qnn2n\") pod \"octavia-c1c0-account-create-update-cjqxk\" (UID: \"7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4\") " pod="openstack/octavia-c1c0-account-create-update-cjqxk" Jan 27 16:43:59 crc kubenswrapper[4772]: I0127 16:43:59.077177 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4-operator-scripts\") pod \"octavia-c1c0-account-create-update-cjqxk\" (UID: \"7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4\") " pod="openstack/octavia-c1c0-account-create-update-cjqxk" Jan 27 16:43:59 crc kubenswrapper[4772]: I0127 16:43:59.077846 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4-operator-scripts\") pod \"octavia-c1c0-account-create-update-cjqxk\" (UID: \"7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4\") " pod="openstack/octavia-c1c0-account-create-update-cjqxk" Jan 27 16:43:59 crc kubenswrapper[4772]: I0127 16:43:59.094132 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnn2n\" (UniqueName: \"kubernetes.io/projected/7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4-kube-api-access-qnn2n\") pod \"octavia-c1c0-account-create-update-cjqxk\" (UID: \"7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4\") " pod="openstack/octavia-c1c0-account-create-update-cjqxk" Jan 27 16:43:59 crc kubenswrapper[4772]: I0127 16:43:59.238795 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-4mfnz"] Jan 27 16:43:59 crc kubenswrapper[4772]: I0127 16:43:59.256462 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-c1c0-account-create-update-cjqxk" Jan 27 16:43:59 crc kubenswrapper[4772]: I0127 16:43:59.681442 4772 generic.go:334] "Generic (PLEG): container finished" podID="c203bc37-6216-4aa6-8a51-1e0a2f01bb43" containerID="ce5afc895546186c87fa545a3ff11c6c821cf7ba305b10a05525fc458e3f4be7" exitCode=0 Jan 27 16:43:59 crc kubenswrapper[4772]: I0127 16:43:59.681685 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-4mfnz" event={"ID":"c203bc37-6216-4aa6-8a51-1e0a2f01bb43","Type":"ContainerDied","Data":"ce5afc895546186c87fa545a3ff11c6c821cf7ba305b10a05525fc458e3f4be7"} Jan 27 16:43:59 crc kubenswrapper[4772]: I0127 16:43:59.681709 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-4mfnz" event={"ID":"c203bc37-6216-4aa6-8a51-1e0a2f01bb43","Type":"ContainerStarted","Data":"80faa54f23772f274c42b0c93fe478a812eadfb51f692f0e40220291b27183c9"} Jan 27 16:43:59 crc kubenswrapper[4772]: I0127 16:43:59.743986 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-c1c0-account-create-update-cjqxk"] Jan 27 16:43:59 crc kubenswrapper[4772]: W0127 16:43:59.756556 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ff09ddf_04fd_42e9_b6fd_8c9fd9fac0e4.slice/crio-f3e01eab28c27034647a68108d237eb65cb31397f37dce39098e0a9f49ce0695 WatchSource:0}: Error finding container f3e01eab28c27034647a68108d237eb65cb31397f37dce39098e0a9f49ce0695: Status 404 returned error can't find the container with id f3e01eab28c27034647a68108d237eb65cb31397f37dce39098e0a9f49ce0695 Jan 27 16:44:00 crc kubenswrapper[4772]: I0127 16:44:00.693943 4772 generic.go:334] "Generic (PLEG): container finished" podID="7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4" containerID="5f81345bb4c2b5ce3e497a908767c4ee00a2cf35d4b6e4a9ea4e1fe6b2891391" exitCode=0 Jan 27 16:44:00 crc kubenswrapper[4772]: I0127 16:44:00.694027 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-c1c0-account-create-update-cjqxk" event={"ID":"7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4","Type":"ContainerDied","Data":"5f81345bb4c2b5ce3e497a908767c4ee00a2cf35d4b6e4a9ea4e1fe6b2891391"} Jan 27 16:44:00 crc kubenswrapper[4772]: I0127 16:44:00.694063 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-c1c0-account-create-update-cjqxk" event={"ID":"7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4","Type":"ContainerStarted","Data":"f3e01eab28c27034647a68108d237eb65cb31397f37dce39098e0a9f49ce0695"} Jan 27 16:44:01 crc kubenswrapper[4772]: I0127 16:44:01.055132 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-4mfnz" Jan 27 16:44:01 crc kubenswrapper[4772]: I0127 16:44:01.115845 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xv4r4\" (UniqueName: \"kubernetes.io/projected/c203bc37-6216-4aa6-8a51-1e0a2f01bb43-kube-api-access-xv4r4\") pod \"c203bc37-6216-4aa6-8a51-1e0a2f01bb43\" (UID: \"c203bc37-6216-4aa6-8a51-1e0a2f01bb43\") " Jan 27 16:44:01 crc kubenswrapper[4772]: I0127 16:44:01.116092 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c203bc37-6216-4aa6-8a51-1e0a2f01bb43-operator-scripts\") pod \"c203bc37-6216-4aa6-8a51-1e0a2f01bb43\" (UID: \"c203bc37-6216-4aa6-8a51-1e0a2f01bb43\") " Jan 27 16:44:01 crc kubenswrapper[4772]: I0127 16:44:01.117143 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c203bc37-6216-4aa6-8a51-1e0a2f01bb43-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c203bc37-6216-4aa6-8a51-1e0a2f01bb43" (UID: "c203bc37-6216-4aa6-8a51-1e0a2f01bb43"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:44:01 crc kubenswrapper[4772]: I0127 16:44:01.122813 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c203bc37-6216-4aa6-8a51-1e0a2f01bb43-kube-api-access-xv4r4" (OuterVolumeSpecName: "kube-api-access-xv4r4") pod "c203bc37-6216-4aa6-8a51-1e0a2f01bb43" (UID: "c203bc37-6216-4aa6-8a51-1e0a2f01bb43"). InnerVolumeSpecName "kube-api-access-xv4r4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:44:01 crc kubenswrapper[4772]: I0127 16:44:01.218046 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c203bc37-6216-4aa6-8a51-1e0a2f01bb43-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:44:01 crc kubenswrapper[4772]: I0127 16:44:01.218083 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xv4r4\" (UniqueName: \"kubernetes.io/projected/c203bc37-6216-4aa6-8a51-1e0a2f01bb43-kube-api-access-xv4r4\") on node \"crc\" DevicePath \"\"" Jan 27 16:44:01 crc kubenswrapper[4772]: I0127 16:44:01.704367 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-4mfnz" event={"ID":"c203bc37-6216-4aa6-8a51-1e0a2f01bb43","Type":"ContainerDied","Data":"80faa54f23772f274c42b0c93fe478a812eadfb51f692f0e40220291b27183c9"} Jan 27 16:44:01 crc kubenswrapper[4772]: I0127 16:44:01.704441 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80faa54f23772f274c42b0c93fe478a812eadfb51f692f0e40220291b27183c9" Jan 27 16:44:01 crc kubenswrapper[4772]: I0127 16:44:01.704405 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-4mfnz" Jan 27 16:44:02 crc kubenswrapper[4772]: I0127 16:44:02.086191 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-c1c0-account-create-update-cjqxk" Jan 27 16:44:02 crc kubenswrapper[4772]: I0127 16:44:02.137467 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4-operator-scripts\") pod \"7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4\" (UID: \"7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4\") " Jan 27 16:44:02 crc kubenswrapper[4772]: I0127 16:44:02.137858 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnn2n\" (UniqueName: \"kubernetes.io/projected/7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4-kube-api-access-qnn2n\") pod \"7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4\" (UID: \"7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4\") " Jan 27 16:44:02 crc kubenswrapper[4772]: I0127 16:44:02.138601 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4" (UID: "7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:44:02 crc kubenswrapper[4772]: I0127 16:44:02.143505 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4-kube-api-access-qnn2n" (OuterVolumeSpecName: "kube-api-access-qnn2n") pod "7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4" (UID: "7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4"). InnerVolumeSpecName "kube-api-access-qnn2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:44:02 crc kubenswrapper[4772]: I0127 16:44:02.240161 4772 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:44:02 crc kubenswrapper[4772]: I0127 16:44:02.240218 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnn2n\" (UniqueName: \"kubernetes.io/projected/7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4-kube-api-access-qnn2n\") on node \"crc\" DevicePath \"\"" Jan 27 16:44:02 crc kubenswrapper[4772]: I0127 16:44:02.714958 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-c1c0-account-create-update-cjqxk" event={"ID":"7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4","Type":"ContainerDied","Data":"f3e01eab28c27034647a68108d237eb65cb31397f37dce39098e0a9f49ce0695"} Jan 27 16:44:02 crc kubenswrapper[4772]: I0127 16:44:02.714986 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-c1c0-account-create-update-cjqxk" Jan 27 16:44:02 crc kubenswrapper[4772]: I0127 16:44:02.714996 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3e01eab28c27034647a68108d237eb65cb31397f37dce39098e0a9f49ce0695" Jan 27 16:44:02 crc kubenswrapper[4772]: E0127 16:44:02.891105 4772 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ff09ddf_04fd_42e9_b6fd_8c9fd9fac0e4.slice\": RecentStats: unable to find data in memory cache]" Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.767087 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-api-6d96bf4746-x9c97"] Jan 27 16:44:04 crc kubenswrapper[4772]: E0127 16:44:04.767966 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4" containerName="mariadb-account-create-update" Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.767983 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4" containerName="mariadb-account-create-update" Jan 27 16:44:04 crc kubenswrapper[4772]: E0127 16:44:04.768019 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c203bc37-6216-4aa6-8a51-1e0a2f01bb43" containerName="mariadb-database-create" Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.768027 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="c203bc37-6216-4aa6-8a51-1e0a2f01bb43" containerName="mariadb-database-create" Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.768294 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4" containerName="mariadb-account-create-update" Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.768337 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="c203bc37-6216-4aa6-8a51-1e0a2f01bb43" containerName="mariadb-database-create" Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.769944 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-6d96bf4746-x9c97" Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.773668 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-octavia-dockercfg-bqhmr" Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.786432 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-scripts" Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.786972 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-config-data" Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.794443 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9371d269-02b3-4049-aeea-4fd56c648b89-config-data\") pod \"octavia-api-6d96bf4746-x9c97\" (UID: \"9371d269-02b3-4049-aeea-4fd56c648b89\") " pod="openstack/octavia-api-6d96bf4746-x9c97" Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.794526 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/9371d269-02b3-4049-aeea-4fd56c648b89-octavia-run\") pod \"octavia-api-6d96bf4746-x9c97\" (UID: \"9371d269-02b3-4049-aeea-4fd56c648b89\") " pod="openstack/octavia-api-6d96bf4746-x9c97" Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.794657 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/9371d269-02b3-4049-aeea-4fd56c648b89-config-data-merged\") pod \"octavia-api-6d96bf4746-x9c97\" (UID: \"9371d269-02b3-4049-aeea-4fd56c648b89\") " pod="openstack/octavia-api-6d96bf4746-x9c97" Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.794691 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9371d269-02b3-4049-aeea-4fd56c648b89-scripts\") pod \"octavia-api-6d96bf4746-x9c97\" (UID: \"9371d269-02b3-4049-aeea-4fd56c648b89\") " pod="openstack/octavia-api-6d96bf4746-x9c97" Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.794769 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9371d269-02b3-4049-aeea-4fd56c648b89-combined-ca-bundle\") pod \"octavia-api-6d96bf4746-x9c97\" (UID: \"9371d269-02b3-4049-aeea-4fd56c648b89\") " pod="openstack/octavia-api-6d96bf4746-x9c97" Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.822744 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-6d96bf4746-x9c97"] Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.896347 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9371d269-02b3-4049-aeea-4fd56c648b89-combined-ca-bundle\") pod \"octavia-api-6d96bf4746-x9c97\" (UID: \"9371d269-02b3-4049-aeea-4fd56c648b89\") " pod="openstack/octavia-api-6d96bf4746-x9c97" Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.896452 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9371d269-02b3-4049-aeea-4fd56c648b89-config-data\") pod \"octavia-api-6d96bf4746-x9c97\" (UID: \"9371d269-02b3-4049-aeea-4fd56c648b89\") " pod="openstack/octavia-api-6d96bf4746-x9c97" Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.896484 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/9371d269-02b3-4049-aeea-4fd56c648b89-octavia-run\") pod \"octavia-api-6d96bf4746-x9c97\" (UID: \"9371d269-02b3-4049-aeea-4fd56c648b89\") " pod="openstack/octavia-api-6d96bf4746-x9c97" Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.896625 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/9371d269-02b3-4049-aeea-4fd56c648b89-config-data-merged\") pod \"octavia-api-6d96bf4746-x9c97\" (UID: \"9371d269-02b3-4049-aeea-4fd56c648b89\") " pod="openstack/octavia-api-6d96bf4746-x9c97" Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.896651 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9371d269-02b3-4049-aeea-4fd56c648b89-scripts\") pod \"octavia-api-6d96bf4746-x9c97\" (UID: \"9371d269-02b3-4049-aeea-4fd56c648b89\") " pod="openstack/octavia-api-6d96bf4746-x9c97" Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.897807 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/9371d269-02b3-4049-aeea-4fd56c648b89-octavia-run\") pod \"octavia-api-6d96bf4746-x9c97\" (UID: \"9371d269-02b3-4049-aeea-4fd56c648b89\") " pod="openstack/octavia-api-6d96bf4746-x9c97" Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.897884 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/9371d269-02b3-4049-aeea-4fd56c648b89-config-data-merged\") pod \"octavia-api-6d96bf4746-x9c97\" (UID: \"9371d269-02b3-4049-aeea-4fd56c648b89\") " pod="openstack/octavia-api-6d96bf4746-x9c97" Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.902371 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9371d269-02b3-4049-aeea-4fd56c648b89-config-data\") pod \"octavia-api-6d96bf4746-x9c97\" (UID: \"9371d269-02b3-4049-aeea-4fd56c648b89\") " pod="openstack/octavia-api-6d96bf4746-x9c97" Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.904702 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9371d269-02b3-4049-aeea-4fd56c648b89-scripts\") pod \"octavia-api-6d96bf4746-x9c97\" (UID: \"9371d269-02b3-4049-aeea-4fd56c648b89\") " pod="openstack/octavia-api-6d96bf4746-x9c97" Jan 27 16:44:04 crc kubenswrapper[4772]: I0127 16:44:04.919661 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9371d269-02b3-4049-aeea-4fd56c648b89-combined-ca-bundle\") pod \"octavia-api-6d96bf4746-x9c97\" (UID: \"9371d269-02b3-4049-aeea-4fd56c648b89\") " pod="openstack/octavia-api-6d96bf4746-x9c97" Jan 27 16:44:05 crc kubenswrapper[4772]: I0127 16:44:05.102347 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-6d96bf4746-x9c97" Jan 27 16:44:05 crc kubenswrapper[4772]: I0127 16:44:05.637400 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 16:44:05 crc kubenswrapper[4772]: I0127 16:44:05.649405 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-6d96bf4746-x9c97"] Jan 27 16:44:05 crc kubenswrapper[4772]: I0127 16:44:05.663221 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:44:05 crc kubenswrapper[4772]: E0127 16:44:05.663507 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:44:05 crc kubenswrapper[4772]: I0127 16:44:05.739252 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-6d96bf4746-x9c97" event={"ID":"9371d269-02b3-4049-aeea-4fd56c648b89","Type":"ContainerStarted","Data":"aeee5e34a46468c554b59b9f0711b0d1ef508610bc8952cbbce4d2282ea800f3"} Jan 27 16:44:07 crc kubenswrapper[4772]: I0127 16:44:07.577247 4772 scope.go:117] "RemoveContainer" containerID="6b8b1c4b2ab6a42f7f2fdd3da73f914caa96bb6a3722955fd242601b775105ee" Jan 27 16:44:07 crc kubenswrapper[4772]: I0127 16:44:07.604286 4772 scope.go:117] "RemoveContainer" containerID="5344c5595e4c241d1f3af5be472eb435a5a87e2e74a29b46f8cee0d5b6b1c135" Jan 27 16:44:07 crc kubenswrapper[4772]: I0127 16:44:07.657065 4772 scope.go:117] "RemoveContainer" containerID="058f6eb2f4f5195120a96ef6b4691590bd77114e18b6ed23148cdd3864b329b4" Jan 27 16:44:07 crc kubenswrapper[4772]: I0127 16:44:07.680198 4772 scope.go:117] "RemoveContainer" containerID="761ce723d19214c2204c8e256e0ee21bf82feb31ce0c9f783c4a5819a97623f7" Jan 27 16:44:15 crc kubenswrapper[4772]: I0127 16:44:15.856306 4772 generic.go:334] "Generic (PLEG): container finished" podID="9371d269-02b3-4049-aeea-4fd56c648b89" containerID="43206bad0bf3620a1a3676348a740d052b18a789884d645086fe2a2e3fe3d350" exitCode=0 Jan 27 16:44:15 crc kubenswrapper[4772]: I0127 16:44:15.856471 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-6d96bf4746-x9c97" event={"ID":"9371d269-02b3-4049-aeea-4fd56c648b89","Type":"ContainerDied","Data":"43206bad0bf3620a1a3676348a740d052b18a789884d645086fe2a2e3fe3d350"} Jan 27 16:44:16 crc kubenswrapper[4772]: I0127 16:44:16.869838 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-6d96bf4746-x9c97" event={"ID":"9371d269-02b3-4049-aeea-4fd56c648b89","Type":"ContainerStarted","Data":"672d529986fa767415faf2a5c7d42386d2e5373e03758b384acd2235c1b4fb8c"} Jan 27 16:44:16 crc kubenswrapper[4772]: I0127 16:44:16.870214 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-6d96bf4746-x9c97" event={"ID":"9371d269-02b3-4049-aeea-4fd56c648b89","Type":"ContainerStarted","Data":"70778e96b10191ab7130a4cd6710c7b9b8b55518c8923812b7964a26367367d7"} Jan 27 16:44:16 crc kubenswrapper[4772]: I0127 16:44:16.870483 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-6d96bf4746-x9c97" Jan 27 16:44:16 crc kubenswrapper[4772]: I0127 16:44:16.892595 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-api-6d96bf4746-x9c97" podStartSLOduration=3.341554981 podStartE2EDuration="12.892574091s" podCreationTimestamp="2026-01-27 16:44:04 +0000 UTC" firstStartedPulling="2026-01-27 16:44:05.637138609 +0000 UTC m=+5831.617747707" lastFinishedPulling="2026-01-27 16:44:15.188157699 +0000 UTC m=+5841.168766817" observedRunningTime="2026-01-27 16:44:16.886379804 +0000 UTC m=+5842.866988902" watchObservedRunningTime="2026-01-27 16:44:16.892574091 +0000 UTC m=+5842.873183189" Jan 27 16:44:17 crc kubenswrapper[4772]: I0127 16:44:17.662642 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:44:17 crc kubenswrapper[4772]: E0127 16:44:17.663538 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:44:17 crc kubenswrapper[4772]: I0127 16:44:17.877343 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-6d96bf4746-x9c97" Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.384532 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-s5xbr" Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.388234 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-s5xbr" Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.395916 4772 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-jv694" podUID="febb140e-d26e-43db-9924-0f06739b9a4a" containerName="ovn-controller" probeResult="failure" output=< Jan 27 16:44:23 crc kubenswrapper[4772]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Jan 27 16:44:23 crc kubenswrapper[4772]: > Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.520318 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jv694-config-dfbjt"] Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.522111 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jv694-config-dfbjt" Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.524499 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.541010 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jv694-config-dfbjt"] Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.590849 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a86d738-a01c-4772-a150-7cc01a6503eb-scripts\") pod \"ovn-controller-jv694-config-dfbjt\" (UID: \"7a86d738-a01c-4772-a150-7cc01a6503eb\") " pod="openstack/ovn-controller-jv694-config-dfbjt" Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.590912 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7a86d738-a01c-4772-a150-7cc01a6503eb-additional-scripts\") pod \"ovn-controller-jv694-config-dfbjt\" (UID: \"7a86d738-a01c-4772-a150-7cc01a6503eb\") " pod="openstack/ovn-controller-jv694-config-dfbjt" Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.590953 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7a86d738-a01c-4772-a150-7cc01a6503eb-var-run-ovn\") pod \"ovn-controller-jv694-config-dfbjt\" (UID: \"7a86d738-a01c-4772-a150-7cc01a6503eb\") " pod="openstack/ovn-controller-jv694-config-dfbjt" Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.591034 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl6kc\" (UniqueName: \"kubernetes.io/projected/7a86d738-a01c-4772-a150-7cc01a6503eb-kube-api-access-tl6kc\") pod \"ovn-controller-jv694-config-dfbjt\" (UID: \"7a86d738-a01c-4772-a150-7cc01a6503eb\") " pod="openstack/ovn-controller-jv694-config-dfbjt" Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.591218 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7a86d738-a01c-4772-a150-7cc01a6503eb-var-run\") pod \"ovn-controller-jv694-config-dfbjt\" (UID: \"7a86d738-a01c-4772-a150-7cc01a6503eb\") " pod="openstack/ovn-controller-jv694-config-dfbjt" Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.591244 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7a86d738-a01c-4772-a150-7cc01a6503eb-var-log-ovn\") pod \"ovn-controller-jv694-config-dfbjt\" (UID: \"7a86d738-a01c-4772-a150-7cc01a6503eb\") " pod="openstack/ovn-controller-jv694-config-dfbjt" Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.692778 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7a86d738-a01c-4772-a150-7cc01a6503eb-var-run\") pod \"ovn-controller-jv694-config-dfbjt\" (UID: \"7a86d738-a01c-4772-a150-7cc01a6503eb\") " pod="openstack/ovn-controller-jv694-config-dfbjt" Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.692831 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7a86d738-a01c-4772-a150-7cc01a6503eb-var-log-ovn\") pod \"ovn-controller-jv694-config-dfbjt\" (UID: \"7a86d738-a01c-4772-a150-7cc01a6503eb\") " pod="openstack/ovn-controller-jv694-config-dfbjt" Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.693472 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a86d738-a01c-4772-a150-7cc01a6503eb-scripts\") pod \"ovn-controller-jv694-config-dfbjt\" (UID: \"7a86d738-a01c-4772-a150-7cc01a6503eb\") " pod="openstack/ovn-controller-jv694-config-dfbjt" Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.693533 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7a86d738-a01c-4772-a150-7cc01a6503eb-additional-scripts\") pod \"ovn-controller-jv694-config-dfbjt\" (UID: \"7a86d738-a01c-4772-a150-7cc01a6503eb\") " pod="openstack/ovn-controller-jv694-config-dfbjt" Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.693574 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7a86d738-a01c-4772-a150-7cc01a6503eb-var-run-ovn\") pod \"ovn-controller-jv694-config-dfbjt\" (UID: \"7a86d738-a01c-4772-a150-7cc01a6503eb\") " pod="openstack/ovn-controller-jv694-config-dfbjt" Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.693589 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7a86d738-a01c-4772-a150-7cc01a6503eb-var-log-ovn\") pod \"ovn-controller-jv694-config-dfbjt\" (UID: \"7a86d738-a01c-4772-a150-7cc01a6503eb\") " pod="openstack/ovn-controller-jv694-config-dfbjt" Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.693610 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl6kc\" (UniqueName: \"kubernetes.io/projected/7a86d738-a01c-4772-a150-7cc01a6503eb-kube-api-access-tl6kc\") pod \"ovn-controller-jv694-config-dfbjt\" (UID: \"7a86d738-a01c-4772-a150-7cc01a6503eb\") " pod="openstack/ovn-controller-jv694-config-dfbjt" Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.693615 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7a86d738-a01c-4772-a150-7cc01a6503eb-var-run\") pod \"ovn-controller-jv694-config-dfbjt\" (UID: \"7a86d738-a01c-4772-a150-7cc01a6503eb\") " pod="openstack/ovn-controller-jv694-config-dfbjt" Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.693650 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7a86d738-a01c-4772-a150-7cc01a6503eb-var-run-ovn\") pod \"ovn-controller-jv694-config-dfbjt\" (UID: \"7a86d738-a01c-4772-a150-7cc01a6503eb\") " pod="openstack/ovn-controller-jv694-config-dfbjt" Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.694350 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7a86d738-a01c-4772-a150-7cc01a6503eb-additional-scripts\") pod \"ovn-controller-jv694-config-dfbjt\" (UID: \"7a86d738-a01c-4772-a150-7cc01a6503eb\") " pod="openstack/ovn-controller-jv694-config-dfbjt" Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.695594 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a86d738-a01c-4772-a150-7cc01a6503eb-scripts\") pod \"ovn-controller-jv694-config-dfbjt\" (UID: \"7a86d738-a01c-4772-a150-7cc01a6503eb\") " pod="openstack/ovn-controller-jv694-config-dfbjt" Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.721127 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl6kc\" (UniqueName: \"kubernetes.io/projected/7a86d738-a01c-4772-a150-7cc01a6503eb-kube-api-access-tl6kc\") pod \"ovn-controller-jv694-config-dfbjt\" (UID: \"7a86d738-a01c-4772-a150-7cc01a6503eb\") " pod="openstack/ovn-controller-jv694-config-dfbjt" Jan 27 16:44:23 crc kubenswrapper[4772]: I0127 16:44:23.867846 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jv694-config-dfbjt" Jan 27 16:44:24 crc kubenswrapper[4772]: I0127 16:44:24.354673 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jv694-config-dfbjt"] Jan 27 16:44:24 crc kubenswrapper[4772]: I0127 16:44:24.944544 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jv694-config-dfbjt" event={"ID":"7a86d738-a01c-4772-a150-7cc01a6503eb","Type":"ContainerStarted","Data":"2c24a009396935ad397dec4435d87a19075c63011ad8f23a534d35cc814f6ddc"} Jan 27 16:44:24 crc kubenswrapper[4772]: I0127 16:44:24.944939 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jv694-config-dfbjt" event={"ID":"7a86d738-a01c-4772-a150-7cc01a6503eb","Type":"ContainerStarted","Data":"475062218bf811acbfff5ec92de82f6c43e05cf2fe5ad2412beabfe854fe2374"} Jan 27 16:44:25 crc kubenswrapper[4772]: I0127 16:44:25.956225 4772 generic.go:334] "Generic (PLEG): container finished" podID="7a86d738-a01c-4772-a150-7cc01a6503eb" containerID="2c24a009396935ad397dec4435d87a19075c63011ad8f23a534d35cc814f6ddc" exitCode=0 Jan 27 16:44:25 crc kubenswrapper[4772]: I0127 16:44:25.956264 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jv694-config-dfbjt" event={"ID":"7a86d738-a01c-4772-a150-7cc01a6503eb","Type":"ContainerDied","Data":"2c24a009396935ad397dec4435d87a19075c63011ad8f23a534d35cc814f6ddc"} Jan 27 16:44:27 crc kubenswrapper[4772]: I0127 16:44:27.339125 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jv694-config-dfbjt" Jan 27 16:44:27 crc kubenswrapper[4772]: I0127 16:44:27.470912 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a86d738-a01c-4772-a150-7cc01a6503eb-scripts\") pod \"7a86d738-a01c-4772-a150-7cc01a6503eb\" (UID: \"7a86d738-a01c-4772-a150-7cc01a6503eb\") " Jan 27 16:44:27 crc kubenswrapper[4772]: I0127 16:44:27.471115 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7a86d738-a01c-4772-a150-7cc01a6503eb-var-run\") pod \"7a86d738-a01c-4772-a150-7cc01a6503eb\" (UID: \"7a86d738-a01c-4772-a150-7cc01a6503eb\") " Jan 27 16:44:27 crc kubenswrapper[4772]: I0127 16:44:27.471162 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a86d738-a01c-4772-a150-7cc01a6503eb-var-run" (OuterVolumeSpecName: "var-run") pod "7a86d738-a01c-4772-a150-7cc01a6503eb" (UID: "7a86d738-a01c-4772-a150-7cc01a6503eb"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 16:44:27 crc kubenswrapper[4772]: I0127 16:44:27.471312 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7a86d738-a01c-4772-a150-7cc01a6503eb-var-run-ovn\") pod \"7a86d738-a01c-4772-a150-7cc01a6503eb\" (UID: \"7a86d738-a01c-4772-a150-7cc01a6503eb\") " Jan 27 16:44:27 crc kubenswrapper[4772]: I0127 16:44:27.471382 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tl6kc\" (UniqueName: \"kubernetes.io/projected/7a86d738-a01c-4772-a150-7cc01a6503eb-kube-api-access-tl6kc\") pod \"7a86d738-a01c-4772-a150-7cc01a6503eb\" (UID: \"7a86d738-a01c-4772-a150-7cc01a6503eb\") " Jan 27 16:44:27 crc kubenswrapper[4772]: I0127 16:44:27.471405 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7a86d738-a01c-4772-a150-7cc01a6503eb-var-log-ovn\") pod \"7a86d738-a01c-4772-a150-7cc01a6503eb\" (UID: \"7a86d738-a01c-4772-a150-7cc01a6503eb\") " Jan 27 16:44:27 crc kubenswrapper[4772]: I0127 16:44:27.471397 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a86d738-a01c-4772-a150-7cc01a6503eb-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "7a86d738-a01c-4772-a150-7cc01a6503eb" (UID: "7a86d738-a01c-4772-a150-7cc01a6503eb"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 16:44:27 crc kubenswrapper[4772]: I0127 16:44:27.471518 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7a86d738-a01c-4772-a150-7cc01a6503eb-additional-scripts\") pod \"7a86d738-a01c-4772-a150-7cc01a6503eb\" (UID: \"7a86d738-a01c-4772-a150-7cc01a6503eb\") " Jan 27 16:44:27 crc kubenswrapper[4772]: I0127 16:44:27.471540 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a86d738-a01c-4772-a150-7cc01a6503eb-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "7a86d738-a01c-4772-a150-7cc01a6503eb" (UID: "7a86d738-a01c-4772-a150-7cc01a6503eb"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 16:44:27 crc kubenswrapper[4772]: I0127 16:44:27.472128 4772 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7a86d738-a01c-4772-a150-7cc01a6503eb-var-run\") on node \"crc\" DevicePath \"\"" Jan 27 16:44:27 crc kubenswrapper[4772]: I0127 16:44:27.472152 4772 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7a86d738-a01c-4772-a150-7cc01a6503eb-var-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 27 16:44:27 crc kubenswrapper[4772]: I0127 16:44:27.472177 4772 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7a86d738-a01c-4772-a150-7cc01a6503eb-var-log-ovn\") on node \"crc\" DevicePath \"\"" Jan 27 16:44:27 crc kubenswrapper[4772]: I0127 16:44:27.472206 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a86d738-a01c-4772-a150-7cc01a6503eb-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "7a86d738-a01c-4772-a150-7cc01a6503eb" (UID: "7a86d738-a01c-4772-a150-7cc01a6503eb"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:44:27 crc kubenswrapper[4772]: I0127 16:44:27.472233 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a86d738-a01c-4772-a150-7cc01a6503eb-scripts" (OuterVolumeSpecName: "scripts") pod "7a86d738-a01c-4772-a150-7cc01a6503eb" (UID: "7a86d738-a01c-4772-a150-7cc01a6503eb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:44:27 crc kubenswrapper[4772]: I0127 16:44:27.485449 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a86d738-a01c-4772-a150-7cc01a6503eb-kube-api-access-tl6kc" (OuterVolumeSpecName: "kube-api-access-tl6kc") pod "7a86d738-a01c-4772-a150-7cc01a6503eb" (UID: "7a86d738-a01c-4772-a150-7cc01a6503eb"). InnerVolumeSpecName "kube-api-access-tl6kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:44:27 crc kubenswrapper[4772]: I0127 16:44:27.574018 4772 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7a86d738-a01c-4772-a150-7cc01a6503eb-additional-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:44:27 crc kubenswrapper[4772]: I0127 16:44:27.574055 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a86d738-a01c-4772-a150-7cc01a6503eb-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:44:27 crc kubenswrapper[4772]: I0127 16:44:27.574067 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tl6kc\" (UniqueName: \"kubernetes.io/projected/7a86d738-a01c-4772-a150-7cc01a6503eb-kube-api-access-tl6kc\") on node \"crc\" DevicePath \"\"" Jan 27 16:44:27 crc kubenswrapper[4772]: I0127 16:44:27.765658 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-jv694-config-dfbjt"] Jan 27 16:44:27 crc kubenswrapper[4772]: I0127 16:44:27.774373 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-jv694-config-dfbjt"] Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.015027 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="475062218bf811acbfff5ec92de82f6c43e05cf2fe5ad2412beabfe854fe2374" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.015139 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jv694-config-dfbjt" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.240380 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-rsyslog-4wcv5"] Jan 27 16:44:28 crc kubenswrapper[4772]: E0127 16:44:28.240764 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a86d738-a01c-4772-a150-7cc01a6503eb" containerName="ovn-config" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.240784 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a86d738-a01c-4772-a150-7cc01a6503eb" containerName="ovn-config" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.241000 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a86d738-a01c-4772-a150-7cc01a6503eb" containerName="ovn-config" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.242010 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-4wcv5" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.243593 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-config-data" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.244024 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-scripts" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.244091 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"octavia-hmport-map" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.265717 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-4wcv5"] Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.385117 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-jv694" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.389248 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/c163e8de-ea19-4a1c-8791-8659b9a09ba3-config-data-merged\") pod \"octavia-rsyslog-4wcv5\" (UID: \"c163e8de-ea19-4a1c-8791-8659b9a09ba3\") " pod="openstack/octavia-rsyslog-4wcv5" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.389309 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/c163e8de-ea19-4a1c-8791-8659b9a09ba3-hm-ports\") pod \"octavia-rsyslog-4wcv5\" (UID: \"c163e8de-ea19-4a1c-8791-8659b9a09ba3\") " pod="openstack/octavia-rsyslog-4wcv5" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.389355 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c163e8de-ea19-4a1c-8791-8659b9a09ba3-config-data\") pod \"octavia-rsyslog-4wcv5\" (UID: \"c163e8de-ea19-4a1c-8791-8659b9a09ba3\") " pod="openstack/octavia-rsyslog-4wcv5" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.389378 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c163e8de-ea19-4a1c-8791-8659b9a09ba3-scripts\") pod \"octavia-rsyslog-4wcv5\" (UID: \"c163e8de-ea19-4a1c-8791-8659b9a09ba3\") " pod="openstack/octavia-rsyslog-4wcv5" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.491098 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/c163e8de-ea19-4a1c-8791-8659b9a09ba3-config-data-merged\") pod \"octavia-rsyslog-4wcv5\" (UID: \"c163e8de-ea19-4a1c-8791-8659b9a09ba3\") " pod="openstack/octavia-rsyslog-4wcv5" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.491290 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/c163e8de-ea19-4a1c-8791-8659b9a09ba3-hm-ports\") pod \"octavia-rsyslog-4wcv5\" (UID: \"c163e8de-ea19-4a1c-8791-8659b9a09ba3\") " pod="openstack/octavia-rsyslog-4wcv5" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.491405 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c163e8de-ea19-4a1c-8791-8659b9a09ba3-config-data\") pod \"octavia-rsyslog-4wcv5\" (UID: \"c163e8de-ea19-4a1c-8791-8659b9a09ba3\") " pod="openstack/octavia-rsyslog-4wcv5" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.491612 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/c163e8de-ea19-4a1c-8791-8659b9a09ba3-config-data-merged\") pod \"octavia-rsyslog-4wcv5\" (UID: \"c163e8de-ea19-4a1c-8791-8659b9a09ba3\") " pod="openstack/octavia-rsyslog-4wcv5" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.492040 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/c163e8de-ea19-4a1c-8791-8659b9a09ba3-hm-ports\") pod \"octavia-rsyslog-4wcv5\" (UID: \"c163e8de-ea19-4a1c-8791-8659b9a09ba3\") " pod="openstack/octavia-rsyslog-4wcv5" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.492429 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c163e8de-ea19-4a1c-8791-8659b9a09ba3-scripts\") pod \"octavia-rsyslog-4wcv5\" (UID: \"c163e8de-ea19-4a1c-8791-8659b9a09ba3\") " pod="openstack/octavia-rsyslog-4wcv5" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.496504 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c163e8de-ea19-4a1c-8791-8659b9a09ba3-config-data\") pod \"octavia-rsyslog-4wcv5\" (UID: \"c163e8de-ea19-4a1c-8791-8659b9a09ba3\") " pod="openstack/octavia-rsyslog-4wcv5" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.496717 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c163e8de-ea19-4a1c-8791-8659b9a09ba3-scripts\") pod \"octavia-rsyslog-4wcv5\" (UID: \"c163e8de-ea19-4a1c-8791-8659b9a09ba3\") " pod="openstack/octavia-rsyslog-4wcv5" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.561276 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-4wcv5" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.682439 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a86d738-a01c-4772-a150-7cc01a6503eb" path="/var/lib/kubelet/pods/7a86d738-a01c-4772-a150-7cc01a6503eb/volumes" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.904907 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-59f8cff499-wwbn2"] Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.906667 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-wwbn2" Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.916475 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-wwbn2"] Jan 27 16:44:28 crc kubenswrapper[4772]: I0127 16:44:28.957276 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.019624 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/79cf41b1-8613-4f3f-8e1e-35d0ca7efffe-httpd-config\") pod \"octavia-image-upload-59f8cff499-wwbn2\" (UID: \"79cf41b1-8613-4f3f-8e1e-35d0ca7efffe\") " pod="openstack/octavia-image-upload-59f8cff499-wwbn2" Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.019779 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/79cf41b1-8613-4f3f-8e1e-35d0ca7efffe-amphora-image\") pod \"octavia-image-upload-59f8cff499-wwbn2\" (UID: \"79cf41b1-8613-4f3f-8e1e-35d0ca7efffe\") " pod="openstack/octavia-image-upload-59f8cff499-wwbn2" Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.120698 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/79cf41b1-8613-4f3f-8e1e-35d0ca7efffe-httpd-config\") pod \"octavia-image-upload-59f8cff499-wwbn2\" (UID: \"79cf41b1-8613-4f3f-8e1e-35d0ca7efffe\") " pod="openstack/octavia-image-upload-59f8cff499-wwbn2" Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.120941 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/79cf41b1-8613-4f3f-8e1e-35d0ca7efffe-amphora-image\") pod \"octavia-image-upload-59f8cff499-wwbn2\" (UID: \"79cf41b1-8613-4f3f-8e1e-35d0ca7efffe\") " pod="openstack/octavia-image-upload-59f8cff499-wwbn2" Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.121552 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/79cf41b1-8613-4f3f-8e1e-35d0ca7efffe-amphora-image\") pod \"octavia-image-upload-59f8cff499-wwbn2\" (UID: \"79cf41b1-8613-4f3f-8e1e-35d0ca7efffe\") " pod="openstack/octavia-image-upload-59f8cff499-wwbn2" Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.131963 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/79cf41b1-8613-4f3f-8e1e-35d0ca7efffe-httpd-config\") pod \"octavia-image-upload-59f8cff499-wwbn2\" (UID: \"79cf41b1-8613-4f3f-8e1e-35d0ca7efffe\") " pod="openstack/octavia-image-upload-59f8cff499-wwbn2" Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.145605 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-4wcv5"] Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.277444 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-wwbn2" Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.317870 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-4wcv5"] Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.619741 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-sync-bz6q7"] Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.621959 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-bz6q7" Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.624764 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-scripts" Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.643206 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-bz6q7"] Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.663615 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:44:29 crc kubenswrapper[4772]: E0127 16:44:29.663921 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.734392 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa-config-data-merged\") pod \"octavia-db-sync-bz6q7\" (UID: \"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa\") " pod="openstack/octavia-db-sync-bz6q7" Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.734923 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa-scripts\") pod \"octavia-db-sync-bz6q7\" (UID: \"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa\") " pod="openstack/octavia-db-sync-bz6q7" Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.734981 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa-config-data\") pod \"octavia-db-sync-bz6q7\" (UID: \"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa\") " pod="openstack/octavia-db-sync-bz6q7" Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.735107 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa-combined-ca-bundle\") pod \"octavia-db-sync-bz6q7\" (UID: \"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa\") " pod="openstack/octavia-db-sync-bz6q7" Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.748924 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-wwbn2"] Jan 27 16:44:29 crc kubenswrapper[4772]: W0127 16:44:29.760500 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79cf41b1_8613_4f3f_8e1e_35d0ca7efffe.slice/crio-4fb27a69e55f5ad48bd9d1e4c9841ecd324aac0beb1fbf6032053c07c57d076f WatchSource:0}: Error finding container 4fb27a69e55f5ad48bd9d1e4c9841ecd324aac0beb1fbf6032053c07c57d076f: Status 404 returned error can't find the container with id 4fb27a69e55f5ad48bd9d1e4c9841ecd324aac0beb1fbf6032053c07c57d076f Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.837088 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa-combined-ca-bundle\") pod \"octavia-db-sync-bz6q7\" (UID: \"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa\") " pod="openstack/octavia-db-sync-bz6q7" Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.837190 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa-config-data-merged\") pod \"octavia-db-sync-bz6q7\" (UID: \"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa\") " pod="openstack/octavia-db-sync-bz6q7" Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.837809 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa-config-data-merged\") pod \"octavia-db-sync-bz6q7\" (UID: \"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa\") " pod="openstack/octavia-db-sync-bz6q7" Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.839060 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa-scripts\") pod \"octavia-db-sync-bz6q7\" (UID: \"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa\") " pod="openstack/octavia-db-sync-bz6q7" Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.839107 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa-config-data\") pod \"octavia-db-sync-bz6q7\" (UID: \"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa\") " pod="openstack/octavia-db-sync-bz6q7" Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.843392 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa-combined-ca-bundle\") pod \"octavia-db-sync-bz6q7\" (UID: \"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa\") " pod="openstack/octavia-db-sync-bz6q7" Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.843533 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa-scripts\") pod \"octavia-db-sync-bz6q7\" (UID: \"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa\") " pod="openstack/octavia-db-sync-bz6q7" Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.852227 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa-config-data\") pod \"octavia-db-sync-bz6q7\" (UID: \"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa\") " pod="openstack/octavia-db-sync-bz6q7" Jan 27 16:44:29 crc kubenswrapper[4772]: I0127 16:44:29.944031 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-bz6q7" Jan 27 16:44:30 crc kubenswrapper[4772]: I0127 16:44:30.078083 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-wwbn2" event={"ID":"79cf41b1-8613-4f3f-8e1e-35d0ca7efffe","Type":"ContainerStarted","Data":"4fb27a69e55f5ad48bd9d1e4c9841ecd324aac0beb1fbf6032053c07c57d076f"} Jan 27 16:44:30 crc kubenswrapper[4772]: I0127 16:44:30.085914 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-4wcv5" event={"ID":"c163e8de-ea19-4a1c-8791-8659b9a09ba3","Type":"ContainerStarted","Data":"e08af4172eecfe384e6bdb1a3062be61c027c3f7eb2af34149f6166aeea022d4"} Jan 27 16:44:30 crc kubenswrapper[4772]: I0127 16:44:30.417381 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-bz6q7"] Jan 27 16:44:30 crc kubenswrapper[4772]: W0127 16:44:30.424241 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1152dfc9_a3d1_41a5_92cb_a5a8c481a8fa.slice/crio-1819e653d8ad3c01c8686d552abe73590b799f2ef81d0815681adfb79d2743ba WatchSource:0}: Error finding container 1819e653d8ad3c01c8686d552abe73590b799f2ef81d0815681adfb79d2743ba: Status 404 returned error can't find the container with id 1819e653d8ad3c01c8686d552abe73590b799f2ef81d0815681adfb79d2743ba Jan 27 16:44:31 crc kubenswrapper[4772]: I0127 16:44:31.096029 4772 generic.go:334] "Generic (PLEG): container finished" podID="1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa" containerID="908bfdb73b8d3ebb90afef686e407a3a2da1f2ea295c24e132068119ed919b42" exitCode=0 Jan 27 16:44:31 crc kubenswrapper[4772]: I0127 16:44:31.096075 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-bz6q7" event={"ID":"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa","Type":"ContainerDied","Data":"908bfdb73b8d3ebb90afef686e407a3a2da1f2ea295c24e132068119ed919b42"} Jan 27 16:44:31 crc kubenswrapper[4772]: I0127 16:44:31.096392 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-bz6q7" event={"ID":"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa","Type":"ContainerStarted","Data":"1819e653d8ad3c01c8686d552abe73590b799f2ef81d0815681adfb79d2743ba"} Jan 27 16:44:34 crc kubenswrapper[4772]: I0127 16:44:34.130145 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-4wcv5" event={"ID":"c163e8de-ea19-4a1c-8791-8659b9a09ba3","Type":"ContainerStarted","Data":"7a55c8b7d3f5ef107842b8268aaea45baea4ce51ff9790efdd6cb25f55809a83"} Jan 27 16:44:34 crc kubenswrapper[4772]: I0127 16:44:34.134363 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-bz6q7" event={"ID":"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa","Type":"ContainerStarted","Data":"3c5cbb7ef3f7daca21bab77efa8022a5214525eb8b99dd2483d6689f1db4cb83"} Jan 27 16:44:34 crc kubenswrapper[4772]: I0127 16:44:34.165928 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-db-sync-bz6q7" podStartSLOduration=5.165907717 podStartE2EDuration="5.165907717s" podCreationTimestamp="2026-01-27 16:44:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:44:34.162985484 +0000 UTC m=+5860.143594582" watchObservedRunningTime="2026-01-27 16:44:34.165907717 +0000 UTC m=+5860.146516815" Jan 27 16:44:36 crc kubenswrapper[4772]: I0127 16:44:36.152078 4772 generic.go:334] "Generic (PLEG): container finished" podID="c163e8de-ea19-4a1c-8791-8659b9a09ba3" containerID="7a55c8b7d3f5ef107842b8268aaea45baea4ce51ff9790efdd6cb25f55809a83" exitCode=0 Jan 27 16:44:36 crc kubenswrapper[4772]: I0127 16:44:36.152182 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-4wcv5" event={"ID":"c163e8de-ea19-4a1c-8791-8659b9a09ba3","Type":"ContainerDied","Data":"7a55c8b7d3f5ef107842b8268aaea45baea4ce51ff9790efdd6cb25f55809a83"} Jan 27 16:44:38 crc kubenswrapper[4772]: I0127 16:44:38.209596 4772 generic.go:334] "Generic (PLEG): container finished" podID="1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa" containerID="3c5cbb7ef3f7daca21bab77efa8022a5214525eb8b99dd2483d6689f1db4cb83" exitCode=0 Jan 27 16:44:38 crc kubenswrapper[4772]: I0127 16:44:38.209674 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-bz6q7" event={"ID":"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa","Type":"ContainerDied","Data":"3c5cbb7ef3f7daca21bab77efa8022a5214525eb8b99dd2483d6689f1db4cb83"} Jan 27 16:44:39 crc kubenswrapper[4772]: I0127 16:44:39.245540 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-6d96bf4746-x9c97" Jan 27 16:44:39 crc kubenswrapper[4772]: I0127 16:44:39.492834 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-6d96bf4746-x9c97" Jan 27 16:44:39 crc kubenswrapper[4772]: I0127 16:44:39.643372 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-bz6q7" Jan 27 16:44:39 crc kubenswrapper[4772]: I0127 16:44:39.752775 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa-config-data\") pod \"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa\" (UID: \"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa\") " Jan 27 16:44:39 crc kubenswrapper[4772]: I0127 16:44:39.753428 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa-scripts\") pod \"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa\" (UID: \"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa\") " Jan 27 16:44:39 crc kubenswrapper[4772]: I0127 16:44:39.753544 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa-combined-ca-bundle\") pod \"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa\" (UID: \"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa\") " Jan 27 16:44:39 crc kubenswrapper[4772]: I0127 16:44:39.753691 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa-config-data-merged\") pod \"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa\" (UID: \"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa\") " Jan 27 16:44:39 crc kubenswrapper[4772]: I0127 16:44:39.760854 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa-scripts" (OuterVolumeSpecName: "scripts") pod "1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa" (UID: "1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:44:39 crc kubenswrapper[4772]: I0127 16:44:39.764295 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa-config-data" (OuterVolumeSpecName: "config-data") pod "1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa" (UID: "1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:44:39 crc kubenswrapper[4772]: I0127 16:44:39.789692 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa" (UID: "1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:44:39 crc kubenswrapper[4772]: I0127 16:44:39.791036 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa" (UID: "1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:44:39 crc kubenswrapper[4772]: I0127 16:44:39.855355 4772 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa-config-data-merged\") on node \"crc\" DevicePath \"\"" Jan 27 16:44:39 crc kubenswrapper[4772]: I0127 16:44:39.855380 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 16:44:39 crc kubenswrapper[4772]: I0127 16:44:39.855390 4772 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa-scripts\") on node \"crc\" DevicePath \"\"" Jan 27 16:44:39 crc kubenswrapper[4772]: I0127 16:44:39.855400 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 16:44:40 crc kubenswrapper[4772]: I0127 16:44:40.232047 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-4wcv5" event={"ID":"c163e8de-ea19-4a1c-8791-8659b9a09ba3","Type":"ContainerStarted","Data":"33227eee176068edc8a2071687aaea5f2a55907beefcfe1533e576e25e787ca5"} Jan 27 16:44:40 crc kubenswrapper[4772]: I0127 16:44:40.232898 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-rsyslog-4wcv5" Jan 27 16:44:40 crc kubenswrapper[4772]: I0127 16:44:40.234304 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-bz6q7" Jan 27 16:44:40 crc kubenswrapper[4772]: I0127 16:44:40.234292 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-bz6q7" event={"ID":"1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa","Type":"ContainerDied","Data":"1819e653d8ad3c01c8686d552abe73590b799f2ef81d0815681adfb79d2743ba"} Jan 27 16:44:40 crc kubenswrapper[4772]: I0127 16:44:40.234434 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1819e653d8ad3c01c8686d552abe73590b799f2ef81d0815681adfb79d2743ba" Jan 27 16:44:40 crc kubenswrapper[4772]: I0127 16:44:40.236203 4772 generic.go:334] "Generic (PLEG): container finished" podID="79cf41b1-8613-4f3f-8e1e-35d0ca7efffe" containerID="68f623ccc11f688238dc60d96b2aa3b6231be01063d7a945260daa5b84ec5a0e" exitCode=0 Jan 27 16:44:40 crc kubenswrapper[4772]: I0127 16:44:40.236253 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-wwbn2" event={"ID":"79cf41b1-8613-4f3f-8e1e-35d0ca7efffe","Type":"ContainerDied","Data":"68f623ccc11f688238dc60d96b2aa3b6231be01063d7a945260daa5b84ec5a0e"} Jan 27 16:44:40 crc kubenswrapper[4772]: I0127 16:44:40.253960 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-rsyslog-4wcv5" podStartSLOduration=2.16546444 podStartE2EDuration="12.253932859s" podCreationTimestamp="2026-01-27 16:44:28 +0000 UTC" firstStartedPulling="2026-01-27 16:44:29.155210433 +0000 UTC m=+5855.135819531" lastFinishedPulling="2026-01-27 16:44:39.243678852 +0000 UTC m=+5865.224287950" observedRunningTime="2026-01-27 16:44:40.253798315 +0000 UTC m=+5866.234407423" watchObservedRunningTime="2026-01-27 16:44:40.253932859 +0000 UTC m=+5866.234541967" Jan 27 16:44:42 crc kubenswrapper[4772]: I0127 16:44:42.259447 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-wwbn2" event={"ID":"79cf41b1-8613-4f3f-8e1e-35d0ca7efffe","Type":"ContainerStarted","Data":"be7964b45b1fe07afd2ed607cc55d8b0c5ba7716904f6e9df312a541c0d3a1a9"} Jan 27 16:44:42 crc kubenswrapper[4772]: I0127 16:44:42.297222 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-59f8cff499-wwbn2" podStartSLOduration=2.814339384 podStartE2EDuration="14.297157088s" podCreationTimestamp="2026-01-27 16:44:28 +0000 UTC" firstStartedPulling="2026-01-27 16:44:29.763413509 +0000 UTC m=+5855.744022617" lastFinishedPulling="2026-01-27 16:44:41.246231233 +0000 UTC m=+5867.226840321" observedRunningTime="2026-01-27 16:44:42.281932464 +0000 UTC m=+5868.262541632" watchObservedRunningTime="2026-01-27 16:44:42.297157088 +0000 UTC m=+5868.277766216" Jan 27 16:44:42 crc kubenswrapper[4772]: I0127 16:44:42.663426 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:44:42 crc kubenswrapper[4772]: E0127 16:44:42.663806 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:44:55 crc kubenswrapper[4772]: I0127 16:44:55.662977 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:44:55 crc kubenswrapper[4772]: E0127 16:44:55.663845 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:44:58 crc kubenswrapper[4772]: I0127 16:44:58.595889 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-rsyslog-4wcv5" Jan 27 16:45:00 crc kubenswrapper[4772]: I0127 16:45:00.169113 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492205-9hv9l"] Jan 27 16:45:00 crc kubenswrapper[4772]: E0127 16:45:00.169864 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa" containerName="init" Jan 27 16:45:00 crc kubenswrapper[4772]: I0127 16:45:00.169883 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa" containerName="init" Jan 27 16:45:00 crc kubenswrapper[4772]: E0127 16:45:00.169898 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa" containerName="octavia-db-sync" Jan 27 16:45:00 crc kubenswrapper[4772]: I0127 16:45:00.169906 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa" containerName="octavia-db-sync" Jan 27 16:45:00 crc kubenswrapper[4772]: I0127 16:45:00.170204 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa" containerName="octavia-db-sync" Jan 27 16:45:00 crc kubenswrapper[4772]: I0127 16:45:00.170982 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492205-9hv9l" Jan 27 16:45:00 crc kubenswrapper[4772]: I0127 16:45:00.173235 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 27 16:45:00 crc kubenswrapper[4772]: I0127 16:45:00.173820 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 27 16:45:00 crc kubenswrapper[4772]: I0127 16:45:00.179531 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492205-9hv9l"] Jan 27 16:45:00 crc kubenswrapper[4772]: I0127 16:45:00.283194 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bnlq\" (UniqueName: \"kubernetes.io/projected/eb399a66-1690-4026-9b2e-9e399d3270d2-kube-api-access-8bnlq\") pod \"collect-profiles-29492205-9hv9l\" (UID: \"eb399a66-1690-4026-9b2e-9e399d3270d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492205-9hv9l" Jan 27 16:45:00 crc kubenswrapper[4772]: I0127 16:45:00.283419 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eb399a66-1690-4026-9b2e-9e399d3270d2-secret-volume\") pod \"collect-profiles-29492205-9hv9l\" (UID: \"eb399a66-1690-4026-9b2e-9e399d3270d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492205-9hv9l" Jan 27 16:45:00 crc kubenswrapper[4772]: I0127 16:45:00.283492 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eb399a66-1690-4026-9b2e-9e399d3270d2-config-volume\") pod \"collect-profiles-29492205-9hv9l\" (UID: \"eb399a66-1690-4026-9b2e-9e399d3270d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492205-9hv9l" Jan 27 16:45:00 crc kubenswrapper[4772]: I0127 16:45:00.385699 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eb399a66-1690-4026-9b2e-9e399d3270d2-secret-volume\") pod \"collect-profiles-29492205-9hv9l\" (UID: \"eb399a66-1690-4026-9b2e-9e399d3270d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492205-9hv9l" Jan 27 16:45:00 crc kubenswrapper[4772]: I0127 16:45:00.385785 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eb399a66-1690-4026-9b2e-9e399d3270d2-config-volume\") pod \"collect-profiles-29492205-9hv9l\" (UID: \"eb399a66-1690-4026-9b2e-9e399d3270d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492205-9hv9l" Jan 27 16:45:00 crc kubenswrapper[4772]: I0127 16:45:00.385823 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bnlq\" (UniqueName: \"kubernetes.io/projected/eb399a66-1690-4026-9b2e-9e399d3270d2-kube-api-access-8bnlq\") pod \"collect-profiles-29492205-9hv9l\" (UID: \"eb399a66-1690-4026-9b2e-9e399d3270d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492205-9hv9l" Jan 27 16:45:00 crc kubenswrapper[4772]: I0127 16:45:00.386962 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eb399a66-1690-4026-9b2e-9e399d3270d2-config-volume\") pod \"collect-profiles-29492205-9hv9l\" (UID: \"eb399a66-1690-4026-9b2e-9e399d3270d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492205-9hv9l" Jan 27 16:45:00 crc kubenswrapper[4772]: I0127 16:45:00.393033 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eb399a66-1690-4026-9b2e-9e399d3270d2-secret-volume\") pod \"collect-profiles-29492205-9hv9l\" (UID: \"eb399a66-1690-4026-9b2e-9e399d3270d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492205-9hv9l" Jan 27 16:45:00 crc kubenswrapper[4772]: I0127 16:45:00.403309 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bnlq\" (UniqueName: \"kubernetes.io/projected/eb399a66-1690-4026-9b2e-9e399d3270d2-kube-api-access-8bnlq\") pod \"collect-profiles-29492205-9hv9l\" (UID: \"eb399a66-1690-4026-9b2e-9e399d3270d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492205-9hv9l" Jan 27 16:45:00 crc kubenswrapper[4772]: I0127 16:45:00.495346 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492205-9hv9l" Jan 27 16:45:00 crc kubenswrapper[4772]: I0127 16:45:00.986558 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492205-9hv9l"] Jan 27 16:45:01 crc kubenswrapper[4772]: I0127 16:45:01.470330 4772 generic.go:334] "Generic (PLEG): container finished" podID="eb399a66-1690-4026-9b2e-9e399d3270d2" containerID="3b5c3dfd99ca4b5982c3131c3d5ce465e41cbe4ff9774e156e9f425715410ede" exitCode=0 Jan 27 16:45:01 crc kubenswrapper[4772]: I0127 16:45:01.470389 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492205-9hv9l" event={"ID":"eb399a66-1690-4026-9b2e-9e399d3270d2","Type":"ContainerDied","Data":"3b5c3dfd99ca4b5982c3131c3d5ce465e41cbe4ff9774e156e9f425715410ede"} Jan 27 16:45:01 crc kubenswrapper[4772]: I0127 16:45:01.470646 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492205-9hv9l" event={"ID":"eb399a66-1690-4026-9b2e-9e399d3270d2","Type":"ContainerStarted","Data":"026979e7ba50a908b6032a92cd91e7808fad026d1d1ebe11a08fe7589b3b8a75"} Jan 27 16:45:02 crc kubenswrapper[4772]: I0127 16:45:02.915392 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492205-9hv9l" Jan 27 16:45:03 crc kubenswrapper[4772]: I0127 16:45:03.037642 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eb399a66-1690-4026-9b2e-9e399d3270d2-config-volume\") pod \"eb399a66-1690-4026-9b2e-9e399d3270d2\" (UID: \"eb399a66-1690-4026-9b2e-9e399d3270d2\") " Jan 27 16:45:03 crc kubenswrapper[4772]: I0127 16:45:03.037723 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bnlq\" (UniqueName: \"kubernetes.io/projected/eb399a66-1690-4026-9b2e-9e399d3270d2-kube-api-access-8bnlq\") pod \"eb399a66-1690-4026-9b2e-9e399d3270d2\" (UID: \"eb399a66-1690-4026-9b2e-9e399d3270d2\") " Jan 27 16:45:03 crc kubenswrapper[4772]: I0127 16:45:03.037831 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eb399a66-1690-4026-9b2e-9e399d3270d2-secret-volume\") pod \"eb399a66-1690-4026-9b2e-9e399d3270d2\" (UID: \"eb399a66-1690-4026-9b2e-9e399d3270d2\") " Jan 27 16:45:03 crc kubenswrapper[4772]: I0127 16:45:03.038219 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb399a66-1690-4026-9b2e-9e399d3270d2-config-volume" (OuterVolumeSpecName: "config-volume") pod "eb399a66-1690-4026-9b2e-9e399d3270d2" (UID: "eb399a66-1690-4026-9b2e-9e399d3270d2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 16:45:03 crc kubenswrapper[4772]: I0127 16:45:03.038694 4772 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eb399a66-1690-4026-9b2e-9e399d3270d2-config-volume\") on node \"crc\" DevicePath \"\"" Jan 27 16:45:03 crc kubenswrapper[4772]: I0127 16:45:03.043533 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb399a66-1690-4026-9b2e-9e399d3270d2-kube-api-access-8bnlq" (OuterVolumeSpecName: "kube-api-access-8bnlq") pod "eb399a66-1690-4026-9b2e-9e399d3270d2" (UID: "eb399a66-1690-4026-9b2e-9e399d3270d2"). InnerVolumeSpecName "kube-api-access-8bnlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:45:03 crc kubenswrapper[4772]: I0127 16:45:03.043862 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb399a66-1690-4026-9b2e-9e399d3270d2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "eb399a66-1690-4026-9b2e-9e399d3270d2" (UID: "eb399a66-1690-4026-9b2e-9e399d3270d2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:45:03 crc kubenswrapper[4772]: I0127 16:45:03.141868 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bnlq\" (UniqueName: \"kubernetes.io/projected/eb399a66-1690-4026-9b2e-9e399d3270d2-kube-api-access-8bnlq\") on node \"crc\" DevicePath \"\"" Jan 27 16:45:03 crc kubenswrapper[4772]: I0127 16:45:03.142328 4772 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eb399a66-1690-4026-9b2e-9e399d3270d2-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 27 16:45:03 crc kubenswrapper[4772]: I0127 16:45:03.487773 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492205-9hv9l" event={"ID":"eb399a66-1690-4026-9b2e-9e399d3270d2","Type":"ContainerDied","Data":"026979e7ba50a908b6032a92cd91e7808fad026d1d1ebe11a08fe7589b3b8a75"} Jan 27 16:45:03 crc kubenswrapper[4772]: I0127 16:45:03.487813 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="026979e7ba50a908b6032a92cd91e7808fad026d1d1ebe11a08fe7589b3b8a75" Jan 27 16:45:03 crc kubenswrapper[4772]: I0127 16:45:03.487865 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492205-9hv9l" Jan 27 16:45:03 crc kubenswrapper[4772]: I0127 16:45:03.684246 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-wwbn2"] Jan 27 16:45:03 crc kubenswrapper[4772]: I0127 16:45:03.684806 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-image-upload-59f8cff499-wwbn2" podUID="79cf41b1-8613-4f3f-8e1e-35d0ca7efffe" containerName="octavia-amphora-httpd" containerID="cri-o://be7964b45b1fe07afd2ed607cc55d8b0c5ba7716904f6e9df312a541c0d3a1a9" gracePeriod=30 Jan 27 16:45:03 crc kubenswrapper[4772]: I0127 16:45:03.988763 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492160-4mmgr"] Jan 27 16:45:03 crc kubenswrapper[4772]: I0127 16:45:03.992470 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492160-4mmgr"] Jan 27 16:45:04 crc kubenswrapper[4772]: I0127 16:45:04.199529 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-wwbn2" Jan 27 16:45:04 crc kubenswrapper[4772]: I0127 16:45:04.364479 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/79cf41b1-8613-4f3f-8e1e-35d0ca7efffe-amphora-image\") pod \"79cf41b1-8613-4f3f-8e1e-35d0ca7efffe\" (UID: \"79cf41b1-8613-4f3f-8e1e-35d0ca7efffe\") " Jan 27 16:45:04 crc kubenswrapper[4772]: I0127 16:45:04.364670 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/79cf41b1-8613-4f3f-8e1e-35d0ca7efffe-httpd-config\") pod \"79cf41b1-8613-4f3f-8e1e-35d0ca7efffe\" (UID: \"79cf41b1-8613-4f3f-8e1e-35d0ca7efffe\") " Jan 27 16:45:04 crc kubenswrapper[4772]: I0127 16:45:04.392412 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79cf41b1-8613-4f3f-8e1e-35d0ca7efffe-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "79cf41b1-8613-4f3f-8e1e-35d0ca7efffe" (UID: "79cf41b1-8613-4f3f-8e1e-35d0ca7efffe"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 16:45:04 crc kubenswrapper[4772]: I0127 16:45:04.466954 4772 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/79cf41b1-8613-4f3f-8e1e-35d0ca7efffe-httpd-config\") on node \"crc\" DevicePath \"\"" Jan 27 16:45:04 crc kubenswrapper[4772]: I0127 16:45:04.470033 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79cf41b1-8613-4f3f-8e1e-35d0ca7efffe-amphora-image" (OuterVolumeSpecName: "amphora-image") pod "79cf41b1-8613-4f3f-8e1e-35d0ca7efffe" (UID: "79cf41b1-8613-4f3f-8e1e-35d0ca7efffe"). InnerVolumeSpecName "amphora-image". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:45:04 crc kubenswrapper[4772]: I0127 16:45:04.498961 4772 generic.go:334] "Generic (PLEG): container finished" podID="79cf41b1-8613-4f3f-8e1e-35d0ca7efffe" containerID="be7964b45b1fe07afd2ed607cc55d8b0c5ba7716904f6e9df312a541c0d3a1a9" exitCode=0 Jan 27 16:45:04 crc kubenswrapper[4772]: I0127 16:45:04.499005 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-wwbn2" event={"ID":"79cf41b1-8613-4f3f-8e1e-35d0ca7efffe","Type":"ContainerDied","Data":"be7964b45b1fe07afd2ed607cc55d8b0c5ba7716904f6e9df312a541c0d3a1a9"} Jan 27 16:45:04 crc kubenswrapper[4772]: I0127 16:45:04.499032 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-wwbn2" event={"ID":"79cf41b1-8613-4f3f-8e1e-35d0ca7efffe","Type":"ContainerDied","Data":"4fb27a69e55f5ad48bd9d1e4c9841ecd324aac0beb1fbf6032053c07c57d076f"} Jan 27 16:45:04 crc kubenswrapper[4772]: I0127 16:45:04.499064 4772 scope.go:117] "RemoveContainer" containerID="be7964b45b1fe07afd2ed607cc55d8b0c5ba7716904f6e9df312a541c0d3a1a9" Jan 27 16:45:04 crc kubenswrapper[4772]: I0127 16:45:04.499222 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-wwbn2" Jan 27 16:45:04 crc kubenswrapper[4772]: I0127 16:45:04.533609 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-wwbn2"] Jan 27 16:45:04 crc kubenswrapper[4772]: I0127 16:45:04.538002 4772 scope.go:117] "RemoveContainer" containerID="68f623ccc11f688238dc60d96b2aa3b6231be01063d7a945260daa5b84ec5a0e" Jan 27 16:45:04 crc kubenswrapper[4772]: I0127 16:45:04.541704 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-wwbn2"] Jan 27 16:45:04 crc kubenswrapper[4772]: I0127 16:45:04.556052 4772 scope.go:117] "RemoveContainer" containerID="be7964b45b1fe07afd2ed607cc55d8b0c5ba7716904f6e9df312a541c0d3a1a9" Jan 27 16:45:04 crc kubenswrapper[4772]: E0127 16:45:04.556691 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be7964b45b1fe07afd2ed607cc55d8b0c5ba7716904f6e9df312a541c0d3a1a9\": container with ID starting with be7964b45b1fe07afd2ed607cc55d8b0c5ba7716904f6e9df312a541c0d3a1a9 not found: ID does not exist" containerID="be7964b45b1fe07afd2ed607cc55d8b0c5ba7716904f6e9df312a541c0d3a1a9" Jan 27 16:45:04 crc kubenswrapper[4772]: I0127 16:45:04.556743 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be7964b45b1fe07afd2ed607cc55d8b0c5ba7716904f6e9df312a541c0d3a1a9"} err="failed to get container status \"be7964b45b1fe07afd2ed607cc55d8b0c5ba7716904f6e9df312a541c0d3a1a9\": rpc error: code = NotFound desc = could not find container \"be7964b45b1fe07afd2ed607cc55d8b0c5ba7716904f6e9df312a541c0d3a1a9\": container with ID starting with be7964b45b1fe07afd2ed607cc55d8b0c5ba7716904f6e9df312a541c0d3a1a9 not found: ID does not exist" Jan 27 16:45:04 crc kubenswrapper[4772]: I0127 16:45:04.556772 4772 scope.go:117] "RemoveContainer" containerID="68f623ccc11f688238dc60d96b2aa3b6231be01063d7a945260daa5b84ec5a0e" Jan 27 16:45:04 crc kubenswrapper[4772]: E0127 16:45:04.557142 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68f623ccc11f688238dc60d96b2aa3b6231be01063d7a945260daa5b84ec5a0e\": container with ID starting with 68f623ccc11f688238dc60d96b2aa3b6231be01063d7a945260daa5b84ec5a0e not found: ID does not exist" containerID="68f623ccc11f688238dc60d96b2aa3b6231be01063d7a945260daa5b84ec5a0e" Jan 27 16:45:04 crc kubenswrapper[4772]: I0127 16:45:04.557223 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68f623ccc11f688238dc60d96b2aa3b6231be01063d7a945260daa5b84ec5a0e"} err="failed to get container status \"68f623ccc11f688238dc60d96b2aa3b6231be01063d7a945260daa5b84ec5a0e\": rpc error: code = NotFound desc = could not find container \"68f623ccc11f688238dc60d96b2aa3b6231be01063d7a945260daa5b84ec5a0e\": container with ID starting with 68f623ccc11f688238dc60d96b2aa3b6231be01063d7a945260daa5b84ec5a0e not found: ID does not exist" Jan 27 16:45:04 crc kubenswrapper[4772]: I0127 16:45:04.568943 4772 reconciler_common.go:293] "Volume detached for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/79cf41b1-8613-4f3f-8e1e-35d0ca7efffe-amphora-image\") on node \"crc\" DevicePath \"\"" Jan 27 16:45:04 crc kubenswrapper[4772]: I0127 16:45:04.675457 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79cf41b1-8613-4f3f-8e1e-35d0ca7efffe" path="/var/lib/kubelet/pods/79cf41b1-8613-4f3f-8e1e-35d0ca7efffe/volumes" Jan 27 16:45:04 crc kubenswrapper[4772]: I0127 16:45:04.676472 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b4afc52-82aa-4768-9cc8-5e9236fc4330" path="/var/lib/kubelet/pods/9b4afc52-82aa-4768-9cc8-5e9236fc4330/volumes" Jan 27 16:45:07 crc kubenswrapper[4772]: I0127 16:45:07.834396 4772 scope.go:117] "RemoveContainer" containerID="8dd8add741d2c060daa432ff2f192c7a04c82b2eab3360197f22c851ca7bd6c0" Jan 27 16:45:09 crc kubenswrapper[4772]: I0127 16:45:09.663604 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:45:09 crc kubenswrapper[4772]: E0127 16:45:09.665158 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:45:22 crc kubenswrapper[4772]: I0127 16:45:22.663265 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:45:22 crc kubenswrapper[4772]: E0127 16:45:22.664409 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:45:32 crc kubenswrapper[4772]: I0127 16:45:32.930035 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-healthmanager-jmckq"] Jan 27 16:45:32 crc kubenswrapper[4772]: E0127 16:45:32.931195 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb399a66-1690-4026-9b2e-9e399d3270d2" containerName="collect-profiles" Jan 27 16:45:32 crc kubenswrapper[4772]: I0127 16:45:32.931219 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb399a66-1690-4026-9b2e-9e399d3270d2" containerName="collect-profiles" Jan 27 16:45:32 crc kubenswrapper[4772]: E0127 16:45:32.931237 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79cf41b1-8613-4f3f-8e1e-35d0ca7efffe" containerName="init" Jan 27 16:45:32 crc kubenswrapper[4772]: I0127 16:45:32.931245 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="79cf41b1-8613-4f3f-8e1e-35d0ca7efffe" containerName="init" Jan 27 16:45:32 crc kubenswrapper[4772]: E0127 16:45:32.931260 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79cf41b1-8613-4f3f-8e1e-35d0ca7efffe" containerName="octavia-amphora-httpd" Jan 27 16:45:32 crc kubenswrapper[4772]: I0127 16:45:32.931268 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="79cf41b1-8613-4f3f-8e1e-35d0ca7efffe" containerName="octavia-amphora-httpd" Jan 27 16:45:32 crc kubenswrapper[4772]: I0127 16:45:32.931506 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb399a66-1690-4026-9b2e-9e399d3270d2" containerName="collect-profiles" Jan 27 16:45:32 crc kubenswrapper[4772]: I0127 16:45:32.931530 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="79cf41b1-8613-4f3f-8e1e-35d0ca7efffe" containerName="octavia-amphora-httpd" Jan 27 16:45:32 crc kubenswrapper[4772]: I0127 16:45:32.932909 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-jmckq" Jan 27 16:45:32 crc kubenswrapper[4772]: I0127 16:45:32.936358 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-scripts" Jan 27 16:45:32 crc kubenswrapper[4772]: I0127 16:45:32.936973 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-certs-secret" Jan 27 16:45:32 crc kubenswrapper[4772]: I0127 16:45:32.937060 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-config-data" Jan 27 16:45:32 crc kubenswrapper[4772]: I0127 16:45:32.962653 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-jmckq"] Jan 27 16:45:33 crc kubenswrapper[4772]: I0127 16:45:33.038051 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32d3e24-6f03-46cc-b7ae-61383778b183-combined-ca-bundle\") pod \"octavia-healthmanager-jmckq\" (UID: \"d32d3e24-6f03-46cc-b7ae-61383778b183\") " pod="openstack/octavia-healthmanager-jmckq" Jan 27 16:45:33 crc kubenswrapper[4772]: I0127 16:45:33.038141 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/d32d3e24-6f03-46cc-b7ae-61383778b183-amphora-certs\") pod \"octavia-healthmanager-jmckq\" (UID: \"d32d3e24-6f03-46cc-b7ae-61383778b183\") " pod="openstack/octavia-healthmanager-jmckq" Jan 27 16:45:33 crc kubenswrapper[4772]: I0127 16:45:33.038196 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d32d3e24-6f03-46cc-b7ae-61383778b183-config-data\") pod \"octavia-healthmanager-jmckq\" (UID: \"d32d3e24-6f03-46cc-b7ae-61383778b183\") " pod="openstack/octavia-healthmanager-jmckq" Jan 27 16:45:33 crc kubenswrapper[4772]: I0127 16:45:33.038222 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d32d3e24-6f03-46cc-b7ae-61383778b183-scripts\") pod \"octavia-healthmanager-jmckq\" (UID: \"d32d3e24-6f03-46cc-b7ae-61383778b183\") " pod="openstack/octavia-healthmanager-jmckq" Jan 27 16:45:33 crc kubenswrapper[4772]: I0127 16:45:33.038381 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d32d3e24-6f03-46cc-b7ae-61383778b183-config-data-merged\") pod \"octavia-healthmanager-jmckq\" (UID: \"d32d3e24-6f03-46cc-b7ae-61383778b183\") " pod="openstack/octavia-healthmanager-jmckq" Jan 27 16:45:33 crc kubenswrapper[4772]: I0127 16:45:33.038414 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/d32d3e24-6f03-46cc-b7ae-61383778b183-hm-ports\") pod \"octavia-healthmanager-jmckq\" (UID: \"d32d3e24-6f03-46cc-b7ae-61383778b183\") " pod="openstack/octavia-healthmanager-jmckq" Jan 27 16:45:33 crc kubenswrapper[4772]: I0127 16:45:33.140419 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d32d3e24-6f03-46cc-b7ae-61383778b183-scripts\") pod \"octavia-healthmanager-jmckq\" (UID: \"d32d3e24-6f03-46cc-b7ae-61383778b183\") " pod="openstack/octavia-healthmanager-jmckq" Jan 27 16:45:33 crc kubenswrapper[4772]: I0127 16:45:33.140571 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d32d3e24-6f03-46cc-b7ae-61383778b183-config-data-merged\") pod \"octavia-healthmanager-jmckq\" (UID: \"d32d3e24-6f03-46cc-b7ae-61383778b183\") " pod="openstack/octavia-healthmanager-jmckq" Jan 27 16:45:33 crc kubenswrapper[4772]: I0127 16:45:33.140604 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/d32d3e24-6f03-46cc-b7ae-61383778b183-hm-ports\") pod \"octavia-healthmanager-jmckq\" (UID: \"d32d3e24-6f03-46cc-b7ae-61383778b183\") " pod="openstack/octavia-healthmanager-jmckq" Jan 27 16:45:33 crc kubenswrapper[4772]: I0127 16:45:33.140628 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32d3e24-6f03-46cc-b7ae-61383778b183-combined-ca-bundle\") pod \"octavia-healthmanager-jmckq\" (UID: \"d32d3e24-6f03-46cc-b7ae-61383778b183\") " pod="openstack/octavia-healthmanager-jmckq" Jan 27 16:45:33 crc kubenswrapper[4772]: I0127 16:45:33.140684 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/d32d3e24-6f03-46cc-b7ae-61383778b183-amphora-certs\") pod \"octavia-healthmanager-jmckq\" (UID: \"d32d3e24-6f03-46cc-b7ae-61383778b183\") " pod="openstack/octavia-healthmanager-jmckq" Jan 27 16:45:33 crc kubenswrapper[4772]: I0127 16:45:33.140712 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d32d3e24-6f03-46cc-b7ae-61383778b183-config-data\") pod \"octavia-healthmanager-jmckq\" (UID: \"d32d3e24-6f03-46cc-b7ae-61383778b183\") " pod="openstack/octavia-healthmanager-jmckq" Jan 27 16:45:33 crc kubenswrapper[4772]: I0127 16:45:33.141068 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d32d3e24-6f03-46cc-b7ae-61383778b183-config-data-merged\") pod \"octavia-healthmanager-jmckq\" (UID: \"d32d3e24-6f03-46cc-b7ae-61383778b183\") " pod="openstack/octavia-healthmanager-jmckq" Jan 27 16:45:33 crc kubenswrapper[4772]: I0127 16:45:33.141904 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/d32d3e24-6f03-46cc-b7ae-61383778b183-hm-ports\") pod \"octavia-healthmanager-jmckq\" (UID: \"d32d3e24-6f03-46cc-b7ae-61383778b183\") " pod="openstack/octavia-healthmanager-jmckq" Jan 27 16:45:33 crc kubenswrapper[4772]: I0127 16:45:33.146534 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d32d3e24-6f03-46cc-b7ae-61383778b183-scripts\") pod \"octavia-healthmanager-jmckq\" (UID: \"d32d3e24-6f03-46cc-b7ae-61383778b183\") " pod="openstack/octavia-healthmanager-jmckq" Jan 27 16:45:33 crc kubenswrapper[4772]: I0127 16:45:33.147063 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32d3e24-6f03-46cc-b7ae-61383778b183-combined-ca-bundle\") pod \"octavia-healthmanager-jmckq\" (UID: \"d32d3e24-6f03-46cc-b7ae-61383778b183\") " pod="openstack/octavia-healthmanager-jmckq" Jan 27 16:45:33 crc kubenswrapper[4772]: I0127 16:45:33.147102 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d32d3e24-6f03-46cc-b7ae-61383778b183-config-data\") pod \"octavia-healthmanager-jmckq\" (UID: \"d32d3e24-6f03-46cc-b7ae-61383778b183\") " pod="openstack/octavia-healthmanager-jmckq" Jan 27 16:45:33 crc kubenswrapper[4772]: I0127 16:45:33.147589 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/d32d3e24-6f03-46cc-b7ae-61383778b183-amphora-certs\") pod \"octavia-healthmanager-jmckq\" (UID: \"d32d3e24-6f03-46cc-b7ae-61383778b183\") " pod="openstack/octavia-healthmanager-jmckq" Jan 27 16:45:33 crc kubenswrapper[4772]: I0127 16:45:33.265885 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-jmckq" Jan 27 16:45:33 crc kubenswrapper[4772]: I0127 16:45:33.801520 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-jmckq"] Jan 27 16:45:33 crc kubenswrapper[4772]: W0127 16:45:33.803243 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd32d3e24_6f03_46cc_b7ae_61383778b183.slice/crio-a0320d675c0ddf4fbe99af5532cb287c52a9f435573808fcb1d0149eac42ba29 WatchSource:0}: Error finding container a0320d675c0ddf4fbe99af5532cb287c52a9f435573808fcb1d0149eac42ba29: Status 404 returned error can't find the container with id a0320d675c0ddf4fbe99af5532cb287c52a9f435573808fcb1d0149eac42ba29 Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.425940 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-housekeeping-6v7mr"] Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.430742 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-6v7mr" Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.433125 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-scripts" Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.433142 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-config-data" Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.441048 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-6v7mr"] Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.467136 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6f78da3-da1c-4e27-ab65-581c656f74d9-config-data\") pod \"octavia-housekeeping-6v7mr\" (UID: \"b6f78da3-da1c-4e27-ab65-581c656f74d9\") " pod="openstack/octavia-housekeeping-6v7mr" Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.467226 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/b6f78da3-da1c-4e27-ab65-581c656f74d9-hm-ports\") pod \"octavia-housekeeping-6v7mr\" (UID: \"b6f78da3-da1c-4e27-ab65-581c656f74d9\") " pod="openstack/octavia-housekeeping-6v7mr" Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.467263 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6f78da3-da1c-4e27-ab65-581c656f74d9-combined-ca-bundle\") pod \"octavia-housekeeping-6v7mr\" (UID: \"b6f78da3-da1c-4e27-ab65-581c656f74d9\") " pod="openstack/octavia-housekeeping-6v7mr" Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.467344 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/b6f78da3-da1c-4e27-ab65-581c656f74d9-config-data-merged\") pod \"octavia-housekeeping-6v7mr\" (UID: \"b6f78da3-da1c-4e27-ab65-581c656f74d9\") " pod="openstack/octavia-housekeeping-6v7mr" Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.467388 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/b6f78da3-da1c-4e27-ab65-581c656f74d9-amphora-certs\") pod \"octavia-housekeeping-6v7mr\" (UID: \"b6f78da3-da1c-4e27-ab65-581c656f74d9\") " pod="openstack/octavia-housekeeping-6v7mr" Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.467457 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6f78da3-da1c-4e27-ab65-581c656f74d9-scripts\") pod \"octavia-housekeeping-6v7mr\" (UID: \"b6f78da3-da1c-4e27-ab65-581c656f74d9\") " pod="openstack/octavia-housekeeping-6v7mr" Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.569080 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6f78da3-da1c-4e27-ab65-581c656f74d9-scripts\") pod \"octavia-housekeeping-6v7mr\" (UID: \"b6f78da3-da1c-4e27-ab65-581c656f74d9\") " pod="openstack/octavia-housekeeping-6v7mr" Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.569238 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6f78da3-da1c-4e27-ab65-581c656f74d9-config-data\") pod \"octavia-housekeeping-6v7mr\" (UID: \"b6f78da3-da1c-4e27-ab65-581c656f74d9\") " pod="openstack/octavia-housekeeping-6v7mr" Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.569280 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/b6f78da3-da1c-4e27-ab65-581c656f74d9-hm-ports\") pod \"octavia-housekeeping-6v7mr\" (UID: \"b6f78da3-da1c-4e27-ab65-581c656f74d9\") " pod="openstack/octavia-housekeeping-6v7mr" Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.569321 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6f78da3-da1c-4e27-ab65-581c656f74d9-combined-ca-bundle\") pod \"octavia-housekeeping-6v7mr\" (UID: \"b6f78da3-da1c-4e27-ab65-581c656f74d9\") " pod="openstack/octavia-housekeeping-6v7mr" Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.569382 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/b6f78da3-da1c-4e27-ab65-581c656f74d9-config-data-merged\") pod \"octavia-housekeeping-6v7mr\" (UID: \"b6f78da3-da1c-4e27-ab65-581c656f74d9\") " pod="openstack/octavia-housekeeping-6v7mr" Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.569408 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/b6f78da3-da1c-4e27-ab65-581c656f74d9-amphora-certs\") pod \"octavia-housekeeping-6v7mr\" (UID: \"b6f78da3-da1c-4e27-ab65-581c656f74d9\") " pod="openstack/octavia-housekeeping-6v7mr" Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.571202 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/b6f78da3-da1c-4e27-ab65-581c656f74d9-hm-ports\") pod \"octavia-housekeeping-6v7mr\" (UID: \"b6f78da3-da1c-4e27-ab65-581c656f74d9\") " pod="openstack/octavia-housekeeping-6v7mr" Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.571526 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/b6f78da3-da1c-4e27-ab65-581c656f74d9-config-data-merged\") pod \"octavia-housekeeping-6v7mr\" (UID: \"b6f78da3-da1c-4e27-ab65-581c656f74d9\") " pod="openstack/octavia-housekeeping-6v7mr" Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.575688 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6f78da3-da1c-4e27-ab65-581c656f74d9-config-data\") pod \"octavia-housekeeping-6v7mr\" (UID: \"b6f78da3-da1c-4e27-ab65-581c656f74d9\") " pod="openstack/octavia-housekeeping-6v7mr" Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.575919 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6f78da3-da1c-4e27-ab65-581c656f74d9-scripts\") pod \"octavia-housekeeping-6v7mr\" (UID: \"b6f78da3-da1c-4e27-ab65-581c656f74d9\") " pod="openstack/octavia-housekeeping-6v7mr" Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.576774 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/b6f78da3-da1c-4e27-ab65-581c656f74d9-amphora-certs\") pod \"octavia-housekeeping-6v7mr\" (UID: \"b6f78da3-da1c-4e27-ab65-581c656f74d9\") " pod="openstack/octavia-housekeeping-6v7mr" Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.577048 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6f78da3-da1c-4e27-ab65-581c656f74d9-combined-ca-bundle\") pod \"octavia-housekeeping-6v7mr\" (UID: \"b6f78da3-da1c-4e27-ab65-581c656f74d9\") " pod="openstack/octavia-housekeeping-6v7mr" Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.761709 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-6v7mr" Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.785892 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-jmckq" event={"ID":"d32d3e24-6f03-46cc-b7ae-61383778b183","Type":"ContainerStarted","Data":"b085960b93f23fd7b1cbd5e3d2ceff5f975cbe459c57a4692b506d5fc2410bb8"} Jan 27 16:45:34 crc kubenswrapper[4772]: I0127 16:45:34.785939 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-jmckq" event={"ID":"d32d3e24-6f03-46cc-b7ae-61383778b183","Type":"ContainerStarted","Data":"a0320d675c0ddf4fbe99af5532cb287c52a9f435573808fcb1d0149eac42ba29"} Jan 27 16:45:35 crc kubenswrapper[4772]: I0127 16:45:35.303280 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-6v7mr"] Jan 27 16:45:35 crc kubenswrapper[4772]: W0127 16:45:35.309626 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6f78da3_da1c_4e27_ab65_581c656f74d9.slice/crio-e42e3263b76dbd91f396be8169e112da836e1ef6a187e6a6692dbf394b121366 WatchSource:0}: Error finding container e42e3263b76dbd91f396be8169e112da836e1ef6a187e6a6692dbf394b121366: Status 404 returned error can't find the container with id e42e3263b76dbd91f396be8169e112da836e1ef6a187e6a6692dbf394b121366 Jan 27 16:45:35 crc kubenswrapper[4772]: I0127 16:45:35.796196 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-6v7mr" event={"ID":"b6f78da3-da1c-4e27-ab65-581c656f74d9","Type":"ContainerStarted","Data":"e42e3263b76dbd91f396be8169e112da836e1ef6a187e6a6692dbf394b121366"} Jan 27 16:45:35 crc kubenswrapper[4772]: I0127 16:45:35.932077 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-worker-4q859"] Jan 27 16:45:35 crc kubenswrapper[4772]: I0127 16:45:35.935573 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-4q859" Jan 27 16:45:35 crc kubenswrapper[4772]: I0127 16:45:35.938769 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-config-data" Jan 27 16:45:35 crc kubenswrapper[4772]: I0127 16:45:35.940485 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-scripts" Jan 27 16:45:35 crc kubenswrapper[4772]: I0127 16:45:35.942427 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-4q859"] Jan 27 16:45:36 crc kubenswrapper[4772]: I0127 16:45:36.009455 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab431622-b724-4ed4-be2b-67ec8b5956db-config-data\") pod \"octavia-worker-4q859\" (UID: \"ab431622-b724-4ed4-be2b-67ec8b5956db\") " pod="openstack/octavia-worker-4q859" Jan 27 16:45:36 crc kubenswrapper[4772]: I0127 16:45:36.009509 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab431622-b724-4ed4-be2b-67ec8b5956db-scripts\") pod \"octavia-worker-4q859\" (UID: \"ab431622-b724-4ed4-be2b-67ec8b5956db\") " pod="openstack/octavia-worker-4q859" Jan 27 16:45:36 crc kubenswrapper[4772]: I0127 16:45:36.010266 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/ab431622-b724-4ed4-be2b-67ec8b5956db-hm-ports\") pod \"octavia-worker-4q859\" (UID: \"ab431622-b724-4ed4-be2b-67ec8b5956db\") " pod="openstack/octavia-worker-4q859" Jan 27 16:45:36 crc kubenswrapper[4772]: I0127 16:45:36.010495 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/ab431622-b724-4ed4-be2b-67ec8b5956db-amphora-certs\") pod \"octavia-worker-4q859\" (UID: \"ab431622-b724-4ed4-be2b-67ec8b5956db\") " pod="openstack/octavia-worker-4q859" Jan 27 16:45:36 crc kubenswrapper[4772]: I0127 16:45:36.010627 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/ab431622-b724-4ed4-be2b-67ec8b5956db-config-data-merged\") pod \"octavia-worker-4q859\" (UID: \"ab431622-b724-4ed4-be2b-67ec8b5956db\") " pod="openstack/octavia-worker-4q859" Jan 27 16:45:36 crc kubenswrapper[4772]: I0127 16:45:36.010686 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab431622-b724-4ed4-be2b-67ec8b5956db-combined-ca-bundle\") pod \"octavia-worker-4q859\" (UID: \"ab431622-b724-4ed4-be2b-67ec8b5956db\") " pod="openstack/octavia-worker-4q859" Jan 27 16:45:36 crc kubenswrapper[4772]: I0127 16:45:36.112114 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/ab431622-b724-4ed4-be2b-67ec8b5956db-hm-ports\") pod \"octavia-worker-4q859\" (UID: \"ab431622-b724-4ed4-be2b-67ec8b5956db\") " pod="openstack/octavia-worker-4q859" Jan 27 16:45:36 crc kubenswrapper[4772]: I0127 16:45:36.112245 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/ab431622-b724-4ed4-be2b-67ec8b5956db-amphora-certs\") pod \"octavia-worker-4q859\" (UID: \"ab431622-b724-4ed4-be2b-67ec8b5956db\") " pod="openstack/octavia-worker-4q859" Jan 27 16:45:36 crc kubenswrapper[4772]: I0127 16:45:36.112311 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/ab431622-b724-4ed4-be2b-67ec8b5956db-config-data-merged\") pod \"octavia-worker-4q859\" (UID: \"ab431622-b724-4ed4-be2b-67ec8b5956db\") " pod="openstack/octavia-worker-4q859" Jan 27 16:45:36 crc kubenswrapper[4772]: I0127 16:45:36.112346 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab431622-b724-4ed4-be2b-67ec8b5956db-combined-ca-bundle\") pod \"octavia-worker-4q859\" (UID: \"ab431622-b724-4ed4-be2b-67ec8b5956db\") " pod="openstack/octavia-worker-4q859" Jan 27 16:45:36 crc kubenswrapper[4772]: I0127 16:45:36.112401 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab431622-b724-4ed4-be2b-67ec8b5956db-config-data\") pod \"octavia-worker-4q859\" (UID: \"ab431622-b724-4ed4-be2b-67ec8b5956db\") " pod="openstack/octavia-worker-4q859" Jan 27 16:45:36 crc kubenswrapper[4772]: I0127 16:45:36.112435 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab431622-b724-4ed4-be2b-67ec8b5956db-scripts\") pod \"octavia-worker-4q859\" (UID: \"ab431622-b724-4ed4-be2b-67ec8b5956db\") " pod="openstack/octavia-worker-4q859" Jan 27 16:45:36 crc kubenswrapper[4772]: I0127 16:45:36.113269 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/ab431622-b724-4ed4-be2b-67ec8b5956db-hm-ports\") pod \"octavia-worker-4q859\" (UID: \"ab431622-b724-4ed4-be2b-67ec8b5956db\") " pod="openstack/octavia-worker-4q859" Jan 27 16:45:36 crc kubenswrapper[4772]: I0127 16:45:36.113278 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/ab431622-b724-4ed4-be2b-67ec8b5956db-config-data-merged\") pod \"octavia-worker-4q859\" (UID: \"ab431622-b724-4ed4-be2b-67ec8b5956db\") " pod="openstack/octavia-worker-4q859" Jan 27 16:45:36 crc kubenswrapper[4772]: I0127 16:45:36.118736 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab431622-b724-4ed4-be2b-67ec8b5956db-combined-ca-bundle\") pod \"octavia-worker-4q859\" (UID: \"ab431622-b724-4ed4-be2b-67ec8b5956db\") " pod="openstack/octavia-worker-4q859" Jan 27 16:45:36 crc kubenswrapper[4772]: I0127 16:45:36.118832 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/ab431622-b724-4ed4-be2b-67ec8b5956db-amphora-certs\") pod \"octavia-worker-4q859\" (UID: \"ab431622-b724-4ed4-be2b-67ec8b5956db\") " pod="openstack/octavia-worker-4q859" Jan 27 16:45:36 crc kubenswrapper[4772]: I0127 16:45:36.119225 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab431622-b724-4ed4-be2b-67ec8b5956db-config-data\") pod \"octavia-worker-4q859\" (UID: \"ab431622-b724-4ed4-be2b-67ec8b5956db\") " pod="openstack/octavia-worker-4q859" Jan 27 16:45:36 crc kubenswrapper[4772]: I0127 16:45:36.138565 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab431622-b724-4ed4-be2b-67ec8b5956db-scripts\") pod \"octavia-worker-4q859\" (UID: \"ab431622-b724-4ed4-be2b-67ec8b5956db\") " pod="openstack/octavia-worker-4q859" Jan 27 16:45:36 crc kubenswrapper[4772]: I0127 16:45:36.273621 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-4q859" Jan 27 16:45:36 crc kubenswrapper[4772]: I0127 16:45:36.809012 4772 generic.go:334] "Generic (PLEG): container finished" podID="d32d3e24-6f03-46cc-b7ae-61383778b183" containerID="b085960b93f23fd7b1cbd5e3d2ceff5f975cbe459c57a4692b506d5fc2410bb8" exitCode=0 Jan 27 16:45:36 crc kubenswrapper[4772]: I0127 16:45:36.809286 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-jmckq" event={"ID":"d32d3e24-6f03-46cc-b7ae-61383778b183","Type":"ContainerDied","Data":"b085960b93f23fd7b1cbd5e3d2ceff5f975cbe459c57a4692b506d5fc2410bb8"} Jan 27 16:45:37 crc kubenswrapper[4772]: I0127 16:45:37.206681 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-4q859"] Jan 27 16:45:37 crc kubenswrapper[4772]: W0127 16:45:37.213408 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab431622_b724_4ed4_be2b_67ec8b5956db.slice/crio-79c2b30249c61672ff175f58e10d97ce72283314333870eb78ed30e834d9566c WatchSource:0}: Error finding container 79c2b30249c61672ff175f58e10d97ce72283314333870eb78ed30e834d9566c: Status 404 returned error can't find the container with id 79c2b30249c61672ff175f58e10d97ce72283314333870eb78ed30e834d9566c Jan 27 16:45:37 crc kubenswrapper[4772]: I0127 16:45:37.503018 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-jmckq"] Jan 27 16:45:37 crc kubenswrapper[4772]: I0127 16:45:37.662758 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:45:37 crc kubenswrapper[4772]: E0127 16:45:37.663017 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:45:37 crc kubenswrapper[4772]: I0127 16:45:37.823682 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-jmckq" event={"ID":"d32d3e24-6f03-46cc-b7ae-61383778b183","Type":"ContainerStarted","Data":"753e40c780a5346ae3fdfa2d1a721e15c7d2c633fabd41fa91babb4e476992e8"} Jan 27 16:45:37 crc kubenswrapper[4772]: I0127 16:45:37.823839 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-healthmanager-jmckq" Jan 27 16:45:37 crc kubenswrapper[4772]: I0127 16:45:37.825452 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-6v7mr" event={"ID":"b6f78da3-da1c-4e27-ab65-581c656f74d9","Type":"ContainerStarted","Data":"35bfac577b0ff976d875ae38ad40b268daada422e60612c1b77a9bf132cb18ba"} Jan 27 16:45:37 crc kubenswrapper[4772]: I0127 16:45:37.826594 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-4q859" event={"ID":"ab431622-b724-4ed4-be2b-67ec8b5956db","Type":"ContainerStarted","Data":"79c2b30249c61672ff175f58e10d97ce72283314333870eb78ed30e834d9566c"} Jan 27 16:45:37 crc kubenswrapper[4772]: I0127 16:45:37.859941 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-healthmanager-jmckq" podStartSLOduration=5.8599232610000005 podStartE2EDuration="5.859923261s" podCreationTimestamp="2026-01-27 16:45:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 16:45:37.852707626 +0000 UTC m=+5923.833316734" watchObservedRunningTime="2026-01-27 16:45:37.859923261 +0000 UTC m=+5923.840532359" Jan 27 16:45:38 crc kubenswrapper[4772]: I0127 16:45:38.837395 4772 generic.go:334] "Generic (PLEG): container finished" podID="b6f78da3-da1c-4e27-ab65-581c656f74d9" containerID="35bfac577b0ff976d875ae38ad40b268daada422e60612c1b77a9bf132cb18ba" exitCode=0 Jan 27 16:45:38 crc kubenswrapper[4772]: I0127 16:45:38.837512 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-6v7mr" event={"ID":"b6f78da3-da1c-4e27-ab65-581c656f74d9","Type":"ContainerDied","Data":"35bfac577b0ff976d875ae38ad40b268daada422e60612c1b77a9bf132cb18ba"} Jan 27 16:45:39 crc kubenswrapper[4772]: I0127 16:45:39.852096 4772 generic.go:334] "Generic (PLEG): container finished" podID="ab431622-b724-4ed4-be2b-67ec8b5956db" containerID="66a2fc4b699fdb30694179059bd5f5442f33dad785be85640528cad608c6c9a6" exitCode=0 Jan 27 16:45:39 crc kubenswrapper[4772]: I0127 16:45:39.852152 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-4q859" event={"ID":"ab431622-b724-4ed4-be2b-67ec8b5956db","Type":"ContainerDied","Data":"66a2fc4b699fdb30694179059bd5f5442f33dad785be85640528cad608c6c9a6"} Jan 27 16:45:39 crc kubenswrapper[4772]: I0127 16:45:39.859575 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-6v7mr" event={"ID":"b6f78da3-da1c-4e27-ab65-581c656f74d9","Type":"ContainerStarted","Data":"612c176fae5b04d8bbe36a3dfabc9d240cc5176477b5b9f0f904aef817a2b4a6"} Jan 27 16:45:39 crc kubenswrapper[4772]: I0127 16:45:39.860989 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-housekeeping-6v7mr" Jan 27 16:45:39 crc kubenswrapper[4772]: I0127 16:45:39.970245 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-housekeeping-6v7mr" podStartSLOduration=4.634420626 podStartE2EDuration="5.970220262s" podCreationTimestamp="2026-01-27 16:45:34 +0000 UTC" firstStartedPulling="2026-01-27 16:45:35.317893313 +0000 UTC m=+5921.298502411" lastFinishedPulling="2026-01-27 16:45:36.653692949 +0000 UTC m=+5922.634302047" observedRunningTime="2026-01-27 16:45:39.931053386 +0000 UTC m=+5925.911662514" watchObservedRunningTime="2026-01-27 16:45:39.970220262 +0000 UTC m=+5925.950829360" Jan 27 16:45:40 crc kubenswrapper[4772]: I0127 16:45:40.870833 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-4q859" event={"ID":"ab431622-b724-4ed4-be2b-67ec8b5956db","Type":"ContainerStarted","Data":"551847f80254beee211eac32b0f14ff96dfe201400eed721c5c9fbf12bef6d1c"} Jan 27 16:45:40 crc kubenswrapper[4772]: I0127 16:45:40.871471 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-worker-4q859" Jan 27 16:45:40 crc kubenswrapper[4772]: I0127 16:45:40.900539 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-worker-4q859" podStartSLOduration=4.712181837 podStartE2EDuration="5.9005165s" podCreationTimestamp="2026-01-27 16:45:35 +0000 UTC" firstStartedPulling="2026-01-27 16:45:37.215738379 +0000 UTC m=+5923.196347477" lastFinishedPulling="2026-01-27 16:45:38.404073032 +0000 UTC m=+5924.384682140" observedRunningTime="2026-01-27 16:45:40.888340833 +0000 UTC m=+5926.868949971" watchObservedRunningTime="2026-01-27 16:45:40.9005165 +0000 UTC m=+5926.881125608" Jan 27 16:45:48 crc kubenswrapper[4772]: I0127 16:45:48.302972 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-healthmanager-jmckq" Jan 27 16:45:49 crc kubenswrapper[4772]: I0127 16:45:49.804062 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-housekeeping-6v7mr" Jan 27 16:45:50 crc kubenswrapper[4772]: I0127 16:45:50.663792 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:45:50 crc kubenswrapper[4772]: E0127 16:45:50.664133 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:45:51 crc kubenswrapper[4772]: I0127 16:45:51.309150 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-worker-4q859" Jan 27 16:46:03 crc kubenswrapper[4772]: I0127 16:46:03.663549 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:46:03 crc kubenswrapper[4772]: E0127 16:46:03.669412 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:46:07 crc kubenswrapper[4772]: I0127 16:46:07.904909 4772 scope.go:117] "RemoveContainer" containerID="fba71bd5b1f04f4ce21ecca46c2028b0aeffaa1f208e5b17978c2d0f906cbf36" Jan 27 16:46:09 crc kubenswrapper[4772]: I0127 16:46:09.041463 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-rxmk9"] Jan 27 16:46:09 crc kubenswrapper[4772]: I0127 16:46:09.051047 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-rxmk9"] Jan 27 16:46:09 crc kubenswrapper[4772]: I0127 16:46:09.061098 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-2bff-account-create-update-nq7qs"] Jan 27 16:46:09 crc kubenswrapper[4772]: I0127 16:46:09.068246 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-2bff-account-create-update-nq7qs"] Jan 27 16:46:10 crc kubenswrapper[4772]: I0127 16:46:10.677277 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fe4269c-3ff7-49b4-82c3-1a419f676f89" path="/var/lib/kubelet/pods/1fe4269c-3ff7-49b4-82c3-1a419f676f89/volumes" Jan 27 16:46:10 crc kubenswrapper[4772]: I0127 16:46:10.679207 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d02ab9dd-317c-4787-aa94-0ad8dff15380" path="/var/lib/kubelet/pods/d02ab9dd-317c-4787-aa94-0ad8dff15380/volumes" Jan 27 16:46:15 crc kubenswrapper[4772]: I0127 16:46:15.038569 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-2jctn"] Jan 27 16:46:15 crc kubenswrapper[4772]: I0127 16:46:15.047352 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-2jctn"] Jan 27 16:46:15 crc kubenswrapper[4772]: I0127 16:46:15.663007 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:46:15 crc kubenswrapper[4772]: E0127 16:46:15.663306 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:46:16 crc kubenswrapper[4772]: I0127 16:46:16.684623 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="370ff587-7186-4f81-83a2-886a15900229" path="/var/lib/kubelet/pods/370ff587-7186-4f81-83a2-886a15900229/volumes" Jan 27 16:46:26 crc kubenswrapper[4772]: I0127 16:46:26.662735 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:46:26 crc kubenswrapper[4772]: E0127 16:46:26.663660 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:46:37 crc kubenswrapper[4772]: I0127 16:46:37.663953 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:46:37 crc kubenswrapper[4772]: E0127 16:46:37.664754 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:46:45 crc kubenswrapper[4772]: I0127 16:46:45.055821 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-fbd6-account-create-update-kvjsf"] Jan 27 16:46:45 crc kubenswrapper[4772]: I0127 16:46:45.064813 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-49fxh"] Jan 27 16:46:45 crc kubenswrapper[4772]: I0127 16:46:45.073594 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-fbd6-account-create-update-kvjsf"] Jan 27 16:46:45 crc kubenswrapper[4772]: I0127 16:46:45.106109 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-49fxh"] Jan 27 16:46:46 crc kubenswrapper[4772]: I0127 16:46:46.681555 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b" path="/var/lib/kubelet/pods/0c10ae3d-af19-4efe-a5b0-783ccf7c5f1b/volumes" Jan 27 16:46:46 crc kubenswrapper[4772]: I0127 16:46:46.684893 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dd740e1-b6ee-444d-b71e-e18d4837ef8a" path="/var/lib/kubelet/pods/6dd740e1-b6ee-444d-b71e-e18d4837ef8a/volumes" Jan 27 16:46:48 crc kubenswrapper[4772]: I0127 16:46:48.663782 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:46:49 crc kubenswrapper[4772]: I0127 16:46:49.578114 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"fad112d2b60be41180a280804db9237c789adfcc03cd7342f9bc2b818aa8e8b4"} Jan 27 16:46:58 crc kubenswrapper[4772]: I0127 16:46:58.041240 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-xvkmn"] Jan 27 16:46:58 crc kubenswrapper[4772]: I0127 16:46:58.057151 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-xvkmn"] Jan 27 16:46:58 crc kubenswrapper[4772]: I0127 16:46:58.677644 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2" path="/var/lib/kubelet/pods/fce2770e-b90c-4d9c-8a0c-40fe0f48aeb2/volumes" Jan 27 16:47:08 crc kubenswrapper[4772]: I0127 16:47:08.016968 4772 scope.go:117] "RemoveContainer" containerID="5a8e2d8cff2b48fb4540fd3d3c996a83107f6a91a01287bf66fdb29b10e52c6b" Jan 27 16:47:08 crc kubenswrapper[4772]: I0127 16:47:08.063257 4772 scope.go:117] "RemoveContainer" containerID="47404bfd4ce996befd817152654b651725e0f51b6b3b55ee5ee110296d25e0c8" Jan 27 16:47:08 crc kubenswrapper[4772]: I0127 16:47:08.111543 4772 scope.go:117] "RemoveContainer" containerID="0eb151bf9a1bfafe986b27f8d85f2a3fcbd2a1f6be731bcd2ce95359e6a6e136" Jan 27 16:47:08 crc kubenswrapper[4772]: I0127 16:47:08.141231 4772 scope.go:117] "RemoveContainer" containerID="ccfecf4ef23c50d400b8ca7553bbf1eff4d9d62e8caff28f5a3562a7630cf393" Jan 27 16:47:08 crc kubenswrapper[4772]: I0127 16:47:08.201487 4772 scope.go:117] "RemoveContainer" containerID="7820202ed4fe89d2fb42f752ab53ea23a8554f7e3772282ffa645c3878d8acde" Jan 27 16:47:08 crc kubenswrapper[4772]: I0127 16:47:08.224716 4772 scope.go:117] "RemoveContainer" containerID="39e5a537d87039cbc6538b5672d3a7484ab8d88c163808f5672ea54c55115098" Jan 27 16:47:35 crc kubenswrapper[4772]: I0127 16:47:35.277293 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-68vzn"] Jan 27 16:47:35 crc kubenswrapper[4772]: I0127 16:47:35.281317 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-68vzn" Jan 27 16:47:35 crc kubenswrapper[4772]: I0127 16:47:35.284047 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-68vzn"] Jan 27 16:47:35 crc kubenswrapper[4772]: I0127 16:47:35.351591 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cc67403-20b9-4476-be2c-26639283d648-catalog-content\") pod \"redhat-operators-68vzn\" (UID: \"9cc67403-20b9-4476-be2c-26639283d648\") " pod="openshift-marketplace/redhat-operators-68vzn" Jan 27 16:47:35 crc kubenswrapper[4772]: I0127 16:47:35.351679 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk77k\" (UniqueName: \"kubernetes.io/projected/9cc67403-20b9-4476-be2c-26639283d648-kube-api-access-lk77k\") pod \"redhat-operators-68vzn\" (UID: \"9cc67403-20b9-4476-be2c-26639283d648\") " pod="openshift-marketplace/redhat-operators-68vzn" Jan 27 16:47:35 crc kubenswrapper[4772]: I0127 16:47:35.351826 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cc67403-20b9-4476-be2c-26639283d648-utilities\") pod \"redhat-operators-68vzn\" (UID: \"9cc67403-20b9-4476-be2c-26639283d648\") " pod="openshift-marketplace/redhat-operators-68vzn" Jan 27 16:47:35 crc kubenswrapper[4772]: I0127 16:47:35.453080 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cc67403-20b9-4476-be2c-26639283d648-catalog-content\") pod \"redhat-operators-68vzn\" (UID: \"9cc67403-20b9-4476-be2c-26639283d648\") " pod="openshift-marketplace/redhat-operators-68vzn" Jan 27 16:47:35 crc kubenswrapper[4772]: I0127 16:47:35.453137 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk77k\" (UniqueName: \"kubernetes.io/projected/9cc67403-20b9-4476-be2c-26639283d648-kube-api-access-lk77k\") pod \"redhat-operators-68vzn\" (UID: \"9cc67403-20b9-4476-be2c-26639283d648\") " pod="openshift-marketplace/redhat-operators-68vzn" Jan 27 16:47:35 crc kubenswrapper[4772]: I0127 16:47:35.453243 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cc67403-20b9-4476-be2c-26639283d648-utilities\") pod \"redhat-operators-68vzn\" (UID: \"9cc67403-20b9-4476-be2c-26639283d648\") " pod="openshift-marketplace/redhat-operators-68vzn" Jan 27 16:47:35 crc kubenswrapper[4772]: I0127 16:47:35.453593 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cc67403-20b9-4476-be2c-26639283d648-catalog-content\") pod \"redhat-operators-68vzn\" (UID: \"9cc67403-20b9-4476-be2c-26639283d648\") " pod="openshift-marketplace/redhat-operators-68vzn" Jan 27 16:47:35 crc kubenswrapper[4772]: I0127 16:47:35.453612 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cc67403-20b9-4476-be2c-26639283d648-utilities\") pod \"redhat-operators-68vzn\" (UID: \"9cc67403-20b9-4476-be2c-26639283d648\") " pod="openshift-marketplace/redhat-operators-68vzn" Jan 27 16:47:35 crc kubenswrapper[4772]: I0127 16:47:35.470977 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk77k\" (UniqueName: \"kubernetes.io/projected/9cc67403-20b9-4476-be2c-26639283d648-kube-api-access-lk77k\") pod \"redhat-operators-68vzn\" (UID: \"9cc67403-20b9-4476-be2c-26639283d648\") " pod="openshift-marketplace/redhat-operators-68vzn" Jan 27 16:47:35 crc kubenswrapper[4772]: I0127 16:47:35.625811 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-68vzn" Jan 27 16:47:36 crc kubenswrapper[4772]: I0127 16:47:36.105098 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-68vzn"] Jan 27 16:47:37 crc kubenswrapper[4772]: I0127 16:47:37.096601 4772 generic.go:334] "Generic (PLEG): container finished" podID="9cc67403-20b9-4476-be2c-26639283d648" containerID="bcb1f038aa2b5d55b98fa48261958780ab7273f33a9cbb5d63c0080c7664d451" exitCode=0 Jan 27 16:47:37 crc kubenswrapper[4772]: I0127 16:47:37.096680 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68vzn" event={"ID":"9cc67403-20b9-4476-be2c-26639283d648","Type":"ContainerDied","Data":"bcb1f038aa2b5d55b98fa48261958780ab7273f33a9cbb5d63c0080c7664d451"} Jan 27 16:47:37 crc kubenswrapper[4772]: I0127 16:47:37.096848 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68vzn" event={"ID":"9cc67403-20b9-4476-be2c-26639283d648","Type":"ContainerStarted","Data":"81502ee598efdd39b85faa6c7f5c54aee251d669948a2c94968e20419d6f1166"} Jan 27 16:47:38 crc kubenswrapper[4772]: I0127 16:47:38.115816 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68vzn" event={"ID":"9cc67403-20b9-4476-be2c-26639283d648","Type":"ContainerStarted","Data":"b04b8450f2bdaed53e1022022c6c5bc40b03e9ec1d425383d5676d25979a0357"} Jan 27 16:47:39 crc kubenswrapper[4772]: I0127 16:47:39.139397 4772 generic.go:334] "Generic (PLEG): container finished" podID="9cc67403-20b9-4476-be2c-26639283d648" containerID="b04b8450f2bdaed53e1022022c6c5bc40b03e9ec1d425383d5676d25979a0357" exitCode=0 Jan 27 16:47:39 crc kubenswrapper[4772]: I0127 16:47:39.139489 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68vzn" event={"ID":"9cc67403-20b9-4476-be2c-26639283d648","Type":"ContainerDied","Data":"b04b8450f2bdaed53e1022022c6c5bc40b03e9ec1d425383d5676d25979a0357"} Jan 27 16:47:40 crc kubenswrapper[4772]: I0127 16:47:40.039447 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-ab59-account-create-update-p9dsf"] Jan 27 16:47:40 crc kubenswrapper[4772]: I0127 16:47:40.050746 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-ng4h8"] Jan 27 16:47:40 crc kubenswrapper[4772]: I0127 16:47:40.061825 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-ab59-account-create-update-p9dsf"] Jan 27 16:47:40 crc kubenswrapper[4772]: I0127 16:47:40.071761 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-ng4h8"] Jan 27 16:47:40 crc kubenswrapper[4772]: I0127 16:47:40.153332 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68vzn" event={"ID":"9cc67403-20b9-4476-be2c-26639283d648","Type":"ContainerStarted","Data":"181b8968a78fb7fd292ffd699059c3cf38d97011a6ab52479b4d69804b746f20"} Jan 27 16:47:40 crc kubenswrapper[4772]: I0127 16:47:40.180327 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-68vzn" podStartSLOduration=2.737921604 podStartE2EDuration="5.180309673s" podCreationTimestamp="2026-01-27 16:47:35 +0000 UTC" firstStartedPulling="2026-01-27 16:47:37.099562858 +0000 UTC m=+6043.080171956" lastFinishedPulling="2026-01-27 16:47:39.541950917 +0000 UTC m=+6045.522560025" observedRunningTime="2026-01-27 16:47:40.170448592 +0000 UTC m=+6046.151057720" watchObservedRunningTime="2026-01-27 16:47:40.180309673 +0000 UTC m=+6046.160918771" Jan 27 16:47:40 crc kubenswrapper[4772]: I0127 16:47:40.676848 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b150c73-f6eb-4193-81ef-84941ff1abef" path="/var/lib/kubelet/pods/2b150c73-f6eb-4193-81ef-84941ff1abef/volumes" Jan 27 16:47:40 crc kubenswrapper[4772]: I0127 16:47:40.678266 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="615446cc-6fba-46f2-aad9-434f11519be9" path="/var/lib/kubelet/pods/615446cc-6fba-46f2-aad9-434f11519be9/volumes" Jan 27 16:47:45 crc kubenswrapper[4772]: I0127 16:47:45.626561 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-68vzn" Jan 27 16:47:45 crc kubenswrapper[4772]: I0127 16:47:45.627134 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-68vzn" Jan 27 16:47:46 crc kubenswrapper[4772]: I0127 16:47:46.672946 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-68vzn" podUID="9cc67403-20b9-4476-be2c-26639283d648" containerName="registry-server" probeResult="failure" output=< Jan 27 16:47:46 crc kubenswrapper[4772]: timeout: failed to connect service ":50051" within 1s Jan 27 16:47:46 crc kubenswrapper[4772]: > Jan 27 16:47:48 crc kubenswrapper[4772]: I0127 16:47:48.036272 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-wlbbm"] Jan 27 16:47:48 crc kubenswrapper[4772]: I0127 16:47:48.047476 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-wlbbm"] Jan 27 16:47:48 crc kubenswrapper[4772]: I0127 16:47:48.679791 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac290494-b5ad-4d85-9f14-daf092e3a6ed" path="/var/lib/kubelet/pods/ac290494-b5ad-4d85-9f14-daf092e3a6ed/volumes" Jan 27 16:47:55 crc kubenswrapper[4772]: I0127 16:47:55.688049 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-68vzn" Jan 27 16:47:55 crc kubenswrapper[4772]: I0127 16:47:55.757150 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-68vzn" Jan 27 16:47:55 crc kubenswrapper[4772]: I0127 16:47:55.950946 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-68vzn"] Jan 27 16:47:57 crc kubenswrapper[4772]: I0127 16:47:57.341310 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-68vzn" podUID="9cc67403-20b9-4476-be2c-26639283d648" containerName="registry-server" containerID="cri-o://181b8968a78fb7fd292ffd699059c3cf38d97011a6ab52479b4d69804b746f20" gracePeriod=2 Jan 27 16:47:57 crc kubenswrapper[4772]: I0127 16:47:57.807109 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-68vzn" Jan 27 16:47:57 crc kubenswrapper[4772]: I0127 16:47:57.833877 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cc67403-20b9-4476-be2c-26639283d648-utilities\") pod \"9cc67403-20b9-4476-be2c-26639283d648\" (UID: \"9cc67403-20b9-4476-be2c-26639283d648\") " Jan 27 16:47:57 crc kubenswrapper[4772]: I0127 16:47:57.833950 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cc67403-20b9-4476-be2c-26639283d648-catalog-content\") pod \"9cc67403-20b9-4476-be2c-26639283d648\" (UID: \"9cc67403-20b9-4476-be2c-26639283d648\") " Jan 27 16:47:57 crc kubenswrapper[4772]: I0127 16:47:57.834230 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lk77k\" (UniqueName: \"kubernetes.io/projected/9cc67403-20b9-4476-be2c-26639283d648-kube-api-access-lk77k\") pod \"9cc67403-20b9-4476-be2c-26639283d648\" (UID: \"9cc67403-20b9-4476-be2c-26639283d648\") " Jan 27 16:47:57 crc kubenswrapper[4772]: I0127 16:47:57.834784 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cc67403-20b9-4476-be2c-26639283d648-utilities" (OuterVolumeSpecName: "utilities") pod "9cc67403-20b9-4476-be2c-26639283d648" (UID: "9cc67403-20b9-4476-be2c-26639283d648"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:47:57 crc kubenswrapper[4772]: I0127 16:47:57.834939 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cc67403-20b9-4476-be2c-26639283d648-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 16:47:57 crc kubenswrapper[4772]: I0127 16:47:57.839523 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cc67403-20b9-4476-be2c-26639283d648-kube-api-access-lk77k" (OuterVolumeSpecName: "kube-api-access-lk77k") pod "9cc67403-20b9-4476-be2c-26639283d648" (UID: "9cc67403-20b9-4476-be2c-26639283d648"). InnerVolumeSpecName "kube-api-access-lk77k". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:47:57 crc kubenswrapper[4772]: I0127 16:47:57.936756 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lk77k\" (UniqueName: \"kubernetes.io/projected/9cc67403-20b9-4476-be2c-26639283d648-kube-api-access-lk77k\") on node \"crc\" DevicePath \"\"" Jan 27 16:47:57 crc kubenswrapper[4772]: I0127 16:47:57.945179 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cc67403-20b9-4476-be2c-26639283d648-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9cc67403-20b9-4476-be2c-26639283d648" (UID: "9cc67403-20b9-4476-be2c-26639283d648"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:47:58 crc kubenswrapper[4772]: I0127 16:47:58.040334 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cc67403-20b9-4476-be2c-26639283d648-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 16:47:58 crc kubenswrapper[4772]: I0127 16:47:58.356676 4772 generic.go:334] "Generic (PLEG): container finished" podID="9cc67403-20b9-4476-be2c-26639283d648" containerID="181b8968a78fb7fd292ffd699059c3cf38d97011a6ab52479b4d69804b746f20" exitCode=0 Jan 27 16:47:58 crc kubenswrapper[4772]: I0127 16:47:58.356733 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68vzn" event={"ID":"9cc67403-20b9-4476-be2c-26639283d648","Type":"ContainerDied","Data":"181b8968a78fb7fd292ffd699059c3cf38d97011a6ab52479b4d69804b746f20"} Jan 27 16:47:58 crc kubenswrapper[4772]: I0127 16:47:58.356786 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68vzn" event={"ID":"9cc67403-20b9-4476-be2c-26639283d648","Type":"ContainerDied","Data":"81502ee598efdd39b85faa6c7f5c54aee251d669948a2c94968e20419d6f1166"} Jan 27 16:47:58 crc kubenswrapper[4772]: I0127 16:47:58.356787 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-68vzn" Jan 27 16:47:58 crc kubenswrapper[4772]: I0127 16:47:58.356803 4772 scope.go:117] "RemoveContainer" containerID="181b8968a78fb7fd292ffd699059c3cf38d97011a6ab52479b4d69804b746f20" Jan 27 16:47:58 crc kubenswrapper[4772]: I0127 16:47:58.402811 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-68vzn"] Jan 27 16:47:58 crc kubenswrapper[4772]: I0127 16:47:58.409412 4772 scope.go:117] "RemoveContainer" containerID="b04b8450f2bdaed53e1022022c6c5bc40b03e9ec1d425383d5676d25979a0357" Jan 27 16:47:58 crc kubenswrapper[4772]: I0127 16:47:58.410075 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-68vzn"] Jan 27 16:47:58 crc kubenswrapper[4772]: I0127 16:47:58.432520 4772 scope.go:117] "RemoveContainer" containerID="bcb1f038aa2b5d55b98fa48261958780ab7273f33a9cbb5d63c0080c7664d451" Jan 27 16:47:58 crc kubenswrapper[4772]: I0127 16:47:58.471750 4772 scope.go:117] "RemoveContainer" containerID="181b8968a78fb7fd292ffd699059c3cf38d97011a6ab52479b4d69804b746f20" Jan 27 16:47:58 crc kubenswrapper[4772]: E0127 16:47:58.472349 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"181b8968a78fb7fd292ffd699059c3cf38d97011a6ab52479b4d69804b746f20\": container with ID starting with 181b8968a78fb7fd292ffd699059c3cf38d97011a6ab52479b4d69804b746f20 not found: ID does not exist" containerID="181b8968a78fb7fd292ffd699059c3cf38d97011a6ab52479b4d69804b746f20" Jan 27 16:47:58 crc kubenswrapper[4772]: I0127 16:47:58.472424 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"181b8968a78fb7fd292ffd699059c3cf38d97011a6ab52479b4d69804b746f20"} err="failed to get container status \"181b8968a78fb7fd292ffd699059c3cf38d97011a6ab52479b4d69804b746f20\": rpc error: code = NotFound desc = could not find container \"181b8968a78fb7fd292ffd699059c3cf38d97011a6ab52479b4d69804b746f20\": container with ID starting with 181b8968a78fb7fd292ffd699059c3cf38d97011a6ab52479b4d69804b746f20 not found: ID does not exist" Jan 27 16:47:58 crc kubenswrapper[4772]: I0127 16:47:58.472447 4772 scope.go:117] "RemoveContainer" containerID="b04b8450f2bdaed53e1022022c6c5bc40b03e9ec1d425383d5676d25979a0357" Jan 27 16:47:58 crc kubenswrapper[4772]: E0127 16:47:58.472697 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b04b8450f2bdaed53e1022022c6c5bc40b03e9ec1d425383d5676d25979a0357\": container with ID starting with b04b8450f2bdaed53e1022022c6c5bc40b03e9ec1d425383d5676d25979a0357 not found: ID does not exist" containerID="b04b8450f2bdaed53e1022022c6c5bc40b03e9ec1d425383d5676d25979a0357" Jan 27 16:47:58 crc kubenswrapper[4772]: I0127 16:47:58.472732 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b04b8450f2bdaed53e1022022c6c5bc40b03e9ec1d425383d5676d25979a0357"} err="failed to get container status \"b04b8450f2bdaed53e1022022c6c5bc40b03e9ec1d425383d5676d25979a0357\": rpc error: code = NotFound desc = could not find container \"b04b8450f2bdaed53e1022022c6c5bc40b03e9ec1d425383d5676d25979a0357\": container with ID starting with b04b8450f2bdaed53e1022022c6c5bc40b03e9ec1d425383d5676d25979a0357 not found: ID does not exist" Jan 27 16:47:58 crc kubenswrapper[4772]: I0127 16:47:58.472750 4772 scope.go:117] "RemoveContainer" containerID="bcb1f038aa2b5d55b98fa48261958780ab7273f33a9cbb5d63c0080c7664d451" Jan 27 16:47:58 crc kubenswrapper[4772]: E0127 16:47:58.473027 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcb1f038aa2b5d55b98fa48261958780ab7273f33a9cbb5d63c0080c7664d451\": container with ID starting with bcb1f038aa2b5d55b98fa48261958780ab7273f33a9cbb5d63c0080c7664d451 not found: ID does not exist" containerID="bcb1f038aa2b5d55b98fa48261958780ab7273f33a9cbb5d63c0080c7664d451" Jan 27 16:47:58 crc kubenswrapper[4772]: I0127 16:47:58.473070 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcb1f038aa2b5d55b98fa48261958780ab7273f33a9cbb5d63c0080c7664d451"} err="failed to get container status \"bcb1f038aa2b5d55b98fa48261958780ab7273f33a9cbb5d63c0080c7664d451\": rpc error: code = NotFound desc = could not find container \"bcb1f038aa2b5d55b98fa48261958780ab7273f33a9cbb5d63c0080c7664d451\": container with ID starting with bcb1f038aa2b5d55b98fa48261958780ab7273f33a9cbb5d63c0080c7664d451 not found: ID does not exist" Jan 27 16:47:58 crc kubenswrapper[4772]: I0127 16:47:58.677401 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cc67403-20b9-4476-be2c-26639283d648" path="/var/lib/kubelet/pods/9cc67403-20b9-4476-be2c-26639283d648/volumes" Jan 27 16:48:08 crc kubenswrapper[4772]: I0127 16:48:08.383297 4772 scope.go:117] "RemoveContainer" containerID="39b253bea060fd0ef46113186002866f36924e03fe711afec1f871af51f40edf" Jan 27 16:48:08 crc kubenswrapper[4772]: I0127 16:48:08.417437 4772 scope.go:117] "RemoveContainer" containerID="972404acaa773e237766879c66e6f19a8b3951bb52780f66333fe8405eb0ccb2" Jan 27 16:48:08 crc kubenswrapper[4772]: I0127 16:48:08.466302 4772 scope.go:117] "RemoveContainer" containerID="04b6871605c5a7ec7b4197615c41edaa0e9453b396fcff037ce590df9243c6a0" Jan 27 16:48:22 crc kubenswrapper[4772]: I0127 16:48:22.064606 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-wfhq4"] Jan 27 16:48:22 crc kubenswrapper[4772]: I0127 16:48:22.075023 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-c674-account-create-update-zr9fr"] Jan 27 16:48:22 crc kubenswrapper[4772]: I0127 16:48:22.084972 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-wfhq4"] Jan 27 16:48:22 crc kubenswrapper[4772]: I0127 16:48:22.094094 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-c674-account-create-update-zr9fr"] Jan 27 16:48:22 crc kubenswrapper[4772]: I0127 16:48:22.675792 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a6c1c65-36ca-4017-a8e2-5e22a550d601" path="/var/lib/kubelet/pods/5a6c1c65-36ca-4017-a8e2-5e22a550d601/volumes" Jan 27 16:48:22 crc kubenswrapper[4772]: I0127 16:48:22.677645 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68c87e84-0237-41a3-b248-59f0e0156b81" path="/var/lib/kubelet/pods/68c87e84-0237-41a3-b248-59f0e0156b81/volumes" Jan 27 16:48:27 crc kubenswrapper[4772]: I0127 16:48:27.032615 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-4pvhb"] Jan 27 16:48:27 crc kubenswrapper[4772]: I0127 16:48:27.041827 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-4pvhb"] Jan 27 16:48:28 crc kubenswrapper[4772]: I0127 16:48:28.675374 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b972e003-d915-4c6e-b84e-00d1f53740c1" path="/var/lib/kubelet/pods/b972e003-d915-4c6e-b84e-00d1f53740c1/volumes" Jan 27 16:49:08 crc kubenswrapper[4772]: I0127 16:49:08.563515 4772 scope.go:117] "RemoveContainer" containerID="159f9ff911847523ab0387be1212efb17cae848a6dcdc1e80961565a39d1eac9" Jan 27 16:49:08 crc kubenswrapper[4772]: I0127 16:49:08.587540 4772 scope.go:117] "RemoveContainer" containerID="fffc3d88e7cbd76e4f8c55e9d4f80e7dfb6325460b6429ed67f89d060f3c380c" Jan 27 16:49:08 crc kubenswrapper[4772]: I0127 16:49:08.631968 4772 scope.go:117] "RemoveContainer" containerID="897a6f6215480fb2b302f194b142ae62f3461ba459140cef6dbb5530febc39e7" Jan 27 16:49:12 crc kubenswrapper[4772]: I0127 16:49:12.058907 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:49:12 crc kubenswrapper[4772]: I0127 16:49:12.059598 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:49:24 crc kubenswrapper[4772]: I0127 16:49:24.048999 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-5g8c9"] Jan 27 16:49:24 crc kubenswrapper[4772]: I0127 16:49:24.058723 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-wmvjd"] Jan 27 16:49:24 crc kubenswrapper[4772]: I0127 16:49:24.078881 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-5g8c9"] Jan 27 16:49:24 crc kubenswrapper[4772]: I0127 16:49:24.095346 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-wmvjd"] Jan 27 16:49:24 crc kubenswrapper[4772]: I0127 16:49:24.681208 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81b3d773-720e-42c5-af9e-abddc2180ac7" path="/var/lib/kubelet/pods/81b3d773-720e-42c5-af9e-abddc2180ac7/volumes" Jan 27 16:49:24 crc kubenswrapper[4772]: I0127 16:49:24.683131 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="952d9a1e-efbf-4617-94af-b5ad42cce494" path="/var/lib/kubelet/pods/952d9a1e-efbf-4617-94af-b5ad42cce494/volumes" Jan 27 16:49:25 crc kubenswrapper[4772]: I0127 16:49:25.034205 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-a20c-account-create-update-7pf8l"] Jan 27 16:49:25 crc kubenswrapper[4772]: I0127 16:49:25.046056 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-a20c-account-create-update-7pf8l"] Jan 27 16:49:25 crc kubenswrapper[4772]: I0127 16:49:25.057335 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-a831-account-create-update-bq6jl"] Jan 27 16:49:25 crc kubenswrapper[4772]: I0127 16:49:25.066700 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-fxhzl"] Jan 27 16:49:25 crc kubenswrapper[4772]: I0127 16:49:25.076738 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-e098-account-create-update-s8qnw"] Jan 27 16:49:25 crc kubenswrapper[4772]: I0127 16:49:25.086526 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-a831-account-create-update-bq6jl"] Jan 27 16:49:25 crc kubenswrapper[4772]: I0127 16:49:25.096433 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-fxhzl"] Jan 27 16:49:25 crc kubenswrapper[4772]: I0127 16:49:25.106611 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-e098-account-create-update-s8qnw"] Jan 27 16:49:26 crc kubenswrapper[4772]: I0127 16:49:26.674987 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4545581d-5f56-406d-938f-c3b073fdcbce" path="/var/lib/kubelet/pods/4545581d-5f56-406d-938f-c3b073fdcbce/volumes" Jan 27 16:49:26 crc kubenswrapper[4772]: I0127 16:49:26.676234 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84c86af5-fd1f-4c53-8978-2b436db59b2a" path="/var/lib/kubelet/pods/84c86af5-fd1f-4c53-8978-2b436db59b2a/volumes" Jan 27 16:49:26 crc kubenswrapper[4772]: I0127 16:49:26.677481 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86f6e427-99ce-4873-bacc-697edca3d34e" path="/var/lib/kubelet/pods/86f6e427-99ce-4873-bacc-697edca3d34e/volumes" Jan 27 16:49:26 crc kubenswrapper[4772]: I0127 16:49:26.678774 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e808813a-e588-4fb9-a15d-588d94a4cd59" path="/var/lib/kubelet/pods/e808813a-e588-4fb9-a15d-588d94a4cd59/volumes" Jan 27 16:49:39 crc kubenswrapper[4772]: I0127 16:49:39.049718 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k56t7"] Jan 27 16:49:39 crc kubenswrapper[4772]: I0127 16:49:39.061949 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k56t7"] Jan 27 16:49:40 crc kubenswrapper[4772]: I0127 16:49:40.675226 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eedd23ad-e532-401a-a991-4bca54fc2711" path="/var/lib/kubelet/pods/eedd23ad-e532-401a-a991-4bca54fc2711/volumes" Jan 27 16:49:42 crc kubenswrapper[4772]: I0127 16:49:42.058105 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:49:42 crc kubenswrapper[4772]: I0127 16:49:42.058648 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:49:57 crc kubenswrapper[4772]: I0127 16:49:57.236432 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gwt4f"] Jan 27 16:49:57 crc kubenswrapper[4772]: E0127 16:49:57.237429 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cc67403-20b9-4476-be2c-26639283d648" containerName="extract-content" Jan 27 16:49:57 crc kubenswrapper[4772]: I0127 16:49:57.237447 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cc67403-20b9-4476-be2c-26639283d648" containerName="extract-content" Jan 27 16:49:57 crc kubenswrapper[4772]: E0127 16:49:57.237460 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cc67403-20b9-4476-be2c-26639283d648" containerName="extract-utilities" Jan 27 16:49:57 crc kubenswrapper[4772]: I0127 16:49:57.237468 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cc67403-20b9-4476-be2c-26639283d648" containerName="extract-utilities" Jan 27 16:49:57 crc kubenswrapper[4772]: E0127 16:49:57.237484 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cc67403-20b9-4476-be2c-26639283d648" containerName="registry-server" Jan 27 16:49:57 crc kubenswrapper[4772]: I0127 16:49:57.237493 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cc67403-20b9-4476-be2c-26639283d648" containerName="registry-server" Jan 27 16:49:57 crc kubenswrapper[4772]: I0127 16:49:57.237716 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cc67403-20b9-4476-be2c-26639283d648" containerName="registry-server" Jan 27 16:49:57 crc kubenswrapper[4772]: I0127 16:49:57.239263 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gwt4f" Jan 27 16:49:57 crc kubenswrapper[4772]: I0127 16:49:57.252678 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gwt4f"] Jan 27 16:49:57 crc kubenswrapper[4772]: I0127 16:49:57.358088 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4rmz\" (UniqueName: \"kubernetes.io/projected/f667e2c1-1507-4b30-9483-15848686b8b6-kube-api-access-x4rmz\") pod \"community-operators-gwt4f\" (UID: \"f667e2c1-1507-4b30-9483-15848686b8b6\") " pod="openshift-marketplace/community-operators-gwt4f" Jan 27 16:49:57 crc kubenswrapper[4772]: I0127 16:49:57.358374 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f667e2c1-1507-4b30-9483-15848686b8b6-catalog-content\") pod \"community-operators-gwt4f\" (UID: \"f667e2c1-1507-4b30-9483-15848686b8b6\") " pod="openshift-marketplace/community-operators-gwt4f" Jan 27 16:49:57 crc kubenswrapper[4772]: I0127 16:49:57.358614 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f667e2c1-1507-4b30-9483-15848686b8b6-utilities\") pod \"community-operators-gwt4f\" (UID: \"f667e2c1-1507-4b30-9483-15848686b8b6\") " pod="openshift-marketplace/community-operators-gwt4f" Jan 27 16:49:57 crc kubenswrapper[4772]: I0127 16:49:57.460634 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4rmz\" (UniqueName: \"kubernetes.io/projected/f667e2c1-1507-4b30-9483-15848686b8b6-kube-api-access-x4rmz\") pod \"community-operators-gwt4f\" (UID: \"f667e2c1-1507-4b30-9483-15848686b8b6\") " pod="openshift-marketplace/community-operators-gwt4f" Jan 27 16:49:57 crc kubenswrapper[4772]: I0127 16:49:57.460760 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f667e2c1-1507-4b30-9483-15848686b8b6-catalog-content\") pod \"community-operators-gwt4f\" (UID: \"f667e2c1-1507-4b30-9483-15848686b8b6\") " pod="openshift-marketplace/community-operators-gwt4f" Jan 27 16:49:57 crc kubenswrapper[4772]: I0127 16:49:57.460796 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f667e2c1-1507-4b30-9483-15848686b8b6-utilities\") pod \"community-operators-gwt4f\" (UID: \"f667e2c1-1507-4b30-9483-15848686b8b6\") " pod="openshift-marketplace/community-operators-gwt4f" Jan 27 16:49:57 crc kubenswrapper[4772]: I0127 16:49:57.461263 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f667e2c1-1507-4b30-9483-15848686b8b6-utilities\") pod \"community-operators-gwt4f\" (UID: \"f667e2c1-1507-4b30-9483-15848686b8b6\") " pod="openshift-marketplace/community-operators-gwt4f" Jan 27 16:49:57 crc kubenswrapper[4772]: I0127 16:49:57.461355 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f667e2c1-1507-4b30-9483-15848686b8b6-catalog-content\") pod \"community-operators-gwt4f\" (UID: \"f667e2c1-1507-4b30-9483-15848686b8b6\") " pod="openshift-marketplace/community-operators-gwt4f" Jan 27 16:49:57 crc kubenswrapper[4772]: I0127 16:49:57.482732 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4rmz\" (UniqueName: \"kubernetes.io/projected/f667e2c1-1507-4b30-9483-15848686b8b6-kube-api-access-x4rmz\") pod \"community-operators-gwt4f\" (UID: \"f667e2c1-1507-4b30-9483-15848686b8b6\") " pod="openshift-marketplace/community-operators-gwt4f" Jan 27 16:49:57 crc kubenswrapper[4772]: I0127 16:49:57.560159 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gwt4f" Jan 27 16:49:58 crc kubenswrapper[4772]: I0127 16:49:58.030308 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9dk47"] Jan 27 16:49:58 crc kubenswrapper[4772]: I0127 16:49:58.040349 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9dk47"] Jan 27 16:49:58 crc kubenswrapper[4772]: I0127 16:49:58.125127 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gwt4f"] Jan 27 16:49:58 crc kubenswrapper[4772]: W0127 16:49:58.134811 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf667e2c1_1507_4b30_9483_15848686b8b6.slice/crio-3b5d228719dda3b1a910a73f9e85ee82e7079ab00ea680955c59d6f4b081932e WatchSource:0}: Error finding container 3b5d228719dda3b1a910a73f9e85ee82e7079ab00ea680955c59d6f4b081932e: Status 404 returned error can't find the container with id 3b5d228719dda3b1a910a73f9e85ee82e7079ab00ea680955c59d6f4b081932e Jan 27 16:49:58 crc kubenswrapper[4772]: I0127 16:49:58.674251 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c73df4be-e448-4930-ae5e-d74fde1b4b6d" path="/var/lib/kubelet/pods/c73df4be-e448-4930-ae5e-d74fde1b4b6d/volumes" Jan 27 16:49:58 crc kubenswrapper[4772]: I0127 16:49:58.886392 4772 generic.go:334] "Generic (PLEG): container finished" podID="f667e2c1-1507-4b30-9483-15848686b8b6" containerID="888b7a4fd3e63152f4d8fa6e54cda7273f183ea1cf263815331a13ae71b4412d" exitCode=0 Jan 27 16:49:58 crc kubenswrapper[4772]: I0127 16:49:58.886447 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gwt4f" event={"ID":"f667e2c1-1507-4b30-9483-15848686b8b6","Type":"ContainerDied","Data":"888b7a4fd3e63152f4d8fa6e54cda7273f183ea1cf263815331a13ae71b4412d"} Jan 27 16:49:58 crc kubenswrapper[4772]: I0127 16:49:58.886505 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gwt4f" event={"ID":"f667e2c1-1507-4b30-9483-15848686b8b6","Type":"ContainerStarted","Data":"3b5d228719dda3b1a910a73f9e85ee82e7079ab00ea680955c59d6f4b081932e"} Jan 27 16:49:58 crc kubenswrapper[4772]: I0127 16:49:58.888628 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 16:49:59 crc kubenswrapper[4772]: I0127 16:49:59.044891 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-9vmft"] Jan 27 16:49:59 crc kubenswrapper[4772]: I0127 16:49:59.052066 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-9vmft"] Jan 27 16:49:59 crc kubenswrapper[4772]: I0127 16:49:59.898040 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gwt4f" event={"ID":"f667e2c1-1507-4b30-9483-15848686b8b6","Type":"ContainerStarted","Data":"43917f23ef47a782c97321d1bc6408d80aea5b744655f841e67b5391ae8c40f0"} Jan 27 16:50:00 crc kubenswrapper[4772]: I0127 16:50:00.683893 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4c79395-a929-4f0d-8aa7-05f24412baed" path="/var/lib/kubelet/pods/f4c79395-a929-4f0d-8aa7-05f24412baed/volumes" Jan 27 16:50:00 crc kubenswrapper[4772]: I0127 16:50:00.914465 4772 generic.go:334] "Generic (PLEG): container finished" podID="f667e2c1-1507-4b30-9483-15848686b8b6" containerID="43917f23ef47a782c97321d1bc6408d80aea5b744655f841e67b5391ae8c40f0" exitCode=0 Jan 27 16:50:00 crc kubenswrapper[4772]: I0127 16:50:00.914554 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gwt4f" event={"ID":"f667e2c1-1507-4b30-9483-15848686b8b6","Type":"ContainerDied","Data":"43917f23ef47a782c97321d1bc6408d80aea5b744655f841e67b5391ae8c40f0"} Jan 27 16:50:01 crc kubenswrapper[4772]: I0127 16:50:01.942144 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gwt4f" event={"ID":"f667e2c1-1507-4b30-9483-15848686b8b6","Type":"ContainerStarted","Data":"9b2105b198f7a6cee59ace3055430b991acc811a7d60365d1268136259fcd850"} Jan 27 16:50:01 crc kubenswrapper[4772]: I0127 16:50:01.968209 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gwt4f" podStartSLOduration=2.546167335 podStartE2EDuration="4.96818838s" podCreationTimestamp="2026-01-27 16:49:57 +0000 UTC" firstStartedPulling="2026-01-27 16:49:58.888406446 +0000 UTC m=+6184.869015544" lastFinishedPulling="2026-01-27 16:50:01.310427491 +0000 UTC m=+6187.291036589" observedRunningTime="2026-01-27 16:50:01.959640306 +0000 UTC m=+6187.940249414" watchObservedRunningTime="2026-01-27 16:50:01.96818838 +0000 UTC m=+6187.948797498" Jan 27 16:50:07 crc kubenswrapper[4772]: I0127 16:50:07.561285 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gwt4f" Jan 27 16:50:07 crc kubenswrapper[4772]: I0127 16:50:07.562255 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gwt4f" Jan 27 16:50:07 crc kubenswrapper[4772]: I0127 16:50:07.603216 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gwt4f" Jan 27 16:50:08 crc kubenswrapper[4772]: I0127 16:50:08.042679 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gwt4f" Jan 27 16:50:08 crc kubenswrapper[4772]: I0127 16:50:08.768510 4772 scope.go:117] "RemoveContainer" containerID="f05adee94e87980f08389d2716dd0d5a92a148aff23a1b26c057f06fd19c6f9b" Jan 27 16:50:08 crc kubenswrapper[4772]: I0127 16:50:08.799560 4772 scope.go:117] "RemoveContainer" containerID="b2dc1b1a95dcd405b12ae74c124c7458f4aa13c48525ad20496373458b83b670" Jan 27 16:50:08 crc kubenswrapper[4772]: I0127 16:50:08.855508 4772 scope.go:117] "RemoveContainer" containerID="4799caa746e9ac611c5360c3f4fb6d88e8100346e94f7137cac433d747b98815" Jan 27 16:50:08 crc kubenswrapper[4772]: I0127 16:50:08.929525 4772 scope.go:117] "RemoveContainer" containerID="63117e003669f061743dd5211454454c043cc51173da6608feb5766b651070d3" Jan 27 16:50:08 crc kubenswrapper[4772]: I0127 16:50:08.976691 4772 scope.go:117] "RemoveContainer" containerID="cba54a4b30bd539d4b80fc206c2394f0309c6182e722300037332e6e39b8dedb" Jan 27 16:50:09 crc kubenswrapper[4772]: I0127 16:50:09.016743 4772 scope.go:117] "RemoveContainer" containerID="0f3d5e3b05300094485e382bc00ab51b2c741fade4c1a775474788ecee8633d6" Jan 27 16:50:09 crc kubenswrapper[4772]: I0127 16:50:09.073475 4772 scope.go:117] "RemoveContainer" containerID="83a6d65ba439c93f6ff663d0a68308cf85e00e3f86836413a3077a6bf72f351a" Jan 27 16:50:09 crc kubenswrapper[4772]: I0127 16:50:09.095324 4772 scope.go:117] "RemoveContainer" containerID="cd101790786da89124bf4fdc7b4bace7b135a40af114a9eeb7dbc7a4372fd732" Jan 27 16:50:09 crc kubenswrapper[4772]: I0127 16:50:09.135768 4772 scope.go:117] "RemoveContainer" containerID="92b70acb37a2142424102ba84a18ba6908a56b707c2a540be2311cd899ea872a" Jan 27 16:50:10 crc kubenswrapper[4772]: I0127 16:50:10.026511 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gwt4f"] Jan 27 16:50:11 crc kubenswrapper[4772]: I0127 16:50:11.036605 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gwt4f" podUID="f667e2c1-1507-4b30-9483-15848686b8b6" containerName="registry-server" containerID="cri-o://9b2105b198f7a6cee59ace3055430b991acc811a7d60365d1268136259fcd850" gracePeriod=2 Jan 27 16:50:11 crc kubenswrapper[4772]: I0127 16:50:11.489963 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gwt4f" Jan 27 16:50:11 crc kubenswrapper[4772]: I0127 16:50:11.566351 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f667e2c1-1507-4b30-9483-15848686b8b6-catalog-content\") pod \"f667e2c1-1507-4b30-9483-15848686b8b6\" (UID: \"f667e2c1-1507-4b30-9483-15848686b8b6\") " Jan 27 16:50:11 crc kubenswrapper[4772]: I0127 16:50:11.566406 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4rmz\" (UniqueName: \"kubernetes.io/projected/f667e2c1-1507-4b30-9483-15848686b8b6-kube-api-access-x4rmz\") pod \"f667e2c1-1507-4b30-9483-15848686b8b6\" (UID: \"f667e2c1-1507-4b30-9483-15848686b8b6\") " Jan 27 16:50:11 crc kubenswrapper[4772]: I0127 16:50:11.566613 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f667e2c1-1507-4b30-9483-15848686b8b6-utilities\") pod \"f667e2c1-1507-4b30-9483-15848686b8b6\" (UID: \"f667e2c1-1507-4b30-9483-15848686b8b6\") " Jan 27 16:50:11 crc kubenswrapper[4772]: I0127 16:50:11.567594 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f667e2c1-1507-4b30-9483-15848686b8b6-utilities" (OuterVolumeSpecName: "utilities") pod "f667e2c1-1507-4b30-9483-15848686b8b6" (UID: "f667e2c1-1507-4b30-9483-15848686b8b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:50:11 crc kubenswrapper[4772]: I0127 16:50:11.572876 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f667e2c1-1507-4b30-9483-15848686b8b6-kube-api-access-x4rmz" (OuterVolumeSpecName: "kube-api-access-x4rmz") pod "f667e2c1-1507-4b30-9483-15848686b8b6" (UID: "f667e2c1-1507-4b30-9483-15848686b8b6"). InnerVolumeSpecName "kube-api-access-x4rmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:50:11 crc kubenswrapper[4772]: I0127 16:50:11.620417 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f667e2c1-1507-4b30-9483-15848686b8b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f667e2c1-1507-4b30-9483-15848686b8b6" (UID: "f667e2c1-1507-4b30-9483-15848686b8b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:50:11 crc kubenswrapper[4772]: I0127 16:50:11.668955 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f667e2c1-1507-4b30-9483-15848686b8b6-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 16:50:11 crc kubenswrapper[4772]: I0127 16:50:11.668991 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f667e2c1-1507-4b30-9483-15848686b8b6-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 16:50:11 crc kubenswrapper[4772]: I0127 16:50:11.669001 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4rmz\" (UniqueName: \"kubernetes.io/projected/f667e2c1-1507-4b30-9483-15848686b8b6-kube-api-access-x4rmz\") on node \"crc\" DevicePath \"\"" Jan 27 16:50:12 crc kubenswrapper[4772]: I0127 16:50:12.050422 4772 generic.go:334] "Generic (PLEG): container finished" podID="f667e2c1-1507-4b30-9483-15848686b8b6" containerID="9b2105b198f7a6cee59ace3055430b991acc811a7d60365d1268136259fcd850" exitCode=0 Jan 27 16:50:12 crc kubenswrapper[4772]: I0127 16:50:12.050478 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gwt4f" event={"ID":"f667e2c1-1507-4b30-9483-15848686b8b6","Type":"ContainerDied","Data":"9b2105b198f7a6cee59ace3055430b991acc811a7d60365d1268136259fcd850"} Jan 27 16:50:12 crc kubenswrapper[4772]: I0127 16:50:12.050503 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gwt4f" Jan 27 16:50:12 crc kubenswrapper[4772]: I0127 16:50:12.050520 4772 scope.go:117] "RemoveContainer" containerID="9b2105b198f7a6cee59ace3055430b991acc811a7d60365d1268136259fcd850" Jan 27 16:50:12 crc kubenswrapper[4772]: I0127 16:50:12.050509 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gwt4f" event={"ID":"f667e2c1-1507-4b30-9483-15848686b8b6","Type":"ContainerDied","Data":"3b5d228719dda3b1a910a73f9e85ee82e7079ab00ea680955c59d6f4b081932e"} Jan 27 16:50:12 crc kubenswrapper[4772]: I0127 16:50:12.058372 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:50:12 crc kubenswrapper[4772]: I0127 16:50:12.058420 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:50:12 crc kubenswrapper[4772]: I0127 16:50:12.058460 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 16:50:12 crc kubenswrapper[4772]: I0127 16:50:12.059762 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fad112d2b60be41180a280804db9237c789adfcc03cd7342f9bc2b818aa8e8b4"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 16:50:12 crc kubenswrapper[4772]: I0127 16:50:12.059827 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://fad112d2b60be41180a280804db9237c789adfcc03cd7342f9bc2b818aa8e8b4" gracePeriod=600 Jan 27 16:50:12 crc kubenswrapper[4772]: I0127 16:50:12.075677 4772 scope.go:117] "RemoveContainer" containerID="43917f23ef47a782c97321d1bc6408d80aea5b744655f841e67b5391ae8c40f0" Jan 27 16:50:12 crc kubenswrapper[4772]: I0127 16:50:12.083479 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gwt4f"] Jan 27 16:50:12 crc kubenswrapper[4772]: I0127 16:50:12.095096 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gwt4f"] Jan 27 16:50:12 crc kubenswrapper[4772]: I0127 16:50:12.106636 4772 scope.go:117] "RemoveContainer" containerID="888b7a4fd3e63152f4d8fa6e54cda7273f183ea1cf263815331a13ae71b4412d" Jan 27 16:50:12 crc kubenswrapper[4772]: I0127 16:50:12.139538 4772 scope.go:117] "RemoveContainer" containerID="9b2105b198f7a6cee59ace3055430b991acc811a7d60365d1268136259fcd850" Jan 27 16:50:12 crc kubenswrapper[4772]: E0127 16:50:12.139931 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b2105b198f7a6cee59ace3055430b991acc811a7d60365d1268136259fcd850\": container with ID starting with 9b2105b198f7a6cee59ace3055430b991acc811a7d60365d1268136259fcd850 not found: ID does not exist" containerID="9b2105b198f7a6cee59ace3055430b991acc811a7d60365d1268136259fcd850" Jan 27 16:50:12 crc kubenswrapper[4772]: I0127 16:50:12.139988 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b2105b198f7a6cee59ace3055430b991acc811a7d60365d1268136259fcd850"} err="failed to get container status \"9b2105b198f7a6cee59ace3055430b991acc811a7d60365d1268136259fcd850\": rpc error: code = NotFound desc = could not find container \"9b2105b198f7a6cee59ace3055430b991acc811a7d60365d1268136259fcd850\": container with ID starting with 9b2105b198f7a6cee59ace3055430b991acc811a7d60365d1268136259fcd850 not found: ID does not exist" Jan 27 16:50:12 crc kubenswrapper[4772]: I0127 16:50:12.140019 4772 scope.go:117] "RemoveContainer" containerID="43917f23ef47a782c97321d1bc6408d80aea5b744655f841e67b5391ae8c40f0" Jan 27 16:50:12 crc kubenswrapper[4772]: E0127 16:50:12.140370 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43917f23ef47a782c97321d1bc6408d80aea5b744655f841e67b5391ae8c40f0\": container with ID starting with 43917f23ef47a782c97321d1bc6408d80aea5b744655f841e67b5391ae8c40f0 not found: ID does not exist" containerID="43917f23ef47a782c97321d1bc6408d80aea5b744655f841e67b5391ae8c40f0" Jan 27 16:50:12 crc kubenswrapper[4772]: I0127 16:50:12.140403 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43917f23ef47a782c97321d1bc6408d80aea5b744655f841e67b5391ae8c40f0"} err="failed to get container status \"43917f23ef47a782c97321d1bc6408d80aea5b744655f841e67b5391ae8c40f0\": rpc error: code = NotFound desc = could not find container \"43917f23ef47a782c97321d1bc6408d80aea5b744655f841e67b5391ae8c40f0\": container with ID starting with 43917f23ef47a782c97321d1bc6408d80aea5b744655f841e67b5391ae8c40f0 not found: ID does not exist" Jan 27 16:50:12 crc kubenswrapper[4772]: I0127 16:50:12.140425 4772 scope.go:117] "RemoveContainer" containerID="888b7a4fd3e63152f4d8fa6e54cda7273f183ea1cf263815331a13ae71b4412d" Jan 27 16:50:12 crc kubenswrapper[4772]: E0127 16:50:12.140684 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"888b7a4fd3e63152f4d8fa6e54cda7273f183ea1cf263815331a13ae71b4412d\": container with ID starting with 888b7a4fd3e63152f4d8fa6e54cda7273f183ea1cf263815331a13ae71b4412d not found: ID does not exist" containerID="888b7a4fd3e63152f4d8fa6e54cda7273f183ea1cf263815331a13ae71b4412d" Jan 27 16:50:12 crc kubenswrapper[4772]: I0127 16:50:12.140712 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"888b7a4fd3e63152f4d8fa6e54cda7273f183ea1cf263815331a13ae71b4412d"} err="failed to get container status \"888b7a4fd3e63152f4d8fa6e54cda7273f183ea1cf263815331a13ae71b4412d\": rpc error: code = NotFound desc = could not find container \"888b7a4fd3e63152f4d8fa6e54cda7273f183ea1cf263815331a13ae71b4412d\": container with ID starting with 888b7a4fd3e63152f4d8fa6e54cda7273f183ea1cf263815331a13ae71b4412d not found: ID does not exist" Jan 27 16:50:12 crc kubenswrapper[4772]: I0127 16:50:12.673681 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f667e2c1-1507-4b30-9483-15848686b8b6" path="/var/lib/kubelet/pods/f667e2c1-1507-4b30-9483-15848686b8b6/volumes" Jan 27 16:50:13 crc kubenswrapper[4772]: I0127 16:50:13.093713 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="fad112d2b60be41180a280804db9237c789adfcc03cd7342f9bc2b818aa8e8b4" exitCode=0 Jan 27 16:50:13 crc kubenswrapper[4772]: I0127 16:50:13.093772 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"fad112d2b60be41180a280804db9237c789adfcc03cd7342f9bc2b818aa8e8b4"} Jan 27 16:50:13 crc kubenswrapper[4772]: I0127 16:50:13.094090 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317"} Jan 27 16:50:13 crc kubenswrapper[4772]: I0127 16:50:13.094119 4772 scope.go:117] "RemoveContainer" containerID="96a40f4ae71e3b6b4ac45c7d87f99fc2edaa1544245388fae41f53b32b3f5a69" Jan 27 16:50:17 crc kubenswrapper[4772]: I0127 16:50:17.072662 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-p6mbs"] Jan 27 16:50:17 crc kubenswrapper[4772]: I0127 16:50:17.085854 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-p6mbs"] Jan 27 16:50:18 crc kubenswrapper[4772]: I0127 16:50:18.677023 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaefb4fd-175a-4431-bd2f-7fc3205684b9" path="/var/lib/kubelet/pods/eaefb4fd-175a-4431-bd2f-7fc3205684b9/volumes" Jan 27 16:51:01 crc kubenswrapper[4772]: I0127 16:51:01.039301 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-hk6d6"] Jan 27 16:51:01 crc kubenswrapper[4772]: I0127 16:51:01.048196 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-e17c-account-create-update-4ccmc"] Jan 27 16:51:01 crc kubenswrapper[4772]: I0127 16:51:01.057989 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-e17c-account-create-update-4ccmc"] Jan 27 16:51:01 crc kubenswrapper[4772]: I0127 16:51:01.066443 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-hk6d6"] Jan 27 16:51:02 crc kubenswrapper[4772]: I0127 16:51:02.675033 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="357c50eb-1246-4dd8-975c-b10d09439cbd" path="/var/lib/kubelet/pods/357c50eb-1246-4dd8-975c-b10d09439cbd/volumes" Jan 27 16:51:02 crc kubenswrapper[4772]: I0127 16:51:02.675791 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d67b98cc-0659-4ebd-a96a-025044731558" path="/var/lib/kubelet/pods/d67b98cc-0659-4ebd-a96a-025044731558/volumes" Jan 27 16:51:09 crc kubenswrapper[4772]: I0127 16:51:09.289081 4772 scope.go:117] "RemoveContainer" containerID="2746702aa9709faf5d661e469190b51cee57d01ba7b40596d6751ad98441dd4d" Jan 27 16:51:09 crc kubenswrapper[4772]: I0127 16:51:09.323971 4772 scope.go:117] "RemoveContainer" containerID="f57fe908a2ad59651426657b881a5c35f1371cdc115dc93b86dbcb58952ce8a9" Jan 27 16:51:09 crc kubenswrapper[4772]: I0127 16:51:09.419885 4772 scope.go:117] "RemoveContainer" containerID="2c24a009396935ad397dec4435d87a19075c63011ad8f23a534d35cc814f6ddc" Jan 27 16:51:09 crc kubenswrapper[4772]: I0127 16:51:09.454656 4772 scope.go:117] "RemoveContainer" containerID="dab4307441a9ab5d1178a5547c03edae0ead3f783db4ef2bf29e8414026bb08f" Jan 27 16:51:12 crc kubenswrapper[4772]: I0127 16:51:12.056478 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-66rdf"] Jan 27 16:51:12 crc kubenswrapper[4772]: I0127 16:51:12.068295 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-66rdf"] Jan 27 16:51:12 crc kubenswrapper[4772]: I0127 16:51:12.680005 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9bc1548-ca21-4230-a5db-a9321ab69a37" path="/var/lib/kubelet/pods/e9bc1548-ca21-4230-a5db-a9321ab69a37/volumes" Jan 27 16:51:21 crc kubenswrapper[4772]: I0127 16:51:21.136526 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bsdgv"] Jan 27 16:51:21 crc kubenswrapper[4772]: E0127 16:51:21.137521 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f667e2c1-1507-4b30-9483-15848686b8b6" containerName="registry-server" Jan 27 16:51:21 crc kubenswrapper[4772]: I0127 16:51:21.137533 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f667e2c1-1507-4b30-9483-15848686b8b6" containerName="registry-server" Jan 27 16:51:21 crc kubenswrapper[4772]: E0127 16:51:21.137549 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f667e2c1-1507-4b30-9483-15848686b8b6" containerName="extract-utilities" Jan 27 16:51:21 crc kubenswrapper[4772]: I0127 16:51:21.137557 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f667e2c1-1507-4b30-9483-15848686b8b6" containerName="extract-utilities" Jan 27 16:51:21 crc kubenswrapper[4772]: E0127 16:51:21.137585 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f667e2c1-1507-4b30-9483-15848686b8b6" containerName="extract-content" Jan 27 16:51:21 crc kubenswrapper[4772]: I0127 16:51:21.137591 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f667e2c1-1507-4b30-9483-15848686b8b6" containerName="extract-content" Jan 27 16:51:21 crc kubenswrapper[4772]: I0127 16:51:21.137764 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="f667e2c1-1507-4b30-9483-15848686b8b6" containerName="registry-server" Jan 27 16:51:21 crc kubenswrapper[4772]: I0127 16:51:21.139025 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bsdgv" Jan 27 16:51:21 crc kubenswrapper[4772]: I0127 16:51:21.162659 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bsdgv"] Jan 27 16:51:21 crc kubenswrapper[4772]: I0127 16:51:21.218977 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e094ebf0-a014-4071-974b-f8cf8bf8a1d3-catalog-content\") pod \"redhat-marketplace-bsdgv\" (UID: \"e094ebf0-a014-4071-974b-f8cf8bf8a1d3\") " pod="openshift-marketplace/redhat-marketplace-bsdgv" Jan 27 16:51:21 crc kubenswrapper[4772]: I0127 16:51:21.219122 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e094ebf0-a014-4071-974b-f8cf8bf8a1d3-utilities\") pod \"redhat-marketplace-bsdgv\" (UID: \"e094ebf0-a014-4071-974b-f8cf8bf8a1d3\") " pod="openshift-marketplace/redhat-marketplace-bsdgv" Jan 27 16:51:21 crc kubenswrapper[4772]: I0127 16:51:21.219280 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7xck\" (UniqueName: \"kubernetes.io/projected/e094ebf0-a014-4071-974b-f8cf8bf8a1d3-kube-api-access-h7xck\") pod \"redhat-marketplace-bsdgv\" (UID: \"e094ebf0-a014-4071-974b-f8cf8bf8a1d3\") " pod="openshift-marketplace/redhat-marketplace-bsdgv" Jan 27 16:51:21 crc kubenswrapper[4772]: I0127 16:51:21.321716 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7xck\" (UniqueName: \"kubernetes.io/projected/e094ebf0-a014-4071-974b-f8cf8bf8a1d3-kube-api-access-h7xck\") pod \"redhat-marketplace-bsdgv\" (UID: \"e094ebf0-a014-4071-974b-f8cf8bf8a1d3\") " pod="openshift-marketplace/redhat-marketplace-bsdgv" Jan 27 16:51:21 crc kubenswrapper[4772]: I0127 16:51:21.322251 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e094ebf0-a014-4071-974b-f8cf8bf8a1d3-catalog-content\") pod \"redhat-marketplace-bsdgv\" (UID: \"e094ebf0-a014-4071-974b-f8cf8bf8a1d3\") " pod="openshift-marketplace/redhat-marketplace-bsdgv" Jan 27 16:51:21 crc kubenswrapper[4772]: I0127 16:51:21.322382 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e094ebf0-a014-4071-974b-f8cf8bf8a1d3-utilities\") pod \"redhat-marketplace-bsdgv\" (UID: \"e094ebf0-a014-4071-974b-f8cf8bf8a1d3\") " pod="openshift-marketplace/redhat-marketplace-bsdgv" Jan 27 16:51:21 crc kubenswrapper[4772]: I0127 16:51:21.322963 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e094ebf0-a014-4071-974b-f8cf8bf8a1d3-catalog-content\") pod \"redhat-marketplace-bsdgv\" (UID: \"e094ebf0-a014-4071-974b-f8cf8bf8a1d3\") " pod="openshift-marketplace/redhat-marketplace-bsdgv" Jan 27 16:51:21 crc kubenswrapper[4772]: I0127 16:51:21.323110 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e094ebf0-a014-4071-974b-f8cf8bf8a1d3-utilities\") pod \"redhat-marketplace-bsdgv\" (UID: \"e094ebf0-a014-4071-974b-f8cf8bf8a1d3\") " pod="openshift-marketplace/redhat-marketplace-bsdgv" Jan 27 16:51:21 crc kubenswrapper[4772]: I0127 16:51:21.343907 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7xck\" (UniqueName: \"kubernetes.io/projected/e094ebf0-a014-4071-974b-f8cf8bf8a1d3-kube-api-access-h7xck\") pod \"redhat-marketplace-bsdgv\" (UID: \"e094ebf0-a014-4071-974b-f8cf8bf8a1d3\") " pod="openshift-marketplace/redhat-marketplace-bsdgv" Jan 27 16:51:21 crc kubenswrapper[4772]: I0127 16:51:21.463741 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bsdgv" Jan 27 16:51:21 crc kubenswrapper[4772]: I0127 16:51:21.928879 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bsdgv"] Jan 27 16:51:21 crc kubenswrapper[4772]: W0127 16:51:21.930293 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode094ebf0_a014_4071_974b_f8cf8bf8a1d3.slice/crio-457497aa77a154894b046aead34072922305c6805acc56f00078d0211b9d58de WatchSource:0}: Error finding container 457497aa77a154894b046aead34072922305c6805acc56f00078d0211b9d58de: Status 404 returned error can't find the container with id 457497aa77a154894b046aead34072922305c6805acc56f00078d0211b9d58de Jan 27 16:51:22 crc kubenswrapper[4772]: I0127 16:51:22.795758 4772 generic.go:334] "Generic (PLEG): container finished" podID="e094ebf0-a014-4071-974b-f8cf8bf8a1d3" containerID="d780cae7ced2ecb0446c90d88daae5a5876a08e50de72a90ffd7cbde562106a0" exitCode=0 Jan 27 16:51:22 crc kubenswrapper[4772]: I0127 16:51:22.795809 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bsdgv" event={"ID":"e094ebf0-a014-4071-974b-f8cf8bf8a1d3","Type":"ContainerDied","Data":"d780cae7ced2ecb0446c90d88daae5a5876a08e50de72a90ffd7cbde562106a0"} Jan 27 16:51:22 crc kubenswrapper[4772]: I0127 16:51:22.796285 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bsdgv" event={"ID":"e094ebf0-a014-4071-974b-f8cf8bf8a1d3","Type":"ContainerStarted","Data":"457497aa77a154894b046aead34072922305c6805acc56f00078d0211b9d58de"} Jan 27 16:51:23 crc kubenswrapper[4772]: I0127 16:51:23.806641 4772 generic.go:334] "Generic (PLEG): container finished" podID="e094ebf0-a014-4071-974b-f8cf8bf8a1d3" containerID="96d3398aa45e14a1610aa742949aabdd7940c244b6689f7f5e0aecbb8e1a59ef" exitCode=0 Jan 27 16:51:23 crc kubenswrapper[4772]: I0127 16:51:23.806731 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bsdgv" event={"ID":"e094ebf0-a014-4071-974b-f8cf8bf8a1d3","Type":"ContainerDied","Data":"96d3398aa45e14a1610aa742949aabdd7940c244b6689f7f5e0aecbb8e1a59ef"} Jan 27 16:51:24 crc kubenswrapper[4772]: I0127 16:51:24.823273 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bsdgv" event={"ID":"e094ebf0-a014-4071-974b-f8cf8bf8a1d3","Type":"ContainerStarted","Data":"924cb21c049d7e2b564fcdeda1cd8c7d8b841546eb43ece1a79a291088be7efe"} Jan 27 16:51:24 crc kubenswrapper[4772]: I0127 16:51:24.849698 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bsdgv" podStartSLOduration=2.406793976 podStartE2EDuration="3.849677539s" podCreationTimestamp="2026-01-27 16:51:21 +0000 UTC" firstStartedPulling="2026-01-27 16:51:22.79921052 +0000 UTC m=+6268.779819628" lastFinishedPulling="2026-01-27 16:51:24.242094063 +0000 UTC m=+6270.222703191" observedRunningTime="2026-01-27 16:51:24.845604853 +0000 UTC m=+6270.826213981" watchObservedRunningTime="2026-01-27 16:51:24.849677539 +0000 UTC m=+6270.830286657" Jan 27 16:51:31 crc kubenswrapper[4772]: I0127 16:51:31.464046 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bsdgv" Jan 27 16:51:31 crc kubenswrapper[4772]: I0127 16:51:31.464749 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bsdgv" Jan 27 16:51:31 crc kubenswrapper[4772]: I0127 16:51:31.511727 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bsdgv" Jan 27 16:51:31 crc kubenswrapper[4772]: I0127 16:51:31.959590 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bsdgv" Jan 27 16:51:32 crc kubenswrapper[4772]: I0127 16:51:32.007452 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bsdgv"] Jan 27 16:51:33 crc kubenswrapper[4772]: I0127 16:51:33.926701 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bsdgv" podUID="e094ebf0-a014-4071-974b-f8cf8bf8a1d3" containerName="registry-server" containerID="cri-o://924cb21c049d7e2b564fcdeda1cd8c7d8b841546eb43ece1a79a291088be7efe" gracePeriod=2 Jan 27 16:51:34 crc kubenswrapper[4772]: I0127 16:51:34.879561 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bsdgv" Jan 27 16:51:34 crc kubenswrapper[4772]: I0127 16:51:34.900432 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e094ebf0-a014-4071-974b-f8cf8bf8a1d3-catalog-content\") pod \"e094ebf0-a014-4071-974b-f8cf8bf8a1d3\" (UID: \"e094ebf0-a014-4071-974b-f8cf8bf8a1d3\") " Jan 27 16:51:34 crc kubenswrapper[4772]: I0127 16:51:34.900783 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7xck\" (UniqueName: \"kubernetes.io/projected/e094ebf0-a014-4071-974b-f8cf8bf8a1d3-kube-api-access-h7xck\") pod \"e094ebf0-a014-4071-974b-f8cf8bf8a1d3\" (UID: \"e094ebf0-a014-4071-974b-f8cf8bf8a1d3\") " Jan 27 16:51:34 crc kubenswrapper[4772]: I0127 16:51:34.902024 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e094ebf0-a014-4071-974b-f8cf8bf8a1d3-utilities\") pod \"e094ebf0-a014-4071-974b-f8cf8bf8a1d3\" (UID: \"e094ebf0-a014-4071-974b-f8cf8bf8a1d3\") " Jan 27 16:51:34 crc kubenswrapper[4772]: I0127 16:51:34.902676 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e094ebf0-a014-4071-974b-f8cf8bf8a1d3-utilities" (OuterVolumeSpecName: "utilities") pod "e094ebf0-a014-4071-974b-f8cf8bf8a1d3" (UID: "e094ebf0-a014-4071-974b-f8cf8bf8a1d3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:51:34 crc kubenswrapper[4772]: I0127 16:51:34.907744 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e094ebf0-a014-4071-974b-f8cf8bf8a1d3-kube-api-access-h7xck" (OuterVolumeSpecName: "kube-api-access-h7xck") pod "e094ebf0-a014-4071-974b-f8cf8bf8a1d3" (UID: "e094ebf0-a014-4071-974b-f8cf8bf8a1d3"). InnerVolumeSpecName "kube-api-access-h7xck". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:51:34 crc kubenswrapper[4772]: I0127 16:51:34.927655 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e094ebf0-a014-4071-974b-f8cf8bf8a1d3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e094ebf0-a014-4071-974b-f8cf8bf8a1d3" (UID: "e094ebf0-a014-4071-974b-f8cf8bf8a1d3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:51:34 crc kubenswrapper[4772]: I0127 16:51:34.939045 4772 generic.go:334] "Generic (PLEG): container finished" podID="e094ebf0-a014-4071-974b-f8cf8bf8a1d3" containerID="924cb21c049d7e2b564fcdeda1cd8c7d8b841546eb43ece1a79a291088be7efe" exitCode=0 Jan 27 16:51:34 crc kubenswrapper[4772]: I0127 16:51:34.939083 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bsdgv" event={"ID":"e094ebf0-a014-4071-974b-f8cf8bf8a1d3","Type":"ContainerDied","Data":"924cb21c049d7e2b564fcdeda1cd8c7d8b841546eb43ece1a79a291088be7efe"} Jan 27 16:51:34 crc kubenswrapper[4772]: I0127 16:51:34.939111 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bsdgv" Jan 27 16:51:34 crc kubenswrapper[4772]: I0127 16:51:34.939125 4772 scope.go:117] "RemoveContainer" containerID="924cb21c049d7e2b564fcdeda1cd8c7d8b841546eb43ece1a79a291088be7efe" Jan 27 16:51:34 crc kubenswrapper[4772]: I0127 16:51:34.939115 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bsdgv" event={"ID":"e094ebf0-a014-4071-974b-f8cf8bf8a1d3","Type":"ContainerDied","Data":"457497aa77a154894b046aead34072922305c6805acc56f00078d0211b9d58de"} Jan 27 16:51:34 crc kubenswrapper[4772]: I0127 16:51:34.975179 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bsdgv"] Jan 27 16:51:34 crc kubenswrapper[4772]: I0127 16:51:34.975673 4772 scope.go:117] "RemoveContainer" containerID="96d3398aa45e14a1610aa742949aabdd7940c244b6689f7f5e0aecbb8e1a59ef" Jan 27 16:51:34 crc kubenswrapper[4772]: I0127 16:51:34.986325 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bsdgv"] Jan 27 16:51:35 crc kubenswrapper[4772]: I0127 16:51:35.000597 4772 scope.go:117] "RemoveContainer" containerID="d780cae7ced2ecb0446c90d88daae5a5876a08e50de72a90ffd7cbde562106a0" Jan 27 16:51:35 crc kubenswrapper[4772]: I0127 16:51:35.004419 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7xck\" (UniqueName: \"kubernetes.io/projected/e094ebf0-a014-4071-974b-f8cf8bf8a1d3-kube-api-access-h7xck\") on node \"crc\" DevicePath \"\"" Jan 27 16:51:35 crc kubenswrapper[4772]: I0127 16:51:35.004443 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e094ebf0-a014-4071-974b-f8cf8bf8a1d3-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 16:51:35 crc kubenswrapper[4772]: I0127 16:51:35.004457 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e094ebf0-a014-4071-974b-f8cf8bf8a1d3-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 16:51:35 crc kubenswrapper[4772]: I0127 16:51:35.045021 4772 scope.go:117] "RemoveContainer" containerID="924cb21c049d7e2b564fcdeda1cd8c7d8b841546eb43ece1a79a291088be7efe" Jan 27 16:51:35 crc kubenswrapper[4772]: E0127 16:51:35.045521 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"924cb21c049d7e2b564fcdeda1cd8c7d8b841546eb43ece1a79a291088be7efe\": container with ID starting with 924cb21c049d7e2b564fcdeda1cd8c7d8b841546eb43ece1a79a291088be7efe not found: ID does not exist" containerID="924cb21c049d7e2b564fcdeda1cd8c7d8b841546eb43ece1a79a291088be7efe" Jan 27 16:51:35 crc kubenswrapper[4772]: I0127 16:51:35.045557 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"924cb21c049d7e2b564fcdeda1cd8c7d8b841546eb43ece1a79a291088be7efe"} err="failed to get container status \"924cb21c049d7e2b564fcdeda1cd8c7d8b841546eb43ece1a79a291088be7efe\": rpc error: code = NotFound desc = could not find container \"924cb21c049d7e2b564fcdeda1cd8c7d8b841546eb43ece1a79a291088be7efe\": container with ID starting with 924cb21c049d7e2b564fcdeda1cd8c7d8b841546eb43ece1a79a291088be7efe not found: ID does not exist" Jan 27 16:51:35 crc kubenswrapper[4772]: I0127 16:51:35.045581 4772 scope.go:117] "RemoveContainer" containerID="96d3398aa45e14a1610aa742949aabdd7940c244b6689f7f5e0aecbb8e1a59ef" Jan 27 16:51:35 crc kubenswrapper[4772]: E0127 16:51:35.046010 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96d3398aa45e14a1610aa742949aabdd7940c244b6689f7f5e0aecbb8e1a59ef\": container with ID starting with 96d3398aa45e14a1610aa742949aabdd7940c244b6689f7f5e0aecbb8e1a59ef not found: ID does not exist" containerID="96d3398aa45e14a1610aa742949aabdd7940c244b6689f7f5e0aecbb8e1a59ef" Jan 27 16:51:35 crc kubenswrapper[4772]: I0127 16:51:35.046046 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96d3398aa45e14a1610aa742949aabdd7940c244b6689f7f5e0aecbb8e1a59ef"} err="failed to get container status \"96d3398aa45e14a1610aa742949aabdd7940c244b6689f7f5e0aecbb8e1a59ef\": rpc error: code = NotFound desc = could not find container \"96d3398aa45e14a1610aa742949aabdd7940c244b6689f7f5e0aecbb8e1a59ef\": container with ID starting with 96d3398aa45e14a1610aa742949aabdd7940c244b6689f7f5e0aecbb8e1a59ef not found: ID does not exist" Jan 27 16:51:35 crc kubenswrapper[4772]: I0127 16:51:35.046087 4772 scope.go:117] "RemoveContainer" containerID="d780cae7ced2ecb0446c90d88daae5a5876a08e50de72a90ffd7cbde562106a0" Jan 27 16:51:35 crc kubenswrapper[4772]: E0127 16:51:35.046524 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d780cae7ced2ecb0446c90d88daae5a5876a08e50de72a90ffd7cbde562106a0\": container with ID starting with d780cae7ced2ecb0446c90d88daae5a5876a08e50de72a90ffd7cbde562106a0 not found: ID does not exist" containerID="d780cae7ced2ecb0446c90d88daae5a5876a08e50de72a90ffd7cbde562106a0" Jan 27 16:51:35 crc kubenswrapper[4772]: I0127 16:51:35.046546 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d780cae7ced2ecb0446c90d88daae5a5876a08e50de72a90ffd7cbde562106a0"} err="failed to get container status \"d780cae7ced2ecb0446c90d88daae5a5876a08e50de72a90ffd7cbde562106a0\": rpc error: code = NotFound desc = could not find container \"d780cae7ced2ecb0446c90d88daae5a5876a08e50de72a90ffd7cbde562106a0\": container with ID starting with d780cae7ced2ecb0446c90d88daae5a5876a08e50de72a90ffd7cbde562106a0 not found: ID does not exist" Jan 27 16:51:36 crc kubenswrapper[4772]: I0127 16:51:36.677078 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e094ebf0-a014-4071-974b-f8cf8bf8a1d3" path="/var/lib/kubelet/pods/e094ebf0-a014-4071-974b-f8cf8bf8a1d3/volumes" Jan 27 16:52:09 crc kubenswrapper[4772]: I0127 16:52:09.588158 4772 scope.go:117] "RemoveContainer" containerID="d0051faf5f33fa9d044b3023d9e4654d63902fac62af135831e6d6e9a248c7b6" Jan 27 16:52:12 crc kubenswrapper[4772]: I0127 16:52:12.058872 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:52:12 crc kubenswrapper[4772]: I0127 16:52:12.059190 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:52:42 crc kubenswrapper[4772]: I0127 16:52:42.059305 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:52:42 crc kubenswrapper[4772]: I0127 16:52:42.060127 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:53:12 crc kubenswrapper[4772]: I0127 16:53:12.058655 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 16:53:12 crc kubenswrapper[4772]: I0127 16:53:12.059566 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 16:53:12 crc kubenswrapper[4772]: I0127 16:53:12.059633 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 16:53:12 crc kubenswrapper[4772]: I0127 16:53:12.061200 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 16:53:12 crc kubenswrapper[4772]: I0127 16:53:12.061312 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" gracePeriod=600 Jan 27 16:53:12 crc kubenswrapper[4772]: E0127 16:53:12.197368 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:53:12 crc kubenswrapper[4772]: I0127 16:53:12.926849 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" exitCode=0 Jan 27 16:53:12 crc kubenswrapper[4772]: I0127 16:53:12.926910 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317"} Jan 27 16:53:12 crc kubenswrapper[4772]: I0127 16:53:12.926985 4772 scope.go:117] "RemoveContainer" containerID="fad112d2b60be41180a280804db9237c789adfcc03cd7342f9bc2b818aa8e8b4" Jan 27 16:53:12 crc kubenswrapper[4772]: I0127 16:53:12.928012 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 16:53:12 crc kubenswrapper[4772]: E0127 16:53:12.928495 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:53:19 crc kubenswrapper[4772]: I0127 16:53:19.297863 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-59lfv"] Jan 27 16:53:19 crc kubenswrapper[4772]: E0127 16:53:19.298787 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e094ebf0-a014-4071-974b-f8cf8bf8a1d3" containerName="extract-utilities" Jan 27 16:53:19 crc kubenswrapper[4772]: I0127 16:53:19.298802 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e094ebf0-a014-4071-974b-f8cf8bf8a1d3" containerName="extract-utilities" Jan 27 16:53:19 crc kubenswrapper[4772]: E0127 16:53:19.298839 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e094ebf0-a014-4071-974b-f8cf8bf8a1d3" containerName="registry-server" Jan 27 16:53:19 crc kubenswrapper[4772]: I0127 16:53:19.298847 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e094ebf0-a014-4071-974b-f8cf8bf8a1d3" containerName="registry-server" Jan 27 16:53:19 crc kubenswrapper[4772]: E0127 16:53:19.298866 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e094ebf0-a014-4071-974b-f8cf8bf8a1d3" containerName="extract-content" Jan 27 16:53:19 crc kubenswrapper[4772]: I0127 16:53:19.298874 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e094ebf0-a014-4071-974b-f8cf8bf8a1d3" containerName="extract-content" Jan 27 16:53:19 crc kubenswrapper[4772]: I0127 16:53:19.299078 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="e094ebf0-a014-4071-974b-f8cf8bf8a1d3" containerName="registry-server" Jan 27 16:53:19 crc kubenswrapper[4772]: I0127 16:53:19.300704 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-59lfv" Jan 27 16:53:19 crc kubenswrapper[4772]: I0127 16:53:19.336884 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-59lfv"] Jan 27 16:53:19 crc kubenswrapper[4772]: I0127 16:53:19.416198 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5bw2\" (UniqueName: \"kubernetes.io/projected/e447eca7-2085-4dde-84ff-0e85dccea5e9-kube-api-access-g5bw2\") pod \"certified-operators-59lfv\" (UID: \"e447eca7-2085-4dde-84ff-0e85dccea5e9\") " pod="openshift-marketplace/certified-operators-59lfv" Jan 27 16:53:19 crc kubenswrapper[4772]: I0127 16:53:19.416244 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e447eca7-2085-4dde-84ff-0e85dccea5e9-catalog-content\") pod \"certified-operators-59lfv\" (UID: \"e447eca7-2085-4dde-84ff-0e85dccea5e9\") " pod="openshift-marketplace/certified-operators-59lfv" Jan 27 16:53:19 crc kubenswrapper[4772]: I0127 16:53:19.416280 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e447eca7-2085-4dde-84ff-0e85dccea5e9-utilities\") pod \"certified-operators-59lfv\" (UID: \"e447eca7-2085-4dde-84ff-0e85dccea5e9\") " pod="openshift-marketplace/certified-operators-59lfv" Jan 27 16:53:19 crc kubenswrapper[4772]: I0127 16:53:19.517696 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e447eca7-2085-4dde-84ff-0e85dccea5e9-utilities\") pod \"certified-operators-59lfv\" (UID: \"e447eca7-2085-4dde-84ff-0e85dccea5e9\") " pod="openshift-marketplace/certified-operators-59lfv" Jan 27 16:53:19 crc kubenswrapper[4772]: I0127 16:53:19.518133 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5bw2\" (UniqueName: \"kubernetes.io/projected/e447eca7-2085-4dde-84ff-0e85dccea5e9-kube-api-access-g5bw2\") pod \"certified-operators-59lfv\" (UID: \"e447eca7-2085-4dde-84ff-0e85dccea5e9\") " pod="openshift-marketplace/certified-operators-59lfv" Jan 27 16:53:19 crc kubenswrapper[4772]: I0127 16:53:19.518273 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e447eca7-2085-4dde-84ff-0e85dccea5e9-catalog-content\") pod \"certified-operators-59lfv\" (UID: \"e447eca7-2085-4dde-84ff-0e85dccea5e9\") " pod="openshift-marketplace/certified-operators-59lfv" Jan 27 16:53:19 crc kubenswrapper[4772]: I0127 16:53:19.518179 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e447eca7-2085-4dde-84ff-0e85dccea5e9-utilities\") pod \"certified-operators-59lfv\" (UID: \"e447eca7-2085-4dde-84ff-0e85dccea5e9\") " pod="openshift-marketplace/certified-operators-59lfv" Jan 27 16:53:19 crc kubenswrapper[4772]: I0127 16:53:19.518746 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e447eca7-2085-4dde-84ff-0e85dccea5e9-catalog-content\") pod \"certified-operators-59lfv\" (UID: \"e447eca7-2085-4dde-84ff-0e85dccea5e9\") " pod="openshift-marketplace/certified-operators-59lfv" Jan 27 16:53:19 crc kubenswrapper[4772]: I0127 16:53:19.543153 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5bw2\" (UniqueName: \"kubernetes.io/projected/e447eca7-2085-4dde-84ff-0e85dccea5e9-kube-api-access-g5bw2\") pod \"certified-operators-59lfv\" (UID: \"e447eca7-2085-4dde-84ff-0e85dccea5e9\") " pod="openshift-marketplace/certified-operators-59lfv" Jan 27 16:53:19 crc kubenswrapper[4772]: I0127 16:53:19.672950 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-59lfv" Jan 27 16:53:20 crc kubenswrapper[4772]: W0127 16:53:20.228484 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode447eca7_2085_4dde_84ff_0e85dccea5e9.slice/crio-d822540001ec1fd39e430aef97cccf0e570b9021e0151dc4cfcb80bcdd01b888 WatchSource:0}: Error finding container d822540001ec1fd39e430aef97cccf0e570b9021e0151dc4cfcb80bcdd01b888: Status 404 returned error can't find the container with id d822540001ec1fd39e430aef97cccf0e570b9021e0151dc4cfcb80bcdd01b888 Jan 27 16:53:20 crc kubenswrapper[4772]: I0127 16:53:20.228545 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-59lfv"] Jan 27 16:53:21 crc kubenswrapper[4772]: I0127 16:53:21.026063 4772 generic.go:334] "Generic (PLEG): container finished" podID="e447eca7-2085-4dde-84ff-0e85dccea5e9" containerID="59cabea9546ba64b883bf93fa4862411c29e790935ca2098d56582afa6542b6d" exitCode=0 Jan 27 16:53:21 crc kubenswrapper[4772]: I0127 16:53:21.026333 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-59lfv" event={"ID":"e447eca7-2085-4dde-84ff-0e85dccea5e9","Type":"ContainerDied","Data":"59cabea9546ba64b883bf93fa4862411c29e790935ca2098d56582afa6542b6d"} Jan 27 16:53:21 crc kubenswrapper[4772]: I0127 16:53:21.026480 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-59lfv" event={"ID":"e447eca7-2085-4dde-84ff-0e85dccea5e9","Type":"ContainerStarted","Data":"d822540001ec1fd39e430aef97cccf0e570b9021e0151dc4cfcb80bcdd01b888"} Jan 27 16:53:22 crc kubenswrapper[4772]: I0127 16:53:22.039699 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-59lfv" event={"ID":"e447eca7-2085-4dde-84ff-0e85dccea5e9","Type":"ContainerStarted","Data":"4598336c5b7617228bcf6cd156c66ab183d12ee169a48a9e366dd5f0b274d0bf"} Jan 27 16:53:23 crc kubenswrapper[4772]: I0127 16:53:23.057345 4772 generic.go:334] "Generic (PLEG): container finished" podID="e447eca7-2085-4dde-84ff-0e85dccea5e9" containerID="4598336c5b7617228bcf6cd156c66ab183d12ee169a48a9e366dd5f0b274d0bf" exitCode=0 Jan 27 16:53:23 crc kubenswrapper[4772]: I0127 16:53:23.057586 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-59lfv" event={"ID":"e447eca7-2085-4dde-84ff-0e85dccea5e9","Type":"ContainerDied","Data":"4598336c5b7617228bcf6cd156c66ab183d12ee169a48a9e366dd5f0b274d0bf"} Jan 27 16:53:24 crc kubenswrapper[4772]: I0127 16:53:24.066525 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-59lfv" event={"ID":"e447eca7-2085-4dde-84ff-0e85dccea5e9","Type":"ContainerStarted","Data":"e22a8458896bb5d1120cb8870b4e1750f928e8189a6f596227508efe1f8e459f"} Jan 27 16:53:24 crc kubenswrapper[4772]: I0127 16:53:24.090422 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-59lfv" podStartSLOduration=2.6734818860000003 podStartE2EDuration="5.090402047s" podCreationTimestamp="2026-01-27 16:53:19 +0000 UTC" firstStartedPulling="2026-01-27 16:53:21.031015544 +0000 UTC m=+6387.011624672" lastFinishedPulling="2026-01-27 16:53:23.447935735 +0000 UTC m=+6389.428544833" observedRunningTime="2026-01-27 16:53:24.083733267 +0000 UTC m=+6390.064342385" watchObservedRunningTime="2026-01-27 16:53:24.090402047 +0000 UTC m=+6390.071011155" Jan 27 16:53:27 crc kubenswrapper[4772]: I0127 16:53:27.663125 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 16:53:27 crc kubenswrapper[4772]: E0127 16:53:27.664344 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:53:29 crc kubenswrapper[4772]: I0127 16:53:29.673435 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-59lfv" Jan 27 16:53:29 crc kubenswrapper[4772]: I0127 16:53:29.673752 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-59lfv" Jan 27 16:53:29 crc kubenswrapper[4772]: I0127 16:53:29.715499 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-59lfv" Jan 27 16:53:30 crc kubenswrapper[4772]: I0127 16:53:30.198362 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-59lfv" Jan 27 16:53:30 crc kubenswrapper[4772]: I0127 16:53:30.264108 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-59lfv"] Jan 27 16:53:32 crc kubenswrapper[4772]: I0127 16:53:32.153223 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-59lfv" podUID="e447eca7-2085-4dde-84ff-0e85dccea5e9" containerName="registry-server" containerID="cri-o://e22a8458896bb5d1120cb8870b4e1750f928e8189a6f596227508efe1f8e459f" gracePeriod=2 Jan 27 16:53:32 crc kubenswrapper[4772]: I0127 16:53:32.664746 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-59lfv" Jan 27 16:53:32 crc kubenswrapper[4772]: I0127 16:53:32.723521 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5bw2\" (UniqueName: \"kubernetes.io/projected/e447eca7-2085-4dde-84ff-0e85dccea5e9-kube-api-access-g5bw2\") pod \"e447eca7-2085-4dde-84ff-0e85dccea5e9\" (UID: \"e447eca7-2085-4dde-84ff-0e85dccea5e9\") " Jan 27 16:53:32 crc kubenswrapper[4772]: I0127 16:53:32.723608 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e447eca7-2085-4dde-84ff-0e85dccea5e9-catalog-content\") pod \"e447eca7-2085-4dde-84ff-0e85dccea5e9\" (UID: \"e447eca7-2085-4dde-84ff-0e85dccea5e9\") " Jan 27 16:53:32 crc kubenswrapper[4772]: I0127 16:53:32.723780 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e447eca7-2085-4dde-84ff-0e85dccea5e9-utilities\") pod \"e447eca7-2085-4dde-84ff-0e85dccea5e9\" (UID: \"e447eca7-2085-4dde-84ff-0e85dccea5e9\") " Jan 27 16:53:32 crc kubenswrapper[4772]: I0127 16:53:32.725676 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e447eca7-2085-4dde-84ff-0e85dccea5e9-utilities" (OuterVolumeSpecName: "utilities") pod "e447eca7-2085-4dde-84ff-0e85dccea5e9" (UID: "e447eca7-2085-4dde-84ff-0e85dccea5e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:53:32 crc kubenswrapper[4772]: I0127 16:53:32.730959 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e447eca7-2085-4dde-84ff-0e85dccea5e9-kube-api-access-g5bw2" (OuterVolumeSpecName: "kube-api-access-g5bw2") pod "e447eca7-2085-4dde-84ff-0e85dccea5e9" (UID: "e447eca7-2085-4dde-84ff-0e85dccea5e9"). InnerVolumeSpecName "kube-api-access-g5bw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:53:32 crc kubenswrapper[4772]: I0127 16:53:32.774321 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e447eca7-2085-4dde-84ff-0e85dccea5e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e447eca7-2085-4dde-84ff-0e85dccea5e9" (UID: "e447eca7-2085-4dde-84ff-0e85dccea5e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:53:32 crc kubenswrapper[4772]: I0127 16:53:32.826288 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e447eca7-2085-4dde-84ff-0e85dccea5e9-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 16:53:32 crc kubenswrapper[4772]: I0127 16:53:32.826317 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5bw2\" (UniqueName: \"kubernetes.io/projected/e447eca7-2085-4dde-84ff-0e85dccea5e9-kube-api-access-g5bw2\") on node \"crc\" DevicePath \"\"" Jan 27 16:53:32 crc kubenswrapper[4772]: I0127 16:53:32.826328 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e447eca7-2085-4dde-84ff-0e85dccea5e9-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 16:53:33 crc kubenswrapper[4772]: I0127 16:53:33.166445 4772 generic.go:334] "Generic (PLEG): container finished" podID="e447eca7-2085-4dde-84ff-0e85dccea5e9" containerID="e22a8458896bb5d1120cb8870b4e1750f928e8189a6f596227508efe1f8e459f" exitCode=0 Jan 27 16:53:33 crc kubenswrapper[4772]: I0127 16:53:33.166497 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-59lfv" event={"ID":"e447eca7-2085-4dde-84ff-0e85dccea5e9","Type":"ContainerDied","Data":"e22a8458896bb5d1120cb8870b4e1750f928e8189a6f596227508efe1f8e459f"} Jan 27 16:53:33 crc kubenswrapper[4772]: I0127 16:53:33.166559 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-59lfv" event={"ID":"e447eca7-2085-4dde-84ff-0e85dccea5e9","Type":"ContainerDied","Data":"d822540001ec1fd39e430aef97cccf0e570b9021e0151dc4cfcb80bcdd01b888"} Jan 27 16:53:33 crc kubenswrapper[4772]: I0127 16:53:33.166584 4772 scope.go:117] "RemoveContainer" containerID="e22a8458896bb5d1120cb8870b4e1750f928e8189a6f596227508efe1f8e459f" Jan 27 16:53:33 crc kubenswrapper[4772]: I0127 16:53:33.166583 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-59lfv" Jan 27 16:53:33 crc kubenswrapper[4772]: I0127 16:53:33.205202 4772 scope.go:117] "RemoveContainer" containerID="4598336c5b7617228bcf6cd156c66ab183d12ee169a48a9e366dd5f0b274d0bf" Jan 27 16:53:33 crc kubenswrapper[4772]: I0127 16:53:33.207088 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-59lfv"] Jan 27 16:53:33 crc kubenswrapper[4772]: I0127 16:53:33.218210 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-59lfv"] Jan 27 16:53:33 crc kubenswrapper[4772]: I0127 16:53:33.237628 4772 scope.go:117] "RemoveContainer" containerID="59cabea9546ba64b883bf93fa4862411c29e790935ca2098d56582afa6542b6d" Jan 27 16:53:33 crc kubenswrapper[4772]: I0127 16:53:33.294821 4772 scope.go:117] "RemoveContainer" containerID="e22a8458896bb5d1120cb8870b4e1750f928e8189a6f596227508efe1f8e459f" Jan 27 16:53:33 crc kubenswrapper[4772]: E0127 16:53:33.295256 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e22a8458896bb5d1120cb8870b4e1750f928e8189a6f596227508efe1f8e459f\": container with ID starting with e22a8458896bb5d1120cb8870b4e1750f928e8189a6f596227508efe1f8e459f not found: ID does not exist" containerID="e22a8458896bb5d1120cb8870b4e1750f928e8189a6f596227508efe1f8e459f" Jan 27 16:53:33 crc kubenswrapper[4772]: I0127 16:53:33.295297 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e22a8458896bb5d1120cb8870b4e1750f928e8189a6f596227508efe1f8e459f"} err="failed to get container status \"e22a8458896bb5d1120cb8870b4e1750f928e8189a6f596227508efe1f8e459f\": rpc error: code = NotFound desc = could not find container \"e22a8458896bb5d1120cb8870b4e1750f928e8189a6f596227508efe1f8e459f\": container with ID starting with e22a8458896bb5d1120cb8870b4e1750f928e8189a6f596227508efe1f8e459f not found: ID does not exist" Jan 27 16:53:33 crc kubenswrapper[4772]: I0127 16:53:33.295325 4772 scope.go:117] "RemoveContainer" containerID="4598336c5b7617228bcf6cd156c66ab183d12ee169a48a9e366dd5f0b274d0bf" Jan 27 16:53:33 crc kubenswrapper[4772]: E0127 16:53:33.295658 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4598336c5b7617228bcf6cd156c66ab183d12ee169a48a9e366dd5f0b274d0bf\": container with ID starting with 4598336c5b7617228bcf6cd156c66ab183d12ee169a48a9e366dd5f0b274d0bf not found: ID does not exist" containerID="4598336c5b7617228bcf6cd156c66ab183d12ee169a48a9e366dd5f0b274d0bf" Jan 27 16:53:33 crc kubenswrapper[4772]: I0127 16:53:33.295698 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4598336c5b7617228bcf6cd156c66ab183d12ee169a48a9e366dd5f0b274d0bf"} err="failed to get container status \"4598336c5b7617228bcf6cd156c66ab183d12ee169a48a9e366dd5f0b274d0bf\": rpc error: code = NotFound desc = could not find container \"4598336c5b7617228bcf6cd156c66ab183d12ee169a48a9e366dd5f0b274d0bf\": container with ID starting with 4598336c5b7617228bcf6cd156c66ab183d12ee169a48a9e366dd5f0b274d0bf not found: ID does not exist" Jan 27 16:53:33 crc kubenswrapper[4772]: I0127 16:53:33.295726 4772 scope.go:117] "RemoveContainer" containerID="59cabea9546ba64b883bf93fa4862411c29e790935ca2098d56582afa6542b6d" Jan 27 16:53:33 crc kubenswrapper[4772]: E0127 16:53:33.295960 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59cabea9546ba64b883bf93fa4862411c29e790935ca2098d56582afa6542b6d\": container with ID starting with 59cabea9546ba64b883bf93fa4862411c29e790935ca2098d56582afa6542b6d not found: ID does not exist" containerID="59cabea9546ba64b883bf93fa4862411c29e790935ca2098d56582afa6542b6d" Jan 27 16:53:33 crc kubenswrapper[4772]: I0127 16:53:33.295989 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59cabea9546ba64b883bf93fa4862411c29e790935ca2098d56582afa6542b6d"} err="failed to get container status \"59cabea9546ba64b883bf93fa4862411c29e790935ca2098d56582afa6542b6d\": rpc error: code = NotFound desc = could not find container \"59cabea9546ba64b883bf93fa4862411c29e790935ca2098d56582afa6542b6d\": container with ID starting with 59cabea9546ba64b883bf93fa4862411c29e790935ca2098d56582afa6542b6d not found: ID does not exist" Jan 27 16:53:34 crc kubenswrapper[4772]: I0127 16:53:34.680380 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e447eca7-2085-4dde-84ff-0e85dccea5e9" path="/var/lib/kubelet/pods/e447eca7-2085-4dde-84ff-0e85dccea5e9/volumes" Jan 27 16:53:39 crc kubenswrapper[4772]: I0127 16:53:39.663126 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 16:53:39 crc kubenswrapper[4772]: E0127 16:53:39.664044 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:53:51 crc kubenswrapper[4772]: I0127 16:53:51.663255 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 16:53:51 crc kubenswrapper[4772]: E0127 16:53:51.664017 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:53:54 crc kubenswrapper[4772]: I0127 16:53:54.036221 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-create-fw49r"] Jan 27 16:53:54 crc kubenswrapper[4772]: I0127 16:53:54.045203 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-create-fw49r"] Jan 27 16:53:54 crc kubenswrapper[4772]: I0127 16:53:54.676204 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6985c580-efad-46fe-8e20-9f932ce3af7d" path="/var/lib/kubelet/pods/6985c580-efad-46fe-8e20-9f932ce3af7d/volumes" Jan 27 16:53:56 crc kubenswrapper[4772]: I0127 16:53:56.036426 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-c028-account-create-update-x9k2v"] Jan 27 16:53:56 crc kubenswrapper[4772]: I0127 16:53:56.048955 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-c028-account-create-update-x9k2v"] Jan 27 16:53:56 crc kubenswrapper[4772]: I0127 16:53:56.681456 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cd343bc-4a97-4ad2-aa82-12ea527398d8" path="/var/lib/kubelet/pods/1cd343bc-4a97-4ad2-aa82-12ea527398d8/volumes" Jan 27 16:54:02 crc kubenswrapper[4772]: I0127 16:54:02.080153 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-persistence-db-create-4mfnz"] Jan 27 16:54:02 crc kubenswrapper[4772]: I0127 16:54:02.090649 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-persistence-db-create-4mfnz"] Jan 27 16:54:02 crc kubenswrapper[4772]: I0127 16:54:02.664437 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 16:54:02 crc kubenswrapper[4772]: E0127 16:54:02.664947 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:54:02 crc kubenswrapper[4772]: I0127 16:54:02.674745 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c203bc37-6216-4aa6-8a51-1e0a2f01bb43" path="/var/lib/kubelet/pods/c203bc37-6216-4aa6-8a51-1e0a2f01bb43/volumes" Jan 27 16:54:03 crc kubenswrapper[4772]: I0127 16:54:03.054428 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-c1c0-account-create-update-cjqxk"] Jan 27 16:54:03 crc kubenswrapper[4772]: I0127 16:54:03.064964 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-c1c0-account-create-update-cjqxk"] Jan 27 16:54:04 crc kubenswrapper[4772]: I0127 16:54:04.681921 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4" path="/var/lib/kubelet/pods/7ff09ddf-04fd-42e9-b6fd-8c9fd9fac0e4/volumes" Jan 27 16:54:09 crc kubenswrapper[4772]: I0127 16:54:09.696109 4772 scope.go:117] "RemoveContainer" containerID="5f81345bb4c2b5ce3e497a908767c4ee00a2cf35d4b6e4a9ea4e1fe6b2891391" Jan 27 16:54:09 crc kubenswrapper[4772]: I0127 16:54:09.728600 4772 scope.go:117] "RemoveContainer" containerID="5ed8b1fb179ec96a663871b1030549de3060884fefe55bd4b363bd01934e9e74" Jan 27 16:54:09 crc kubenswrapper[4772]: I0127 16:54:09.776071 4772 scope.go:117] "RemoveContainer" containerID="ce5afc895546186c87fa545a3ff11c6c821cf7ba305b10a05525fc458e3f4be7" Jan 27 16:54:09 crc kubenswrapper[4772]: I0127 16:54:09.814185 4772 scope.go:117] "RemoveContainer" containerID="742607c6bfa7c1d2a5f8ceee5a81852f3b14a71713e837fcc9210fdc6dccc556" Jan 27 16:54:15 crc kubenswrapper[4772]: I0127 16:54:15.664562 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 16:54:15 crc kubenswrapper[4772]: E0127 16:54:15.665773 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:54:26 crc kubenswrapper[4772]: I0127 16:54:26.663510 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 16:54:26 crc kubenswrapper[4772]: E0127 16:54:26.664950 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:54:39 crc kubenswrapper[4772]: I0127 16:54:39.664257 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 16:54:39 crc kubenswrapper[4772]: E0127 16:54:39.665278 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:54:40 crc kubenswrapper[4772]: I0127 16:54:40.039427 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-sync-bz6q7"] Jan 27 16:54:40 crc kubenswrapper[4772]: I0127 16:54:40.047146 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-sync-bz6q7"] Jan 27 16:54:40 crc kubenswrapper[4772]: I0127 16:54:40.680970 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa" path="/var/lib/kubelet/pods/1152dfc9-a3d1-41a5-92cb-a5a8c481a8fa/volumes" Jan 27 16:54:53 crc kubenswrapper[4772]: I0127 16:54:53.662842 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 16:54:53 crc kubenswrapper[4772]: E0127 16:54:53.664961 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:55:08 crc kubenswrapper[4772]: I0127 16:55:08.664728 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 16:55:08 crc kubenswrapper[4772]: E0127 16:55:08.665802 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:55:09 crc kubenswrapper[4772]: I0127 16:55:09.957300 4772 scope.go:117] "RemoveContainer" containerID="908bfdb73b8d3ebb90afef686e407a3a2da1f2ea295c24e132068119ed919b42" Jan 27 16:55:09 crc kubenswrapper[4772]: I0127 16:55:09.995510 4772 scope.go:117] "RemoveContainer" containerID="3c5cbb7ef3f7daca21bab77efa8022a5214525eb8b99dd2483d6689f1db4cb83" Jan 27 16:55:20 crc kubenswrapper[4772]: I0127 16:55:20.664024 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 16:55:20 crc kubenswrapper[4772]: E0127 16:55:20.665223 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:55:35 crc kubenswrapper[4772]: I0127 16:55:35.663395 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 16:55:35 crc kubenswrapper[4772]: E0127 16:55:35.664361 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:55:48 crc kubenswrapper[4772]: I0127 16:55:48.663700 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 16:55:48 crc kubenswrapper[4772]: E0127 16:55:48.664619 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:56:01 crc kubenswrapper[4772]: I0127 16:56:01.663778 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 16:56:01 crc kubenswrapper[4772]: E0127 16:56:01.664736 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:56:16 crc kubenswrapper[4772]: I0127 16:56:16.663927 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 16:56:16 crc kubenswrapper[4772]: E0127 16:56:16.665275 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:56:27 crc kubenswrapper[4772]: I0127 16:56:27.664348 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 16:56:27 crc kubenswrapper[4772]: E0127 16:56:27.665354 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:56:41 crc kubenswrapper[4772]: I0127 16:56:41.663906 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 16:56:41 crc kubenswrapper[4772]: E0127 16:56:41.664726 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:56:54 crc kubenswrapper[4772]: I0127 16:56:54.662967 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 16:56:54 crc kubenswrapper[4772]: E0127 16:56:54.663750 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:57:08 crc kubenswrapper[4772]: I0127 16:57:08.663500 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 16:57:08 crc kubenswrapper[4772]: E0127 16:57:08.664656 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:57:22 crc kubenswrapper[4772]: I0127 16:57:22.664985 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 16:57:22 crc kubenswrapper[4772]: E0127 16:57:22.666570 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:57:34 crc kubenswrapper[4772]: I0127 16:57:34.669880 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 16:57:34 crc kubenswrapper[4772]: E0127 16:57:34.670801 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:57:49 crc kubenswrapper[4772]: I0127 16:57:49.663304 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 16:57:49 crc kubenswrapper[4772]: E0127 16:57:49.664627 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:58:02 crc kubenswrapper[4772]: I0127 16:58:02.663102 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 16:58:02 crc kubenswrapper[4772]: E0127 16:58:02.664099 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 16:58:15 crc kubenswrapper[4772]: I0127 16:58:15.663855 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 16:58:16 crc kubenswrapper[4772]: I0127 16:58:16.441305 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"0c934f11ffcfb51be7cc650e76f8b239868b5820a22f2783555e83c31ae7ef8b"} Jan 27 16:58:16 crc kubenswrapper[4772]: I0127 16:58:16.637464 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-89vmp"] Jan 27 16:58:16 crc kubenswrapper[4772]: E0127 16:58:16.637913 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e447eca7-2085-4dde-84ff-0e85dccea5e9" containerName="extract-utilities" Jan 27 16:58:16 crc kubenswrapper[4772]: I0127 16:58:16.637934 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e447eca7-2085-4dde-84ff-0e85dccea5e9" containerName="extract-utilities" Jan 27 16:58:16 crc kubenswrapper[4772]: E0127 16:58:16.637957 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e447eca7-2085-4dde-84ff-0e85dccea5e9" containerName="registry-server" Jan 27 16:58:16 crc kubenswrapper[4772]: I0127 16:58:16.637965 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e447eca7-2085-4dde-84ff-0e85dccea5e9" containerName="registry-server" Jan 27 16:58:16 crc kubenswrapper[4772]: E0127 16:58:16.637985 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e447eca7-2085-4dde-84ff-0e85dccea5e9" containerName="extract-content" Jan 27 16:58:16 crc kubenswrapper[4772]: I0127 16:58:16.637992 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="e447eca7-2085-4dde-84ff-0e85dccea5e9" containerName="extract-content" Jan 27 16:58:16 crc kubenswrapper[4772]: I0127 16:58:16.638279 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="e447eca7-2085-4dde-84ff-0e85dccea5e9" containerName="registry-server" Jan 27 16:58:16 crc kubenswrapper[4772]: I0127 16:58:16.640099 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-89vmp" Jan 27 16:58:16 crc kubenswrapper[4772]: I0127 16:58:16.722923 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-89vmp"] Jan 27 16:58:16 crc kubenswrapper[4772]: I0127 16:58:16.783565 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwlb6\" (UniqueName: \"kubernetes.io/projected/f91cc5e4-df6f-47e0-b9ce-795e207a8561-kube-api-access-fwlb6\") pod \"redhat-operators-89vmp\" (UID: \"f91cc5e4-df6f-47e0-b9ce-795e207a8561\") " pod="openshift-marketplace/redhat-operators-89vmp" Jan 27 16:58:16 crc kubenswrapper[4772]: I0127 16:58:16.783726 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f91cc5e4-df6f-47e0-b9ce-795e207a8561-utilities\") pod \"redhat-operators-89vmp\" (UID: \"f91cc5e4-df6f-47e0-b9ce-795e207a8561\") " pod="openshift-marketplace/redhat-operators-89vmp" Jan 27 16:58:16 crc kubenswrapper[4772]: I0127 16:58:16.783758 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f91cc5e4-df6f-47e0-b9ce-795e207a8561-catalog-content\") pod \"redhat-operators-89vmp\" (UID: \"f91cc5e4-df6f-47e0-b9ce-795e207a8561\") " pod="openshift-marketplace/redhat-operators-89vmp" Jan 27 16:58:16 crc kubenswrapper[4772]: I0127 16:58:16.885066 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwlb6\" (UniqueName: \"kubernetes.io/projected/f91cc5e4-df6f-47e0-b9ce-795e207a8561-kube-api-access-fwlb6\") pod \"redhat-operators-89vmp\" (UID: \"f91cc5e4-df6f-47e0-b9ce-795e207a8561\") " pod="openshift-marketplace/redhat-operators-89vmp" Jan 27 16:58:16 crc kubenswrapper[4772]: I0127 16:58:16.885186 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f91cc5e4-df6f-47e0-b9ce-795e207a8561-utilities\") pod \"redhat-operators-89vmp\" (UID: \"f91cc5e4-df6f-47e0-b9ce-795e207a8561\") " pod="openshift-marketplace/redhat-operators-89vmp" Jan 27 16:58:16 crc kubenswrapper[4772]: I0127 16:58:16.885208 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f91cc5e4-df6f-47e0-b9ce-795e207a8561-catalog-content\") pod \"redhat-operators-89vmp\" (UID: \"f91cc5e4-df6f-47e0-b9ce-795e207a8561\") " pod="openshift-marketplace/redhat-operators-89vmp" Jan 27 16:58:16 crc kubenswrapper[4772]: I0127 16:58:16.885667 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f91cc5e4-df6f-47e0-b9ce-795e207a8561-catalog-content\") pod \"redhat-operators-89vmp\" (UID: \"f91cc5e4-df6f-47e0-b9ce-795e207a8561\") " pod="openshift-marketplace/redhat-operators-89vmp" Jan 27 16:58:16 crc kubenswrapper[4772]: I0127 16:58:16.885767 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f91cc5e4-df6f-47e0-b9ce-795e207a8561-utilities\") pod \"redhat-operators-89vmp\" (UID: \"f91cc5e4-df6f-47e0-b9ce-795e207a8561\") " pod="openshift-marketplace/redhat-operators-89vmp" Jan 27 16:58:16 crc kubenswrapper[4772]: I0127 16:58:16.904752 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwlb6\" (UniqueName: \"kubernetes.io/projected/f91cc5e4-df6f-47e0-b9ce-795e207a8561-kube-api-access-fwlb6\") pod \"redhat-operators-89vmp\" (UID: \"f91cc5e4-df6f-47e0-b9ce-795e207a8561\") " pod="openshift-marketplace/redhat-operators-89vmp" Jan 27 16:58:17 crc kubenswrapper[4772]: I0127 16:58:17.022506 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-89vmp" Jan 27 16:58:17 crc kubenswrapper[4772]: I0127 16:58:17.527593 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-89vmp"] Jan 27 16:58:17 crc kubenswrapper[4772]: W0127 16:58:17.532374 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf91cc5e4_df6f_47e0_b9ce_795e207a8561.slice/crio-909dee473f295614475f73b896a6fcc55c3055397f22ce88d988578e94255531 WatchSource:0}: Error finding container 909dee473f295614475f73b896a6fcc55c3055397f22ce88d988578e94255531: Status 404 returned error can't find the container with id 909dee473f295614475f73b896a6fcc55c3055397f22ce88d988578e94255531 Jan 27 16:58:18 crc kubenswrapper[4772]: I0127 16:58:18.458588 4772 generic.go:334] "Generic (PLEG): container finished" podID="f91cc5e4-df6f-47e0-b9ce-795e207a8561" containerID="8b5be62588d532a863b7d2cb4a30bdd9901e4a378dd25611d18afdd47f45e4a0" exitCode=0 Jan 27 16:58:18 crc kubenswrapper[4772]: I0127 16:58:18.458705 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-89vmp" event={"ID":"f91cc5e4-df6f-47e0-b9ce-795e207a8561","Type":"ContainerDied","Data":"8b5be62588d532a863b7d2cb4a30bdd9901e4a378dd25611d18afdd47f45e4a0"} Jan 27 16:58:18 crc kubenswrapper[4772]: I0127 16:58:18.459103 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-89vmp" event={"ID":"f91cc5e4-df6f-47e0-b9ce-795e207a8561","Type":"ContainerStarted","Data":"909dee473f295614475f73b896a6fcc55c3055397f22ce88d988578e94255531"} Jan 27 16:58:18 crc kubenswrapper[4772]: I0127 16:58:18.461576 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 16:58:19 crc kubenswrapper[4772]: I0127 16:58:19.475996 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-89vmp" event={"ID":"f91cc5e4-df6f-47e0-b9ce-795e207a8561","Type":"ContainerStarted","Data":"8ce2ec520d6e47d51f04b217aa2ebed5a969d9d59a7c977fcf91c69467c21d2b"} Jan 27 16:58:20 crc kubenswrapper[4772]: I0127 16:58:20.492210 4772 generic.go:334] "Generic (PLEG): container finished" podID="f91cc5e4-df6f-47e0-b9ce-795e207a8561" containerID="8ce2ec520d6e47d51f04b217aa2ebed5a969d9d59a7c977fcf91c69467c21d2b" exitCode=0 Jan 27 16:58:20 crc kubenswrapper[4772]: I0127 16:58:20.492859 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-89vmp" event={"ID":"f91cc5e4-df6f-47e0-b9ce-795e207a8561","Type":"ContainerDied","Data":"8ce2ec520d6e47d51f04b217aa2ebed5a969d9d59a7c977fcf91c69467c21d2b"} Jan 27 16:58:21 crc kubenswrapper[4772]: I0127 16:58:21.505554 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-89vmp" event={"ID":"f91cc5e4-df6f-47e0-b9ce-795e207a8561","Type":"ContainerStarted","Data":"340c878f6ca56d802bba3e1f2d501805058c80c44e6d3a585e2176c49b0e4816"} Jan 27 16:58:21 crc kubenswrapper[4772]: I0127 16:58:21.534316 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-89vmp" podStartSLOduration=3.090597674 podStartE2EDuration="5.534295549s" podCreationTimestamp="2026-01-27 16:58:16 +0000 UTC" firstStartedPulling="2026-01-27 16:58:18.461240808 +0000 UTC m=+6684.441849926" lastFinishedPulling="2026-01-27 16:58:20.904938703 +0000 UTC m=+6686.885547801" observedRunningTime="2026-01-27 16:58:21.525277482 +0000 UTC m=+6687.505886610" watchObservedRunningTime="2026-01-27 16:58:21.534295549 +0000 UTC m=+6687.514904657" Jan 27 16:58:27 crc kubenswrapper[4772]: I0127 16:58:27.023261 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-89vmp" Jan 27 16:58:27 crc kubenswrapper[4772]: I0127 16:58:27.024888 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-89vmp" Jan 27 16:58:28 crc kubenswrapper[4772]: I0127 16:58:28.080995 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-89vmp" podUID="f91cc5e4-df6f-47e0-b9ce-795e207a8561" containerName="registry-server" probeResult="failure" output=< Jan 27 16:58:28 crc kubenswrapper[4772]: timeout: failed to connect service ":50051" within 1s Jan 27 16:58:28 crc kubenswrapper[4772]: > Jan 27 16:58:37 crc kubenswrapper[4772]: I0127 16:58:37.101686 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-89vmp" Jan 27 16:58:37 crc kubenswrapper[4772]: I0127 16:58:37.182106 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-89vmp" Jan 27 16:58:37 crc kubenswrapper[4772]: I0127 16:58:37.354871 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-89vmp"] Jan 27 16:58:38 crc kubenswrapper[4772]: I0127 16:58:38.686976 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-89vmp" podUID="f91cc5e4-df6f-47e0-b9ce-795e207a8561" containerName="registry-server" containerID="cri-o://340c878f6ca56d802bba3e1f2d501805058c80c44e6d3a585e2176c49b0e4816" gracePeriod=2 Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.170795 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-89vmp" Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.296803 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f91cc5e4-df6f-47e0-b9ce-795e207a8561-utilities\") pod \"f91cc5e4-df6f-47e0-b9ce-795e207a8561\" (UID: \"f91cc5e4-df6f-47e0-b9ce-795e207a8561\") " Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.296888 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwlb6\" (UniqueName: \"kubernetes.io/projected/f91cc5e4-df6f-47e0-b9ce-795e207a8561-kube-api-access-fwlb6\") pod \"f91cc5e4-df6f-47e0-b9ce-795e207a8561\" (UID: \"f91cc5e4-df6f-47e0-b9ce-795e207a8561\") " Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.296950 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f91cc5e4-df6f-47e0-b9ce-795e207a8561-catalog-content\") pod \"f91cc5e4-df6f-47e0-b9ce-795e207a8561\" (UID: \"f91cc5e4-df6f-47e0-b9ce-795e207a8561\") " Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.298649 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f91cc5e4-df6f-47e0-b9ce-795e207a8561-utilities" (OuterVolumeSpecName: "utilities") pod "f91cc5e4-df6f-47e0-b9ce-795e207a8561" (UID: "f91cc5e4-df6f-47e0-b9ce-795e207a8561"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.303237 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f91cc5e4-df6f-47e0-b9ce-795e207a8561-kube-api-access-fwlb6" (OuterVolumeSpecName: "kube-api-access-fwlb6") pod "f91cc5e4-df6f-47e0-b9ce-795e207a8561" (UID: "f91cc5e4-df6f-47e0-b9ce-795e207a8561"). InnerVolumeSpecName "kube-api-access-fwlb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.400084 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f91cc5e4-df6f-47e0-b9ce-795e207a8561-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.400149 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwlb6\" (UniqueName: \"kubernetes.io/projected/f91cc5e4-df6f-47e0-b9ce-795e207a8561-kube-api-access-fwlb6\") on node \"crc\" DevicePath \"\"" Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.425256 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f91cc5e4-df6f-47e0-b9ce-795e207a8561-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f91cc5e4-df6f-47e0-b9ce-795e207a8561" (UID: "f91cc5e4-df6f-47e0-b9ce-795e207a8561"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.502505 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f91cc5e4-df6f-47e0-b9ce-795e207a8561-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.723158 4772 generic.go:334] "Generic (PLEG): container finished" podID="f91cc5e4-df6f-47e0-b9ce-795e207a8561" containerID="340c878f6ca56d802bba3e1f2d501805058c80c44e6d3a585e2176c49b0e4816" exitCode=0 Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.723246 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-89vmp" event={"ID":"f91cc5e4-df6f-47e0-b9ce-795e207a8561","Type":"ContainerDied","Data":"340c878f6ca56d802bba3e1f2d501805058c80c44e6d3a585e2176c49b0e4816"} Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.723295 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-89vmp" event={"ID":"f91cc5e4-df6f-47e0-b9ce-795e207a8561","Type":"ContainerDied","Data":"909dee473f295614475f73b896a6fcc55c3055397f22ce88d988578e94255531"} Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.723310 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-89vmp" Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.723325 4772 scope.go:117] "RemoveContainer" containerID="340c878f6ca56d802bba3e1f2d501805058c80c44e6d3a585e2176c49b0e4816" Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.772642 4772 scope.go:117] "RemoveContainer" containerID="8ce2ec520d6e47d51f04b217aa2ebed5a969d9d59a7c977fcf91c69467c21d2b" Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.777560 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-89vmp"] Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.789753 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-89vmp"] Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.797779 4772 scope.go:117] "RemoveContainer" containerID="8b5be62588d532a863b7d2cb4a30bdd9901e4a378dd25611d18afdd47f45e4a0" Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.844446 4772 scope.go:117] "RemoveContainer" containerID="340c878f6ca56d802bba3e1f2d501805058c80c44e6d3a585e2176c49b0e4816" Jan 27 16:58:39 crc kubenswrapper[4772]: E0127 16:58:39.844860 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"340c878f6ca56d802bba3e1f2d501805058c80c44e6d3a585e2176c49b0e4816\": container with ID starting with 340c878f6ca56d802bba3e1f2d501805058c80c44e6d3a585e2176c49b0e4816 not found: ID does not exist" containerID="340c878f6ca56d802bba3e1f2d501805058c80c44e6d3a585e2176c49b0e4816" Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.844906 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"340c878f6ca56d802bba3e1f2d501805058c80c44e6d3a585e2176c49b0e4816"} err="failed to get container status \"340c878f6ca56d802bba3e1f2d501805058c80c44e6d3a585e2176c49b0e4816\": rpc error: code = NotFound desc = could not find container \"340c878f6ca56d802bba3e1f2d501805058c80c44e6d3a585e2176c49b0e4816\": container with ID starting with 340c878f6ca56d802bba3e1f2d501805058c80c44e6d3a585e2176c49b0e4816 not found: ID does not exist" Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.844931 4772 scope.go:117] "RemoveContainer" containerID="8ce2ec520d6e47d51f04b217aa2ebed5a969d9d59a7c977fcf91c69467c21d2b" Jan 27 16:58:39 crc kubenswrapper[4772]: E0127 16:58:39.845197 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ce2ec520d6e47d51f04b217aa2ebed5a969d9d59a7c977fcf91c69467c21d2b\": container with ID starting with 8ce2ec520d6e47d51f04b217aa2ebed5a969d9d59a7c977fcf91c69467c21d2b not found: ID does not exist" containerID="8ce2ec520d6e47d51f04b217aa2ebed5a969d9d59a7c977fcf91c69467c21d2b" Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.845218 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ce2ec520d6e47d51f04b217aa2ebed5a969d9d59a7c977fcf91c69467c21d2b"} err="failed to get container status \"8ce2ec520d6e47d51f04b217aa2ebed5a969d9d59a7c977fcf91c69467c21d2b\": rpc error: code = NotFound desc = could not find container \"8ce2ec520d6e47d51f04b217aa2ebed5a969d9d59a7c977fcf91c69467c21d2b\": container with ID starting with 8ce2ec520d6e47d51f04b217aa2ebed5a969d9d59a7c977fcf91c69467c21d2b not found: ID does not exist" Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.845231 4772 scope.go:117] "RemoveContainer" containerID="8b5be62588d532a863b7d2cb4a30bdd9901e4a378dd25611d18afdd47f45e4a0" Jan 27 16:58:39 crc kubenswrapper[4772]: E0127 16:58:39.845430 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b5be62588d532a863b7d2cb4a30bdd9901e4a378dd25611d18afdd47f45e4a0\": container with ID starting with 8b5be62588d532a863b7d2cb4a30bdd9901e4a378dd25611d18afdd47f45e4a0 not found: ID does not exist" containerID="8b5be62588d532a863b7d2cb4a30bdd9901e4a378dd25611d18afdd47f45e4a0" Jan 27 16:58:39 crc kubenswrapper[4772]: I0127 16:58:39.845455 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b5be62588d532a863b7d2cb4a30bdd9901e4a378dd25611d18afdd47f45e4a0"} err="failed to get container status \"8b5be62588d532a863b7d2cb4a30bdd9901e4a378dd25611d18afdd47f45e4a0\": rpc error: code = NotFound desc = could not find container \"8b5be62588d532a863b7d2cb4a30bdd9901e4a378dd25611d18afdd47f45e4a0\": container with ID starting with 8b5be62588d532a863b7d2cb4a30bdd9901e4a378dd25611d18afdd47f45e4a0 not found: ID does not exist" Jan 27 16:58:40 crc kubenswrapper[4772]: I0127 16:58:40.683517 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f91cc5e4-df6f-47e0-b9ce-795e207a8561" path="/var/lib/kubelet/pods/f91cc5e4-df6f-47e0-b9ce-795e207a8561/volumes" Jan 27 17:00:00 crc kubenswrapper[4772]: I0127 17:00:00.171148 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492220-sqfps"] Jan 27 17:00:00 crc kubenswrapper[4772]: E0127 17:00:00.172270 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f91cc5e4-df6f-47e0-b9ce-795e207a8561" containerName="extract-content" Jan 27 17:00:00 crc kubenswrapper[4772]: I0127 17:00:00.172288 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f91cc5e4-df6f-47e0-b9ce-795e207a8561" containerName="extract-content" Jan 27 17:00:00 crc kubenswrapper[4772]: E0127 17:00:00.172322 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f91cc5e4-df6f-47e0-b9ce-795e207a8561" containerName="extract-utilities" Jan 27 17:00:00 crc kubenswrapper[4772]: I0127 17:00:00.172331 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f91cc5e4-df6f-47e0-b9ce-795e207a8561" containerName="extract-utilities" Jan 27 17:00:00 crc kubenswrapper[4772]: E0127 17:00:00.172355 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f91cc5e4-df6f-47e0-b9ce-795e207a8561" containerName="registry-server" Jan 27 17:00:00 crc kubenswrapper[4772]: I0127 17:00:00.172364 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="f91cc5e4-df6f-47e0-b9ce-795e207a8561" containerName="registry-server" Jan 27 17:00:00 crc kubenswrapper[4772]: I0127 17:00:00.172593 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="f91cc5e4-df6f-47e0-b9ce-795e207a8561" containerName="registry-server" Jan 27 17:00:00 crc kubenswrapper[4772]: I0127 17:00:00.173409 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492220-sqfps" Jan 27 17:00:00 crc kubenswrapper[4772]: I0127 17:00:00.176453 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 27 17:00:00 crc kubenswrapper[4772]: I0127 17:00:00.181387 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 27 17:00:00 crc kubenswrapper[4772]: I0127 17:00:00.198308 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492220-sqfps"] Jan 27 17:00:00 crc kubenswrapper[4772]: I0127 17:00:00.302520 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/73038a7f-6c26-47b7-ad06-bd235e268224-secret-volume\") pod \"collect-profiles-29492220-sqfps\" (UID: \"73038a7f-6c26-47b7-ad06-bd235e268224\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492220-sqfps" Jan 27 17:00:00 crc kubenswrapper[4772]: I0127 17:00:00.302653 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mc69\" (UniqueName: \"kubernetes.io/projected/73038a7f-6c26-47b7-ad06-bd235e268224-kube-api-access-6mc69\") pod \"collect-profiles-29492220-sqfps\" (UID: \"73038a7f-6c26-47b7-ad06-bd235e268224\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492220-sqfps" Jan 27 17:00:00 crc kubenswrapper[4772]: I0127 17:00:00.302682 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/73038a7f-6c26-47b7-ad06-bd235e268224-config-volume\") pod \"collect-profiles-29492220-sqfps\" (UID: \"73038a7f-6c26-47b7-ad06-bd235e268224\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492220-sqfps" Jan 27 17:00:00 crc kubenswrapper[4772]: I0127 17:00:00.404746 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/73038a7f-6c26-47b7-ad06-bd235e268224-secret-volume\") pod \"collect-profiles-29492220-sqfps\" (UID: \"73038a7f-6c26-47b7-ad06-bd235e268224\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492220-sqfps" Jan 27 17:00:00 crc kubenswrapper[4772]: I0127 17:00:00.404845 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mc69\" (UniqueName: \"kubernetes.io/projected/73038a7f-6c26-47b7-ad06-bd235e268224-kube-api-access-6mc69\") pod \"collect-profiles-29492220-sqfps\" (UID: \"73038a7f-6c26-47b7-ad06-bd235e268224\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492220-sqfps" Jan 27 17:00:00 crc kubenswrapper[4772]: I0127 17:00:00.404864 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/73038a7f-6c26-47b7-ad06-bd235e268224-config-volume\") pod \"collect-profiles-29492220-sqfps\" (UID: \"73038a7f-6c26-47b7-ad06-bd235e268224\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492220-sqfps" Jan 27 17:00:00 crc kubenswrapper[4772]: I0127 17:00:00.405782 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/73038a7f-6c26-47b7-ad06-bd235e268224-config-volume\") pod \"collect-profiles-29492220-sqfps\" (UID: \"73038a7f-6c26-47b7-ad06-bd235e268224\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492220-sqfps" Jan 27 17:00:00 crc kubenswrapper[4772]: I0127 17:00:00.412946 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/73038a7f-6c26-47b7-ad06-bd235e268224-secret-volume\") pod \"collect-profiles-29492220-sqfps\" (UID: \"73038a7f-6c26-47b7-ad06-bd235e268224\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492220-sqfps" Jan 27 17:00:00 crc kubenswrapper[4772]: I0127 17:00:00.425867 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mc69\" (UniqueName: \"kubernetes.io/projected/73038a7f-6c26-47b7-ad06-bd235e268224-kube-api-access-6mc69\") pod \"collect-profiles-29492220-sqfps\" (UID: \"73038a7f-6c26-47b7-ad06-bd235e268224\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492220-sqfps" Jan 27 17:00:00 crc kubenswrapper[4772]: I0127 17:00:00.500637 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492220-sqfps" Jan 27 17:00:00 crc kubenswrapper[4772]: I0127 17:00:00.975926 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492220-sqfps"] Jan 27 17:00:01 crc kubenswrapper[4772]: I0127 17:00:01.556920 4772 generic.go:334] "Generic (PLEG): container finished" podID="73038a7f-6c26-47b7-ad06-bd235e268224" containerID="21fbf772d614ea3a35cb7d6244635ba0574a7b4a610726fba6575e200d3d3209" exitCode=0 Jan 27 17:00:01 crc kubenswrapper[4772]: I0127 17:00:01.556967 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492220-sqfps" event={"ID":"73038a7f-6c26-47b7-ad06-bd235e268224","Type":"ContainerDied","Data":"21fbf772d614ea3a35cb7d6244635ba0574a7b4a610726fba6575e200d3d3209"} Jan 27 17:00:01 crc kubenswrapper[4772]: I0127 17:00:01.557270 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492220-sqfps" event={"ID":"73038a7f-6c26-47b7-ad06-bd235e268224","Type":"ContainerStarted","Data":"c6cd8a12b017031e00e5f372168cdf699372dbce305414437676d1d6580430fc"} Jan 27 17:00:02 crc kubenswrapper[4772]: I0127 17:00:02.934013 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492220-sqfps" Jan 27 17:00:03 crc kubenswrapper[4772]: I0127 17:00:03.055568 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mc69\" (UniqueName: \"kubernetes.io/projected/73038a7f-6c26-47b7-ad06-bd235e268224-kube-api-access-6mc69\") pod \"73038a7f-6c26-47b7-ad06-bd235e268224\" (UID: \"73038a7f-6c26-47b7-ad06-bd235e268224\") " Jan 27 17:00:03 crc kubenswrapper[4772]: I0127 17:00:03.055684 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/73038a7f-6c26-47b7-ad06-bd235e268224-config-volume\") pod \"73038a7f-6c26-47b7-ad06-bd235e268224\" (UID: \"73038a7f-6c26-47b7-ad06-bd235e268224\") " Jan 27 17:00:03 crc kubenswrapper[4772]: I0127 17:00:03.055824 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/73038a7f-6c26-47b7-ad06-bd235e268224-secret-volume\") pod \"73038a7f-6c26-47b7-ad06-bd235e268224\" (UID: \"73038a7f-6c26-47b7-ad06-bd235e268224\") " Jan 27 17:00:03 crc kubenswrapper[4772]: I0127 17:00:03.056369 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73038a7f-6c26-47b7-ad06-bd235e268224-config-volume" (OuterVolumeSpecName: "config-volume") pod "73038a7f-6c26-47b7-ad06-bd235e268224" (UID: "73038a7f-6c26-47b7-ad06-bd235e268224"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 17:00:03 crc kubenswrapper[4772]: I0127 17:00:03.061359 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73038a7f-6c26-47b7-ad06-bd235e268224-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "73038a7f-6c26-47b7-ad06-bd235e268224" (UID: "73038a7f-6c26-47b7-ad06-bd235e268224"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 17:00:03 crc kubenswrapper[4772]: I0127 17:00:03.061388 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73038a7f-6c26-47b7-ad06-bd235e268224-kube-api-access-6mc69" (OuterVolumeSpecName: "kube-api-access-6mc69") pod "73038a7f-6c26-47b7-ad06-bd235e268224" (UID: "73038a7f-6c26-47b7-ad06-bd235e268224"). InnerVolumeSpecName "kube-api-access-6mc69". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:00:03 crc kubenswrapper[4772]: I0127 17:00:03.159156 4772 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/73038a7f-6c26-47b7-ad06-bd235e268224-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 27 17:00:03 crc kubenswrapper[4772]: I0127 17:00:03.159248 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mc69\" (UniqueName: \"kubernetes.io/projected/73038a7f-6c26-47b7-ad06-bd235e268224-kube-api-access-6mc69\") on node \"crc\" DevicePath \"\"" Jan 27 17:00:03 crc kubenswrapper[4772]: I0127 17:00:03.159268 4772 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/73038a7f-6c26-47b7-ad06-bd235e268224-config-volume\") on node \"crc\" DevicePath \"\"" Jan 27 17:00:03 crc kubenswrapper[4772]: I0127 17:00:03.578212 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492220-sqfps" event={"ID":"73038a7f-6c26-47b7-ad06-bd235e268224","Type":"ContainerDied","Data":"c6cd8a12b017031e00e5f372168cdf699372dbce305414437676d1d6580430fc"} Jan 27 17:00:03 crc kubenswrapper[4772]: I0127 17:00:03.578263 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6cd8a12b017031e00e5f372168cdf699372dbce305414437676d1d6580430fc" Jan 27 17:00:03 crc kubenswrapper[4772]: I0127 17:00:03.578292 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492220-sqfps" Jan 27 17:00:04 crc kubenswrapper[4772]: I0127 17:00:04.034655 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492175-fg6x4"] Jan 27 17:00:04 crc kubenswrapper[4772]: I0127 17:00:04.046691 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492175-fg6x4"] Jan 27 17:00:04 crc kubenswrapper[4772]: I0127 17:00:04.683433 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0be908ee-6173-4ee8-80c4-0738697898d2" path="/var/lib/kubelet/pods/0be908ee-6173-4ee8-80c4-0738697898d2/volumes" Jan 27 17:00:10 crc kubenswrapper[4772]: I0127 17:00:10.209706 4772 scope.go:117] "RemoveContainer" containerID="d013dea461e279e8b861558e82f04a509da66ccae91eabf32103d04803eb33bd" Jan 27 17:00:42 crc kubenswrapper[4772]: I0127 17:00:42.058960 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:00:42 crc kubenswrapper[4772]: I0127 17:00:42.059964 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:00:50 crc kubenswrapper[4772]: I0127 17:00:50.070849 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-txfct"] Jan 27 17:00:50 crc kubenswrapper[4772]: E0127 17:00:50.072073 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73038a7f-6c26-47b7-ad06-bd235e268224" containerName="collect-profiles" Jan 27 17:00:50 crc kubenswrapper[4772]: I0127 17:00:50.072094 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="73038a7f-6c26-47b7-ad06-bd235e268224" containerName="collect-profiles" Jan 27 17:00:50 crc kubenswrapper[4772]: I0127 17:00:50.072367 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="73038a7f-6c26-47b7-ad06-bd235e268224" containerName="collect-profiles" Jan 27 17:00:50 crc kubenswrapper[4772]: I0127 17:00:50.074262 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-txfct" Jan 27 17:00:50 crc kubenswrapper[4772]: I0127 17:00:50.084095 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-txfct"] Jan 27 17:00:50 crc kubenswrapper[4772]: I0127 17:00:50.116337 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkhfw\" (UniqueName: \"kubernetes.io/projected/10eef819-4355-4b65-bb81-95c055327034-kube-api-access-tkhfw\") pod \"community-operators-txfct\" (UID: \"10eef819-4355-4b65-bb81-95c055327034\") " pod="openshift-marketplace/community-operators-txfct" Jan 27 17:00:50 crc kubenswrapper[4772]: I0127 17:00:50.116378 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10eef819-4355-4b65-bb81-95c055327034-utilities\") pod \"community-operators-txfct\" (UID: \"10eef819-4355-4b65-bb81-95c055327034\") " pod="openshift-marketplace/community-operators-txfct" Jan 27 17:00:50 crc kubenswrapper[4772]: I0127 17:00:50.116403 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10eef819-4355-4b65-bb81-95c055327034-catalog-content\") pod \"community-operators-txfct\" (UID: \"10eef819-4355-4b65-bb81-95c055327034\") " pod="openshift-marketplace/community-operators-txfct" Jan 27 17:00:50 crc kubenswrapper[4772]: I0127 17:00:50.217658 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkhfw\" (UniqueName: \"kubernetes.io/projected/10eef819-4355-4b65-bb81-95c055327034-kube-api-access-tkhfw\") pod \"community-operators-txfct\" (UID: \"10eef819-4355-4b65-bb81-95c055327034\") " pod="openshift-marketplace/community-operators-txfct" Jan 27 17:00:50 crc kubenswrapper[4772]: I0127 17:00:50.217720 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10eef819-4355-4b65-bb81-95c055327034-utilities\") pod \"community-operators-txfct\" (UID: \"10eef819-4355-4b65-bb81-95c055327034\") " pod="openshift-marketplace/community-operators-txfct" Jan 27 17:00:50 crc kubenswrapper[4772]: I0127 17:00:50.217749 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10eef819-4355-4b65-bb81-95c055327034-catalog-content\") pod \"community-operators-txfct\" (UID: \"10eef819-4355-4b65-bb81-95c055327034\") " pod="openshift-marketplace/community-operators-txfct" Jan 27 17:00:50 crc kubenswrapper[4772]: I0127 17:00:50.218210 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10eef819-4355-4b65-bb81-95c055327034-utilities\") pod \"community-operators-txfct\" (UID: \"10eef819-4355-4b65-bb81-95c055327034\") " pod="openshift-marketplace/community-operators-txfct" Jan 27 17:00:50 crc kubenswrapper[4772]: I0127 17:00:50.218255 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10eef819-4355-4b65-bb81-95c055327034-catalog-content\") pod \"community-operators-txfct\" (UID: \"10eef819-4355-4b65-bb81-95c055327034\") " pod="openshift-marketplace/community-operators-txfct" Jan 27 17:00:50 crc kubenswrapper[4772]: I0127 17:00:50.240087 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkhfw\" (UniqueName: \"kubernetes.io/projected/10eef819-4355-4b65-bb81-95c055327034-kube-api-access-tkhfw\") pod \"community-operators-txfct\" (UID: \"10eef819-4355-4b65-bb81-95c055327034\") " pod="openshift-marketplace/community-operators-txfct" Jan 27 17:00:50 crc kubenswrapper[4772]: I0127 17:00:50.437309 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-txfct" Jan 27 17:00:50 crc kubenswrapper[4772]: I0127 17:00:50.948546 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-txfct"] Jan 27 17:00:50 crc kubenswrapper[4772]: W0127 17:00:50.951654 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10eef819_4355_4b65_bb81_95c055327034.slice/crio-23557e0504a3fe336145be5a361937c216f6e8cad04bfe40c3131be90c233123 WatchSource:0}: Error finding container 23557e0504a3fe336145be5a361937c216f6e8cad04bfe40c3131be90c233123: Status 404 returned error can't find the container with id 23557e0504a3fe336145be5a361937c216f6e8cad04bfe40c3131be90c233123 Jan 27 17:00:51 crc kubenswrapper[4772]: I0127 17:00:51.128123 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txfct" event={"ID":"10eef819-4355-4b65-bb81-95c055327034","Type":"ContainerStarted","Data":"23557e0504a3fe336145be5a361937c216f6e8cad04bfe40c3131be90c233123"} Jan 27 17:00:52 crc kubenswrapper[4772]: I0127 17:00:52.136759 4772 generic.go:334] "Generic (PLEG): container finished" podID="10eef819-4355-4b65-bb81-95c055327034" containerID="7236c5d9dd7b59e4e1b4af9645ea7ed61b907d3619dcb83d87301be5c3316f7c" exitCode=0 Jan 27 17:00:52 crc kubenswrapper[4772]: I0127 17:00:52.136858 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txfct" event={"ID":"10eef819-4355-4b65-bb81-95c055327034","Type":"ContainerDied","Data":"7236c5d9dd7b59e4e1b4af9645ea7ed61b907d3619dcb83d87301be5c3316f7c"} Jan 27 17:00:56 crc kubenswrapper[4772]: I0127 17:00:56.177069 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txfct" event={"ID":"10eef819-4355-4b65-bb81-95c055327034","Type":"ContainerStarted","Data":"490d9faa657579105d3e49ba461cb9e24406692936441c0207c58ab54ad55605"} Jan 27 17:00:57 crc kubenswrapper[4772]: I0127 17:00:57.188048 4772 generic.go:334] "Generic (PLEG): container finished" podID="10eef819-4355-4b65-bb81-95c055327034" containerID="490d9faa657579105d3e49ba461cb9e24406692936441c0207c58ab54ad55605" exitCode=0 Jan 27 17:00:57 crc kubenswrapper[4772]: I0127 17:00:57.188099 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txfct" event={"ID":"10eef819-4355-4b65-bb81-95c055327034","Type":"ContainerDied","Data":"490d9faa657579105d3e49ba461cb9e24406692936441c0207c58ab54ad55605"} Jan 27 17:00:58 crc kubenswrapper[4772]: I0127 17:00:58.200132 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txfct" event={"ID":"10eef819-4355-4b65-bb81-95c055327034","Type":"ContainerStarted","Data":"114bf805859534c9af2af0c87c4604c2095e959428fa0c7464d58e26e36fef93"} Jan 27 17:00:58 crc kubenswrapper[4772]: I0127 17:00:58.226149 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-txfct" podStartSLOduration=2.680170077 podStartE2EDuration="8.226126016s" podCreationTimestamp="2026-01-27 17:00:50 +0000 UTC" firstStartedPulling="2026-01-27 17:00:52.1384179 +0000 UTC m=+6838.119027018" lastFinishedPulling="2026-01-27 17:00:57.684373849 +0000 UTC m=+6843.664982957" observedRunningTime="2026-01-27 17:00:58.221127443 +0000 UTC m=+6844.201736551" watchObservedRunningTime="2026-01-27 17:00:58.226126016 +0000 UTC m=+6844.206735104" Jan 27 17:01:00 crc kubenswrapper[4772]: I0127 17:01:00.169643 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29492221-sbxjp"] Jan 27 17:01:00 crc kubenswrapper[4772]: I0127 17:01:00.172419 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29492221-sbxjp" Jan 27 17:01:00 crc kubenswrapper[4772]: I0127 17:01:00.194787 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29492221-sbxjp"] Jan 27 17:01:00 crc kubenswrapper[4772]: I0127 17:01:00.235974 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnldn\" (UniqueName: \"kubernetes.io/projected/8459d055-62d3-4699-b477-ea15946b982c-kube-api-access-qnldn\") pod \"keystone-cron-29492221-sbxjp\" (UID: \"8459d055-62d3-4699-b477-ea15946b982c\") " pod="openstack/keystone-cron-29492221-sbxjp" Jan 27 17:01:00 crc kubenswrapper[4772]: I0127 17:01:00.236051 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8459d055-62d3-4699-b477-ea15946b982c-config-data\") pod \"keystone-cron-29492221-sbxjp\" (UID: \"8459d055-62d3-4699-b477-ea15946b982c\") " pod="openstack/keystone-cron-29492221-sbxjp" Jan 27 17:01:00 crc kubenswrapper[4772]: I0127 17:01:00.236098 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8459d055-62d3-4699-b477-ea15946b982c-combined-ca-bundle\") pod \"keystone-cron-29492221-sbxjp\" (UID: \"8459d055-62d3-4699-b477-ea15946b982c\") " pod="openstack/keystone-cron-29492221-sbxjp" Jan 27 17:01:00 crc kubenswrapper[4772]: I0127 17:01:00.236154 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8459d055-62d3-4699-b477-ea15946b982c-fernet-keys\") pod \"keystone-cron-29492221-sbxjp\" (UID: \"8459d055-62d3-4699-b477-ea15946b982c\") " pod="openstack/keystone-cron-29492221-sbxjp" Jan 27 17:01:00 crc kubenswrapper[4772]: I0127 17:01:00.337997 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8459d055-62d3-4699-b477-ea15946b982c-combined-ca-bundle\") pod \"keystone-cron-29492221-sbxjp\" (UID: \"8459d055-62d3-4699-b477-ea15946b982c\") " pod="openstack/keystone-cron-29492221-sbxjp" Jan 27 17:01:00 crc kubenswrapper[4772]: I0127 17:01:00.338482 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8459d055-62d3-4699-b477-ea15946b982c-fernet-keys\") pod \"keystone-cron-29492221-sbxjp\" (UID: \"8459d055-62d3-4699-b477-ea15946b982c\") " pod="openstack/keystone-cron-29492221-sbxjp" Jan 27 17:01:00 crc kubenswrapper[4772]: I0127 17:01:00.338704 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnldn\" (UniqueName: \"kubernetes.io/projected/8459d055-62d3-4699-b477-ea15946b982c-kube-api-access-qnldn\") pod \"keystone-cron-29492221-sbxjp\" (UID: \"8459d055-62d3-4699-b477-ea15946b982c\") " pod="openstack/keystone-cron-29492221-sbxjp" Jan 27 17:01:00 crc kubenswrapper[4772]: I0127 17:01:00.338879 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8459d055-62d3-4699-b477-ea15946b982c-config-data\") pod \"keystone-cron-29492221-sbxjp\" (UID: \"8459d055-62d3-4699-b477-ea15946b982c\") " pod="openstack/keystone-cron-29492221-sbxjp" Jan 27 17:01:00 crc kubenswrapper[4772]: I0127 17:01:00.344587 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8459d055-62d3-4699-b477-ea15946b982c-combined-ca-bundle\") pod \"keystone-cron-29492221-sbxjp\" (UID: \"8459d055-62d3-4699-b477-ea15946b982c\") " pod="openstack/keystone-cron-29492221-sbxjp" Jan 27 17:01:00 crc kubenswrapper[4772]: I0127 17:01:00.344781 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8459d055-62d3-4699-b477-ea15946b982c-config-data\") pod \"keystone-cron-29492221-sbxjp\" (UID: \"8459d055-62d3-4699-b477-ea15946b982c\") " pod="openstack/keystone-cron-29492221-sbxjp" Jan 27 17:01:00 crc kubenswrapper[4772]: I0127 17:01:00.346110 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8459d055-62d3-4699-b477-ea15946b982c-fernet-keys\") pod \"keystone-cron-29492221-sbxjp\" (UID: \"8459d055-62d3-4699-b477-ea15946b982c\") " pod="openstack/keystone-cron-29492221-sbxjp" Jan 27 17:01:00 crc kubenswrapper[4772]: I0127 17:01:00.355074 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnldn\" (UniqueName: \"kubernetes.io/projected/8459d055-62d3-4699-b477-ea15946b982c-kube-api-access-qnldn\") pod \"keystone-cron-29492221-sbxjp\" (UID: \"8459d055-62d3-4699-b477-ea15946b982c\") " pod="openstack/keystone-cron-29492221-sbxjp" Jan 27 17:01:00 crc kubenswrapper[4772]: I0127 17:01:00.438057 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-txfct" Jan 27 17:01:00 crc kubenswrapper[4772]: I0127 17:01:00.438124 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-txfct" Jan 27 17:01:00 crc kubenswrapper[4772]: I0127 17:01:00.494504 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29492221-sbxjp" Jan 27 17:01:00 crc kubenswrapper[4772]: I0127 17:01:00.501932 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-txfct" Jan 27 17:01:01 crc kubenswrapper[4772]: I0127 17:01:01.271342 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29492221-sbxjp"] Jan 27 17:01:01 crc kubenswrapper[4772]: W0127 17:01:01.286650 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8459d055_62d3_4699_b477_ea15946b982c.slice/crio-9e569490788ee5825687fe890769c82986f397a62be24308cfba122d04118a20 WatchSource:0}: Error finding container 9e569490788ee5825687fe890769c82986f397a62be24308cfba122d04118a20: Status 404 returned error can't find the container with id 9e569490788ee5825687fe890769c82986f397a62be24308cfba122d04118a20 Jan 27 17:01:02 crc kubenswrapper[4772]: I0127 17:01:02.244097 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29492221-sbxjp" event={"ID":"8459d055-62d3-4699-b477-ea15946b982c","Type":"ContainerStarted","Data":"9e569490788ee5825687fe890769c82986f397a62be24308cfba122d04118a20"} Jan 27 17:01:05 crc kubenswrapper[4772]: I0127 17:01:05.274449 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29492221-sbxjp" event={"ID":"8459d055-62d3-4699-b477-ea15946b982c","Type":"ContainerStarted","Data":"4cd94d61ef219342980e5481c59103d0260264e8f26556ff01423fdd9d5863ee"} Jan 27 17:01:07 crc kubenswrapper[4772]: I0127 17:01:07.302238 4772 generic.go:334] "Generic (PLEG): container finished" podID="8459d055-62d3-4699-b477-ea15946b982c" containerID="4cd94d61ef219342980e5481c59103d0260264e8f26556ff01423fdd9d5863ee" exitCode=0 Jan 27 17:01:07 crc kubenswrapper[4772]: I0127 17:01:07.302344 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29492221-sbxjp" event={"ID":"8459d055-62d3-4699-b477-ea15946b982c","Type":"ContainerDied","Data":"4cd94d61ef219342980e5481c59103d0260264e8f26556ff01423fdd9d5863ee"} Jan 27 17:01:08 crc kubenswrapper[4772]: I0127 17:01:08.655383 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29492221-sbxjp" Jan 27 17:01:08 crc kubenswrapper[4772]: I0127 17:01:08.750777 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8459d055-62d3-4699-b477-ea15946b982c-fernet-keys\") pod \"8459d055-62d3-4699-b477-ea15946b982c\" (UID: \"8459d055-62d3-4699-b477-ea15946b982c\") " Jan 27 17:01:08 crc kubenswrapper[4772]: I0127 17:01:08.750872 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8459d055-62d3-4699-b477-ea15946b982c-config-data\") pod \"8459d055-62d3-4699-b477-ea15946b982c\" (UID: \"8459d055-62d3-4699-b477-ea15946b982c\") " Jan 27 17:01:08 crc kubenswrapper[4772]: I0127 17:01:08.751043 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnldn\" (UniqueName: \"kubernetes.io/projected/8459d055-62d3-4699-b477-ea15946b982c-kube-api-access-qnldn\") pod \"8459d055-62d3-4699-b477-ea15946b982c\" (UID: \"8459d055-62d3-4699-b477-ea15946b982c\") " Jan 27 17:01:08 crc kubenswrapper[4772]: I0127 17:01:08.751147 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8459d055-62d3-4699-b477-ea15946b982c-combined-ca-bundle\") pod \"8459d055-62d3-4699-b477-ea15946b982c\" (UID: \"8459d055-62d3-4699-b477-ea15946b982c\") " Jan 27 17:01:08 crc kubenswrapper[4772]: I0127 17:01:08.757490 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8459d055-62d3-4699-b477-ea15946b982c-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8459d055-62d3-4699-b477-ea15946b982c" (UID: "8459d055-62d3-4699-b477-ea15946b982c"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 17:01:08 crc kubenswrapper[4772]: I0127 17:01:08.758697 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8459d055-62d3-4699-b477-ea15946b982c-kube-api-access-qnldn" (OuterVolumeSpecName: "kube-api-access-qnldn") pod "8459d055-62d3-4699-b477-ea15946b982c" (UID: "8459d055-62d3-4699-b477-ea15946b982c"). InnerVolumeSpecName "kube-api-access-qnldn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:01:08 crc kubenswrapper[4772]: I0127 17:01:08.779499 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8459d055-62d3-4699-b477-ea15946b982c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8459d055-62d3-4699-b477-ea15946b982c" (UID: "8459d055-62d3-4699-b477-ea15946b982c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 17:01:08 crc kubenswrapper[4772]: I0127 17:01:08.804920 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8459d055-62d3-4699-b477-ea15946b982c-config-data" (OuterVolumeSpecName: "config-data") pod "8459d055-62d3-4699-b477-ea15946b982c" (UID: "8459d055-62d3-4699-b477-ea15946b982c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 17:01:08 crc kubenswrapper[4772]: I0127 17:01:08.854109 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnldn\" (UniqueName: \"kubernetes.io/projected/8459d055-62d3-4699-b477-ea15946b982c-kube-api-access-qnldn\") on node \"crc\" DevicePath \"\"" Jan 27 17:01:08 crc kubenswrapper[4772]: I0127 17:01:08.854436 4772 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8459d055-62d3-4699-b477-ea15946b982c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 27 17:01:08 crc kubenswrapper[4772]: I0127 17:01:08.854451 4772 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8459d055-62d3-4699-b477-ea15946b982c-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 27 17:01:08 crc kubenswrapper[4772]: I0127 17:01:08.854464 4772 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8459d055-62d3-4699-b477-ea15946b982c-config-data\") on node \"crc\" DevicePath \"\"" Jan 27 17:01:09 crc kubenswrapper[4772]: I0127 17:01:09.322418 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29492221-sbxjp" event={"ID":"8459d055-62d3-4699-b477-ea15946b982c","Type":"ContainerDied","Data":"9e569490788ee5825687fe890769c82986f397a62be24308cfba122d04118a20"} Jan 27 17:01:09 crc kubenswrapper[4772]: I0127 17:01:09.322481 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e569490788ee5825687fe890769c82986f397a62be24308cfba122d04118a20" Jan 27 17:01:09 crc kubenswrapper[4772]: I0127 17:01:09.322482 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29492221-sbxjp" Jan 27 17:01:10 crc kubenswrapper[4772]: I0127 17:01:10.554725 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-txfct" Jan 27 17:01:10 crc kubenswrapper[4772]: I0127 17:01:10.619293 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-txfct"] Jan 27 17:01:10 crc kubenswrapper[4772]: I0127 17:01:10.661685 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4vbvz"] Jan 27 17:01:10 crc kubenswrapper[4772]: I0127 17:01:10.661903 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4vbvz" podUID="20e4371a-8bd2-4405-bb18-861923bfd37e" containerName="registry-server" containerID="cri-o://facc4381b6ccbc800e2ee9d053a4ca1a05dc9ad0f250dcd95e8e1ea0e652318f" gracePeriod=2 Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.152290 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4vbvz" Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.304301 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20e4371a-8bd2-4405-bb18-861923bfd37e-utilities\") pod \"20e4371a-8bd2-4405-bb18-861923bfd37e\" (UID: \"20e4371a-8bd2-4405-bb18-861923bfd37e\") " Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.304442 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20e4371a-8bd2-4405-bb18-861923bfd37e-catalog-content\") pod \"20e4371a-8bd2-4405-bb18-861923bfd37e\" (UID: \"20e4371a-8bd2-4405-bb18-861923bfd37e\") " Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.304656 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csq5c\" (UniqueName: \"kubernetes.io/projected/20e4371a-8bd2-4405-bb18-861923bfd37e-kube-api-access-csq5c\") pod \"20e4371a-8bd2-4405-bb18-861923bfd37e\" (UID: \"20e4371a-8bd2-4405-bb18-861923bfd37e\") " Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.306555 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20e4371a-8bd2-4405-bb18-861923bfd37e-utilities" (OuterVolumeSpecName: "utilities") pod "20e4371a-8bd2-4405-bb18-861923bfd37e" (UID: "20e4371a-8bd2-4405-bb18-861923bfd37e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.328370 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20e4371a-8bd2-4405-bb18-861923bfd37e-kube-api-access-csq5c" (OuterVolumeSpecName: "kube-api-access-csq5c") pod "20e4371a-8bd2-4405-bb18-861923bfd37e" (UID: "20e4371a-8bd2-4405-bb18-861923bfd37e"). InnerVolumeSpecName "kube-api-access-csq5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.345268 4772 generic.go:334] "Generic (PLEG): container finished" podID="20e4371a-8bd2-4405-bb18-861923bfd37e" containerID="facc4381b6ccbc800e2ee9d053a4ca1a05dc9ad0f250dcd95e8e1ea0e652318f" exitCode=0 Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.345357 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4vbvz" Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.345433 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4vbvz" event={"ID":"20e4371a-8bd2-4405-bb18-861923bfd37e","Type":"ContainerDied","Data":"facc4381b6ccbc800e2ee9d053a4ca1a05dc9ad0f250dcd95e8e1ea0e652318f"} Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.345550 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4vbvz" event={"ID":"20e4371a-8bd2-4405-bb18-861923bfd37e","Type":"ContainerDied","Data":"f541cf85cdae422e49f9b3df32c8dd0416ecc007711177dc8c1d7d7680929e43"} Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.346141 4772 scope.go:117] "RemoveContainer" containerID="facc4381b6ccbc800e2ee9d053a4ca1a05dc9ad0f250dcd95e8e1ea0e652318f" Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.407766 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20e4371a-8bd2-4405-bb18-861923bfd37e-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.407808 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csq5c\" (UniqueName: \"kubernetes.io/projected/20e4371a-8bd2-4405-bb18-861923bfd37e-kube-api-access-csq5c\") on node \"crc\" DevicePath \"\"" Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.416054 4772 scope.go:117] "RemoveContainer" containerID="3e6786fd605112a5af975912e2a7080680ebac9d9bf41ad3486e573e6d6d2f4c" Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.421621 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20e4371a-8bd2-4405-bb18-861923bfd37e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "20e4371a-8bd2-4405-bb18-861923bfd37e" (UID: "20e4371a-8bd2-4405-bb18-861923bfd37e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.469690 4772 scope.go:117] "RemoveContainer" containerID="a913cea196ab4b53872fe2376a7200e2beda9ee644b2571a89e275314690904e" Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.503252 4772 scope.go:117] "RemoveContainer" containerID="facc4381b6ccbc800e2ee9d053a4ca1a05dc9ad0f250dcd95e8e1ea0e652318f" Jan 27 17:01:11 crc kubenswrapper[4772]: E0127 17:01:11.503777 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"facc4381b6ccbc800e2ee9d053a4ca1a05dc9ad0f250dcd95e8e1ea0e652318f\": container with ID starting with facc4381b6ccbc800e2ee9d053a4ca1a05dc9ad0f250dcd95e8e1ea0e652318f not found: ID does not exist" containerID="facc4381b6ccbc800e2ee9d053a4ca1a05dc9ad0f250dcd95e8e1ea0e652318f" Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.503820 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"facc4381b6ccbc800e2ee9d053a4ca1a05dc9ad0f250dcd95e8e1ea0e652318f"} err="failed to get container status \"facc4381b6ccbc800e2ee9d053a4ca1a05dc9ad0f250dcd95e8e1ea0e652318f\": rpc error: code = NotFound desc = could not find container \"facc4381b6ccbc800e2ee9d053a4ca1a05dc9ad0f250dcd95e8e1ea0e652318f\": container with ID starting with facc4381b6ccbc800e2ee9d053a4ca1a05dc9ad0f250dcd95e8e1ea0e652318f not found: ID does not exist" Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.503844 4772 scope.go:117] "RemoveContainer" containerID="3e6786fd605112a5af975912e2a7080680ebac9d9bf41ad3486e573e6d6d2f4c" Jan 27 17:01:11 crc kubenswrapper[4772]: E0127 17:01:11.504139 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e6786fd605112a5af975912e2a7080680ebac9d9bf41ad3486e573e6d6d2f4c\": container with ID starting with 3e6786fd605112a5af975912e2a7080680ebac9d9bf41ad3486e573e6d6d2f4c not found: ID does not exist" containerID="3e6786fd605112a5af975912e2a7080680ebac9d9bf41ad3486e573e6d6d2f4c" Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.504176 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e6786fd605112a5af975912e2a7080680ebac9d9bf41ad3486e573e6d6d2f4c"} err="failed to get container status \"3e6786fd605112a5af975912e2a7080680ebac9d9bf41ad3486e573e6d6d2f4c\": rpc error: code = NotFound desc = could not find container \"3e6786fd605112a5af975912e2a7080680ebac9d9bf41ad3486e573e6d6d2f4c\": container with ID starting with 3e6786fd605112a5af975912e2a7080680ebac9d9bf41ad3486e573e6d6d2f4c not found: ID does not exist" Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.504190 4772 scope.go:117] "RemoveContainer" containerID="a913cea196ab4b53872fe2376a7200e2beda9ee644b2571a89e275314690904e" Jan 27 17:01:11 crc kubenswrapper[4772]: E0127 17:01:11.504454 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a913cea196ab4b53872fe2376a7200e2beda9ee644b2571a89e275314690904e\": container with ID starting with a913cea196ab4b53872fe2376a7200e2beda9ee644b2571a89e275314690904e not found: ID does not exist" containerID="a913cea196ab4b53872fe2376a7200e2beda9ee644b2571a89e275314690904e" Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.504526 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a913cea196ab4b53872fe2376a7200e2beda9ee644b2571a89e275314690904e"} err="failed to get container status \"a913cea196ab4b53872fe2376a7200e2beda9ee644b2571a89e275314690904e\": rpc error: code = NotFound desc = could not find container \"a913cea196ab4b53872fe2376a7200e2beda9ee644b2571a89e275314690904e\": container with ID starting with a913cea196ab4b53872fe2376a7200e2beda9ee644b2571a89e275314690904e not found: ID does not exist" Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.509780 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20e4371a-8bd2-4405-bb18-861923bfd37e-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.678480 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4vbvz"] Jan 27 17:01:11 crc kubenswrapper[4772]: I0127 17:01:11.688879 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4vbvz"] Jan 27 17:01:12 crc kubenswrapper[4772]: I0127 17:01:12.058272 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:01:12 crc kubenswrapper[4772]: I0127 17:01:12.058341 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:01:12 crc kubenswrapper[4772]: I0127 17:01:12.687441 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20e4371a-8bd2-4405-bb18-861923bfd37e" path="/var/lib/kubelet/pods/20e4371a-8bd2-4405-bb18-861923bfd37e/volumes" Jan 27 17:01:32 crc kubenswrapper[4772]: I0127 17:01:32.747131 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wt2ph"] Jan 27 17:01:32 crc kubenswrapper[4772]: E0127 17:01:32.751012 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20e4371a-8bd2-4405-bb18-861923bfd37e" containerName="extract-utilities" Jan 27 17:01:32 crc kubenswrapper[4772]: I0127 17:01:32.751204 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="20e4371a-8bd2-4405-bb18-861923bfd37e" containerName="extract-utilities" Jan 27 17:01:32 crc kubenswrapper[4772]: E0127 17:01:32.751331 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20e4371a-8bd2-4405-bb18-861923bfd37e" containerName="registry-server" Jan 27 17:01:32 crc kubenswrapper[4772]: I0127 17:01:32.751419 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="20e4371a-8bd2-4405-bb18-861923bfd37e" containerName="registry-server" Jan 27 17:01:32 crc kubenswrapper[4772]: E0127 17:01:32.751529 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20e4371a-8bd2-4405-bb18-861923bfd37e" containerName="extract-content" Jan 27 17:01:32 crc kubenswrapper[4772]: I0127 17:01:32.751704 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="20e4371a-8bd2-4405-bb18-861923bfd37e" containerName="extract-content" Jan 27 17:01:32 crc kubenswrapper[4772]: E0127 17:01:32.751881 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8459d055-62d3-4699-b477-ea15946b982c" containerName="keystone-cron" Jan 27 17:01:32 crc kubenswrapper[4772]: I0127 17:01:32.751989 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="8459d055-62d3-4699-b477-ea15946b982c" containerName="keystone-cron" Jan 27 17:01:32 crc kubenswrapper[4772]: I0127 17:01:32.752439 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="20e4371a-8bd2-4405-bb18-861923bfd37e" containerName="registry-server" Jan 27 17:01:32 crc kubenswrapper[4772]: I0127 17:01:32.752788 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="8459d055-62d3-4699-b477-ea15946b982c" containerName="keystone-cron" Jan 27 17:01:32 crc kubenswrapper[4772]: I0127 17:01:32.754941 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wt2ph" Jan 27 17:01:32 crc kubenswrapper[4772]: I0127 17:01:32.778327 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wt2ph"] Jan 27 17:01:32 crc kubenswrapper[4772]: I0127 17:01:32.905451 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b18dbed-da5b-4cc5-adff-4ea91a19d097-catalog-content\") pod \"redhat-marketplace-wt2ph\" (UID: \"5b18dbed-da5b-4cc5-adff-4ea91a19d097\") " pod="openshift-marketplace/redhat-marketplace-wt2ph" Jan 27 17:01:32 crc kubenswrapper[4772]: I0127 17:01:32.905818 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b18dbed-da5b-4cc5-adff-4ea91a19d097-utilities\") pod \"redhat-marketplace-wt2ph\" (UID: \"5b18dbed-da5b-4cc5-adff-4ea91a19d097\") " pod="openshift-marketplace/redhat-marketplace-wt2ph" Jan 27 17:01:32 crc kubenswrapper[4772]: I0127 17:01:32.905881 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgn7d\" (UniqueName: \"kubernetes.io/projected/5b18dbed-da5b-4cc5-adff-4ea91a19d097-kube-api-access-mgn7d\") pod \"redhat-marketplace-wt2ph\" (UID: \"5b18dbed-da5b-4cc5-adff-4ea91a19d097\") " pod="openshift-marketplace/redhat-marketplace-wt2ph" Jan 27 17:01:33 crc kubenswrapper[4772]: I0127 17:01:33.007262 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b18dbed-da5b-4cc5-adff-4ea91a19d097-catalog-content\") pod \"redhat-marketplace-wt2ph\" (UID: \"5b18dbed-da5b-4cc5-adff-4ea91a19d097\") " pod="openshift-marketplace/redhat-marketplace-wt2ph" Jan 27 17:01:33 crc kubenswrapper[4772]: I0127 17:01:33.007447 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b18dbed-da5b-4cc5-adff-4ea91a19d097-utilities\") pod \"redhat-marketplace-wt2ph\" (UID: \"5b18dbed-da5b-4cc5-adff-4ea91a19d097\") " pod="openshift-marketplace/redhat-marketplace-wt2ph" Jan 27 17:01:33 crc kubenswrapper[4772]: I0127 17:01:33.007474 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgn7d\" (UniqueName: \"kubernetes.io/projected/5b18dbed-da5b-4cc5-adff-4ea91a19d097-kube-api-access-mgn7d\") pod \"redhat-marketplace-wt2ph\" (UID: \"5b18dbed-da5b-4cc5-adff-4ea91a19d097\") " pod="openshift-marketplace/redhat-marketplace-wt2ph" Jan 27 17:01:33 crc kubenswrapper[4772]: I0127 17:01:33.007807 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b18dbed-da5b-4cc5-adff-4ea91a19d097-catalog-content\") pod \"redhat-marketplace-wt2ph\" (UID: \"5b18dbed-da5b-4cc5-adff-4ea91a19d097\") " pod="openshift-marketplace/redhat-marketplace-wt2ph" Jan 27 17:01:33 crc kubenswrapper[4772]: I0127 17:01:33.007900 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b18dbed-da5b-4cc5-adff-4ea91a19d097-utilities\") pod \"redhat-marketplace-wt2ph\" (UID: \"5b18dbed-da5b-4cc5-adff-4ea91a19d097\") " pod="openshift-marketplace/redhat-marketplace-wt2ph" Jan 27 17:01:33 crc kubenswrapper[4772]: I0127 17:01:33.027181 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgn7d\" (UniqueName: \"kubernetes.io/projected/5b18dbed-da5b-4cc5-adff-4ea91a19d097-kube-api-access-mgn7d\") pod \"redhat-marketplace-wt2ph\" (UID: \"5b18dbed-da5b-4cc5-adff-4ea91a19d097\") " pod="openshift-marketplace/redhat-marketplace-wt2ph" Jan 27 17:01:33 crc kubenswrapper[4772]: I0127 17:01:33.089876 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wt2ph" Jan 27 17:01:33 crc kubenswrapper[4772]: I0127 17:01:33.605290 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wt2ph"] Jan 27 17:01:34 crc kubenswrapper[4772]: I0127 17:01:34.604108 4772 generic.go:334] "Generic (PLEG): container finished" podID="5b18dbed-da5b-4cc5-adff-4ea91a19d097" containerID="2bada1bcab38f9662557d5435a8e993461513d6d69ba69f3f231aad6221846b4" exitCode=0 Jan 27 17:01:34 crc kubenswrapper[4772]: I0127 17:01:34.604157 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wt2ph" event={"ID":"5b18dbed-da5b-4cc5-adff-4ea91a19d097","Type":"ContainerDied","Data":"2bada1bcab38f9662557d5435a8e993461513d6d69ba69f3f231aad6221846b4"} Jan 27 17:01:34 crc kubenswrapper[4772]: I0127 17:01:34.604486 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wt2ph" event={"ID":"5b18dbed-da5b-4cc5-adff-4ea91a19d097","Type":"ContainerStarted","Data":"acdeb111e40eb59f1abc80768c71ed2578e0a0139b01996558c0b753b7b98050"} Jan 27 17:01:36 crc kubenswrapper[4772]: I0127 17:01:36.630401 4772 generic.go:334] "Generic (PLEG): container finished" podID="5b18dbed-da5b-4cc5-adff-4ea91a19d097" containerID="c7710f16831fa2297e6c94f73e15418fba26c2b35e5006130588a2939a816e3b" exitCode=0 Jan 27 17:01:36 crc kubenswrapper[4772]: I0127 17:01:36.630547 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wt2ph" event={"ID":"5b18dbed-da5b-4cc5-adff-4ea91a19d097","Type":"ContainerDied","Data":"c7710f16831fa2297e6c94f73e15418fba26c2b35e5006130588a2939a816e3b"} Jan 27 17:01:37 crc kubenswrapper[4772]: I0127 17:01:37.645628 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wt2ph" event={"ID":"5b18dbed-da5b-4cc5-adff-4ea91a19d097","Type":"ContainerStarted","Data":"e79dd4d96df89caf0c641fb436c853eeaebb87cdbc1ffe88dfe6a8b470494139"} Jan 27 17:01:37 crc kubenswrapper[4772]: I0127 17:01:37.677979 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wt2ph" podStartSLOduration=2.981748966 podStartE2EDuration="5.67794794s" podCreationTimestamp="2026-01-27 17:01:32 +0000 UTC" firstStartedPulling="2026-01-27 17:01:34.607997285 +0000 UTC m=+6880.588606383" lastFinishedPulling="2026-01-27 17:01:37.304196219 +0000 UTC m=+6883.284805357" observedRunningTime="2026-01-27 17:01:37.671849105 +0000 UTC m=+6883.652458243" watchObservedRunningTime="2026-01-27 17:01:37.67794794 +0000 UTC m=+6883.658557078" Jan 27 17:01:42 crc kubenswrapper[4772]: I0127 17:01:42.058800 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:01:42 crc kubenswrapper[4772]: I0127 17:01:42.059478 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:01:42 crc kubenswrapper[4772]: I0127 17:01:42.059546 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 17:01:42 crc kubenswrapper[4772]: I0127 17:01:42.060419 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0c934f11ffcfb51be7cc650e76f8b239868b5820a22f2783555e83c31ae7ef8b"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 17:01:42 crc kubenswrapper[4772]: I0127 17:01:42.060498 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://0c934f11ffcfb51be7cc650e76f8b239868b5820a22f2783555e83c31ae7ef8b" gracePeriod=600 Jan 27 17:01:42 crc kubenswrapper[4772]: I0127 17:01:42.705139 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="0c934f11ffcfb51be7cc650e76f8b239868b5820a22f2783555e83c31ae7ef8b" exitCode=0 Jan 27 17:01:42 crc kubenswrapper[4772]: I0127 17:01:42.705237 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"0c934f11ffcfb51be7cc650e76f8b239868b5820a22f2783555e83c31ae7ef8b"} Jan 27 17:01:42 crc kubenswrapper[4772]: I0127 17:01:42.705467 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7"} Jan 27 17:01:42 crc kubenswrapper[4772]: I0127 17:01:42.705490 4772 scope.go:117] "RemoveContainer" containerID="20d6f541f5f3fc25f3b782ee3a329f52bc226dbfbe21a2a0bc1c99608d7d2317" Jan 27 17:01:43 crc kubenswrapper[4772]: I0127 17:01:43.090342 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wt2ph" Jan 27 17:01:43 crc kubenswrapper[4772]: I0127 17:01:43.090408 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wt2ph" Jan 27 17:01:43 crc kubenswrapper[4772]: I0127 17:01:43.150535 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wt2ph" Jan 27 17:01:43 crc kubenswrapper[4772]: I0127 17:01:43.796958 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wt2ph" Jan 27 17:01:43 crc kubenswrapper[4772]: I0127 17:01:43.848863 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wt2ph"] Jan 27 17:01:45 crc kubenswrapper[4772]: I0127 17:01:45.761752 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wt2ph" podUID="5b18dbed-da5b-4cc5-adff-4ea91a19d097" containerName="registry-server" containerID="cri-o://e79dd4d96df89caf0c641fb436c853eeaebb87cdbc1ffe88dfe6a8b470494139" gracePeriod=2 Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.275128 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wt2ph" Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.455668 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgn7d\" (UniqueName: \"kubernetes.io/projected/5b18dbed-da5b-4cc5-adff-4ea91a19d097-kube-api-access-mgn7d\") pod \"5b18dbed-da5b-4cc5-adff-4ea91a19d097\" (UID: \"5b18dbed-da5b-4cc5-adff-4ea91a19d097\") " Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.456227 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b18dbed-da5b-4cc5-adff-4ea91a19d097-catalog-content\") pod \"5b18dbed-da5b-4cc5-adff-4ea91a19d097\" (UID: \"5b18dbed-da5b-4cc5-adff-4ea91a19d097\") " Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.456511 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b18dbed-da5b-4cc5-adff-4ea91a19d097-utilities\") pod \"5b18dbed-da5b-4cc5-adff-4ea91a19d097\" (UID: \"5b18dbed-da5b-4cc5-adff-4ea91a19d097\") " Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.457891 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b18dbed-da5b-4cc5-adff-4ea91a19d097-utilities" (OuterVolumeSpecName: "utilities") pod "5b18dbed-da5b-4cc5-adff-4ea91a19d097" (UID: "5b18dbed-da5b-4cc5-adff-4ea91a19d097"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.464924 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b18dbed-da5b-4cc5-adff-4ea91a19d097-kube-api-access-mgn7d" (OuterVolumeSpecName: "kube-api-access-mgn7d") pod "5b18dbed-da5b-4cc5-adff-4ea91a19d097" (UID: "5b18dbed-da5b-4cc5-adff-4ea91a19d097"). InnerVolumeSpecName "kube-api-access-mgn7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.487888 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b18dbed-da5b-4cc5-adff-4ea91a19d097-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b18dbed-da5b-4cc5-adff-4ea91a19d097" (UID: "5b18dbed-da5b-4cc5-adff-4ea91a19d097"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.560479 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b18dbed-da5b-4cc5-adff-4ea91a19d097-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.560537 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgn7d\" (UniqueName: \"kubernetes.io/projected/5b18dbed-da5b-4cc5-adff-4ea91a19d097-kube-api-access-mgn7d\") on node \"crc\" DevicePath \"\"" Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.560559 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b18dbed-da5b-4cc5-adff-4ea91a19d097-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.778591 4772 generic.go:334] "Generic (PLEG): container finished" podID="5b18dbed-da5b-4cc5-adff-4ea91a19d097" containerID="e79dd4d96df89caf0c641fb436c853eeaebb87cdbc1ffe88dfe6a8b470494139" exitCode=0 Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.778662 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wt2ph" event={"ID":"5b18dbed-da5b-4cc5-adff-4ea91a19d097","Type":"ContainerDied","Data":"e79dd4d96df89caf0c641fb436c853eeaebb87cdbc1ffe88dfe6a8b470494139"} Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.778707 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wt2ph" event={"ID":"5b18dbed-da5b-4cc5-adff-4ea91a19d097","Type":"ContainerDied","Data":"acdeb111e40eb59f1abc80768c71ed2578e0a0139b01996558c0b753b7b98050"} Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.778744 4772 scope.go:117] "RemoveContainer" containerID="e79dd4d96df89caf0c641fb436c853eeaebb87cdbc1ffe88dfe6a8b470494139" Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.778979 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wt2ph" Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.822607 4772 scope.go:117] "RemoveContainer" containerID="c7710f16831fa2297e6c94f73e15418fba26c2b35e5006130588a2939a816e3b" Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.832962 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wt2ph"] Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.842418 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wt2ph"] Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.856310 4772 scope.go:117] "RemoveContainer" containerID="2bada1bcab38f9662557d5435a8e993461513d6d69ba69f3f231aad6221846b4" Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.894303 4772 scope.go:117] "RemoveContainer" containerID="e79dd4d96df89caf0c641fb436c853eeaebb87cdbc1ffe88dfe6a8b470494139" Jan 27 17:01:46 crc kubenswrapper[4772]: E0127 17:01:46.894782 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e79dd4d96df89caf0c641fb436c853eeaebb87cdbc1ffe88dfe6a8b470494139\": container with ID starting with e79dd4d96df89caf0c641fb436c853eeaebb87cdbc1ffe88dfe6a8b470494139 not found: ID does not exist" containerID="e79dd4d96df89caf0c641fb436c853eeaebb87cdbc1ffe88dfe6a8b470494139" Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.894903 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e79dd4d96df89caf0c641fb436c853eeaebb87cdbc1ffe88dfe6a8b470494139"} err="failed to get container status \"e79dd4d96df89caf0c641fb436c853eeaebb87cdbc1ffe88dfe6a8b470494139\": rpc error: code = NotFound desc = could not find container \"e79dd4d96df89caf0c641fb436c853eeaebb87cdbc1ffe88dfe6a8b470494139\": container with ID starting with e79dd4d96df89caf0c641fb436c853eeaebb87cdbc1ffe88dfe6a8b470494139 not found: ID does not exist" Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.894938 4772 scope.go:117] "RemoveContainer" containerID="c7710f16831fa2297e6c94f73e15418fba26c2b35e5006130588a2939a816e3b" Jan 27 17:01:46 crc kubenswrapper[4772]: E0127 17:01:46.895321 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7710f16831fa2297e6c94f73e15418fba26c2b35e5006130588a2939a816e3b\": container with ID starting with c7710f16831fa2297e6c94f73e15418fba26c2b35e5006130588a2939a816e3b not found: ID does not exist" containerID="c7710f16831fa2297e6c94f73e15418fba26c2b35e5006130588a2939a816e3b" Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.895356 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7710f16831fa2297e6c94f73e15418fba26c2b35e5006130588a2939a816e3b"} err="failed to get container status \"c7710f16831fa2297e6c94f73e15418fba26c2b35e5006130588a2939a816e3b\": rpc error: code = NotFound desc = could not find container \"c7710f16831fa2297e6c94f73e15418fba26c2b35e5006130588a2939a816e3b\": container with ID starting with c7710f16831fa2297e6c94f73e15418fba26c2b35e5006130588a2939a816e3b not found: ID does not exist" Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.895374 4772 scope.go:117] "RemoveContainer" containerID="2bada1bcab38f9662557d5435a8e993461513d6d69ba69f3f231aad6221846b4" Jan 27 17:01:46 crc kubenswrapper[4772]: E0127 17:01:46.895654 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bada1bcab38f9662557d5435a8e993461513d6d69ba69f3f231aad6221846b4\": container with ID starting with 2bada1bcab38f9662557d5435a8e993461513d6d69ba69f3f231aad6221846b4 not found: ID does not exist" containerID="2bada1bcab38f9662557d5435a8e993461513d6d69ba69f3f231aad6221846b4" Jan 27 17:01:46 crc kubenswrapper[4772]: I0127 17:01:46.895682 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bada1bcab38f9662557d5435a8e993461513d6d69ba69f3f231aad6221846b4"} err="failed to get container status \"2bada1bcab38f9662557d5435a8e993461513d6d69ba69f3f231aad6221846b4\": rpc error: code = NotFound desc = could not find container \"2bada1bcab38f9662557d5435a8e993461513d6d69ba69f3f231aad6221846b4\": container with ID starting with 2bada1bcab38f9662557d5435a8e993461513d6d69ba69f3f231aad6221846b4 not found: ID does not exist" Jan 27 17:01:48 crc kubenswrapper[4772]: I0127 17:01:48.676402 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b18dbed-da5b-4cc5-adff-4ea91a19d097" path="/var/lib/kubelet/pods/5b18dbed-da5b-4cc5-adff-4ea91a19d097/volumes" Jan 27 17:03:20 crc kubenswrapper[4772]: I0127 17:03:20.493275 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-b5nn7"] Jan 27 17:03:20 crc kubenswrapper[4772]: E0127 17:03:20.498973 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b18dbed-da5b-4cc5-adff-4ea91a19d097" containerName="registry-server" Jan 27 17:03:20 crc kubenswrapper[4772]: I0127 17:03:20.499286 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b18dbed-da5b-4cc5-adff-4ea91a19d097" containerName="registry-server" Jan 27 17:03:20 crc kubenswrapper[4772]: E0127 17:03:20.499324 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b18dbed-da5b-4cc5-adff-4ea91a19d097" containerName="extract-utilities" Jan 27 17:03:20 crc kubenswrapper[4772]: I0127 17:03:20.499333 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b18dbed-da5b-4cc5-adff-4ea91a19d097" containerName="extract-utilities" Jan 27 17:03:20 crc kubenswrapper[4772]: E0127 17:03:20.499360 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b18dbed-da5b-4cc5-adff-4ea91a19d097" containerName="extract-content" Jan 27 17:03:20 crc kubenswrapper[4772]: I0127 17:03:20.499369 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b18dbed-da5b-4cc5-adff-4ea91a19d097" containerName="extract-content" Jan 27 17:03:20 crc kubenswrapper[4772]: I0127 17:03:20.499618 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b18dbed-da5b-4cc5-adff-4ea91a19d097" containerName="registry-server" Jan 27 17:03:20 crc kubenswrapper[4772]: I0127 17:03:20.501271 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b5nn7" Jan 27 17:03:20 crc kubenswrapper[4772]: I0127 17:03:20.506451 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b5nn7"] Jan 27 17:03:20 crc kubenswrapper[4772]: I0127 17:03:20.581695 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq22p\" (UniqueName: \"kubernetes.io/projected/a6d8c5f6-5840-4996-8edf-b0cdd3ab833a-kube-api-access-nq22p\") pod \"certified-operators-b5nn7\" (UID: \"a6d8c5f6-5840-4996-8edf-b0cdd3ab833a\") " pod="openshift-marketplace/certified-operators-b5nn7" Jan 27 17:03:20 crc kubenswrapper[4772]: I0127 17:03:20.582104 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6d8c5f6-5840-4996-8edf-b0cdd3ab833a-utilities\") pod \"certified-operators-b5nn7\" (UID: \"a6d8c5f6-5840-4996-8edf-b0cdd3ab833a\") " pod="openshift-marketplace/certified-operators-b5nn7" Jan 27 17:03:20 crc kubenswrapper[4772]: I0127 17:03:20.582453 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6d8c5f6-5840-4996-8edf-b0cdd3ab833a-catalog-content\") pod \"certified-operators-b5nn7\" (UID: \"a6d8c5f6-5840-4996-8edf-b0cdd3ab833a\") " pod="openshift-marketplace/certified-operators-b5nn7" Jan 27 17:03:20 crc kubenswrapper[4772]: I0127 17:03:20.684843 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq22p\" (UniqueName: \"kubernetes.io/projected/a6d8c5f6-5840-4996-8edf-b0cdd3ab833a-kube-api-access-nq22p\") pod \"certified-operators-b5nn7\" (UID: \"a6d8c5f6-5840-4996-8edf-b0cdd3ab833a\") " pod="openshift-marketplace/certified-operators-b5nn7" Jan 27 17:03:20 crc kubenswrapper[4772]: I0127 17:03:20.684924 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6d8c5f6-5840-4996-8edf-b0cdd3ab833a-utilities\") pod \"certified-operators-b5nn7\" (UID: \"a6d8c5f6-5840-4996-8edf-b0cdd3ab833a\") " pod="openshift-marketplace/certified-operators-b5nn7" Jan 27 17:03:20 crc kubenswrapper[4772]: I0127 17:03:20.684960 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6d8c5f6-5840-4996-8edf-b0cdd3ab833a-catalog-content\") pod \"certified-operators-b5nn7\" (UID: \"a6d8c5f6-5840-4996-8edf-b0cdd3ab833a\") " pod="openshift-marketplace/certified-operators-b5nn7" Jan 27 17:03:20 crc kubenswrapper[4772]: I0127 17:03:20.685446 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6d8c5f6-5840-4996-8edf-b0cdd3ab833a-catalog-content\") pod \"certified-operators-b5nn7\" (UID: \"a6d8c5f6-5840-4996-8edf-b0cdd3ab833a\") " pod="openshift-marketplace/certified-operators-b5nn7" Jan 27 17:03:20 crc kubenswrapper[4772]: I0127 17:03:20.685459 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6d8c5f6-5840-4996-8edf-b0cdd3ab833a-utilities\") pod \"certified-operators-b5nn7\" (UID: \"a6d8c5f6-5840-4996-8edf-b0cdd3ab833a\") " pod="openshift-marketplace/certified-operators-b5nn7" Jan 27 17:03:20 crc kubenswrapper[4772]: I0127 17:03:20.703566 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq22p\" (UniqueName: \"kubernetes.io/projected/a6d8c5f6-5840-4996-8edf-b0cdd3ab833a-kube-api-access-nq22p\") pod \"certified-operators-b5nn7\" (UID: \"a6d8c5f6-5840-4996-8edf-b0cdd3ab833a\") " pod="openshift-marketplace/certified-operators-b5nn7" Jan 27 17:03:20 crc kubenswrapper[4772]: I0127 17:03:20.875870 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b5nn7" Jan 27 17:03:21 crc kubenswrapper[4772]: I0127 17:03:21.452034 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b5nn7"] Jan 27 17:03:21 crc kubenswrapper[4772]: I0127 17:03:21.851248 4772 generic.go:334] "Generic (PLEG): container finished" podID="a6d8c5f6-5840-4996-8edf-b0cdd3ab833a" containerID="fc3b4387c7de25d81bfd6644e1b09a028e3623cf29b8e426464b8e6357b85402" exitCode=0 Jan 27 17:03:21 crc kubenswrapper[4772]: I0127 17:03:21.851320 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b5nn7" event={"ID":"a6d8c5f6-5840-4996-8edf-b0cdd3ab833a","Type":"ContainerDied","Data":"fc3b4387c7de25d81bfd6644e1b09a028e3623cf29b8e426464b8e6357b85402"} Jan 27 17:03:21 crc kubenswrapper[4772]: I0127 17:03:21.851358 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b5nn7" event={"ID":"a6d8c5f6-5840-4996-8edf-b0cdd3ab833a","Type":"ContainerStarted","Data":"868827379c2358104fc3ee3fd215040de500d1cf0eddf9c817b9a3388c61a59b"} Jan 27 17:03:21 crc kubenswrapper[4772]: I0127 17:03:21.855092 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 17:03:23 crc kubenswrapper[4772]: I0127 17:03:23.878198 4772 generic.go:334] "Generic (PLEG): container finished" podID="a6d8c5f6-5840-4996-8edf-b0cdd3ab833a" containerID="b4ebb1482a887c740701c8024c6fde60a92ddf64448d43fc59fdc8d0125b04af" exitCode=0 Jan 27 17:03:23 crc kubenswrapper[4772]: I0127 17:03:23.878316 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b5nn7" event={"ID":"a6d8c5f6-5840-4996-8edf-b0cdd3ab833a","Type":"ContainerDied","Data":"b4ebb1482a887c740701c8024c6fde60a92ddf64448d43fc59fdc8d0125b04af"} Jan 27 17:03:24 crc kubenswrapper[4772]: I0127 17:03:24.895680 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b5nn7" event={"ID":"a6d8c5f6-5840-4996-8edf-b0cdd3ab833a","Type":"ContainerStarted","Data":"446397c85d820ac08eaaff34c1f773386f0e789b43f7fcda38339ade6491fa4e"} Jan 27 17:03:24 crc kubenswrapper[4772]: I0127 17:03:24.919124 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-b5nn7" podStartSLOduration=2.4662688790000002 podStartE2EDuration="4.919104964s" podCreationTimestamp="2026-01-27 17:03:20 +0000 UTC" firstStartedPulling="2026-01-27 17:03:21.854648298 +0000 UTC m=+6987.835257416" lastFinishedPulling="2026-01-27 17:03:24.307484383 +0000 UTC m=+6990.288093501" observedRunningTime="2026-01-27 17:03:24.917134818 +0000 UTC m=+6990.897743916" watchObservedRunningTime="2026-01-27 17:03:24.919104964 +0000 UTC m=+6990.899714062" Jan 27 17:03:30 crc kubenswrapper[4772]: I0127 17:03:30.876699 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-b5nn7" Jan 27 17:03:30 crc kubenswrapper[4772]: I0127 17:03:30.878357 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-b5nn7" Jan 27 17:03:30 crc kubenswrapper[4772]: I0127 17:03:30.956823 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-b5nn7" Jan 27 17:03:31 crc kubenswrapper[4772]: I0127 17:03:31.009948 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-b5nn7" Jan 27 17:03:31 crc kubenswrapper[4772]: I0127 17:03:31.206259 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b5nn7"] Jan 27 17:03:32 crc kubenswrapper[4772]: I0127 17:03:32.990423 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-b5nn7" podUID="a6d8c5f6-5840-4996-8edf-b0cdd3ab833a" containerName="registry-server" containerID="cri-o://446397c85d820ac08eaaff34c1f773386f0e789b43f7fcda38339ade6491fa4e" gracePeriod=2 Jan 27 17:03:33 crc kubenswrapper[4772]: I0127 17:03:33.550098 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b5nn7" Jan 27 17:03:33 crc kubenswrapper[4772]: I0127 17:03:33.568485 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6d8c5f6-5840-4996-8edf-b0cdd3ab833a-catalog-content\") pod \"a6d8c5f6-5840-4996-8edf-b0cdd3ab833a\" (UID: \"a6d8c5f6-5840-4996-8edf-b0cdd3ab833a\") " Jan 27 17:03:33 crc kubenswrapper[4772]: I0127 17:03:33.569488 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6d8c5f6-5840-4996-8edf-b0cdd3ab833a-utilities\") pod \"a6d8c5f6-5840-4996-8edf-b0cdd3ab833a\" (UID: \"a6d8c5f6-5840-4996-8edf-b0cdd3ab833a\") " Jan 27 17:03:33 crc kubenswrapper[4772]: I0127 17:03:33.570255 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nq22p\" (UniqueName: \"kubernetes.io/projected/a6d8c5f6-5840-4996-8edf-b0cdd3ab833a-kube-api-access-nq22p\") pod \"a6d8c5f6-5840-4996-8edf-b0cdd3ab833a\" (UID: \"a6d8c5f6-5840-4996-8edf-b0cdd3ab833a\") " Jan 27 17:03:33 crc kubenswrapper[4772]: I0127 17:03:33.570710 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6d8c5f6-5840-4996-8edf-b0cdd3ab833a-utilities" (OuterVolumeSpecName: "utilities") pod "a6d8c5f6-5840-4996-8edf-b0cdd3ab833a" (UID: "a6d8c5f6-5840-4996-8edf-b0cdd3ab833a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:03:33 crc kubenswrapper[4772]: I0127 17:03:33.576605 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6d8c5f6-5840-4996-8edf-b0cdd3ab833a-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 17:03:33 crc kubenswrapper[4772]: I0127 17:03:33.586508 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6d8c5f6-5840-4996-8edf-b0cdd3ab833a-kube-api-access-nq22p" (OuterVolumeSpecName: "kube-api-access-nq22p") pod "a6d8c5f6-5840-4996-8edf-b0cdd3ab833a" (UID: "a6d8c5f6-5840-4996-8edf-b0cdd3ab833a"). InnerVolumeSpecName "kube-api-access-nq22p". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:03:33 crc kubenswrapper[4772]: I0127 17:03:33.651385 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6d8c5f6-5840-4996-8edf-b0cdd3ab833a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a6d8c5f6-5840-4996-8edf-b0cdd3ab833a" (UID: "a6d8c5f6-5840-4996-8edf-b0cdd3ab833a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:03:33 crc kubenswrapper[4772]: I0127 17:03:33.678781 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nq22p\" (UniqueName: \"kubernetes.io/projected/a6d8c5f6-5840-4996-8edf-b0cdd3ab833a-kube-api-access-nq22p\") on node \"crc\" DevicePath \"\"" Jan 27 17:03:33 crc kubenswrapper[4772]: I0127 17:03:33.678838 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6d8c5f6-5840-4996-8edf-b0cdd3ab833a-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 17:03:34 crc kubenswrapper[4772]: I0127 17:03:34.020370 4772 generic.go:334] "Generic (PLEG): container finished" podID="a6d8c5f6-5840-4996-8edf-b0cdd3ab833a" containerID="446397c85d820ac08eaaff34c1f773386f0e789b43f7fcda38339ade6491fa4e" exitCode=0 Jan 27 17:03:34 crc kubenswrapper[4772]: I0127 17:03:34.020438 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b5nn7" event={"ID":"a6d8c5f6-5840-4996-8edf-b0cdd3ab833a","Type":"ContainerDied","Data":"446397c85d820ac08eaaff34c1f773386f0e789b43f7fcda38339ade6491fa4e"} Jan 27 17:03:34 crc kubenswrapper[4772]: I0127 17:03:34.020489 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b5nn7" Jan 27 17:03:34 crc kubenswrapper[4772]: I0127 17:03:34.020511 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b5nn7" event={"ID":"a6d8c5f6-5840-4996-8edf-b0cdd3ab833a","Type":"ContainerDied","Data":"868827379c2358104fc3ee3fd215040de500d1cf0eddf9c817b9a3388c61a59b"} Jan 27 17:03:34 crc kubenswrapper[4772]: I0127 17:03:34.020558 4772 scope.go:117] "RemoveContainer" containerID="446397c85d820ac08eaaff34c1f773386f0e789b43f7fcda38339ade6491fa4e" Jan 27 17:03:34 crc kubenswrapper[4772]: I0127 17:03:34.042958 4772 scope.go:117] "RemoveContainer" containerID="b4ebb1482a887c740701c8024c6fde60a92ddf64448d43fc59fdc8d0125b04af" Jan 27 17:03:34 crc kubenswrapper[4772]: I0127 17:03:34.067603 4772 scope.go:117] "RemoveContainer" containerID="fc3b4387c7de25d81bfd6644e1b09a028e3623cf29b8e426464b8e6357b85402" Jan 27 17:03:34 crc kubenswrapper[4772]: I0127 17:03:34.071418 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b5nn7"] Jan 27 17:03:34 crc kubenswrapper[4772]: I0127 17:03:34.079245 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-b5nn7"] Jan 27 17:03:34 crc kubenswrapper[4772]: I0127 17:03:34.105446 4772 scope.go:117] "RemoveContainer" containerID="446397c85d820ac08eaaff34c1f773386f0e789b43f7fcda38339ade6491fa4e" Jan 27 17:03:34 crc kubenswrapper[4772]: E0127 17:03:34.105883 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"446397c85d820ac08eaaff34c1f773386f0e789b43f7fcda38339ade6491fa4e\": container with ID starting with 446397c85d820ac08eaaff34c1f773386f0e789b43f7fcda38339ade6491fa4e not found: ID does not exist" containerID="446397c85d820ac08eaaff34c1f773386f0e789b43f7fcda38339ade6491fa4e" Jan 27 17:03:34 crc kubenswrapper[4772]: I0127 17:03:34.105948 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"446397c85d820ac08eaaff34c1f773386f0e789b43f7fcda38339ade6491fa4e"} err="failed to get container status \"446397c85d820ac08eaaff34c1f773386f0e789b43f7fcda38339ade6491fa4e\": rpc error: code = NotFound desc = could not find container \"446397c85d820ac08eaaff34c1f773386f0e789b43f7fcda38339ade6491fa4e\": container with ID starting with 446397c85d820ac08eaaff34c1f773386f0e789b43f7fcda38339ade6491fa4e not found: ID does not exist" Jan 27 17:03:34 crc kubenswrapper[4772]: I0127 17:03:34.105997 4772 scope.go:117] "RemoveContainer" containerID="b4ebb1482a887c740701c8024c6fde60a92ddf64448d43fc59fdc8d0125b04af" Jan 27 17:03:34 crc kubenswrapper[4772]: E0127 17:03:34.106732 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4ebb1482a887c740701c8024c6fde60a92ddf64448d43fc59fdc8d0125b04af\": container with ID starting with b4ebb1482a887c740701c8024c6fde60a92ddf64448d43fc59fdc8d0125b04af not found: ID does not exist" containerID="b4ebb1482a887c740701c8024c6fde60a92ddf64448d43fc59fdc8d0125b04af" Jan 27 17:03:34 crc kubenswrapper[4772]: I0127 17:03:34.106770 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4ebb1482a887c740701c8024c6fde60a92ddf64448d43fc59fdc8d0125b04af"} err="failed to get container status \"b4ebb1482a887c740701c8024c6fde60a92ddf64448d43fc59fdc8d0125b04af\": rpc error: code = NotFound desc = could not find container \"b4ebb1482a887c740701c8024c6fde60a92ddf64448d43fc59fdc8d0125b04af\": container with ID starting with b4ebb1482a887c740701c8024c6fde60a92ddf64448d43fc59fdc8d0125b04af not found: ID does not exist" Jan 27 17:03:34 crc kubenswrapper[4772]: I0127 17:03:34.106792 4772 scope.go:117] "RemoveContainer" containerID="fc3b4387c7de25d81bfd6644e1b09a028e3623cf29b8e426464b8e6357b85402" Jan 27 17:03:34 crc kubenswrapper[4772]: E0127 17:03:34.107067 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc3b4387c7de25d81bfd6644e1b09a028e3623cf29b8e426464b8e6357b85402\": container with ID starting with fc3b4387c7de25d81bfd6644e1b09a028e3623cf29b8e426464b8e6357b85402 not found: ID does not exist" containerID="fc3b4387c7de25d81bfd6644e1b09a028e3623cf29b8e426464b8e6357b85402" Jan 27 17:03:34 crc kubenswrapper[4772]: I0127 17:03:34.107120 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc3b4387c7de25d81bfd6644e1b09a028e3623cf29b8e426464b8e6357b85402"} err="failed to get container status \"fc3b4387c7de25d81bfd6644e1b09a028e3623cf29b8e426464b8e6357b85402\": rpc error: code = NotFound desc = could not find container \"fc3b4387c7de25d81bfd6644e1b09a028e3623cf29b8e426464b8e6357b85402\": container with ID starting with fc3b4387c7de25d81bfd6644e1b09a028e3623cf29b8e426464b8e6357b85402 not found: ID does not exist" Jan 27 17:03:34 crc kubenswrapper[4772]: I0127 17:03:34.694716 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6d8c5f6-5840-4996-8edf-b0cdd3ab833a" path="/var/lib/kubelet/pods/a6d8c5f6-5840-4996-8edf-b0cdd3ab833a/volumes" Jan 27 17:03:42 crc kubenswrapper[4772]: I0127 17:03:42.058690 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:03:42 crc kubenswrapper[4772]: I0127 17:03:42.059576 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:04:12 crc kubenswrapper[4772]: I0127 17:04:12.058104 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:04:12 crc kubenswrapper[4772]: I0127 17:04:12.058779 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:04:42 crc kubenswrapper[4772]: I0127 17:04:42.058285 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:04:42 crc kubenswrapper[4772]: I0127 17:04:42.058976 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:04:42 crc kubenswrapper[4772]: I0127 17:04:42.059027 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 17:04:42 crc kubenswrapper[4772]: I0127 17:04:42.059840 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 17:04:42 crc kubenswrapper[4772]: I0127 17:04:42.059895 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" gracePeriod=600 Jan 27 17:04:42 crc kubenswrapper[4772]: E0127 17:04:42.186363 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:04:42 crc kubenswrapper[4772]: I0127 17:04:42.773742 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" exitCode=0 Jan 27 17:04:42 crc kubenswrapper[4772]: I0127 17:04:42.773810 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7"} Jan 27 17:04:42 crc kubenswrapper[4772]: I0127 17:04:42.773912 4772 scope.go:117] "RemoveContainer" containerID="0c934f11ffcfb51be7cc650e76f8b239868b5820a22f2783555e83c31ae7ef8b" Jan 27 17:04:42 crc kubenswrapper[4772]: I0127 17:04:42.774905 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:04:42 crc kubenswrapper[4772]: E0127 17:04:42.775238 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:04:54 crc kubenswrapper[4772]: I0127 17:04:54.677927 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:04:54 crc kubenswrapper[4772]: E0127 17:04:54.679218 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:05:09 crc kubenswrapper[4772]: I0127 17:05:09.674883 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:05:09 crc kubenswrapper[4772]: E0127 17:05:09.677312 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:05:21 crc kubenswrapper[4772]: I0127 17:05:21.663231 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:05:21 crc kubenswrapper[4772]: E0127 17:05:21.664287 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:05:32 crc kubenswrapper[4772]: I0127 17:05:32.667337 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:05:32 crc kubenswrapper[4772]: E0127 17:05:32.668152 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:05:46 crc kubenswrapper[4772]: I0127 17:05:46.663371 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:05:46 crc kubenswrapper[4772]: E0127 17:05:46.664062 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:06:01 crc kubenswrapper[4772]: I0127 17:06:01.663381 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:06:01 crc kubenswrapper[4772]: E0127 17:06:01.664557 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:06:12 crc kubenswrapper[4772]: I0127 17:06:12.663673 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:06:12 crc kubenswrapper[4772]: E0127 17:06:12.664583 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:06:26 crc kubenswrapper[4772]: I0127 17:06:26.663959 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:06:26 crc kubenswrapper[4772]: E0127 17:06:26.665228 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:06:38 crc kubenswrapper[4772]: I0127 17:06:38.663078 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:06:38 crc kubenswrapper[4772]: E0127 17:06:38.663864 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:06:51 crc kubenswrapper[4772]: I0127 17:06:51.662948 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:06:51 crc kubenswrapper[4772]: E0127 17:06:51.663985 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:07:03 crc kubenswrapper[4772]: I0127 17:07:03.662969 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:07:03 crc kubenswrapper[4772]: E0127 17:07:03.664190 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:07:16 crc kubenswrapper[4772]: I0127 17:07:16.663523 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:07:16 crc kubenswrapper[4772]: E0127 17:07:16.665974 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:07:29 crc kubenswrapper[4772]: I0127 17:07:29.675086 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:07:29 crc kubenswrapper[4772]: E0127 17:07:29.675848 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:07:41 crc kubenswrapper[4772]: I0127 17:07:41.663312 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:07:41 crc kubenswrapper[4772]: E0127 17:07:41.664131 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:07:55 crc kubenswrapper[4772]: I0127 17:07:55.662705 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:07:55 crc kubenswrapper[4772]: E0127 17:07:55.663575 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:08:06 crc kubenswrapper[4772]: I0127 17:08:06.663775 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:08:06 crc kubenswrapper[4772]: E0127 17:08:06.664889 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:08:17 crc kubenswrapper[4772]: I0127 17:08:17.663780 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:08:17 crc kubenswrapper[4772]: E0127 17:08:17.664861 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:08:32 crc kubenswrapper[4772]: I0127 17:08:32.664235 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:08:32 crc kubenswrapper[4772]: E0127 17:08:32.665110 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:08:39 crc kubenswrapper[4772]: I0127 17:08:39.959004 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-v67fk"] Jan 27 17:08:39 crc kubenswrapper[4772]: E0127 17:08:39.960099 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6d8c5f6-5840-4996-8edf-b0cdd3ab833a" containerName="extract-utilities" Jan 27 17:08:39 crc kubenswrapper[4772]: I0127 17:08:39.960116 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6d8c5f6-5840-4996-8edf-b0cdd3ab833a" containerName="extract-utilities" Jan 27 17:08:39 crc kubenswrapper[4772]: E0127 17:08:39.960195 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6d8c5f6-5840-4996-8edf-b0cdd3ab833a" containerName="extract-content" Jan 27 17:08:39 crc kubenswrapper[4772]: I0127 17:08:39.960204 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6d8c5f6-5840-4996-8edf-b0cdd3ab833a" containerName="extract-content" Jan 27 17:08:39 crc kubenswrapper[4772]: E0127 17:08:39.960218 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6d8c5f6-5840-4996-8edf-b0cdd3ab833a" containerName="registry-server" Jan 27 17:08:39 crc kubenswrapper[4772]: I0127 17:08:39.960227 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6d8c5f6-5840-4996-8edf-b0cdd3ab833a" containerName="registry-server" Jan 27 17:08:39 crc kubenswrapper[4772]: I0127 17:08:39.960472 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6d8c5f6-5840-4996-8edf-b0cdd3ab833a" containerName="registry-server" Jan 27 17:08:39 crc kubenswrapper[4772]: I0127 17:08:39.962299 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v67fk" Jan 27 17:08:39 crc kubenswrapper[4772]: I0127 17:08:39.986493 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v67fk"] Jan 27 17:08:39 crc kubenswrapper[4772]: I0127 17:08:39.996366 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww2ls\" (UniqueName: \"kubernetes.io/projected/4696c11b-cd7f-4a9b-84d0-b4e59ff4383a-kube-api-access-ww2ls\") pod \"redhat-operators-v67fk\" (UID: \"4696c11b-cd7f-4a9b-84d0-b4e59ff4383a\") " pod="openshift-marketplace/redhat-operators-v67fk" Jan 27 17:08:39 crc kubenswrapper[4772]: I0127 17:08:39.996762 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4696c11b-cd7f-4a9b-84d0-b4e59ff4383a-catalog-content\") pod \"redhat-operators-v67fk\" (UID: \"4696c11b-cd7f-4a9b-84d0-b4e59ff4383a\") " pod="openshift-marketplace/redhat-operators-v67fk" Jan 27 17:08:39 crc kubenswrapper[4772]: I0127 17:08:39.996993 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4696c11b-cd7f-4a9b-84d0-b4e59ff4383a-utilities\") pod \"redhat-operators-v67fk\" (UID: \"4696c11b-cd7f-4a9b-84d0-b4e59ff4383a\") " pod="openshift-marketplace/redhat-operators-v67fk" Jan 27 17:08:40 crc kubenswrapper[4772]: I0127 17:08:40.099579 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4696c11b-cd7f-4a9b-84d0-b4e59ff4383a-utilities\") pod \"redhat-operators-v67fk\" (UID: \"4696c11b-cd7f-4a9b-84d0-b4e59ff4383a\") " pod="openshift-marketplace/redhat-operators-v67fk" Jan 27 17:08:40 crc kubenswrapper[4772]: I0127 17:08:40.099740 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww2ls\" (UniqueName: \"kubernetes.io/projected/4696c11b-cd7f-4a9b-84d0-b4e59ff4383a-kube-api-access-ww2ls\") pod \"redhat-operators-v67fk\" (UID: \"4696c11b-cd7f-4a9b-84d0-b4e59ff4383a\") " pod="openshift-marketplace/redhat-operators-v67fk" Jan 27 17:08:40 crc kubenswrapper[4772]: I0127 17:08:40.099856 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4696c11b-cd7f-4a9b-84d0-b4e59ff4383a-catalog-content\") pod \"redhat-operators-v67fk\" (UID: \"4696c11b-cd7f-4a9b-84d0-b4e59ff4383a\") " pod="openshift-marketplace/redhat-operators-v67fk" Jan 27 17:08:40 crc kubenswrapper[4772]: I0127 17:08:40.100192 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4696c11b-cd7f-4a9b-84d0-b4e59ff4383a-utilities\") pod \"redhat-operators-v67fk\" (UID: \"4696c11b-cd7f-4a9b-84d0-b4e59ff4383a\") " pod="openshift-marketplace/redhat-operators-v67fk" Jan 27 17:08:40 crc kubenswrapper[4772]: I0127 17:08:40.100398 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4696c11b-cd7f-4a9b-84d0-b4e59ff4383a-catalog-content\") pod \"redhat-operators-v67fk\" (UID: \"4696c11b-cd7f-4a9b-84d0-b4e59ff4383a\") " pod="openshift-marketplace/redhat-operators-v67fk" Jan 27 17:08:40 crc kubenswrapper[4772]: I0127 17:08:40.120994 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww2ls\" (UniqueName: \"kubernetes.io/projected/4696c11b-cd7f-4a9b-84d0-b4e59ff4383a-kube-api-access-ww2ls\") pod \"redhat-operators-v67fk\" (UID: \"4696c11b-cd7f-4a9b-84d0-b4e59ff4383a\") " pod="openshift-marketplace/redhat-operators-v67fk" Jan 27 17:08:40 crc kubenswrapper[4772]: I0127 17:08:40.288345 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v67fk" Jan 27 17:08:40 crc kubenswrapper[4772]: I0127 17:08:40.771508 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v67fk"] Jan 27 17:08:41 crc kubenswrapper[4772]: I0127 17:08:41.327329 4772 generic.go:334] "Generic (PLEG): container finished" podID="4696c11b-cd7f-4a9b-84d0-b4e59ff4383a" containerID="12499ba818facb5728a5e2af322ab50a280b5ba47cb8956bb6a0e6872c81b057" exitCode=0 Jan 27 17:08:41 crc kubenswrapper[4772]: I0127 17:08:41.327551 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v67fk" event={"ID":"4696c11b-cd7f-4a9b-84d0-b4e59ff4383a","Type":"ContainerDied","Data":"12499ba818facb5728a5e2af322ab50a280b5ba47cb8956bb6a0e6872c81b057"} Jan 27 17:08:41 crc kubenswrapper[4772]: I0127 17:08:41.327694 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v67fk" event={"ID":"4696c11b-cd7f-4a9b-84d0-b4e59ff4383a","Type":"ContainerStarted","Data":"d4542b704bdfa74b42beaef5f64efafc7ed30121010c3bdd793a4aaa21d8cfa5"} Jan 27 17:08:41 crc kubenswrapper[4772]: I0127 17:08:41.329922 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 17:08:42 crc kubenswrapper[4772]: I0127 17:08:42.339529 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v67fk" event={"ID":"4696c11b-cd7f-4a9b-84d0-b4e59ff4383a","Type":"ContainerStarted","Data":"06a1f585dd0be4dfae7eee556971dc6e4e153c2aa01f033489b777fe2060b96c"} Jan 27 17:08:43 crc kubenswrapper[4772]: I0127 17:08:43.360050 4772 generic.go:334] "Generic (PLEG): container finished" podID="4696c11b-cd7f-4a9b-84d0-b4e59ff4383a" containerID="06a1f585dd0be4dfae7eee556971dc6e4e153c2aa01f033489b777fe2060b96c" exitCode=0 Jan 27 17:08:43 crc kubenswrapper[4772]: I0127 17:08:43.360126 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v67fk" event={"ID":"4696c11b-cd7f-4a9b-84d0-b4e59ff4383a","Type":"ContainerDied","Data":"06a1f585dd0be4dfae7eee556971dc6e4e153c2aa01f033489b777fe2060b96c"} Jan 27 17:08:43 crc kubenswrapper[4772]: I0127 17:08:43.664025 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:08:43 crc kubenswrapper[4772]: E0127 17:08:43.664928 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:08:45 crc kubenswrapper[4772]: I0127 17:08:45.406388 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v67fk" event={"ID":"4696c11b-cd7f-4a9b-84d0-b4e59ff4383a","Type":"ContainerStarted","Data":"670c3dd5d1eadcc4c7af0be85bbbd52745f9845379d48bccf45b532e278f530f"} Jan 27 17:08:45 crc kubenswrapper[4772]: I0127 17:08:45.434622 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-v67fk" podStartSLOduration=3.1334585600000002 podStartE2EDuration="6.434605184s" podCreationTimestamp="2026-01-27 17:08:39 +0000 UTC" firstStartedPulling="2026-01-27 17:08:41.329638392 +0000 UTC m=+7307.310247480" lastFinishedPulling="2026-01-27 17:08:44.630784996 +0000 UTC m=+7310.611394104" observedRunningTime="2026-01-27 17:08:45.433025579 +0000 UTC m=+7311.413634687" watchObservedRunningTime="2026-01-27 17:08:45.434605184 +0000 UTC m=+7311.415214282" Jan 27 17:08:50 crc kubenswrapper[4772]: I0127 17:08:50.289557 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-v67fk" Jan 27 17:08:50 crc kubenswrapper[4772]: I0127 17:08:50.291004 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-v67fk" Jan 27 17:08:51 crc kubenswrapper[4772]: I0127 17:08:51.347092 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v67fk" podUID="4696c11b-cd7f-4a9b-84d0-b4e59ff4383a" containerName="registry-server" probeResult="failure" output=< Jan 27 17:08:51 crc kubenswrapper[4772]: timeout: failed to connect service ":50051" within 1s Jan 27 17:08:51 crc kubenswrapper[4772]: > Jan 27 17:08:54 crc kubenswrapper[4772]: I0127 17:08:54.676563 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:08:54 crc kubenswrapper[4772]: E0127 17:08:54.677478 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:09:00 crc kubenswrapper[4772]: I0127 17:09:00.358803 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-v67fk" Jan 27 17:09:00 crc kubenswrapper[4772]: I0127 17:09:00.422390 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-v67fk" Jan 27 17:09:00 crc kubenswrapper[4772]: I0127 17:09:00.606217 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v67fk"] Jan 27 17:09:01 crc kubenswrapper[4772]: I0127 17:09:01.568536 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-v67fk" podUID="4696c11b-cd7f-4a9b-84d0-b4e59ff4383a" containerName="registry-server" containerID="cri-o://670c3dd5d1eadcc4c7af0be85bbbd52745f9845379d48bccf45b532e278f530f" gracePeriod=2 Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.129570 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v67fk" Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.274651 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ww2ls\" (UniqueName: \"kubernetes.io/projected/4696c11b-cd7f-4a9b-84d0-b4e59ff4383a-kube-api-access-ww2ls\") pod \"4696c11b-cd7f-4a9b-84d0-b4e59ff4383a\" (UID: \"4696c11b-cd7f-4a9b-84d0-b4e59ff4383a\") " Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.274782 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4696c11b-cd7f-4a9b-84d0-b4e59ff4383a-catalog-content\") pod \"4696c11b-cd7f-4a9b-84d0-b4e59ff4383a\" (UID: \"4696c11b-cd7f-4a9b-84d0-b4e59ff4383a\") " Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.275023 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4696c11b-cd7f-4a9b-84d0-b4e59ff4383a-utilities\") pod \"4696c11b-cd7f-4a9b-84d0-b4e59ff4383a\" (UID: \"4696c11b-cd7f-4a9b-84d0-b4e59ff4383a\") " Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.275743 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4696c11b-cd7f-4a9b-84d0-b4e59ff4383a-utilities" (OuterVolumeSpecName: "utilities") pod "4696c11b-cd7f-4a9b-84d0-b4e59ff4383a" (UID: "4696c11b-cd7f-4a9b-84d0-b4e59ff4383a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.284558 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4696c11b-cd7f-4a9b-84d0-b4e59ff4383a-kube-api-access-ww2ls" (OuterVolumeSpecName: "kube-api-access-ww2ls") pod "4696c11b-cd7f-4a9b-84d0-b4e59ff4383a" (UID: "4696c11b-cd7f-4a9b-84d0-b4e59ff4383a"). InnerVolumeSpecName "kube-api-access-ww2ls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.377508 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4696c11b-cd7f-4a9b-84d0-b4e59ff4383a-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.377747 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ww2ls\" (UniqueName: \"kubernetes.io/projected/4696c11b-cd7f-4a9b-84d0-b4e59ff4383a-kube-api-access-ww2ls\") on node \"crc\" DevicePath \"\"" Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.402548 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4696c11b-cd7f-4a9b-84d0-b4e59ff4383a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4696c11b-cd7f-4a9b-84d0-b4e59ff4383a" (UID: "4696c11b-cd7f-4a9b-84d0-b4e59ff4383a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.479481 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4696c11b-cd7f-4a9b-84d0-b4e59ff4383a-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.586540 4772 generic.go:334] "Generic (PLEG): container finished" podID="4696c11b-cd7f-4a9b-84d0-b4e59ff4383a" containerID="670c3dd5d1eadcc4c7af0be85bbbd52745f9845379d48bccf45b532e278f530f" exitCode=0 Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.586600 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v67fk" event={"ID":"4696c11b-cd7f-4a9b-84d0-b4e59ff4383a","Type":"ContainerDied","Data":"670c3dd5d1eadcc4c7af0be85bbbd52745f9845379d48bccf45b532e278f530f"} Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.586641 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v67fk" event={"ID":"4696c11b-cd7f-4a9b-84d0-b4e59ff4383a","Type":"ContainerDied","Data":"d4542b704bdfa74b42beaef5f64efafc7ed30121010c3bdd793a4aaa21d8cfa5"} Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.586645 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v67fk" Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.586699 4772 scope.go:117] "RemoveContainer" containerID="670c3dd5d1eadcc4c7af0be85bbbd52745f9845379d48bccf45b532e278f530f" Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.624854 4772 scope.go:117] "RemoveContainer" containerID="06a1f585dd0be4dfae7eee556971dc6e4e153c2aa01f033489b777fe2060b96c" Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.659992 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v67fk"] Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.678329 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-v67fk"] Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.682157 4772 scope.go:117] "RemoveContainer" containerID="12499ba818facb5728a5e2af322ab50a280b5ba47cb8956bb6a0e6872c81b057" Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.724517 4772 scope.go:117] "RemoveContainer" containerID="670c3dd5d1eadcc4c7af0be85bbbd52745f9845379d48bccf45b532e278f530f" Jan 27 17:09:02 crc kubenswrapper[4772]: E0127 17:09:02.725328 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"670c3dd5d1eadcc4c7af0be85bbbd52745f9845379d48bccf45b532e278f530f\": container with ID starting with 670c3dd5d1eadcc4c7af0be85bbbd52745f9845379d48bccf45b532e278f530f not found: ID does not exist" containerID="670c3dd5d1eadcc4c7af0be85bbbd52745f9845379d48bccf45b532e278f530f" Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.725384 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"670c3dd5d1eadcc4c7af0be85bbbd52745f9845379d48bccf45b532e278f530f"} err="failed to get container status \"670c3dd5d1eadcc4c7af0be85bbbd52745f9845379d48bccf45b532e278f530f\": rpc error: code = NotFound desc = could not find container \"670c3dd5d1eadcc4c7af0be85bbbd52745f9845379d48bccf45b532e278f530f\": container with ID starting with 670c3dd5d1eadcc4c7af0be85bbbd52745f9845379d48bccf45b532e278f530f not found: ID does not exist" Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.725421 4772 scope.go:117] "RemoveContainer" containerID="06a1f585dd0be4dfae7eee556971dc6e4e153c2aa01f033489b777fe2060b96c" Jan 27 17:09:02 crc kubenswrapper[4772]: E0127 17:09:02.725870 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06a1f585dd0be4dfae7eee556971dc6e4e153c2aa01f033489b777fe2060b96c\": container with ID starting with 06a1f585dd0be4dfae7eee556971dc6e4e153c2aa01f033489b777fe2060b96c not found: ID does not exist" containerID="06a1f585dd0be4dfae7eee556971dc6e4e153c2aa01f033489b777fe2060b96c" Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.726083 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06a1f585dd0be4dfae7eee556971dc6e4e153c2aa01f033489b777fe2060b96c"} err="failed to get container status \"06a1f585dd0be4dfae7eee556971dc6e4e153c2aa01f033489b777fe2060b96c\": rpc error: code = NotFound desc = could not find container \"06a1f585dd0be4dfae7eee556971dc6e4e153c2aa01f033489b777fe2060b96c\": container with ID starting with 06a1f585dd0be4dfae7eee556971dc6e4e153c2aa01f033489b777fe2060b96c not found: ID does not exist" Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.726325 4772 scope.go:117] "RemoveContainer" containerID="12499ba818facb5728a5e2af322ab50a280b5ba47cb8956bb6a0e6872c81b057" Jan 27 17:09:02 crc kubenswrapper[4772]: E0127 17:09:02.727123 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12499ba818facb5728a5e2af322ab50a280b5ba47cb8956bb6a0e6872c81b057\": container with ID starting with 12499ba818facb5728a5e2af322ab50a280b5ba47cb8956bb6a0e6872c81b057 not found: ID does not exist" containerID="12499ba818facb5728a5e2af322ab50a280b5ba47cb8956bb6a0e6872c81b057" Jan 27 17:09:02 crc kubenswrapper[4772]: I0127 17:09:02.727220 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12499ba818facb5728a5e2af322ab50a280b5ba47cb8956bb6a0e6872c81b057"} err="failed to get container status \"12499ba818facb5728a5e2af322ab50a280b5ba47cb8956bb6a0e6872c81b057\": rpc error: code = NotFound desc = could not find container \"12499ba818facb5728a5e2af322ab50a280b5ba47cb8956bb6a0e6872c81b057\": container with ID starting with 12499ba818facb5728a5e2af322ab50a280b5ba47cb8956bb6a0e6872c81b057 not found: ID does not exist" Jan 27 17:09:04 crc kubenswrapper[4772]: I0127 17:09:04.685677 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4696c11b-cd7f-4a9b-84d0-b4e59ff4383a" path="/var/lib/kubelet/pods/4696c11b-cd7f-4a9b-84d0-b4e59ff4383a/volumes" Jan 27 17:09:09 crc kubenswrapper[4772]: I0127 17:09:09.662756 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:09:09 crc kubenswrapper[4772]: E0127 17:09:09.664555 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:09:23 crc kubenswrapper[4772]: I0127 17:09:23.662797 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:09:23 crc kubenswrapper[4772]: E0127 17:09:23.663557 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:09:38 crc kubenswrapper[4772]: I0127 17:09:38.664513 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:09:38 crc kubenswrapper[4772]: E0127 17:09:38.665525 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:09:52 crc kubenswrapper[4772]: I0127 17:09:52.663343 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:09:53 crc kubenswrapper[4772]: I0127 17:09:53.129744 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"ea5d30742e90c5fe244c125256ab670ebb26e9206885008781626f7b363771b3"} Jan 27 17:11:11 crc kubenswrapper[4772]: I0127 17:11:11.973355 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kb62q"] Jan 27 17:11:11 crc kubenswrapper[4772]: E0127 17:11:11.974531 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4696c11b-cd7f-4a9b-84d0-b4e59ff4383a" containerName="extract-utilities" Jan 27 17:11:11 crc kubenswrapper[4772]: I0127 17:11:11.974550 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4696c11b-cd7f-4a9b-84d0-b4e59ff4383a" containerName="extract-utilities" Jan 27 17:11:11 crc kubenswrapper[4772]: E0127 17:11:11.974568 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4696c11b-cd7f-4a9b-84d0-b4e59ff4383a" containerName="extract-content" Jan 27 17:11:11 crc kubenswrapper[4772]: I0127 17:11:11.974578 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4696c11b-cd7f-4a9b-84d0-b4e59ff4383a" containerName="extract-content" Jan 27 17:11:11 crc kubenswrapper[4772]: E0127 17:11:11.974595 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4696c11b-cd7f-4a9b-84d0-b4e59ff4383a" containerName="registry-server" Jan 27 17:11:11 crc kubenswrapper[4772]: I0127 17:11:11.974604 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4696c11b-cd7f-4a9b-84d0-b4e59ff4383a" containerName="registry-server" Jan 27 17:11:11 crc kubenswrapper[4772]: I0127 17:11:11.974835 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="4696c11b-cd7f-4a9b-84d0-b4e59ff4383a" containerName="registry-server" Jan 27 17:11:11 crc kubenswrapper[4772]: I0127 17:11:11.976773 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kb62q" Jan 27 17:11:11 crc kubenswrapper[4772]: I0127 17:11:11.997269 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kb62q"] Jan 27 17:11:12 crc kubenswrapper[4772]: I0127 17:11:12.075619 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgjx6\" (UniqueName: \"kubernetes.io/projected/e840c358-8d41-4381-b643-3bd35f0716a2-kube-api-access-wgjx6\") pod \"community-operators-kb62q\" (UID: \"e840c358-8d41-4381-b643-3bd35f0716a2\") " pod="openshift-marketplace/community-operators-kb62q" Jan 27 17:11:12 crc kubenswrapper[4772]: I0127 17:11:12.075692 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e840c358-8d41-4381-b643-3bd35f0716a2-catalog-content\") pod \"community-operators-kb62q\" (UID: \"e840c358-8d41-4381-b643-3bd35f0716a2\") " pod="openshift-marketplace/community-operators-kb62q" Jan 27 17:11:12 crc kubenswrapper[4772]: I0127 17:11:12.076065 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e840c358-8d41-4381-b643-3bd35f0716a2-utilities\") pod \"community-operators-kb62q\" (UID: \"e840c358-8d41-4381-b643-3bd35f0716a2\") " pod="openshift-marketplace/community-operators-kb62q" Jan 27 17:11:12 crc kubenswrapper[4772]: I0127 17:11:12.178346 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e840c358-8d41-4381-b643-3bd35f0716a2-utilities\") pod \"community-operators-kb62q\" (UID: \"e840c358-8d41-4381-b643-3bd35f0716a2\") " pod="openshift-marketplace/community-operators-kb62q" Jan 27 17:11:12 crc kubenswrapper[4772]: I0127 17:11:12.178446 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgjx6\" (UniqueName: \"kubernetes.io/projected/e840c358-8d41-4381-b643-3bd35f0716a2-kube-api-access-wgjx6\") pod \"community-operators-kb62q\" (UID: \"e840c358-8d41-4381-b643-3bd35f0716a2\") " pod="openshift-marketplace/community-operators-kb62q" Jan 27 17:11:12 crc kubenswrapper[4772]: I0127 17:11:12.178476 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e840c358-8d41-4381-b643-3bd35f0716a2-catalog-content\") pod \"community-operators-kb62q\" (UID: \"e840c358-8d41-4381-b643-3bd35f0716a2\") " pod="openshift-marketplace/community-operators-kb62q" Jan 27 17:11:12 crc kubenswrapper[4772]: I0127 17:11:12.178928 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e840c358-8d41-4381-b643-3bd35f0716a2-utilities\") pod \"community-operators-kb62q\" (UID: \"e840c358-8d41-4381-b643-3bd35f0716a2\") " pod="openshift-marketplace/community-operators-kb62q" Jan 27 17:11:12 crc kubenswrapper[4772]: I0127 17:11:12.178967 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e840c358-8d41-4381-b643-3bd35f0716a2-catalog-content\") pod \"community-operators-kb62q\" (UID: \"e840c358-8d41-4381-b643-3bd35f0716a2\") " pod="openshift-marketplace/community-operators-kb62q" Jan 27 17:11:12 crc kubenswrapper[4772]: I0127 17:11:12.197970 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgjx6\" (UniqueName: \"kubernetes.io/projected/e840c358-8d41-4381-b643-3bd35f0716a2-kube-api-access-wgjx6\") pod \"community-operators-kb62q\" (UID: \"e840c358-8d41-4381-b643-3bd35f0716a2\") " pod="openshift-marketplace/community-operators-kb62q" Jan 27 17:11:12 crc kubenswrapper[4772]: I0127 17:11:12.319522 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kb62q" Jan 27 17:11:12 crc kubenswrapper[4772]: I0127 17:11:12.850741 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kb62q"] Jan 27 17:11:12 crc kubenswrapper[4772]: I0127 17:11:12.958138 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kb62q" event={"ID":"e840c358-8d41-4381-b643-3bd35f0716a2","Type":"ContainerStarted","Data":"2d27b8d9a06391c00a3262da5e1b5a9e4e974a9ab2faf27c3443e60c44269845"} Jan 27 17:11:13 crc kubenswrapper[4772]: I0127 17:11:13.972304 4772 generic.go:334] "Generic (PLEG): container finished" podID="e840c358-8d41-4381-b643-3bd35f0716a2" containerID="8c0024f9c25988180a3cb150988e4b597cfed85d7eb62aa9274ef355d1d88296" exitCode=0 Jan 27 17:11:13 crc kubenswrapper[4772]: I0127 17:11:13.972345 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kb62q" event={"ID":"e840c358-8d41-4381-b643-3bd35f0716a2","Type":"ContainerDied","Data":"8c0024f9c25988180a3cb150988e4b597cfed85d7eb62aa9274ef355d1d88296"} Jan 27 17:11:18 crc kubenswrapper[4772]: I0127 17:11:18.016687 4772 generic.go:334] "Generic (PLEG): container finished" podID="e840c358-8d41-4381-b643-3bd35f0716a2" containerID="7949054364bb39054181aad8e729a07d19b95a294c767430f2b49200025c69c8" exitCode=0 Jan 27 17:11:18 crc kubenswrapper[4772]: I0127 17:11:18.017009 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kb62q" event={"ID":"e840c358-8d41-4381-b643-3bd35f0716a2","Type":"ContainerDied","Data":"7949054364bb39054181aad8e729a07d19b95a294c767430f2b49200025c69c8"} Jan 27 17:11:19 crc kubenswrapper[4772]: I0127 17:11:19.031966 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kb62q" event={"ID":"e840c358-8d41-4381-b643-3bd35f0716a2","Type":"ContainerStarted","Data":"3cf860b49847ad6631c20f8f8a781d32c0a504d63e63a22b870a53c993267fd3"} Jan 27 17:11:19 crc kubenswrapper[4772]: I0127 17:11:19.073282 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kb62q" podStartSLOduration=3.59528929 podStartE2EDuration="8.07325119s" podCreationTimestamp="2026-01-27 17:11:11 +0000 UTC" firstStartedPulling="2026-01-27 17:11:13.974781272 +0000 UTC m=+7459.955390380" lastFinishedPulling="2026-01-27 17:11:18.452743182 +0000 UTC m=+7464.433352280" observedRunningTime="2026-01-27 17:11:19.055350002 +0000 UTC m=+7465.035959140" watchObservedRunningTime="2026-01-27 17:11:19.07325119 +0000 UTC m=+7465.053860328" Jan 27 17:11:22 crc kubenswrapper[4772]: I0127 17:11:22.320914 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kb62q" Jan 27 17:11:22 crc kubenswrapper[4772]: I0127 17:11:22.321516 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kb62q" Jan 27 17:11:22 crc kubenswrapper[4772]: I0127 17:11:22.418848 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kb62q" Jan 27 17:11:23 crc kubenswrapper[4772]: I0127 17:11:23.156094 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kb62q" Jan 27 17:11:23 crc kubenswrapper[4772]: I0127 17:11:23.283314 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kb62q"] Jan 27 17:11:23 crc kubenswrapper[4772]: I0127 17:11:23.352346 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-txfct"] Jan 27 17:11:23 crc kubenswrapper[4772]: I0127 17:11:23.352640 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-txfct" podUID="10eef819-4355-4b65-bb81-95c055327034" containerName="registry-server" containerID="cri-o://114bf805859534c9af2af0c87c4604c2095e959428fa0c7464d58e26e36fef93" gracePeriod=2 Jan 27 17:11:23 crc kubenswrapper[4772]: I0127 17:11:23.912031 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-txfct" Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.031452 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10eef819-4355-4b65-bb81-95c055327034-catalog-content\") pod \"10eef819-4355-4b65-bb81-95c055327034\" (UID: \"10eef819-4355-4b65-bb81-95c055327034\") " Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.031638 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10eef819-4355-4b65-bb81-95c055327034-utilities\") pod \"10eef819-4355-4b65-bb81-95c055327034\" (UID: \"10eef819-4355-4b65-bb81-95c055327034\") " Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.031698 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkhfw\" (UniqueName: \"kubernetes.io/projected/10eef819-4355-4b65-bb81-95c055327034-kube-api-access-tkhfw\") pod \"10eef819-4355-4b65-bb81-95c055327034\" (UID: \"10eef819-4355-4b65-bb81-95c055327034\") " Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.032224 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10eef819-4355-4b65-bb81-95c055327034-utilities" (OuterVolumeSpecName: "utilities") pod "10eef819-4355-4b65-bb81-95c055327034" (UID: "10eef819-4355-4b65-bb81-95c055327034"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.041514 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10eef819-4355-4b65-bb81-95c055327034-kube-api-access-tkhfw" (OuterVolumeSpecName: "kube-api-access-tkhfw") pod "10eef819-4355-4b65-bb81-95c055327034" (UID: "10eef819-4355-4b65-bb81-95c055327034"). InnerVolumeSpecName "kube-api-access-tkhfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.084033 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10eef819-4355-4b65-bb81-95c055327034-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "10eef819-4355-4b65-bb81-95c055327034" (UID: "10eef819-4355-4b65-bb81-95c055327034"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.091305 4772 generic.go:334] "Generic (PLEG): container finished" podID="10eef819-4355-4b65-bb81-95c055327034" containerID="114bf805859534c9af2af0c87c4604c2095e959428fa0c7464d58e26e36fef93" exitCode=0 Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.092146 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-txfct" Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.092372 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txfct" event={"ID":"10eef819-4355-4b65-bb81-95c055327034","Type":"ContainerDied","Data":"114bf805859534c9af2af0c87c4604c2095e959428fa0c7464d58e26e36fef93"} Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.092445 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txfct" event={"ID":"10eef819-4355-4b65-bb81-95c055327034","Type":"ContainerDied","Data":"23557e0504a3fe336145be5a361937c216f6e8cad04bfe40c3131be90c233123"} Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.092477 4772 scope.go:117] "RemoveContainer" containerID="114bf805859534c9af2af0c87c4604c2095e959428fa0c7464d58e26e36fef93" Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.120431 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-txfct"] Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.127240 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-txfct"] Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.130037 4772 scope.go:117] "RemoveContainer" containerID="490d9faa657579105d3e49ba461cb9e24406692936441c0207c58ab54ad55605" Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.133586 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10eef819-4355-4b65-bb81-95c055327034-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.133613 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkhfw\" (UniqueName: \"kubernetes.io/projected/10eef819-4355-4b65-bb81-95c055327034-kube-api-access-tkhfw\") on node \"crc\" DevicePath \"\"" Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.133625 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10eef819-4355-4b65-bb81-95c055327034-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.150682 4772 scope.go:117] "RemoveContainer" containerID="7236c5d9dd7b59e4e1b4af9645ea7ed61b907d3619dcb83d87301be5c3316f7c" Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.191275 4772 scope.go:117] "RemoveContainer" containerID="114bf805859534c9af2af0c87c4604c2095e959428fa0c7464d58e26e36fef93" Jan 27 17:11:24 crc kubenswrapper[4772]: E0127 17:11:24.191802 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"114bf805859534c9af2af0c87c4604c2095e959428fa0c7464d58e26e36fef93\": container with ID starting with 114bf805859534c9af2af0c87c4604c2095e959428fa0c7464d58e26e36fef93 not found: ID does not exist" containerID="114bf805859534c9af2af0c87c4604c2095e959428fa0c7464d58e26e36fef93" Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.191834 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"114bf805859534c9af2af0c87c4604c2095e959428fa0c7464d58e26e36fef93"} err="failed to get container status \"114bf805859534c9af2af0c87c4604c2095e959428fa0c7464d58e26e36fef93\": rpc error: code = NotFound desc = could not find container \"114bf805859534c9af2af0c87c4604c2095e959428fa0c7464d58e26e36fef93\": container with ID starting with 114bf805859534c9af2af0c87c4604c2095e959428fa0c7464d58e26e36fef93 not found: ID does not exist" Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.191855 4772 scope.go:117] "RemoveContainer" containerID="490d9faa657579105d3e49ba461cb9e24406692936441c0207c58ab54ad55605" Jan 27 17:11:24 crc kubenswrapper[4772]: E0127 17:11:24.192094 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"490d9faa657579105d3e49ba461cb9e24406692936441c0207c58ab54ad55605\": container with ID starting with 490d9faa657579105d3e49ba461cb9e24406692936441c0207c58ab54ad55605 not found: ID does not exist" containerID="490d9faa657579105d3e49ba461cb9e24406692936441c0207c58ab54ad55605" Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.192125 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"490d9faa657579105d3e49ba461cb9e24406692936441c0207c58ab54ad55605"} err="failed to get container status \"490d9faa657579105d3e49ba461cb9e24406692936441c0207c58ab54ad55605\": rpc error: code = NotFound desc = could not find container \"490d9faa657579105d3e49ba461cb9e24406692936441c0207c58ab54ad55605\": container with ID starting with 490d9faa657579105d3e49ba461cb9e24406692936441c0207c58ab54ad55605 not found: ID does not exist" Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.192139 4772 scope.go:117] "RemoveContainer" containerID="7236c5d9dd7b59e4e1b4af9645ea7ed61b907d3619dcb83d87301be5c3316f7c" Jan 27 17:11:24 crc kubenswrapper[4772]: E0127 17:11:24.192360 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7236c5d9dd7b59e4e1b4af9645ea7ed61b907d3619dcb83d87301be5c3316f7c\": container with ID starting with 7236c5d9dd7b59e4e1b4af9645ea7ed61b907d3619dcb83d87301be5c3316f7c not found: ID does not exist" containerID="7236c5d9dd7b59e4e1b4af9645ea7ed61b907d3619dcb83d87301be5c3316f7c" Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.192387 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7236c5d9dd7b59e4e1b4af9645ea7ed61b907d3619dcb83d87301be5c3316f7c"} err="failed to get container status \"7236c5d9dd7b59e4e1b4af9645ea7ed61b907d3619dcb83d87301be5c3316f7c\": rpc error: code = NotFound desc = could not find container \"7236c5d9dd7b59e4e1b4af9645ea7ed61b907d3619dcb83d87301be5c3316f7c\": container with ID starting with 7236c5d9dd7b59e4e1b4af9645ea7ed61b907d3619dcb83d87301be5c3316f7c not found: ID does not exist" Jan 27 17:11:24 crc kubenswrapper[4772]: I0127 17:11:24.677076 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10eef819-4355-4b65-bb81-95c055327034" path="/var/lib/kubelet/pods/10eef819-4355-4b65-bb81-95c055327034/volumes" Jan 27 17:12:12 crc kubenswrapper[4772]: I0127 17:12:12.059390 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:12:12 crc kubenswrapper[4772]: I0127 17:12:12.062140 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:12:25 crc kubenswrapper[4772]: I0127 17:12:25.281323 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kch9d"] Jan 27 17:12:25 crc kubenswrapper[4772]: E0127 17:12:25.286713 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10eef819-4355-4b65-bb81-95c055327034" containerName="extract-utilities" Jan 27 17:12:25 crc kubenswrapper[4772]: I0127 17:12:25.286864 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="10eef819-4355-4b65-bb81-95c055327034" containerName="extract-utilities" Jan 27 17:12:25 crc kubenswrapper[4772]: E0127 17:12:25.286964 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10eef819-4355-4b65-bb81-95c055327034" containerName="extract-content" Jan 27 17:12:25 crc kubenswrapper[4772]: I0127 17:12:25.287050 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="10eef819-4355-4b65-bb81-95c055327034" containerName="extract-content" Jan 27 17:12:25 crc kubenswrapper[4772]: E0127 17:12:25.287145 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10eef819-4355-4b65-bb81-95c055327034" containerName="registry-server" Jan 27 17:12:25 crc kubenswrapper[4772]: I0127 17:12:25.287247 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="10eef819-4355-4b65-bb81-95c055327034" containerName="registry-server" Jan 27 17:12:25 crc kubenswrapper[4772]: I0127 17:12:25.287593 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="10eef819-4355-4b65-bb81-95c055327034" containerName="registry-server" Jan 27 17:12:25 crc kubenswrapper[4772]: I0127 17:12:25.289462 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kch9d" Jan 27 17:12:25 crc kubenswrapper[4772]: I0127 17:12:25.315953 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kch9d"] Jan 27 17:12:25 crc kubenswrapper[4772]: I0127 17:12:25.355864 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb293122-6b7f-4867-a949-55eb128d7ed4-utilities\") pod \"redhat-marketplace-kch9d\" (UID: \"bb293122-6b7f-4867-a949-55eb128d7ed4\") " pod="openshift-marketplace/redhat-marketplace-kch9d" Jan 27 17:12:25 crc kubenswrapper[4772]: I0127 17:12:25.356068 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb293122-6b7f-4867-a949-55eb128d7ed4-catalog-content\") pod \"redhat-marketplace-kch9d\" (UID: \"bb293122-6b7f-4867-a949-55eb128d7ed4\") " pod="openshift-marketplace/redhat-marketplace-kch9d" Jan 27 17:12:25 crc kubenswrapper[4772]: I0127 17:12:25.356158 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbjdb\" (UniqueName: \"kubernetes.io/projected/bb293122-6b7f-4867-a949-55eb128d7ed4-kube-api-access-qbjdb\") pod \"redhat-marketplace-kch9d\" (UID: \"bb293122-6b7f-4867-a949-55eb128d7ed4\") " pod="openshift-marketplace/redhat-marketplace-kch9d" Jan 27 17:12:25 crc kubenswrapper[4772]: I0127 17:12:25.458011 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb293122-6b7f-4867-a949-55eb128d7ed4-utilities\") pod \"redhat-marketplace-kch9d\" (UID: \"bb293122-6b7f-4867-a949-55eb128d7ed4\") " pod="openshift-marketplace/redhat-marketplace-kch9d" Jan 27 17:12:25 crc kubenswrapper[4772]: I0127 17:12:25.458148 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb293122-6b7f-4867-a949-55eb128d7ed4-catalog-content\") pod \"redhat-marketplace-kch9d\" (UID: \"bb293122-6b7f-4867-a949-55eb128d7ed4\") " pod="openshift-marketplace/redhat-marketplace-kch9d" Jan 27 17:12:25 crc kubenswrapper[4772]: I0127 17:12:25.458226 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbjdb\" (UniqueName: \"kubernetes.io/projected/bb293122-6b7f-4867-a949-55eb128d7ed4-kube-api-access-qbjdb\") pod \"redhat-marketplace-kch9d\" (UID: \"bb293122-6b7f-4867-a949-55eb128d7ed4\") " pod="openshift-marketplace/redhat-marketplace-kch9d" Jan 27 17:12:25 crc kubenswrapper[4772]: I0127 17:12:25.458789 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb293122-6b7f-4867-a949-55eb128d7ed4-catalog-content\") pod \"redhat-marketplace-kch9d\" (UID: \"bb293122-6b7f-4867-a949-55eb128d7ed4\") " pod="openshift-marketplace/redhat-marketplace-kch9d" Jan 27 17:12:25 crc kubenswrapper[4772]: I0127 17:12:25.459065 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb293122-6b7f-4867-a949-55eb128d7ed4-utilities\") pod \"redhat-marketplace-kch9d\" (UID: \"bb293122-6b7f-4867-a949-55eb128d7ed4\") " pod="openshift-marketplace/redhat-marketplace-kch9d" Jan 27 17:12:25 crc kubenswrapper[4772]: I0127 17:12:25.479589 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbjdb\" (UniqueName: \"kubernetes.io/projected/bb293122-6b7f-4867-a949-55eb128d7ed4-kube-api-access-qbjdb\") pod \"redhat-marketplace-kch9d\" (UID: \"bb293122-6b7f-4867-a949-55eb128d7ed4\") " pod="openshift-marketplace/redhat-marketplace-kch9d" Jan 27 17:12:25 crc kubenswrapper[4772]: I0127 17:12:25.627512 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kch9d" Jan 27 17:12:26 crc kubenswrapper[4772]: I0127 17:12:26.174775 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kch9d"] Jan 27 17:12:26 crc kubenswrapper[4772]: W0127 17:12:26.183498 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb293122_6b7f_4867_a949_55eb128d7ed4.slice/crio-f69b93ed1f2e5c33fe404e803dd9066123c9ea68b33c570a0dd62ead30b49ded WatchSource:0}: Error finding container f69b93ed1f2e5c33fe404e803dd9066123c9ea68b33c570a0dd62ead30b49ded: Status 404 returned error can't find the container with id f69b93ed1f2e5c33fe404e803dd9066123c9ea68b33c570a0dd62ead30b49ded Jan 27 17:12:26 crc kubenswrapper[4772]: I0127 17:12:26.790892 4772 generic.go:334] "Generic (PLEG): container finished" podID="bb293122-6b7f-4867-a949-55eb128d7ed4" containerID="1465504341b8ca1cf6d84aac554bd7d0847a03f9111d50be755b674dab248fde" exitCode=0 Jan 27 17:12:26 crc kubenswrapper[4772]: I0127 17:12:26.790931 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kch9d" event={"ID":"bb293122-6b7f-4867-a949-55eb128d7ed4","Type":"ContainerDied","Data":"1465504341b8ca1cf6d84aac554bd7d0847a03f9111d50be755b674dab248fde"} Jan 27 17:12:26 crc kubenswrapper[4772]: I0127 17:12:26.790957 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kch9d" event={"ID":"bb293122-6b7f-4867-a949-55eb128d7ed4","Type":"ContainerStarted","Data":"f69b93ed1f2e5c33fe404e803dd9066123c9ea68b33c570a0dd62ead30b49ded"} Jan 27 17:12:28 crc kubenswrapper[4772]: I0127 17:12:28.812677 4772 generic.go:334] "Generic (PLEG): container finished" podID="bb293122-6b7f-4867-a949-55eb128d7ed4" containerID="6d808146785900f1a51eaac3c8b810a51d7e6d7fdd2da00ad5472854f5f87dc2" exitCode=0 Jan 27 17:12:28 crc kubenswrapper[4772]: I0127 17:12:28.812806 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kch9d" event={"ID":"bb293122-6b7f-4867-a949-55eb128d7ed4","Type":"ContainerDied","Data":"6d808146785900f1a51eaac3c8b810a51d7e6d7fdd2da00ad5472854f5f87dc2"} Jan 27 17:12:33 crc kubenswrapper[4772]: I0127 17:12:33.856145 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kch9d" event={"ID":"bb293122-6b7f-4867-a949-55eb128d7ed4","Type":"ContainerStarted","Data":"606a545c1b864cd576bd005c79dedbfba1ea800f4253958444bbd989a1dca52d"} Jan 27 17:12:33 crc kubenswrapper[4772]: I0127 17:12:33.877052 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kch9d" podStartSLOduration=2.395932153 podStartE2EDuration="8.877037175s" podCreationTimestamp="2026-01-27 17:12:25 +0000 UTC" firstStartedPulling="2026-01-27 17:12:26.792234651 +0000 UTC m=+7532.772843749" lastFinishedPulling="2026-01-27 17:12:33.273339673 +0000 UTC m=+7539.253948771" observedRunningTime="2026-01-27 17:12:33.875693857 +0000 UTC m=+7539.856302955" watchObservedRunningTime="2026-01-27 17:12:33.877037175 +0000 UTC m=+7539.857646273" Jan 27 17:12:35 crc kubenswrapper[4772]: I0127 17:12:35.628565 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kch9d" Jan 27 17:12:35 crc kubenswrapper[4772]: I0127 17:12:35.629083 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kch9d" Jan 27 17:12:35 crc kubenswrapper[4772]: I0127 17:12:35.711456 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kch9d" Jan 27 17:12:42 crc kubenswrapper[4772]: I0127 17:12:42.059056 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:12:42 crc kubenswrapper[4772]: I0127 17:12:42.060442 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:12:45 crc kubenswrapper[4772]: I0127 17:12:45.687833 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kch9d" Jan 27 17:12:45 crc kubenswrapper[4772]: I0127 17:12:45.747828 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kch9d"] Jan 27 17:12:45 crc kubenswrapper[4772]: I0127 17:12:45.987644 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kch9d" podUID="bb293122-6b7f-4867-a949-55eb128d7ed4" containerName="registry-server" containerID="cri-o://606a545c1b864cd576bd005c79dedbfba1ea800f4253958444bbd989a1dca52d" gracePeriod=2 Jan 27 17:12:46 crc kubenswrapper[4772]: I0127 17:12:46.501057 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kch9d" Jan 27 17:12:46 crc kubenswrapper[4772]: I0127 17:12:46.628434 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb293122-6b7f-4867-a949-55eb128d7ed4-catalog-content\") pod \"bb293122-6b7f-4867-a949-55eb128d7ed4\" (UID: \"bb293122-6b7f-4867-a949-55eb128d7ed4\") " Jan 27 17:12:46 crc kubenswrapper[4772]: I0127 17:12:46.628633 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb293122-6b7f-4867-a949-55eb128d7ed4-utilities\") pod \"bb293122-6b7f-4867-a949-55eb128d7ed4\" (UID: \"bb293122-6b7f-4867-a949-55eb128d7ed4\") " Jan 27 17:12:46 crc kubenswrapper[4772]: I0127 17:12:46.628665 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbjdb\" (UniqueName: \"kubernetes.io/projected/bb293122-6b7f-4867-a949-55eb128d7ed4-kube-api-access-qbjdb\") pod \"bb293122-6b7f-4867-a949-55eb128d7ed4\" (UID: \"bb293122-6b7f-4867-a949-55eb128d7ed4\") " Jan 27 17:12:46 crc kubenswrapper[4772]: I0127 17:12:46.629858 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb293122-6b7f-4867-a949-55eb128d7ed4-utilities" (OuterVolumeSpecName: "utilities") pod "bb293122-6b7f-4867-a949-55eb128d7ed4" (UID: "bb293122-6b7f-4867-a949-55eb128d7ed4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:12:46 crc kubenswrapper[4772]: I0127 17:12:46.641463 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb293122-6b7f-4867-a949-55eb128d7ed4-kube-api-access-qbjdb" (OuterVolumeSpecName: "kube-api-access-qbjdb") pod "bb293122-6b7f-4867-a949-55eb128d7ed4" (UID: "bb293122-6b7f-4867-a949-55eb128d7ed4"). InnerVolumeSpecName "kube-api-access-qbjdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:12:46 crc kubenswrapper[4772]: I0127 17:12:46.662228 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb293122-6b7f-4867-a949-55eb128d7ed4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bb293122-6b7f-4867-a949-55eb128d7ed4" (UID: "bb293122-6b7f-4867-a949-55eb128d7ed4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:12:46 crc kubenswrapper[4772]: I0127 17:12:46.732324 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb293122-6b7f-4867-a949-55eb128d7ed4-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 17:12:46 crc kubenswrapper[4772]: I0127 17:12:46.733145 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbjdb\" (UniqueName: \"kubernetes.io/projected/bb293122-6b7f-4867-a949-55eb128d7ed4-kube-api-access-qbjdb\") on node \"crc\" DevicePath \"\"" Jan 27 17:12:46 crc kubenswrapper[4772]: I0127 17:12:46.733220 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb293122-6b7f-4867-a949-55eb128d7ed4-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 17:12:47 crc kubenswrapper[4772]: I0127 17:12:47.005574 4772 generic.go:334] "Generic (PLEG): container finished" podID="bb293122-6b7f-4867-a949-55eb128d7ed4" containerID="606a545c1b864cd576bd005c79dedbfba1ea800f4253958444bbd989a1dca52d" exitCode=0 Jan 27 17:12:47 crc kubenswrapper[4772]: I0127 17:12:47.005626 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kch9d" event={"ID":"bb293122-6b7f-4867-a949-55eb128d7ed4","Type":"ContainerDied","Data":"606a545c1b864cd576bd005c79dedbfba1ea800f4253958444bbd989a1dca52d"} Jan 27 17:12:47 crc kubenswrapper[4772]: I0127 17:12:47.005658 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kch9d" event={"ID":"bb293122-6b7f-4867-a949-55eb128d7ed4","Type":"ContainerDied","Data":"f69b93ed1f2e5c33fe404e803dd9066123c9ea68b33c570a0dd62ead30b49ded"} Jan 27 17:12:47 crc kubenswrapper[4772]: I0127 17:12:47.005682 4772 scope.go:117] "RemoveContainer" containerID="606a545c1b864cd576bd005c79dedbfba1ea800f4253958444bbd989a1dca52d" Jan 27 17:12:47 crc kubenswrapper[4772]: I0127 17:12:47.005725 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kch9d" Jan 27 17:12:47 crc kubenswrapper[4772]: I0127 17:12:47.049163 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kch9d"] Jan 27 17:12:47 crc kubenswrapper[4772]: I0127 17:12:47.054475 4772 scope.go:117] "RemoveContainer" containerID="6d808146785900f1a51eaac3c8b810a51d7e6d7fdd2da00ad5472854f5f87dc2" Jan 27 17:12:47 crc kubenswrapper[4772]: I0127 17:12:47.059414 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kch9d"] Jan 27 17:12:47 crc kubenswrapper[4772]: I0127 17:12:47.080201 4772 scope.go:117] "RemoveContainer" containerID="1465504341b8ca1cf6d84aac554bd7d0847a03f9111d50be755b674dab248fde" Jan 27 17:12:47 crc kubenswrapper[4772]: I0127 17:12:47.138686 4772 scope.go:117] "RemoveContainer" containerID="606a545c1b864cd576bd005c79dedbfba1ea800f4253958444bbd989a1dca52d" Jan 27 17:12:47 crc kubenswrapper[4772]: E0127 17:12:47.139238 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"606a545c1b864cd576bd005c79dedbfba1ea800f4253958444bbd989a1dca52d\": container with ID starting with 606a545c1b864cd576bd005c79dedbfba1ea800f4253958444bbd989a1dca52d not found: ID does not exist" containerID="606a545c1b864cd576bd005c79dedbfba1ea800f4253958444bbd989a1dca52d" Jan 27 17:12:47 crc kubenswrapper[4772]: I0127 17:12:47.139281 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"606a545c1b864cd576bd005c79dedbfba1ea800f4253958444bbd989a1dca52d"} err="failed to get container status \"606a545c1b864cd576bd005c79dedbfba1ea800f4253958444bbd989a1dca52d\": rpc error: code = NotFound desc = could not find container \"606a545c1b864cd576bd005c79dedbfba1ea800f4253958444bbd989a1dca52d\": container with ID starting with 606a545c1b864cd576bd005c79dedbfba1ea800f4253958444bbd989a1dca52d not found: ID does not exist" Jan 27 17:12:47 crc kubenswrapper[4772]: I0127 17:12:47.139310 4772 scope.go:117] "RemoveContainer" containerID="6d808146785900f1a51eaac3c8b810a51d7e6d7fdd2da00ad5472854f5f87dc2" Jan 27 17:12:47 crc kubenswrapper[4772]: E0127 17:12:47.139678 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d808146785900f1a51eaac3c8b810a51d7e6d7fdd2da00ad5472854f5f87dc2\": container with ID starting with 6d808146785900f1a51eaac3c8b810a51d7e6d7fdd2da00ad5472854f5f87dc2 not found: ID does not exist" containerID="6d808146785900f1a51eaac3c8b810a51d7e6d7fdd2da00ad5472854f5f87dc2" Jan 27 17:12:47 crc kubenswrapper[4772]: I0127 17:12:47.139718 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d808146785900f1a51eaac3c8b810a51d7e6d7fdd2da00ad5472854f5f87dc2"} err="failed to get container status \"6d808146785900f1a51eaac3c8b810a51d7e6d7fdd2da00ad5472854f5f87dc2\": rpc error: code = NotFound desc = could not find container \"6d808146785900f1a51eaac3c8b810a51d7e6d7fdd2da00ad5472854f5f87dc2\": container with ID starting with 6d808146785900f1a51eaac3c8b810a51d7e6d7fdd2da00ad5472854f5f87dc2 not found: ID does not exist" Jan 27 17:12:47 crc kubenswrapper[4772]: I0127 17:12:47.139740 4772 scope.go:117] "RemoveContainer" containerID="1465504341b8ca1cf6d84aac554bd7d0847a03f9111d50be755b674dab248fde" Jan 27 17:12:47 crc kubenswrapper[4772]: E0127 17:12:47.140085 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1465504341b8ca1cf6d84aac554bd7d0847a03f9111d50be755b674dab248fde\": container with ID starting with 1465504341b8ca1cf6d84aac554bd7d0847a03f9111d50be755b674dab248fde not found: ID does not exist" containerID="1465504341b8ca1cf6d84aac554bd7d0847a03f9111d50be755b674dab248fde" Jan 27 17:12:47 crc kubenswrapper[4772]: I0127 17:12:47.140111 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1465504341b8ca1cf6d84aac554bd7d0847a03f9111d50be755b674dab248fde"} err="failed to get container status \"1465504341b8ca1cf6d84aac554bd7d0847a03f9111d50be755b674dab248fde\": rpc error: code = NotFound desc = could not find container \"1465504341b8ca1cf6d84aac554bd7d0847a03f9111d50be755b674dab248fde\": container with ID starting with 1465504341b8ca1cf6d84aac554bd7d0847a03f9111d50be755b674dab248fde not found: ID does not exist" Jan 27 17:12:48 crc kubenswrapper[4772]: I0127 17:12:48.683945 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb293122-6b7f-4867-a949-55eb128d7ed4" path="/var/lib/kubelet/pods/bb293122-6b7f-4867-a949-55eb128d7ed4/volumes" Jan 27 17:13:12 crc kubenswrapper[4772]: I0127 17:13:12.059109 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:13:12 crc kubenswrapper[4772]: I0127 17:13:12.060112 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:13:12 crc kubenswrapper[4772]: I0127 17:13:12.060229 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 17:13:12 crc kubenswrapper[4772]: I0127 17:13:12.061492 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ea5d30742e90c5fe244c125256ab670ebb26e9206885008781626f7b363771b3"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 17:13:12 crc kubenswrapper[4772]: I0127 17:13:12.061598 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://ea5d30742e90c5fe244c125256ab670ebb26e9206885008781626f7b363771b3" gracePeriod=600 Jan 27 17:13:12 crc kubenswrapper[4772]: I0127 17:13:12.742607 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="ea5d30742e90c5fe244c125256ab670ebb26e9206885008781626f7b363771b3" exitCode=0 Jan 27 17:13:12 crc kubenswrapper[4772]: I0127 17:13:12.742699 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"ea5d30742e90c5fe244c125256ab670ebb26e9206885008781626f7b363771b3"} Jan 27 17:13:12 crc kubenswrapper[4772]: I0127 17:13:12.743008 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246"} Jan 27 17:13:12 crc kubenswrapper[4772]: I0127 17:13:12.743050 4772 scope.go:117] "RemoveContainer" containerID="9bc4d0691b7d281178157feda7a06246e962bb43aa27764495f3ea77eef906b7" Jan 27 17:14:17 crc kubenswrapper[4772]: I0127 17:14:17.877880 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dlnbw"] Jan 27 17:14:17 crc kubenswrapper[4772]: E0127 17:14:17.879238 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb293122-6b7f-4867-a949-55eb128d7ed4" containerName="registry-server" Jan 27 17:14:17 crc kubenswrapper[4772]: I0127 17:14:17.879261 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb293122-6b7f-4867-a949-55eb128d7ed4" containerName="registry-server" Jan 27 17:14:17 crc kubenswrapper[4772]: E0127 17:14:17.879286 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb293122-6b7f-4867-a949-55eb128d7ed4" containerName="extract-content" Jan 27 17:14:17 crc kubenswrapper[4772]: I0127 17:14:17.879297 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb293122-6b7f-4867-a949-55eb128d7ed4" containerName="extract-content" Jan 27 17:14:17 crc kubenswrapper[4772]: E0127 17:14:17.879322 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb293122-6b7f-4867-a949-55eb128d7ed4" containerName="extract-utilities" Jan 27 17:14:17 crc kubenswrapper[4772]: I0127 17:14:17.879334 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb293122-6b7f-4867-a949-55eb128d7ed4" containerName="extract-utilities" Jan 27 17:14:17 crc kubenswrapper[4772]: I0127 17:14:17.879635 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb293122-6b7f-4867-a949-55eb128d7ed4" containerName="registry-server" Jan 27 17:14:17 crc kubenswrapper[4772]: I0127 17:14:17.882076 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dlnbw" Jan 27 17:14:17 crc kubenswrapper[4772]: I0127 17:14:17.888144 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dlnbw"] Jan 27 17:14:18 crc kubenswrapper[4772]: I0127 17:14:18.016731 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9192220-7559-4548-a923-ff5096e93763-catalog-content\") pod \"certified-operators-dlnbw\" (UID: \"a9192220-7559-4548-a923-ff5096e93763\") " pod="openshift-marketplace/certified-operators-dlnbw" Jan 27 17:14:18 crc kubenswrapper[4772]: I0127 17:14:18.017070 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9192220-7559-4548-a923-ff5096e93763-utilities\") pod \"certified-operators-dlnbw\" (UID: \"a9192220-7559-4548-a923-ff5096e93763\") " pod="openshift-marketplace/certified-operators-dlnbw" Jan 27 17:14:18 crc kubenswrapper[4772]: I0127 17:14:18.017092 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6ndg\" (UniqueName: \"kubernetes.io/projected/a9192220-7559-4548-a923-ff5096e93763-kube-api-access-c6ndg\") pod \"certified-operators-dlnbw\" (UID: \"a9192220-7559-4548-a923-ff5096e93763\") " pod="openshift-marketplace/certified-operators-dlnbw" Jan 27 17:14:18 crc kubenswrapper[4772]: I0127 17:14:18.119160 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9192220-7559-4548-a923-ff5096e93763-catalog-content\") pod \"certified-operators-dlnbw\" (UID: \"a9192220-7559-4548-a923-ff5096e93763\") " pod="openshift-marketplace/certified-operators-dlnbw" Jan 27 17:14:18 crc kubenswrapper[4772]: I0127 17:14:18.119342 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9192220-7559-4548-a923-ff5096e93763-utilities\") pod \"certified-operators-dlnbw\" (UID: \"a9192220-7559-4548-a923-ff5096e93763\") " pod="openshift-marketplace/certified-operators-dlnbw" Jan 27 17:14:18 crc kubenswrapper[4772]: I0127 17:14:18.119393 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6ndg\" (UniqueName: \"kubernetes.io/projected/a9192220-7559-4548-a923-ff5096e93763-kube-api-access-c6ndg\") pod \"certified-operators-dlnbw\" (UID: \"a9192220-7559-4548-a923-ff5096e93763\") " pod="openshift-marketplace/certified-operators-dlnbw" Jan 27 17:14:18 crc kubenswrapper[4772]: I0127 17:14:18.119697 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9192220-7559-4548-a923-ff5096e93763-catalog-content\") pod \"certified-operators-dlnbw\" (UID: \"a9192220-7559-4548-a923-ff5096e93763\") " pod="openshift-marketplace/certified-operators-dlnbw" Jan 27 17:14:18 crc kubenswrapper[4772]: I0127 17:14:18.120086 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9192220-7559-4548-a923-ff5096e93763-utilities\") pod \"certified-operators-dlnbw\" (UID: \"a9192220-7559-4548-a923-ff5096e93763\") " pod="openshift-marketplace/certified-operators-dlnbw" Jan 27 17:14:18 crc kubenswrapper[4772]: I0127 17:14:18.142714 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6ndg\" (UniqueName: \"kubernetes.io/projected/a9192220-7559-4548-a923-ff5096e93763-kube-api-access-c6ndg\") pod \"certified-operators-dlnbw\" (UID: \"a9192220-7559-4548-a923-ff5096e93763\") " pod="openshift-marketplace/certified-operators-dlnbw" Jan 27 17:14:18 crc kubenswrapper[4772]: I0127 17:14:18.213391 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dlnbw" Jan 27 17:14:18 crc kubenswrapper[4772]: I0127 17:14:18.548535 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dlnbw"] Jan 27 17:14:19 crc kubenswrapper[4772]: I0127 17:14:19.502083 4772 generic.go:334] "Generic (PLEG): container finished" podID="a9192220-7559-4548-a923-ff5096e93763" containerID="dbaaaa681dd7df9d44b3fd95409448538cdae1fcb4228db607f07ec31e16f8aa" exitCode=0 Jan 27 17:14:19 crc kubenswrapper[4772]: I0127 17:14:19.502205 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dlnbw" event={"ID":"a9192220-7559-4548-a923-ff5096e93763","Type":"ContainerDied","Data":"dbaaaa681dd7df9d44b3fd95409448538cdae1fcb4228db607f07ec31e16f8aa"} Jan 27 17:14:19 crc kubenswrapper[4772]: I0127 17:14:19.502485 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dlnbw" event={"ID":"a9192220-7559-4548-a923-ff5096e93763","Type":"ContainerStarted","Data":"ce84b67e6b038a4d23f97f5b2ddc4a454987ce4f78ab313a4e04e788041f4d13"} Jan 27 17:14:19 crc kubenswrapper[4772]: I0127 17:14:19.506717 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 17:14:21 crc kubenswrapper[4772]: I0127 17:14:21.526348 4772 generic.go:334] "Generic (PLEG): container finished" podID="a9192220-7559-4548-a923-ff5096e93763" containerID="3d5249bc1ee6cc2a2ba5b5dd83f216326e35b40984cb41a3156586b5873760f9" exitCode=0 Jan 27 17:14:21 crc kubenswrapper[4772]: I0127 17:14:21.527249 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dlnbw" event={"ID":"a9192220-7559-4548-a923-ff5096e93763","Type":"ContainerDied","Data":"3d5249bc1ee6cc2a2ba5b5dd83f216326e35b40984cb41a3156586b5873760f9"} Jan 27 17:14:22 crc kubenswrapper[4772]: I0127 17:14:22.538007 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dlnbw" event={"ID":"a9192220-7559-4548-a923-ff5096e93763","Type":"ContainerStarted","Data":"3fa8645b6717e26a0c6cd856f0a02932bce52294060a03d033d72ec5b711534a"} Jan 27 17:14:22 crc kubenswrapper[4772]: I0127 17:14:22.563158 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dlnbw" podStartSLOduration=3.069485953 podStartE2EDuration="5.563143755s" podCreationTimestamp="2026-01-27 17:14:17 +0000 UTC" firstStartedPulling="2026-01-27 17:14:19.506504015 +0000 UTC m=+7645.487113113" lastFinishedPulling="2026-01-27 17:14:22.000161807 +0000 UTC m=+7647.980770915" observedRunningTime="2026-01-27 17:14:22.554067817 +0000 UTC m=+7648.534676915" watchObservedRunningTime="2026-01-27 17:14:22.563143755 +0000 UTC m=+7648.543752853" Jan 27 17:14:28 crc kubenswrapper[4772]: I0127 17:14:28.214047 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dlnbw" Jan 27 17:14:28 crc kubenswrapper[4772]: I0127 17:14:28.215735 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dlnbw" Jan 27 17:14:28 crc kubenswrapper[4772]: I0127 17:14:28.296552 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dlnbw" Jan 27 17:14:28 crc kubenswrapper[4772]: I0127 17:14:28.643082 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dlnbw" Jan 27 17:14:28 crc kubenswrapper[4772]: I0127 17:14:28.702466 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dlnbw"] Jan 27 17:14:30 crc kubenswrapper[4772]: I0127 17:14:30.616433 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dlnbw" podUID="a9192220-7559-4548-a923-ff5096e93763" containerName="registry-server" containerID="cri-o://3fa8645b6717e26a0c6cd856f0a02932bce52294060a03d033d72ec5b711534a" gracePeriod=2 Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.134796 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dlnbw" Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.211642 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9192220-7559-4548-a923-ff5096e93763-utilities\") pod \"a9192220-7559-4548-a923-ff5096e93763\" (UID: \"a9192220-7559-4548-a923-ff5096e93763\") " Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.211733 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9192220-7559-4548-a923-ff5096e93763-catalog-content\") pod \"a9192220-7559-4548-a923-ff5096e93763\" (UID: \"a9192220-7559-4548-a923-ff5096e93763\") " Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.211774 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6ndg\" (UniqueName: \"kubernetes.io/projected/a9192220-7559-4548-a923-ff5096e93763-kube-api-access-c6ndg\") pod \"a9192220-7559-4548-a923-ff5096e93763\" (UID: \"a9192220-7559-4548-a923-ff5096e93763\") " Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.213406 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9192220-7559-4548-a923-ff5096e93763-utilities" (OuterVolumeSpecName: "utilities") pod "a9192220-7559-4548-a923-ff5096e93763" (UID: "a9192220-7559-4548-a923-ff5096e93763"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.218241 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9192220-7559-4548-a923-ff5096e93763-kube-api-access-c6ndg" (OuterVolumeSpecName: "kube-api-access-c6ndg") pod "a9192220-7559-4548-a923-ff5096e93763" (UID: "a9192220-7559-4548-a923-ff5096e93763"). InnerVolumeSpecName "kube-api-access-c6ndg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.314063 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9192220-7559-4548-a923-ff5096e93763-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.314112 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6ndg\" (UniqueName: \"kubernetes.io/projected/a9192220-7559-4548-a923-ff5096e93763-kube-api-access-c6ndg\") on node \"crc\" DevicePath \"\"" Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.629967 4772 generic.go:334] "Generic (PLEG): container finished" podID="a9192220-7559-4548-a923-ff5096e93763" containerID="3fa8645b6717e26a0c6cd856f0a02932bce52294060a03d033d72ec5b711534a" exitCode=0 Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.630040 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dlnbw" event={"ID":"a9192220-7559-4548-a923-ff5096e93763","Type":"ContainerDied","Data":"3fa8645b6717e26a0c6cd856f0a02932bce52294060a03d033d72ec5b711534a"} Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.630085 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dlnbw" Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.630115 4772 scope.go:117] "RemoveContainer" containerID="3fa8645b6717e26a0c6cd856f0a02932bce52294060a03d033d72ec5b711534a" Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.630095 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dlnbw" event={"ID":"a9192220-7559-4548-a923-ff5096e93763","Type":"ContainerDied","Data":"ce84b67e6b038a4d23f97f5b2ddc4a454987ce4f78ab313a4e04e788041f4d13"} Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.663531 4772 scope.go:117] "RemoveContainer" containerID="3d5249bc1ee6cc2a2ba5b5dd83f216326e35b40984cb41a3156586b5873760f9" Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.698013 4772 scope.go:117] "RemoveContainer" containerID="dbaaaa681dd7df9d44b3fd95409448538cdae1fcb4228db607f07ec31e16f8aa" Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.739691 4772 scope.go:117] "RemoveContainer" containerID="3fa8645b6717e26a0c6cd856f0a02932bce52294060a03d033d72ec5b711534a" Jan 27 17:14:31 crc kubenswrapper[4772]: E0127 17:14:31.740254 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fa8645b6717e26a0c6cd856f0a02932bce52294060a03d033d72ec5b711534a\": container with ID starting with 3fa8645b6717e26a0c6cd856f0a02932bce52294060a03d033d72ec5b711534a not found: ID does not exist" containerID="3fa8645b6717e26a0c6cd856f0a02932bce52294060a03d033d72ec5b711534a" Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.740322 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fa8645b6717e26a0c6cd856f0a02932bce52294060a03d033d72ec5b711534a"} err="failed to get container status \"3fa8645b6717e26a0c6cd856f0a02932bce52294060a03d033d72ec5b711534a\": rpc error: code = NotFound desc = could not find container \"3fa8645b6717e26a0c6cd856f0a02932bce52294060a03d033d72ec5b711534a\": container with ID starting with 3fa8645b6717e26a0c6cd856f0a02932bce52294060a03d033d72ec5b711534a not found: ID does not exist" Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.740366 4772 scope.go:117] "RemoveContainer" containerID="3d5249bc1ee6cc2a2ba5b5dd83f216326e35b40984cb41a3156586b5873760f9" Jan 27 17:14:31 crc kubenswrapper[4772]: E0127 17:14:31.741066 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d5249bc1ee6cc2a2ba5b5dd83f216326e35b40984cb41a3156586b5873760f9\": container with ID starting with 3d5249bc1ee6cc2a2ba5b5dd83f216326e35b40984cb41a3156586b5873760f9 not found: ID does not exist" containerID="3d5249bc1ee6cc2a2ba5b5dd83f216326e35b40984cb41a3156586b5873760f9" Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.741103 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d5249bc1ee6cc2a2ba5b5dd83f216326e35b40984cb41a3156586b5873760f9"} err="failed to get container status \"3d5249bc1ee6cc2a2ba5b5dd83f216326e35b40984cb41a3156586b5873760f9\": rpc error: code = NotFound desc = could not find container \"3d5249bc1ee6cc2a2ba5b5dd83f216326e35b40984cb41a3156586b5873760f9\": container with ID starting with 3d5249bc1ee6cc2a2ba5b5dd83f216326e35b40984cb41a3156586b5873760f9 not found: ID does not exist" Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.741131 4772 scope.go:117] "RemoveContainer" containerID="dbaaaa681dd7df9d44b3fd95409448538cdae1fcb4228db607f07ec31e16f8aa" Jan 27 17:14:31 crc kubenswrapper[4772]: E0127 17:14:31.741529 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbaaaa681dd7df9d44b3fd95409448538cdae1fcb4228db607f07ec31e16f8aa\": container with ID starting with dbaaaa681dd7df9d44b3fd95409448538cdae1fcb4228db607f07ec31e16f8aa not found: ID does not exist" containerID="dbaaaa681dd7df9d44b3fd95409448538cdae1fcb4228db607f07ec31e16f8aa" Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.741566 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbaaaa681dd7df9d44b3fd95409448538cdae1fcb4228db607f07ec31e16f8aa"} err="failed to get container status \"dbaaaa681dd7df9d44b3fd95409448538cdae1fcb4228db607f07ec31e16f8aa\": rpc error: code = NotFound desc = could not find container \"dbaaaa681dd7df9d44b3fd95409448538cdae1fcb4228db607f07ec31e16f8aa\": container with ID starting with dbaaaa681dd7df9d44b3fd95409448538cdae1fcb4228db607f07ec31e16f8aa not found: ID does not exist" Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.895824 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9192220-7559-4548-a923-ff5096e93763-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a9192220-7559-4548-a923-ff5096e93763" (UID: "a9192220-7559-4548-a923-ff5096e93763"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.927102 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9192220-7559-4548-a923-ff5096e93763-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.976600 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dlnbw"] Jan 27 17:14:31 crc kubenswrapper[4772]: I0127 17:14:31.983770 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dlnbw"] Jan 27 17:14:32 crc kubenswrapper[4772]: I0127 17:14:32.676656 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9192220-7559-4548-a923-ff5096e93763" path="/var/lib/kubelet/pods/a9192220-7559-4548-a923-ff5096e93763/volumes" Jan 27 17:15:00 crc kubenswrapper[4772]: I0127 17:15:00.199208 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492235-blstm"] Jan 27 17:15:00 crc kubenswrapper[4772]: E0127 17:15:00.200188 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9192220-7559-4548-a923-ff5096e93763" containerName="extract-content" Jan 27 17:15:00 crc kubenswrapper[4772]: I0127 17:15:00.200203 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9192220-7559-4548-a923-ff5096e93763" containerName="extract-content" Jan 27 17:15:00 crc kubenswrapper[4772]: E0127 17:15:00.200238 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9192220-7559-4548-a923-ff5096e93763" containerName="registry-server" Jan 27 17:15:00 crc kubenswrapper[4772]: I0127 17:15:00.200244 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9192220-7559-4548-a923-ff5096e93763" containerName="registry-server" Jan 27 17:15:00 crc kubenswrapper[4772]: E0127 17:15:00.200260 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9192220-7559-4548-a923-ff5096e93763" containerName="extract-utilities" Jan 27 17:15:00 crc kubenswrapper[4772]: I0127 17:15:00.200267 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9192220-7559-4548-a923-ff5096e93763" containerName="extract-utilities" Jan 27 17:15:00 crc kubenswrapper[4772]: I0127 17:15:00.200449 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9192220-7559-4548-a923-ff5096e93763" containerName="registry-server" Jan 27 17:15:00 crc kubenswrapper[4772]: I0127 17:15:00.201112 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492235-blstm" Jan 27 17:15:00 crc kubenswrapper[4772]: I0127 17:15:00.204669 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 27 17:15:00 crc kubenswrapper[4772]: I0127 17:15:00.204950 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 27 17:15:00 crc kubenswrapper[4772]: I0127 17:15:00.229123 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492235-blstm"] Jan 27 17:15:00 crc kubenswrapper[4772]: I0127 17:15:00.305221 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0db9aea1-8a66-4447-9150-812172da2a26-secret-volume\") pod \"collect-profiles-29492235-blstm\" (UID: \"0db9aea1-8a66-4447-9150-812172da2a26\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492235-blstm" Jan 27 17:15:00 crc kubenswrapper[4772]: I0127 17:15:00.305353 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qmx7\" (UniqueName: \"kubernetes.io/projected/0db9aea1-8a66-4447-9150-812172da2a26-kube-api-access-6qmx7\") pod \"collect-profiles-29492235-blstm\" (UID: \"0db9aea1-8a66-4447-9150-812172da2a26\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492235-blstm" Jan 27 17:15:00 crc kubenswrapper[4772]: I0127 17:15:00.305396 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0db9aea1-8a66-4447-9150-812172da2a26-config-volume\") pod \"collect-profiles-29492235-blstm\" (UID: \"0db9aea1-8a66-4447-9150-812172da2a26\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492235-blstm" Jan 27 17:15:00 crc kubenswrapper[4772]: I0127 17:15:00.407151 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0db9aea1-8a66-4447-9150-812172da2a26-secret-volume\") pod \"collect-profiles-29492235-blstm\" (UID: \"0db9aea1-8a66-4447-9150-812172da2a26\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492235-blstm" Jan 27 17:15:00 crc kubenswrapper[4772]: I0127 17:15:00.407328 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qmx7\" (UniqueName: \"kubernetes.io/projected/0db9aea1-8a66-4447-9150-812172da2a26-kube-api-access-6qmx7\") pod \"collect-profiles-29492235-blstm\" (UID: \"0db9aea1-8a66-4447-9150-812172da2a26\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492235-blstm" Jan 27 17:15:00 crc kubenswrapper[4772]: I0127 17:15:00.407373 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0db9aea1-8a66-4447-9150-812172da2a26-config-volume\") pod \"collect-profiles-29492235-blstm\" (UID: \"0db9aea1-8a66-4447-9150-812172da2a26\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492235-blstm" Jan 27 17:15:00 crc kubenswrapper[4772]: I0127 17:15:00.408496 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0db9aea1-8a66-4447-9150-812172da2a26-config-volume\") pod \"collect-profiles-29492235-blstm\" (UID: \"0db9aea1-8a66-4447-9150-812172da2a26\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492235-blstm" Jan 27 17:15:00 crc kubenswrapper[4772]: I0127 17:15:00.416270 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0db9aea1-8a66-4447-9150-812172da2a26-secret-volume\") pod \"collect-profiles-29492235-blstm\" (UID: \"0db9aea1-8a66-4447-9150-812172da2a26\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492235-blstm" Jan 27 17:15:00 crc kubenswrapper[4772]: I0127 17:15:00.425594 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qmx7\" (UniqueName: \"kubernetes.io/projected/0db9aea1-8a66-4447-9150-812172da2a26-kube-api-access-6qmx7\") pod \"collect-profiles-29492235-blstm\" (UID: \"0db9aea1-8a66-4447-9150-812172da2a26\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492235-blstm" Jan 27 17:15:00 crc kubenswrapper[4772]: I0127 17:15:00.534432 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492235-blstm" Jan 27 17:15:01 crc kubenswrapper[4772]: I0127 17:15:01.066642 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492235-blstm"] Jan 27 17:15:01 crc kubenswrapper[4772]: I0127 17:15:01.256783 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492235-blstm" event={"ID":"0db9aea1-8a66-4447-9150-812172da2a26","Type":"ContainerStarted","Data":"8cac6045fccfbe1e377f14b47042375c5f42ee5f5f23e9bb7094c2ffd4201d8c"} Jan 27 17:15:01 crc kubenswrapper[4772]: I0127 17:15:01.259573 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492235-blstm" event={"ID":"0db9aea1-8a66-4447-9150-812172da2a26","Type":"ContainerStarted","Data":"336e829070d890d5276fa9fc7334182d780e1d99b11bda8599d5a4ab4962b513"} Jan 27 17:15:01 crc kubenswrapper[4772]: I0127 17:15:01.275375 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29492235-blstm" podStartSLOduration=1.275351768 podStartE2EDuration="1.275351768s" podCreationTimestamp="2026-01-27 17:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 17:15:01.270600664 +0000 UTC m=+7687.251209772" watchObservedRunningTime="2026-01-27 17:15:01.275351768 +0000 UTC m=+7687.255960866" Jan 27 17:15:02 crc kubenswrapper[4772]: I0127 17:15:02.268063 4772 generic.go:334] "Generic (PLEG): container finished" podID="0db9aea1-8a66-4447-9150-812172da2a26" containerID="8cac6045fccfbe1e377f14b47042375c5f42ee5f5f23e9bb7094c2ffd4201d8c" exitCode=0 Jan 27 17:15:02 crc kubenswrapper[4772]: I0127 17:15:02.268334 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492235-blstm" event={"ID":"0db9aea1-8a66-4447-9150-812172da2a26","Type":"ContainerDied","Data":"8cac6045fccfbe1e377f14b47042375c5f42ee5f5f23e9bb7094c2ffd4201d8c"} Jan 27 17:15:03 crc kubenswrapper[4772]: I0127 17:15:03.689967 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492235-blstm" Jan 27 17:15:03 crc kubenswrapper[4772]: I0127 17:15:03.802289 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0db9aea1-8a66-4447-9150-812172da2a26-secret-volume\") pod \"0db9aea1-8a66-4447-9150-812172da2a26\" (UID: \"0db9aea1-8a66-4447-9150-812172da2a26\") " Jan 27 17:15:03 crc kubenswrapper[4772]: I0127 17:15:03.802369 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0db9aea1-8a66-4447-9150-812172da2a26-config-volume\") pod \"0db9aea1-8a66-4447-9150-812172da2a26\" (UID: \"0db9aea1-8a66-4447-9150-812172da2a26\") " Jan 27 17:15:03 crc kubenswrapper[4772]: I0127 17:15:03.802475 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qmx7\" (UniqueName: \"kubernetes.io/projected/0db9aea1-8a66-4447-9150-812172da2a26-kube-api-access-6qmx7\") pod \"0db9aea1-8a66-4447-9150-812172da2a26\" (UID: \"0db9aea1-8a66-4447-9150-812172da2a26\") " Jan 27 17:15:03 crc kubenswrapper[4772]: I0127 17:15:03.804572 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0db9aea1-8a66-4447-9150-812172da2a26-config-volume" (OuterVolumeSpecName: "config-volume") pod "0db9aea1-8a66-4447-9150-812172da2a26" (UID: "0db9aea1-8a66-4447-9150-812172da2a26"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 17:15:03 crc kubenswrapper[4772]: I0127 17:15:03.809106 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0db9aea1-8a66-4447-9150-812172da2a26-kube-api-access-6qmx7" (OuterVolumeSpecName: "kube-api-access-6qmx7") pod "0db9aea1-8a66-4447-9150-812172da2a26" (UID: "0db9aea1-8a66-4447-9150-812172da2a26"). InnerVolumeSpecName "kube-api-access-6qmx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:15:03 crc kubenswrapper[4772]: I0127 17:15:03.809959 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0db9aea1-8a66-4447-9150-812172da2a26-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0db9aea1-8a66-4447-9150-812172da2a26" (UID: "0db9aea1-8a66-4447-9150-812172da2a26"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 17:15:03 crc kubenswrapper[4772]: I0127 17:15:03.905413 4772 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0db9aea1-8a66-4447-9150-812172da2a26-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 27 17:15:03 crc kubenswrapper[4772]: I0127 17:15:03.905696 4772 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0db9aea1-8a66-4447-9150-812172da2a26-config-volume\") on node \"crc\" DevicePath \"\"" Jan 27 17:15:03 crc kubenswrapper[4772]: I0127 17:15:03.905707 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qmx7\" (UniqueName: \"kubernetes.io/projected/0db9aea1-8a66-4447-9150-812172da2a26-kube-api-access-6qmx7\") on node \"crc\" DevicePath \"\"" Jan 27 17:15:04 crc kubenswrapper[4772]: I0127 17:15:04.296652 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492235-blstm" event={"ID":"0db9aea1-8a66-4447-9150-812172da2a26","Type":"ContainerDied","Data":"336e829070d890d5276fa9fc7334182d780e1d99b11bda8599d5a4ab4962b513"} Jan 27 17:15:04 crc kubenswrapper[4772]: I0127 17:15:04.297118 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="336e829070d890d5276fa9fc7334182d780e1d99b11bda8599d5a4ab4962b513" Jan 27 17:15:04 crc kubenswrapper[4772]: I0127 17:15:04.297275 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492235-blstm" Jan 27 17:15:04 crc kubenswrapper[4772]: I0127 17:15:04.371971 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492190-nkwl9"] Jan 27 17:15:04 crc kubenswrapper[4772]: I0127 17:15:04.382282 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492190-nkwl9"] Jan 27 17:15:04 crc kubenswrapper[4772]: I0127 17:15:04.686301 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aad11684-a5b7-4df1-9d18-5179c6113f66" path="/var/lib/kubelet/pods/aad11684-a5b7-4df1-9d18-5179c6113f66/volumes" Jan 27 17:15:10 crc kubenswrapper[4772]: I0127 17:15:10.713958 4772 scope.go:117] "RemoveContainer" containerID="114edebee04cbeb82762a8f7e28bf44b5665934fa2746ec43b3a0a20d9084515" Jan 27 17:15:12 crc kubenswrapper[4772]: I0127 17:15:12.068304 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:15:12 crc kubenswrapper[4772]: I0127 17:15:12.069137 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:15:42 crc kubenswrapper[4772]: I0127 17:15:42.058350 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:15:42 crc kubenswrapper[4772]: I0127 17:15:42.058933 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:16:12 crc kubenswrapper[4772]: I0127 17:16:12.058925 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:16:12 crc kubenswrapper[4772]: I0127 17:16:12.061254 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:16:12 crc kubenswrapper[4772]: I0127 17:16:12.061493 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 17:16:12 crc kubenswrapper[4772]: I0127 17:16:12.062763 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 17:16:12 crc kubenswrapper[4772]: I0127 17:16:12.063030 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" gracePeriod=600 Jan 27 17:16:12 crc kubenswrapper[4772]: E0127 17:16:12.211710 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:16:13 crc kubenswrapper[4772]: I0127 17:16:13.063960 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" exitCode=0 Jan 27 17:16:13 crc kubenswrapper[4772]: I0127 17:16:13.064011 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246"} Jan 27 17:16:13 crc kubenswrapper[4772]: I0127 17:16:13.064054 4772 scope.go:117] "RemoveContainer" containerID="ea5d30742e90c5fe244c125256ab670ebb26e9206885008781626f7b363771b3" Jan 27 17:16:13 crc kubenswrapper[4772]: I0127 17:16:13.064729 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:16:13 crc kubenswrapper[4772]: E0127 17:16:13.065001 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:16:25 crc kubenswrapper[4772]: I0127 17:16:25.663672 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:16:25 crc kubenswrapper[4772]: E0127 17:16:25.664507 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:16:36 crc kubenswrapper[4772]: I0127 17:16:36.663201 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:16:36 crc kubenswrapper[4772]: E0127 17:16:36.664605 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:16:51 crc kubenswrapper[4772]: I0127 17:16:51.664021 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:16:51 crc kubenswrapper[4772]: E0127 17:16:51.665752 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:17:03 crc kubenswrapper[4772]: I0127 17:17:03.664586 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:17:03 crc kubenswrapper[4772]: E0127 17:17:03.665736 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:17:14 crc kubenswrapper[4772]: I0127 17:17:14.673275 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:17:14 crc kubenswrapper[4772]: E0127 17:17:14.686014 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:17:28 crc kubenswrapper[4772]: I0127 17:17:28.664246 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:17:28 crc kubenswrapper[4772]: E0127 17:17:28.665460 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:17:43 crc kubenswrapper[4772]: I0127 17:17:43.663210 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:17:43 crc kubenswrapper[4772]: E0127 17:17:43.663964 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:17:57 crc kubenswrapper[4772]: I0127 17:17:57.662695 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:17:57 crc kubenswrapper[4772]: E0127 17:17:57.663444 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:18:09 crc kubenswrapper[4772]: I0127 17:18:09.703412 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:18:09 crc kubenswrapper[4772]: E0127 17:18:09.705790 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:18:20 crc kubenswrapper[4772]: I0127 17:18:20.664140 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:18:20 crc kubenswrapper[4772]: E0127 17:18:20.665381 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:18:33 crc kubenswrapper[4772]: I0127 17:18:33.663906 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:18:33 crc kubenswrapper[4772]: E0127 17:18:33.664822 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:18:44 crc kubenswrapper[4772]: I0127 17:18:44.676072 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:18:44 crc kubenswrapper[4772]: E0127 17:18:44.677465 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:18:58 crc kubenswrapper[4772]: I0127 17:18:58.664119 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:18:58 crc kubenswrapper[4772]: E0127 17:18:58.665361 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:19:08 crc kubenswrapper[4772]: I0127 17:19:08.738606 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sdbgg"] Jan 27 17:19:08 crc kubenswrapper[4772]: E0127 17:19:08.740951 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0db9aea1-8a66-4447-9150-812172da2a26" containerName="collect-profiles" Jan 27 17:19:08 crc kubenswrapper[4772]: I0127 17:19:08.741058 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0db9aea1-8a66-4447-9150-812172da2a26" containerName="collect-profiles" Jan 27 17:19:08 crc kubenswrapper[4772]: I0127 17:19:08.741407 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="0db9aea1-8a66-4447-9150-812172da2a26" containerName="collect-profiles" Jan 27 17:19:08 crc kubenswrapper[4772]: I0127 17:19:08.743673 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdbgg" Jan 27 17:19:08 crc kubenswrapper[4772]: I0127 17:19:08.750752 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sdbgg"] Jan 27 17:19:08 crc kubenswrapper[4772]: I0127 17:19:08.758733 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09211095-3894-4db1-bcea-29d1c2064979-utilities\") pod \"redhat-operators-sdbgg\" (UID: \"09211095-3894-4db1-bcea-29d1c2064979\") " pod="openshift-marketplace/redhat-operators-sdbgg" Jan 27 17:19:08 crc kubenswrapper[4772]: I0127 17:19:08.758798 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpgqr\" (UniqueName: \"kubernetes.io/projected/09211095-3894-4db1-bcea-29d1c2064979-kube-api-access-bpgqr\") pod \"redhat-operators-sdbgg\" (UID: \"09211095-3894-4db1-bcea-29d1c2064979\") " pod="openshift-marketplace/redhat-operators-sdbgg" Jan 27 17:19:08 crc kubenswrapper[4772]: I0127 17:19:08.758945 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09211095-3894-4db1-bcea-29d1c2064979-catalog-content\") pod \"redhat-operators-sdbgg\" (UID: \"09211095-3894-4db1-bcea-29d1c2064979\") " pod="openshift-marketplace/redhat-operators-sdbgg" Jan 27 17:19:08 crc kubenswrapper[4772]: I0127 17:19:08.860997 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09211095-3894-4db1-bcea-29d1c2064979-utilities\") pod \"redhat-operators-sdbgg\" (UID: \"09211095-3894-4db1-bcea-29d1c2064979\") " pod="openshift-marketplace/redhat-operators-sdbgg" Jan 27 17:19:08 crc kubenswrapper[4772]: I0127 17:19:08.861052 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpgqr\" (UniqueName: \"kubernetes.io/projected/09211095-3894-4db1-bcea-29d1c2064979-kube-api-access-bpgqr\") pod \"redhat-operators-sdbgg\" (UID: \"09211095-3894-4db1-bcea-29d1c2064979\") " pod="openshift-marketplace/redhat-operators-sdbgg" Jan 27 17:19:08 crc kubenswrapper[4772]: I0127 17:19:08.861129 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09211095-3894-4db1-bcea-29d1c2064979-catalog-content\") pod \"redhat-operators-sdbgg\" (UID: \"09211095-3894-4db1-bcea-29d1c2064979\") " pod="openshift-marketplace/redhat-operators-sdbgg" Jan 27 17:19:08 crc kubenswrapper[4772]: I0127 17:19:08.861641 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09211095-3894-4db1-bcea-29d1c2064979-utilities\") pod \"redhat-operators-sdbgg\" (UID: \"09211095-3894-4db1-bcea-29d1c2064979\") " pod="openshift-marketplace/redhat-operators-sdbgg" Jan 27 17:19:08 crc kubenswrapper[4772]: I0127 17:19:08.861688 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09211095-3894-4db1-bcea-29d1c2064979-catalog-content\") pod \"redhat-operators-sdbgg\" (UID: \"09211095-3894-4db1-bcea-29d1c2064979\") " pod="openshift-marketplace/redhat-operators-sdbgg" Jan 27 17:19:08 crc kubenswrapper[4772]: I0127 17:19:08.882526 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpgqr\" (UniqueName: \"kubernetes.io/projected/09211095-3894-4db1-bcea-29d1c2064979-kube-api-access-bpgqr\") pod \"redhat-operators-sdbgg\" (UID: \"09211095-3894-4db1-bcea-29d1c2064979\") " pod="openshift-marketplace/redhat-operators-sdbgg" Jan 27 17:19:09 crc kubenswrapper[4772]: I0127 17:19:09.081084 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdbgg" Jan 27 17:19:09 crc kubenswrapper[4772]: I0127 17:19:09.559038 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sdbgg"] Jan 27 17:19:09 crc kubenswrapper[4772]: I0127 17:19:09.993490 4772 generic.go:334] "Generic (PLEG): container finished" podID="09211095-3894-4db1-bcea-29d1c2064979" containerID="c7d1aeb8760ccc3859f31c865946184e19dcfb95f1b5a51ede194d6b5159fd54" exitCode=0 Jan 27 17:19:09 crc kubenswrapper[4772]: I0127 17:19:09.993587 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdbgg" event={"ID":"09211095-3894-4db1-bcea-29d1c2064979","Type":"ContainerDied","Data":"c7d1aeb8760ccc3859f31c865946184e19dcfb95f1b5a51ede194d6b5159fd54"} Jan 27 17:19:09 crc kubenswrapper[4772]: I0127 17:19:09.993904 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdbgg" event={"ID":"09211095-3894-4db1-bcea-29d1c2064979","Type":"ContainerStarted","Data":"7b9104e2450ae40c8ccedbf7fb5718b06fc3fd6643498b6b6b5fcd64d1d1eed1"} Jan 27 17:19:10 crc kubenswrapper[4772]: I0127 17:19:10.664894 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:19:10 crc kubenswrapper[4772]: E0127 17:19:10.665979 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:19:19 crc kubenswrapper[4772]: I0127 17:19:19.079048 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdbgg" event={"ID":"09211095-3894-4db1-bcea-29d1c2064979","Type":"ContainerStarted","Data":"03288c1630ff3a4aaf5cd2be5b75940caff281cc957f7c9a00544bea93505a5f"} Jan 27 17:19:21 crc kubenswrapper[4772]: I0127 17:19:21.099443 4772 generic.go:334] "Generic (PLEG): container finished" podID="09211095-3894-4db1-bcea-29d1c2064979" containerID="03288c1630ff3a4aaf5cd2be5b75940caff281cc957f7c9a00544bea93505a5f" exitCode=0 Jan 27 17:19:21 crc kubenswrapper[4772]: I0127 17:19:21.099502 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdbgg" event={"ID":"09211095-3894-4db1-bcea-29d1c2064979","Type":"ContainerDied","Data":"03288c1630ff3a4aaf5cd2be5b75940caff281cc957f7c9a00544bea93505a5f"} Jan 27 17:19:21 crc kubenswrapper[4772]: I0127 17:19:21.102691 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 17:19:22 crc kubenswrapper[4772]: I0127 17:19:22.110035 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdbgg" event={"ID":"09211095-3894-4db1-bcea-29d1c2064979","Type":"ContainerStarted","Data":"9a0674be9539d95a9bb4caa1bb6e380e96fd25b87da2cf087c1d397de11bd988"} Jan 27 17:19:22 crc kubenswrapper[4772]: I0127 17:19:22.132060 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sdbgg" podStartSLOduration=2.382977571 podStartE2EDuration="14.132037977s" podCreationTimestamp="2026-01-27 17:19:08 +0000 UTC" firstStartedPulling="2026-01-27 17:19:09.994950226 +0000 UTC m=+7935.975559324" lastFinishedPulling="2026-01-27 17:19:21.744010632 +0000 UTC m=+7947.724619730" observedRunningTime="2026-01-27 17:19:22.127105597 +0000 UTC m=+7948.107714685" watchObservedRunningTime="2026-01-27 17:19:22.132037977 +0000 UTC m=+7948.112647065" Jan 27 17:19:24 crc kubenswrapper[4772]: I0127 17:19:24.675925 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:19:24 crc kubenswrapper[4772]: E0127 17:19:24.676545 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:19:29 crc kubenswrapper[4772]: I0127 17:19:29.081626 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sdbgg" Jan 27 17:19:29 crc kubenswrapper[4772]: I0127 17:19:29.082135 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sdbgg" Jan 27 17:19:29 crc kubenswrapper[4772]: I0127 17:19:29.128403 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sdbgg" Jan 27 17:19:29 crc kubenswrapper[4772]: I0127 17:19:29.223576 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sdbgg" Jan 27 17:19:29 crc kubenswrapper[4772]: I0127 17:19:29.286838 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sdbgg"] Jan 27 17:19:29 crc kubenswrapper[4772]: I0127 17:19:29.364216 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5whzm"] Jan 27 17:19:29 crc kubenswrapper[4772]: I0127 17:19:29.364491 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5whzm" podUID="79b85747-dcbc-462d-85d1-3d00801b5106" containerName="registry-server" containerID="cri-o://a2d56215a8255cd6fe0e2caab61dc965898bd5a40e9690f3cc4fda7a7884bd0c" gracePeriod=2 Jan 27 17:19:29 crc kubenswrapper[4772]: E0127 17:19:29.758674 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a2d56215a8255cd6fe0e2caab61dc965898bd5a40e9690f3cc4fda7a7884bd0c is running failed: container process not found" containerID="a2d56215a8255cd6fe0e2caab61dc965898bd5a40e9690f3cc4fda7a7884bd0c" cmd=["grpc_health_probe","-addr=:50051"] Jan 27 17:19:29 crc kubenswrapper[4772]: E0127 17:19:29.759495 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a2d56215a8255cd6fe0e2caab61dc965898bd5a40e9690f3cc4fda7a7884bd0c is running failed: container process not found" containerID="a2d56215a8255cd6fe0e2caab61dc965898bd5a40e9690f3cc4fda7a7884bd0c" cmd=["grpc_health_probe","-addr=:50051"] Jan 27 17:19:29 crc kubenswrapper[4772]: E0127 17:19:29.759861 4772 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a2d56215a8255cd6fe0e2caab61dc965898bd5a40e9690f3cc4fda7a7884bd0c is running failed: container process not found" containerID="a2d56215a8255cd6fe0e2caab61dc965898bd5a40e9690f3cc4fda7a7884bd0c" cmd=["grpc_health_probe","-addr=:50051"] Jan 27 17:19:29 crc kubenswrapper[4772]: E0127 17:19:29.759933 4772 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a2d56215a8255cd6fe0e2caab61dc965898bd5a40e9690f3cc4fda7a7884bd0c is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-5whzm" podUID="79b85747-dcbc-462d-85d1-3d00801b5106" containerName="registry-server" Jan 27 17:19:29 crc kubenswrapper[4772]: I0127 17:19:29.813820 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5whzm" Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.014558 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79b85747-dcbc-462d-85d1-3d00801b5106-utilities\") pod \"79b85747-dcbc-462d-85d1-3d00801b5106\" (UID: \"79b85747-dcbc-462d-85d1-3d00801b5106\") " Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.014638 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79b85747-dcbc-462d-85d1-3d00801b5106-catalog-content\") pod \"79b85747-dcbc-462d-85d1-3d00801b5106\" (UID: \"79b85747-dcbc-462d-85d1-3d00801b5106\") " Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.014685 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vr9pf\" (UniqueName: \"kubernetes.io/projected/79b85747-dcbc-462d-85d1-3d00801b5106-kube-api-access-vr9pf\") pod \"79b85747-dcbc-462d-85d1-3d00801b5106\" (UID: \"79b85747-dcbc-462d-85d1-3d00801b5106\") " Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.015093 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79b85747-dcbc-462d-85d1-3d00801b5106-utilities" (OuterVolumeSpecName: "utilities") pod "79b85747-dcbc-462d-85d1-3d00801b5106" (UID: "79b85747-dcbc-462d-85d1-3d00801b5106"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.015238 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79b85747-dcbc-462d-85d1-3d00801b5106-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.021221 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79b85747-dcbc-462d-85d1-3d00801b5106-kube-api-access-vr9pf" (OuterVolumeSpecName: "kube-api-access-vr9pf") pod "79b85747-dcbc-462d-85d1-3d00801b5106" (UID: "79b85747-dcbc-462d-85d1-3d00801b5106"). InnerVolumeSpecName "kube-api-access-vr9pf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.116505 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vr9pf\" (UniqueName: \"kubernetes.io/projected/79b85747-dcbc-462d-85d1-3d00801b5106-kube-api-access-vr9pf\") on node \"crc\" DevicePath \"\"" Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.133854 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79b85747-dcbc-462d-85d1-3d00801b5106-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "79b85747-dcbc-462d-85d1-3d00801b5106" (UID: "79b85747-dcbc-462d-85d1-3d00801b5106"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.183027 4772 generic.go:334] "Generic (PLEG): container finished" podID="79b85747-dcbc-462d-85d1-3d00801b5106" containerID="a2d56215a8255cd6fe0e2caab61dc965898bd5a40e9690f3cc4fda7a7884bd0c" exitCode=0 Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.183070 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5whzm" event={"ID":"79b85747-dcbc-462d-85d1-3d00801b5106","Type":"ContainerDied","Data":"a2d56215a8255cd6fe0e2caab61dc965898bd5a40e9690f3cc4fda7a7884bd0c"} Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.183130 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5whzm" Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.183139 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5whzm" event={"ID":"79b85747-dcbc-462d-85d1-3d00801b5106","Type":"ContainerDied","Data":"8e4bc519d0ced9952d6857ff31015675ce4705ed12ce2547ebdba49c33fc4d62"} Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.183161 4772 scope.go:117] "RemoveContainer" containerID="a2d56215a8255cd6fe0e2caab61dc965898bd5a40e9690f3cc4fda7a7884bd0c" Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.212268 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5whzm"] Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.212281 4772 scope.go:117] "RemoveContainer" containerID="eb48a00f974d6ceae0f8e67b809e0845663f47278e4609e58f35d13935712ff7" Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.218404 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79b85747-dcbc-462d-85d1-3d00801b5106-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.219239 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5whzm"] Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.235316 4772 scope.go:117] "RemoveContainer" containerID="d2c52e57f45d6596ec80d58984736144fe9463319975ceee95db9f926e0b38f1" Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.279994 4772 scope.go:117] "RemoveContainer" containerID="a2d56215a8255cd6fe0e2caab61dc965898bd5a40e9690f3cc4fda7a7884bd0c" Jan 27 17:19:30 crc kubenswrapper[4772]: E0127 17:19:30.280471 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2d56215a8255cd6fe0e2caab61dc965898bd5a40e9690f3cc4fda7a7884bd0c\": container with ID starting with a2d56215a8255cd6fe0e2caab61dc965898bd5a40e9690f3cc4fda7a7884bd0c not found: ID does not exist" containerID="a2d56215a8255cd6fe0e2caab61dc965898bd5a40e9690f3cc4fda7a7884bd0c" Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.280514 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2d56215a8255cd6fe0e2caab61dc965898bd5a40e9690f3cc4fda7a7884bd0c"} err="failed to get container status \"a2d56215a8255cd6fe0e2caab61dc965898bd5a40e9690f3cc4fda7a7884bd0c\": rpc error: code = NotFound desc = could not find container \"a2d56215a8255cd6fe0e2caab61dc965898bd5a40e9690f3cc4fda7a7884bd0c\": container with ID starting with a2d56215a8255cd6fe0e2caab61dc965898bd5a40e9690f3cc4fda7a7884bd0c not found: ID does not exist" Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.280543 4772 scope.go:117] "RemoveContainer" containerID="eb48a00f974d6ceae0f8e67b809e0845663f47278e4609e58f35d13935712ff7" Jan 27 17:19:30 crc kubenswrapper[4772]: E0127 17:19:30.281006 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb48a00f974d6ceae0f8e67b809e0845663f47278e4609e58f35d13935712ff7\": container with ID starting with eb48a00f974d6ceae0f8e67b809e0845663f47278e4609e58f35d13935712ff7 not found: ID does not exist" containerID="eb48a00f974d6ceae0f8e67b809e0845663f47278e4609e58f35d13935712ff7" Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.281047 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb48a00f974d6ceae0f8e67b809e0845663f47278e4609e58f35d13935712ff7"} err="failed to get container status \"eb48a00f974d6ceae0f8e67b809e0845663f47278e4609e58f35d13935712ff7\": rpc error: code = NotFound desc = could not find container \"eb48a00f974d6ceae0f8e67b809e0845663f47278e4609e58f35d13935712ff7\": container with ID starting with eb48a00f974d6ceae0f8e67b809e0845663f47278e4609e58f35d13935712ff7 not found: ID does not exist" Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.281074 4772 scope.go:117] "RemoveContainer" containerID="d2c52e57f45d6596ec80d58984736144fe9463319975ceee95db9f926e0b38f1" Jan 27 17:19:30 crc kubenswrapper[4772]: E0127 17:19:30.281378 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2c52e57f45d6596ec80d58984736144fe9463319975ceee95db9f926e0b38f1\": container with ID starting with d2c52e57f45d6596ec80d58984736144fe9463319975ceee95db9f926e0b38f1 not found: ID does not exist" containerID="d2c52e57f45d6596ec80d58984736144fe9463319975ceee95db9f926e0b38f1" Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.281402 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2c52e57f45d6596ec80d58984736144fe9463319975ceee95db9f926e0b38f1"} err="failed to get container status \"d2c52e57f45d6596ec80d58984736144fe9463319975ceee95db9f926e0b38f1\": rpc error: code = NotFound desc = could not find container \"d2c52e57f45d6596ec80d58984736144fe9463319975ceee95db9f926e0b38f1\": container with ID starting with d2c52e57f45d6596ec80d58984736144fe9463319975ceee95db9f926e0b38f1 not found: ID does not exist" Jan 27 17:19:30 crc kubenswrapper[4772]: I0127 17:19:30.674984 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79b85747-dcbc-462d-85d1-3d00801b5106" path="/var/lib/kubelet/pods/79b85747-dcbc-462d-85d1-3d00801b5106/volumes" Jan 27 17:19:39 crc kubenswrapper[4772]: I0127 17:19:39.663570 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:19:39 crc kubenswrapper[4772]: E0127 17:19:39.664897 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:19:52 crc kubenswrapper[4772]: I0127 17:19:52.672070 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:19:52 crc kubenswrapper[4772]: E0127 17:19:52.672987 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:20:03 crc kubenswrapper[4772]: I0127 17:20:03.663711 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:20:03 crc kubenswrapper[4772]: E0127 17:20:03.664857 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:20:16 crc kubenswrapper[4772]: I0127 17:20:16.664057 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:20:16 crc kubenswrapper[4772]: E0127 17:20:16.667039 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:20:31 crc kubenswrapper[4772]: I0127 17:20:31.663567 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:20:31 crc kubenswrapper[4772]: E0127 17:20:31.664787 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:20:43 crc kubenswrapper[4772]: I0127 17:20:43.663023 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:20:43 crc kubenswrapper[4772]: E0127 17:20:43.663895 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:20:57 crc kubenswrapper[4772]: I0127 17:20:57.663581 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:20:57 crc kubenswrapper[4772]: E0127 17:20:57.664478 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:21:11 crc kubenswrapper[4772]: I0127 17:21:11.662557 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:21:11 crc kubenswrapper[4772]: E0127 17:21:11.663218 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:21:26 crc kubenswrapper[4772]: I0127 17:21:26.663724 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:21:27 crc kubenswrapper[4772]: I0127 17:21:27.279926 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"a0596c152e2f2b0802267f33c0f2d3224ef4bcfda5d22940bb1eb3256403bf5f"} Jan 27 17:22:33 crc kubenswrapper[4772]: I0127 17:22:33.634820 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4xvdq"] Jan 27 17:22:33 crc kubenswrapper[4772]: E0127 17:22:33.635862 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79b85747-dcbc-462d-85d1-3d00801b5106" containerName="extract-utilities" Jan 27 17:22:33 crc kubenswrapper[4772]: I0127 17:22:33.635881 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="79b85747-dcbc-462d-85d1-3d00801b5106" containerName="extract-utilities" Jan 27 17:22:33 crc kubenswrapper[4772]: E0127 17:22:33.635911 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79b85747-dcbc-462d-85d1-3d00801b5106" containerName="extract-content" Jan 27 17:22:33 crc kubenswrapper[4772]: I0127 17:22:33.635920 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="79b85747-dcbc-462d-85d1-3d00801b5106" containerName="extract-content" Jan 27 17:22:33 crc kubenswrapper[4772]: E0127 17:22:33.635934 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79b85747-dcbc-462d-85d1-3d00801b5106" containerName="registry-server" Jan 27 17:22:33 crc kubenswrapper[4772]: I0127 17:22:33.635943 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="79b85747-dcbc-462d-85d1-3d00801b5106" containerName="registry-server" Jan 27 17:22:33 crc kubenswrapper[4772]: I0127 17:22:33.636157 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="79b85747-dcbc-462d-85d1-3d00801b5106" containerName="registry-server" Jan 27 17:22:33 crc kubenswrapper[4772]: I0127 17:22:33.637836 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4xvdq" Jan 27 17:22:33 crc kubenswrapper[4772]: I0127 17:22:33.646380 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4xvdq"] Jan 27 17:22:33 crc kubenswrapper[4772]: I0127 17:22:33.728411 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b2a1877-960b-4e44-8e6d-47744d3e764b-utilities\") pod \"redhat-marketplace-4xvdq\" (UID: \"6b2a1877-960b-4e44-8e6d-47744d3e764b\") " pod="openshift-marketplace/redhat-marketplace-4xvdq" Jan 27 17:22:33 crc kubenswrapper[4772]: I0127 17:22:33.728504 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fcfk\" (UniqueName: \"kubernetes.io/projected/6b2a1877-960b-4e44-8e6d-47744d3e764b-kube-api-access-9fcfk\") pod \"redhat-marketplace-4xvdq\" (UID: \"6b2a1877-960b-4e44-8e6d-47744d3e764b\") " pod="openshift-marketplace/redhat-marketplace-4xvdq" Jan 27 17:22:33 crc kubenswrapper[4772]: I0127 17:22:33.728594 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b2a1877-960b-4e44-8e6d-47744d3e764b-catalog-content\") pod \"redhat-marketplace-4xvdq\" (UID: \"6b2a1877-960b-4e44-8e6d-47744d3e764b\") " pod="openshift-marketplace/redhat-marketplace-4xvdq" Jan 27 17:22:33 crc kubenswrapper[4772]: I0127 17:22:33.831006 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b2a1877-960b-4e44-8e6d-47744d3e764b-utilities\") pod \"redhat-marketplace-4xvdq\" (UID: \"6b2a1877-960b-4e44-8e6d-47744d3e764b\") " pod="openshift-marketplace/redhat-marketplace-4xvdq" Jan 27 17:22:33 crc kubenswrapper[4772]: I0127 17:22:33.831087 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fcfk\" (UniqueName: \"kubernetes.io/projected/6b2a1877-960b-4e44-8e6d-47744d3e764b-kube-api-access-9fcfk\") pod \"redhat-marketplace-4xvdq\" (UID: \"6b2a1877-960b-4e44-8e6d-47744d3e764b\") " pod="openshift-marketplace/redhat-marketplace-4xvdq" Jan 27 17:22:33 crc kubenswrapper[4772]: I0127 17:22:33.831146 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b2a1877-960b-4e44-8e6d-47744d3e764b-catalog-content\") pod \"redhat-marketplace-4xvdq\" (UID: \"6b2a1877-960b-4e44-8e6d-47744d3e764b\") " pod="openshift-marketplace/redhat-marketplace-4xvdq" Jan 27 17:22:33 crc kubenswrapper[4772]: I0127 17:22:33.831688 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b2a1877-960b-4e44-8e6d-47744d3e764b-catalog-content\") pod \"redhat-marketplace-4xvdq\" (UID: \"6b2a1877-960b-4e44-8e6d-47744d3e764b\") " pod="openshift-marketplace/redhat-marketplace-4xvdq" Jan 27 17:22:33 crc kubenswrapper[4772]: I0127 17:22:33.832288 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b2a1877-960b-4e44-8e6d-47744d3e764b-utilities\") pod \"redhat-marketplace-4xvdq\" (UID: \"6b2a1877-960b-4e44-8e6d-47744d3e764b\") " pod="openshift-marketplace/redhat-marketplace-4xvdq" Jan 27 17:22:33 crc kubenswrapper[4772]: I0127 17:22:33.852870 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fcfk\" (UniqueName: \"kubernetes.io/projected/6b2a1877-960b-4e44-8e6d-47744d3e764b-kube-api-access-9fcfk\") pod \"redhat-marketplace-4xvdq\" (UID: \"6b2a1877-960b-4e44-8e6d-47744d3e764b\") " pod="openshift-marketplace/redhat-marketplace-4xvdq" Jan 27 17:22:33 crc kubenswrapper[4772]: I0127 17:22:33.954320 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4xvdq" Jan 27 17:22:34 crc kubenswrapper[4772]: I0127 17:22:34.489937 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4xvdq"] Jan 27 17:22:34 crc kubenswrapper[4772]: I0127 17:22:34.971079 4772 generic.go:334] "Generic (PLEG): container finished" podID="6b2a1877-960b-4e44-8e6d-47744d3e764b" containerID="08f7f6255d2bb42989667d0a5e1713a33ba4ba31840b81d1925d49697b9af8c6" exitCode=0 Jan 27 17:22:34 crc kubenswrapper[4772]: I0127 17:22:34.971124 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4xvdq" event={"ID":"6b2a1877-960b-4e44-8e6d-47744d3e764b","Type":"ContainerDied","Data":"08f7f6255d2bb42989667d0a5e1713a33ba4ba31840b81d1925d49697b9af8c6"} Jan 27 17:22:34 crc kubenswrapper[4772]: I0127 17:22:34.971154 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4xvdq" event={"ID":"6b2a1877-960b-4e44-8e6d-47744d3e764b","Type":"ContainerStarted","Data":"aee6dea538b507f3e6d4df03a2d19ed1bfe49288afa763d30e38e9301cce0362"} Jan 27 17:22:35 crc kubenswrapper[4772]: I0127 17:22:35.982715 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4xvdq" event={"ID":"6b2a1877-960b-4e44-8e6d-47744d3e764b","Type":"ContainerStarted","Data":"e682f9deaaadeb63ad6e4ae31c623a39a070781b1ba79a28bca994caba62c19f"} Jan 27 17:22:36 crc kubenswrapper[4772]: I0127 17:22:36.995907 4772 generic.go:334] "Generic (PLEG): container finished" podID="6b2a1877-960b-4e44-8e6d-47744d3e764b" containerID="e682f9deaaadeb63ad6e4ae31c623a39a070781b1ba79a28bca994caba62c19f" exitCode=0 Jan 27 17:22:36 crc kubenswrapper[4772]: I0127 17:22:36.996161 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4xvdq" event={"ID":"6b2a1877-960b-4e44-8e6d-47744d3e764b","Type":"ContainerDied","Data":"e682f9deaaadeb63ad6e4ae31c623a39a070781b1ba79a28bca994caba62c19f"} Jan 27 17:22:37 crc kubenswrapper[4772]: I0127 17:22:37.829344 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-c894q"] Jan 27 17:22:37 crc kubenswrapper[4772]: I0127 17:22:37.831620 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c894q" Jan 27 17:22:37 crc kubenswrapper[4772]: I0127 17:22:37.843365 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c894q"] Jan 27 17:22:37 crc kubenswrapper[4772]: I0127 17:22:37.918998 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79767c4a-3d92-40a5-8128-a9b7785d4672-catalog-content\") pod \"community-operators-c894q\" (UID: \"79767c4a-3d92-40a5-8128-a9b7785d4672\") " pod="openshift-marketplace/community-operators-c894q" Jan 27 17:22:37 crc kubenswrapper[4772]: I0127 17:22:37.919115 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsv8q\" (UniqueName: \"kubernetes.io/projected/79767c4a-3d92-40a5-8128-a9b7785d4672-kube-api-access-vsv8q\") pod \"community-operators-c894q\" (UID: \"79767c4a-3d92-40a5-8128-a9b7785d4672\") " pod="openshift-marketplace/community-operators-c894q" Jan 27 17:22:37 crc kubenswrapper[4772]: I0127 17:22:37.919151 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79767c4a-3d92-40a5-8128-a9b7785d4672-utilities\") pod \"community-operators-c894q\" (UID: \"79767c4a-3d92-40a5-8128-a9b7785d4672\") " pod="openshift-marketplace/community-operators-c894q" Jan 27 17:22:38 crc kubenswrapper[4772]: I0127 17:22:38.020197 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79767c4a-3d92-40a5-8128-a9b7785d4672-catalog-content\") pod \"community-operators-c894q\" (UID: \"79767c4a-3d92-40a5-8128-a9b7785d4672\") " pod="openshift-marketplace/community-operators-c894q" Jan 27 17:22:38 crc kubenswrapper[4772]: I0127 17:22:38.020237 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4xvdq" event={"ID":"6b2a1877-960b-4e44-8e6d-47744d3e764b","Type":"ContainerStarted","Data":"313eb3ad4fc2fcb56fe6d4828f016dd448a0cc926ce0e656cd2cc46796002ab2"} Jan 27 17:22:38 crc kubenswrapper[4772]: I0127 17:22:38.020600 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsv8q\" (UniqueName: \"kubernetes.io/projected/79767c4a-3d92-40a5-8128-a9b7785d4672-kube-api-access-vsv8q\") pod \"community-operators-c894q\" (UID: \"79767c4a-3d92-40a5-8128-a9b7785d4672\") " pod="openshift-marketplace/community-operators-c894q" Jan 27 17:22:38 crc kubenswrapper[4772]: I0127 17:22:38.020633 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79767c4a-3d92-40a5-8128-a9b7785d4672-utilities\") pod \"community-operators-c894q\" (UID: \"79767c4a-3d92-40a5-8128-a9b7785d4672\") " pod="openshift-marketplace/community-operators-c894q" Jan 27 17:22:38 crc kubenswrapper[4772]: I0127 17:22:38.020671 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79767c4a-3d92-40a5-8128-a9b7785d4672-catalog-content\") pod \"community-operators-c894q\" (UID: \"79767c4a-3d92-40a5-8128-a9b7785d4672\") " pod="openshift-marketplace/community-operators-c894q" Jan 27 17:22:38 crc kubenswrapper[4772]: I0127 17:22:38.021564 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79767c4a-3d92-40a5-8128-a9b7785d4672-utilities\") pod \"community-operators-c894q\" (UID: \"79767c4a-3d92-40a5-8128-a9b7785d4672\") " pod="openshift-marketplace/community-operators-c894q" Jan 27 17:22:38 crc kubenswrapper[4772]: I0127 17:22:38.041936 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsv8q\" (UniqueName: \"kubernetes.io/projected/79767c4a-3d92-40a5-8128-a9b7785d4672-kube-api-access-vsv8q\") pod \"community-operators-c894q\" (UID: \"79767c4a-3d92-40a5-8128-a9b7785d4672\") " pod="openshift-marketplace/community-operators-c894q" Jan 27 17:22:38 crc kubenswrapper[4772]: I0127 17:22:38.048279 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4xvdq" podStartSLOduration=2.5936443909999998 podStartE2EDuration="5.048262827s" podCreationTimestamp="2026-01-27 17:22:33 +0000 UTC" firstStartedPulling="2026-01-27 17:22:34.973435452 +0000 UTC m=+8140.954044560" lastFinishedPulling="2026-01-27 17:22:37.428053858 +0000 UTC m=+8143.408662996" observedRunningTime="2026-01-27 17:22:38.039234861 +0000 UTC m=+8144.019843979" watchObservedRunningTime="2026-01-27 17:22:38.048262827 +0000 UTC m=+8144.028871915" Jan 27 17:22:38 crc kubenswrapper[4772]: I0127 17:22:38.158239 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c894q" Jan 27 17:22:38 crc kubenswrapper[4772]: I0127 17:22:38.726001 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c894q"] Jan 27 17:22:38 crc kubenswrapper[4772]: W0127 17:22:38.726467 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79767c4a_3d92_40a5_8128_a9b7785d4672.slice/crio-64cc0b646724882690474db4430de2f2239d9f2a6e5ce0a0dcddb5c0fc527179 WatchSource:0}: Error finding container 64cc0b646724882690474db4430de2f2239d9f2a6e5ce0a0dcddb5c0fc527179: Status 404 returned error can't find the container with id 64cc0b646724882690474db4430de2f2239d9f2a6e5ce0a0dcddb5c0fc527179 Jan 27 17:22:39 crc kubenswrapper[4772]: I0127 17:22:39.031837 4772 generic.go:334] "Generic (PLEG): container finished" podID="79767c4a-3d92-40a5-8128-a9b7785d4672" containerID="b7996be57a39eb18a216e3d76b272bae729bf867e09dc88e02c79f2bbf29e2c0" exitCode=0 Jan 27 17:22:39 crc kubenswrapper[4772]: I0127 17:22:39.031965 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c894q" event={"ID":"79767c4a-3d92-40a5-8128-a9b7785d4672","Type":"ContainerDied","Data":"b7996be57a39eb18a216e3d76b272bae729bf867e09dc88e02c79f2bbf29e2c0"} Jan 27 17:22:39 crc kubenswrapper[4772]: I0127 17:22:39.032212 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c894q" event={"ID":"79767c4a-3d92-40a5-8128-a9b7785d4672","Type":"ContainerStarted","Data":"64cc0b646724882690474db4430de2f2239d9f2a6e5ce0a0dcddb5c0fc527179"} Jan 27 17:22:41 crc kubenswrapper[4772]: I0127 17:22:41.053926 4772 generic.go:334] "Generic (PLEG): container finished" podID="79767c4a-3d92-40a5-8128-a9b7785d4672" containerID="931a3ea833e76451e7ae9443d7919ff52b6d3c292558ce3eac47358e7f05fdad" exitCode=0 Jan 27 17:22:41 crc kubenswrapper[4772]: I0127 17:22:41.054012 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c894q" event={"ID":"79767c4a-3d92-40a5-8128-a9b7785d4672","Type":"ContainerDied","Data":"931a3ea833e76451e7ae9443d7919ff52b6d3c292558ce3eac47358e7f05fdad"} Jan 27 17:22:42 crc kubenswrapper[4772]: I0127 17:22:42.066571 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c894q" event={"ID":"79767c4a-3d92-40a5-8128-a9b7785d4672","Type":"ContainerStarted","Data":"227375896e50ca3ae2a98fd56581e6e897945772df9f8a7d35dcddb59e16d22a"} Jan 27 17:22:42 crc kubenswrapper[4772]: I0127 17:22:42.099097 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-c894q" podStartSLOduration=2.420509466 podStartE2EDuration="5.099072773s" podCreationTimestamp="2026-01-27 17:22:37 +0000 UTC" firstStartedPulling="2026-01-27 17:22:39.034523379 +0000 UTC m=+8145.015132487" lastFinishedPulling="2026-01-27 17:22:41.713086656 +0000 UTC m=+8147.693695794" observedRunningTime="2026-01-27 17:22:42.096549011 +0000 UTC m=+8148.077158139" watchObservedRunningTime="2026-01-27 17:22:42.099072773 +0000 UTC m=+8148.079681881" Jan 27 17:22:43 crc kubenswrapper[4772]: I0127 17:22:43.956423 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4xvdq" Jan 27 17:22:43 crc kubenswrapper[4772]: I0127 17:22:43.958103 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4xvdq" Jan 27 17:22:44 crc kubenswrapper[4772]: I0127 17:22:44.040757 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4xvdq" Jan 27 17:22:44 crc kubenswrapper[4772]: I0127 17:22:44.136771 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4xvdq" Jan 27 17:22:45 crc kubenswrapper[4772]: I0127 17:22:45.425679 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4xvdq"] Jan 27 17:22:46 crc kubenswrapper[4772]: I0127 17:22:46.102070 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4xvdq" podUID="6b2a1877-960b-4e44-8e6d-47744d3e764b" containerName="registry-server" containerID="cri-o://313eb3ad4fc2fcb56fe6d4828f016dd448a0cc926ce0e656cd2cc46796002ab2" gracePeriod=2 Jan 27 17:22:46 crc kubenswrapper[4772]: I0127 17:22:46.595438 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4xvdq" Jan 27 17:22:46 crc kubenswrapper[4772]: I0127 17:22:46.733506 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b2a1877-960b-4e44-8e6d-47744d3e764b-utilities\") pod \"6b2a1877-960b-4e44-8e6d-47744d3e764b\" (UID: \"6b2a1877-960b-4e44-8e6d-47744d3e764b\") " Jan 27 17:22:46 crc kubenswrapper[4772]: I0127 17:22:46.733590 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b2a1877-960b-4e44-8e6d-47744d3e764b-catalog-content\") pod \"6b2a1877-960b-4e44-8e6d-47744d3e764b\" (UID: \"6b2a1877-960b-4e44-8e6d-47744d3e764b\") " Jan 27 17:22:46 crc kubenswrapper[4772]: I0127 17:22:46.733751 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fcfk\" (UniqueName: \"kubernetes.io/projected/6b2a1877-960b-4e44-8e6d-47744d3e764b-kube-api-access-9fcfk\") pod \"6b2a1877-960b-4e44-8e6d-47744d3e764b\" (UID: \"6b2a1877-960b-4e44-8e6d-47744d3e764b\") " Jan 27 17:22:46 crc kubenswrapper[4772]: I0127 17:22:46.734884 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b2a1877-960b-4e44-8e6d-47744d3e764b-utilities" (OuterVolumeSpecName: "utilities") pod "6b2a1877-960b-4e44-8e6d-47744d3e764b" (UID: "6b2a1877-960b-4e44-8e6d-47744d3e764b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:22:46 crc kubenswrapper[4772]: I0127 17:22:46.742916 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b2a1877-960b-4e44-8e6d-47744d3e764b-kube-api-access-9fcfk" (OuterVolumeSpecName: "kube-api-access-9fcfk") pod "6b2a1877-960b-4e44-8e6d-47744d3e764b" (UID: "6b2a1877-960b-4e44-8e6d-47744d3e764b"). InnerVolumeSpecName "kube-api-access-9fcfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:22:46 crc kubenswrapper[4772]: I0127 17:22:46.782964 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b2a1877-960b-4e44-8e6d-47744d3e764b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b2a1877-960b-4e44-8e6d-47744d3e764b" (UID: "6b2a1877-960b-4e44-8e6d-47744d3e764b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:22:46 crc kubenswrapper[4772]: I0127 17:22:46.839898 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b2a1877-960b-4e44-8e6d-47744d3e764b-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 17:22:46 crc kubenswrapper[4772]: I0127 17:22:46.840139 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b2a1877-960b-4e44-8e6d-47744d3e764b-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 17:22:46 crc kubenswrapper[4772]: I0127 17:22:46.840323 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fcfk\" (UniqueName: \"kubernetes.io/projected/6b2a1877-960b-4e44-8e6d-47744d3e764b-kube-api-access-9fcfk\") on node \"crc\" DevicePath \"\"" Jan 27 17:22:47 crc kubenswrapper[4772]: I0127 17:22:47.117406 4772 generic.go:334] "Generic (PLEG): container finished" podID="6b2a1877-960b-4e44-8e6d-47744d3e764b" containerID="313eb3ad4fc2fcb56fe6d4828f016dd448a0cc926ce0e656cd2cc46796002ab2" exitCode=0 Jan 27 17:22:47 crc kubenswrapper[4772]: I0127 17:22:47.117474 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4xvdq" event={"ID":"6b2a1877-960b-4e44-8e6d-47744d3e764b","Type":"ContainerDied","Data":"313eb3ad4fc2fcb56fe6d4828f016dd448a0cc926ce0e656cd2cc46796002ab2"} Jan 27 17:22:47 crc kubenswrapper[4772]: I0127 17:22:47.117524 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4xvdq" Jan 27 17:22:47 crc kubenswrapper[4772]: I0127 17:22:47.117557 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4xvdq" event={"ID":"6b2a1877-960b-4e44-8e6d-47744d3e764b","Type":"ContainerDied","Data":"aee6dea538b507f3e6d4df03a2d19ed1bfe49288afa763d30e38e9301cce0362"} Jan 27 17:22:47 crc kubenswrapper[4772]: I0127 17:22:47.117589 4772 scope.go:117] "RemoveContainer" containerID="313eb3ad4fc2fcb56fe6d4828f016dd448a0cc926ce0e656cd2cc46796002ab2" Jan 27 17:22:47 crc kubenswrapper[4772]: I0127 17:22:47.146906 4772 scope.go:117] "RemoveContainer" containerID="e682f9deaaadeb63ad6e4ae31c623a39a070781b1ba79a28bca994caba62c19f" Jan 27 17:22:47 crc kubenswrapper[4772]: I0127 17:22:47.173879 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4xvdq"] Jan 27 17:22:47 crc kubenswrapper[4772]: I0127 17:22:47.190749 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4xvdq"] Jan 27 17:22:47 crc kubenswrapper[4772]: I0127 17:22:47.198553 4772 scope.go:117] "RemoveContainer" containerID="08f7f6255d2bb42989667d0a5e1713a33ba4ba31840b81d1925d49697b9af8c6" Jan 27 17:22:47 crc kubenswrapper[4772]: I0127 17:22:47.241630 4772 scope.go:117] "RemoveContainer" containerID="313eb3ad4fc2fcb56fe6d4828f016dd448a0cc926ce0e656cd2cc46796002ab2" Jan 27 17:22:47 crc kubenswrapper[4772]: E0127 17:22:47.242380 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"313eb3ad4fc2fcb56fe6d4828f016dd448a0cc926ce0e656cd2cc46796002ab2\": container with ID starting with 313eb3ad4fc2fcb56fe6d4828f016dd448a0cc926ce0e656cd2cc46796002ab2 not found: ID does not exist" containerID="313eb3ad4fc2fcb56fe6d4828f016dd448a0cc926ce0e656cd2cc46796002ab2" Jan 27 17:22:47 crc kubenswrapper[4772]: I0127 17:22:47.242451 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"313eb3ad4fc2fcb56fe6d4828f016dd448a0cc926ce0e656cd2cc46796002ab2"} err="failed to get container status \"313eb3ad4fc2fcb56fe6d4828f016dd448a0cc926ce0e656cd2cc46796002ab2\": rpc error: code = NotFound desc = could not find container \"313eb3ad4fc2fcb56fe6d4828f016dd448a0cc926ce0e656cd2cc46796002ab2\": container with ID starting with 313eb3ad4fc2fcb56fe6d4828f016dd448a0cc926ce0e656cd2cc46796002ab2 not found: ID does not exist" Jan 27 17:22:47 crc kubenswrapper[4772]: I0127 17:22:47.242479 4772 scope.go:117] "RemoveContainer" containerID="e682f9deaaadeb63ad6e4ae31c623a39a070781b1ba79a28bca994caba62c19f" Jan 27 17:22:47 crc kubenswrapper[4772]: E0127 17:22:47.242999 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e682f9deaaadeb63ad6e4ae31c623a39a070781b1ba79a28bca994caba62c19f\": container with ID starting with e682f9deaaadeb63ad6e4ae31c623a39a070781b1ba79a28bca994caba62c19f not found: ID does not exist" containerID="e682f9deaaadeb63ad6e4ae31c623a39a070781b1ba79a28bca994caba62c19f" Jan 27 17:22:47 crc kubenswrapper[4772]: I0127 17:22:47.243054 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e682f9deaaadeb63ad6e4ae31c623a39a070781b1ba79a28bca994caba62c19f"} err="failed to get container status \"e682f9deaaadeb63ad6e4ae31c623a39a070781b1ba79a28bca994caba62c19f\": rpc error: code = NotFound desc = could not find container \"e682f9deaaadeb63ad6e4ae31c623a39a070781b1ba79a28bca994caba62c19f\": container with ID starting with e682f9deaaadeb63ad6e4ae31c623a39a070781b1ba79a28bca994caba62c19f not found: ID does not exist" Jan 27 17:22:47 crc kubenswrapper[4772]: I0127 17:22:47.243087 4772 scope.go:117] "RemoveContainer" containerID="08f7f6255d2bb42989667d0a5e1713a33ba4ba31840b81d1925d49697b9af8c6" Jan 27 17:22:47 crc kubenswrapper[4772]: E0127 17:22:47.243529 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08f7f6255d2bb42989667d0a5e1713a33ba4ba31840b81d1925d49697b9af8c6\": container with ID starting with 08f7f6255d2bb42989667d0a5e1713a33ba4ba31840b81d1925d49697b9af8c6 not found: ID does not exist" containerID="08f7f6255d2bb42989667d0a5e1713a33ba4ba31840b81d1925d49697b9af8c6" Jan 27 17:22:47 crc kubenswrapper[4772]: I0127 17:22:47.243564 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08f7f6255d2bb42989667d0a5e1713a33ba4ba31840b81d1925d49697b9af8c6"} err="failed to get container status \"08f7f6255d2bb42989667d0a5e1713a33ba4ba31840b81d1925d49697b9af8c6\": rpc error: code = NotFound desc = could not find container \"08f7f6255d2bb42989667d0a5e1713a33ba4ba31840b81d1925d49697b9af8c6\": container with ID starting with 08f7f6255d2bb42989667d0a5e1713a33ba4ba31840b81d1925d49697b9af8c6 not found: ID does not exist" Jan 27 17:22:48 crc kubenswrapper[4772]: I0127 17:22:48.158762 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-c894q" Jan 27 17:22:48 crc kubenswrapper[4772]: I0127 17:22:48.158881 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-c894q" Jan 27 17:22:48 crc kubenswrapper[4772]: I0127 17:22:48.206147 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-c894q" Jan 27 17:22:48 crc kubenswrapper[4772]: I0127 17:22:48.674597 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b2a1877-960b-4e44-8e6d-47744d3e764b" path="/var/lib/kubelet/pods/6b2a1877-960b-4e44-8e6d-47744d3e764b/volumes" Jan 27 17:22:49 crc kubenswrapper[4772]: I0127 17:22:49.184847 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-c894q" Jan 27 17:22:49 crc kubenswrapper[4772]: I0127 17:22:49.429355 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c894q"] Jan 27 17:22:51 crc kubenswrapper[4772]: I0127 17:22:51.164669 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-c894q" podUID="79767c4a-3d92-40a5-8128-a9b7785d4672" containerName="registry-server" containerID="cri-o://227375896e50ca3ae2a98fd56581e6e897945772df9f8a7d35dcddb59e16d22a" gracePeriod=2 Jan 27 17:22:51 crc kubenswrapper[4772]: I0127 17:22:51.712243 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c894q" Jan 27 17:22:51 crc kubenswrapper[4772]: I0127 17:22:51.753892 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsv8q\" (UniqueName: \"kubernetes.io/projected/79767c4a-3d92-40a5-8128-a9b7785d4672-kube-api-access-vsv8q\") pod \"79767c4a-3d92-40a5-8128-a9b7785d4672\" (UID: \"79767c4a-3d92-40a5-8128-a9b7785d4672\") " Jan 27 17:22:51 crc kubenswrapper[4772]: I0127 17:22:51.754002 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79767c4a-3d92-40a5-8128-a9b7785d4672-catalog-content\") pod \"79767c4a-3d92-40a5-8128-a9b7785d4672\" (UID: \"79767c4a-3d92-40a5-8128-a9b7785d4672\") " Jan 27 17:22:51 crc kubenswrapper[4772]: I0127 17:22:51.754089 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79767c4a-3d92-40a5-8128-a9b7785d4672-utilities\") pod \"79767c4a-3d92-40a5-8128-a9b7785d4672\" (UID: \"79767c4a-3d92-40a5-8128-a9b7785d4672\") " Jan 27 17:22:51 crc kubenswrapper[4772]: I0127 17:22:51.755110 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79767c4a-3d92-40a5-8128-a9b7785d4672-utilities" (OuterVolumeSpecName: "utilities") pod "79767c4a-3d92-40a5-8128-a9b7785d4672" (UID: "79767c4a-3d92-40a5-8128-a9b7785d4672"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:22:51 crc kubenswrapper[4772]: I0127 17:22:51.761250 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79767c4a-3d92-40a5-8128-a9b7785d4672-kube-api-access-vsv8q" (OuterVolumeSpecName: "kube-api-access-vsv8q") pod "79767c4a-3d92-40a5-8128-a9b7785d4672" (UID: "79767c4a-3d92-40a5-8128-a9b7785d4672"). InnerVolumeSpecName "kube-api-access-vsv8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:22:51 crc kubenswrapper[4772]: I0127 17:22:51.811524 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79767c4a-3d92-40a5-8128-a9b7785d4672-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "79767c4a-3d92-40a5-8128-a9b7785d4672" (UID: "79767c4a-3d92-40a5-8128-a9b7785d4672"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:22:51 crc kubenswrapper[4772]: I0127 17:22:51.855478 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79767c4a-3d92-40a5-8128-a9b7785d4672-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 17:22:51 crc kubenswrapper[4772]: I0127 17:22:51.855506 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79767c4a-3d92-40a5-8128-a9b7785d4672-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 17:22:51 crc kubenswrapper[4772]: I0127 17:22:51.855515 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsv8q\" (UniqueName: \"kubernetes.io/projected/79767c4a-3d92-40a5-8128-a9b7785d4672-kube-api-access-vsv8q\") on node \"crc\" DevicePath \"\"" Jan 27 17:22:52 crc kubenswrapper[4772]: I0127 17:22:52.216277 4772 generic.go:334] "Generic (PLEG): container finished" podID="79767c4a-3d92-40a5-8128-a9b7785d4672" containerID="227375896e50ca3ae2a98fd56581e6e897945772df9f8a7d35dcddb59e16d22a" exitCode=0 Jan 27 17:22:52 crc kubenswrapper[4772]: I0127 17:22:52.216339 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c894q" event={"ID":"79767c4a-3d92-40a5-8128-a9b7785d4672","Type":"ContainerDied","Data":"227375896e50ca3ae2a98fd56581e6e897945772df9f8a7d35dcddb59e16d22a"} Jan 27 17:22:52 crc kubenswrapper[4772]: I0127 17:22:52.216381 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c894q" event={"ID":"79767c4a-3d92-40a5-8128-a9b7785d4672","Type":"ContainerDied","Data":"64cc0b646724882690474db4430de2f2239d9f2a6e5ce0a0dcddb5c0fc527179"} Jan 27 17:22:52 crc kubenswrapper[4772]: I0127 17:22:52.216412 4772 scope.go:117] "RemoveContainer" containerID="227375896e50ca3ae2a98fd56581e6e897945772df9f8a7d35dcddb59e16d22a" Jan 27 17:22:52 crc kubenswrapper[4772]: I0127 17:22:52.216380 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c894q" Jan 27 17:22:52 crc kubenswrapper[4772]: I0127 17:22:52.265683 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c894q"] Jan 27 17:22:52 crc kubenswrapper[4772]: I0127 17:22:52.275981 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-c894q"] Jan 27 17:22:52 crc kubenswrapper[4772]: I0127 17:22:52.485355 4772 scope.go:117] "RemoveContainer" containerID="931a3ea833e76451e7ae9443d7919ff52b6d3c292558ce3eac47358e7f05fdad" Jan 27 17:22:52 crc kubenswrapper[4772]: I0127 17:22:52.526225 4772 scope.go:117] "RemoveContainer" containerID="b7996be57a39eb18a216e3d76b272bae729bf867e09dc88e02c79f2bbf29e2c0" Jan 27 17:22:52 crc kubenswrapper[4772]: I0127 17:22:52.568123 4772 scope.go:117] "RemoveContainer" containerID="227375896e50ca3ae2a98fd56581e6e897945772df9f8a7d35dcddb59e16d22a" Jan 27 17:22:52 crc kubenswrapper[4772]: E0127 17:22:52.569243 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"227375896e50ca3ae2a98fd56581e6e897945772df9f8a7d35dcddb59e16d22a\": container with ID starting with 227375896e50ca3ae2a98fd56581e6e897945772df9f8a7d35dcddb59e16d22a not found: ID does not exist" containerID="227375896e50ca3ae2a98fd56581e6e897945772df9f8a7d35dcddb59e16d22a" Jan 27 17:22:52 crc kubenswrapper[4772]: I0127 17:22:52.569348 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"227375896e50ca3ae2a98fd56581e6e897945772df9f8a7d35dcddb59e16d22a"} err="failed to get container status \"227375896e50ca3ae2a98fd56581e6e897945772df9f8a7d35dcddb59e16d22a\": rpc error: code = NotFound desc = could not find container \"227375896e50ca3ae2a98fd56581e6e897945772df9f8a7d35dcddb59e16d22a\": container with ID starting with 227375896e50ca3ae2a98fd56581e6e897945772df9f8a7d35dcddb59e16d22a not found: ID does not exist" Jan 27 17:22:52 crc kubenswrapper[4772]: I0127 17:22:52.569408 4772 scope.go:117] "RemoveContainer" containerID="931a3ea833e76451e7ae9443d7919ff52b6d3c292558ce3eac47358e7f05fdad" Jan 27 17:22:52 crc kubenswrapper[4772]: E0127 17:22:52.569886 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"931a3ea833e76451e7ae9443d7919ff52b6d3c292558ce3eac47358e7f05fdad\": container with ID starting with 931a3ea833e76451e7ae9443d7919ff52b6d3c292558ce3eac47358e7f05fdad not found: ID does not exist" containerID="931a3ea833e76451e7ae9443d7919ff52b6d3c292558ce3eac47358e7f05fdad" Jan 27 17:22:52 crc kubenswrapper[4772]: I0127 17:22:52.569929 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"931a3ea833e76451e7ae9443d7919ff52b6d3c292558ce3eac47358e7f05fdad"} err="failed to get container status \"931a3ea833e76451e7ae9443d7919ff52b6d3c292558ce3eac47358e7f05fdad\": rpc error: code = NotFound desc = could not find container \"931a3ea833e76451e7ae9443d7919ff52b6d3c292558ce3eac47358e7f05fdad\": container with ID starting with 931a3ea833e76451e7ae9443d7919ff52b6d3c292558ce3eac47358e7f05fdad not found: ID does not exist" Jan 27 17:22:52 crc kubenswrapper[4772]: I0127 17:22:52.569955 4772 scope.go:117] "RemoveContainer" containerID="b7996be57a39eb18a216e3d76b272bae729bf867e09dc88e02c79f2bbf29e2c0" Jan 27 17:22:52 crc kubenswrapper[4772]: E0127 17:22:52.570375 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7996be57a39eb18a216e3d76b272bae729bf867e09dc88e02c79f2bbf29e2c0\": container with ID starting with b7996be57a39eb18a216e3d76b272bae729bf867e09dc88e02c79f2bbf29e2c0 not found: ID does not exist" containerID="b7996be57a39eb18a216e3d76b272bae729bf867e09dc88e02c79f2bbf29e2c0" Jan 27 17:22:52 crc kubenswrapper[4772]: I0127 17:22:52.570411 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7996be57a39eb18a216e3d76b272bae729bf867e09dc88e02c79f2bbf29e2c0"} err="failed to get container status \"b7996be57a39eb18a216e3d76b272bae729bf867e09dc88e02c79f2bbf29e2c0\": rpc error: code = NotFound desc = could not find container \"b7996be57a39eb18a216e3d76b272bae729bf867e09dc88e02c79f2bbf29e2c0\": container with ID starting with b7996be57a39eb18a216e3d76b272bae729bf867e09dc88e02c79f2bbf29e2c0 not found: ID does not exist" Jan 27 17:22:52 crc kubenswrapper[4772]: I0127 17:22:52.675589 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79767c4a-3d92-40a5-8128-a9b7785d4672" path="/var/lib/kubelet/pods/79767c4a-3d92-40a5-8128-a9b7785d4672/volumes" Jan 27 17:23:42 crc kubenswrapper[4772]: I0127 17:23:42.058936 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:23:42 crc kubenswrapper[4772]: I0127 17:23:42.059571 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:24:12 crc kubenswrapper[4772]: I0127 17:24:12.058274 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:24:12 crc kubenswrapper[4772]: I0127 17:24:12.059160 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:24:42 crc kubenswrapper[4772]: I0127 17:24:42.058880 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:24:42 crc kubenswrapper[4772]: I0127 17:24:42.059341 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:24:42 crc kubenswrapper[4772]: I0127 17:24:42.059383 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 17:24:42 crc kubenswrapper[4772]: I0127 17:24:42.060157 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a0596c152e2f2b0802267f33c0f2d3224ef4bcfda5d22940bb1eb3256403bf5f"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 17:24:42 crc kubenswrapper[4772]: I0127 17:24:42.060292 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://a0596c152e2f2b0802267f33c0f2d3224ef4bcfda5d22940bb1eb3256403bf5f" gracePeriod=600 Jan 27 17:24:42 crc kubenswrapper[4772]: I0127 17:24:42.446041 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="a0596c152e2f2b0802267f33c0f2d3224ef4bcfda5d22940bb1eb3256403bf5f" exitCode=0 Jan 27 17:24:42 crc kubenswrapper[4772]: I0127 17:24:42.446121 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"a0596c152e2f2b0802267f33c0f2d3224ef4bcfda5d22940bb1eb3256403bf5f"} Jan 27 17:24:42 crc kubenswrapper[4772]: I0127 17:24:42.446504 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284"} Jan 27 17:24:42 crc kubenswrapper[4772]: I0127 17:24:42.446531 4772 scope.go:117] "RemoveContainer" containerID="01301e7d30d90ee5d4e73d016fb4f9b1d80c5e9139db297ce836d144473d7246" Jan 27 17:25:19 crc kubenswrapper[4772]: I0127 17:25:19.746201 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vf8r7"] Jan 27 17:25:19 crc kubenswrapper[4772]: E0127 17:25:19.747634 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79767c4a-3d92-40a5-8128-a9b7785d4672" containerName="extract-content" Jan 27 17:25:19 crc kubenswrapper[4772]: I0127 17:25:19.747660 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="79767c4a-3d92-40a5-8128-a9b7785d4672" containerName="extract-content" Jan 27 17:25:19 crc kubenswrapper[4772]: E0127 17:25:19.747696 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b2a1877-960b-4e44-8e6d-47744d3e764b" containerName="extract-utilities" Jan 27 17:25:19 crc kubenswrapper[4772]: I0127 17:25:19.747709 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b2a1877-960b-4e44-8e6d-47744d3e764b" containerName="extract-utilities" Jan 27 17:25:19 crc kubenswrapper[4772]: E0127 17:25:19.747727 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79767c4a-3d92-40a5-8128-a9b7785d4672" containerName="registry-server" Jan 27 17:25:19 crc kubenswrapper[4772]: I0127 17:25:19.747737 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="79767c4a-3d92-40a5-8128-a9b7785d4672" containerName="registry-server" Jan 27 17:25:19 crc kubenswrapper[4772]: E0127 17:25:19.747759 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79767c4a-3d92-40a5-8128-a9b7785d4672" containerName="extract-utilities" Jan 27 17:25:19 crc kubenswrapper[4772]: I0127 17:25:19.747768 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="79767c4a-3d92-40a5-8128-a9b7785d4672" containerName="extract-utilities" Jan 27 17:25:19 crc kubenswrapper[4772]: E0127 17:25:19.747791 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b2a1877-960b-4e44-8e6d-47744d3e764b" containerName="extract-content" Jan 27 17:25:19 crc kubenswrapper[4772]: I0127 17:25:19.747799 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b2a1877-960b-4e44-8e6d-47744d3e764b" containerName="extract-content" Jan 27 17:25:19 crc kubenswrapper[4772]: E0127 17:25:19.747826 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b2a1877-960b-4e44-8e6d-47744d3e764b" containerName="registry-server" Jan 27 17:25:19 crc kubenswrapper[4772]: I0127 17:25:19.747865 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b2a1877-960b-4e44-8e6d-47744d3e764b" containerName="registry-server" Jan 27 17:25:19 crc kubenswrapper[4772]: I0127 17:25:19.748106 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="79767c4a-3d92-40a5-8128-a9b7785d4672" containerName="registry-server" Jan 27 17:25:19 crc kubenswrapper[4772]: I0127 17:25:19.748128 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b2a1877-960b-4e44-8e6d-47744d3e764b" containerName="registry-server" Jan 27 17:25:19 crc kubenswrapper[4772]: I0127 17:25:19.749781 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vf8r7" Jan 27 17:25:19 crc kubenswrapper[4772]: I0127 17:25:19.776218 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vf8r7"] Jan 27 17:25:19 crc kubenswrapper[4772]: I0127 17:25:19.833133 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjq7m\" (UniqueName: \"kubernetes.io/projected/46d7e227-3c40-4173-abf4-7c1ee4aee5d2-kube-api-access-zjq7m\") pod \"certified-operators-vf8r7\" (UID: \"46d7e227-3c40-4173-abf4-7c1ee4aee5d2\") " pod="openshift-marketplace/certified-operators-vf8r7" Jan 27 17:25:19 crc kubenswrapper[4772]: I0127 17:25:19.833403 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46d7e227-3c40-4173-abf4-7c1ee4aee5d2-utilities\") pod \"certified-operators-vf8r7\" (UID: \"46d7e227-3c40-4173-abf4-7c1ee4aee5d2\") " pod="openshift-marketplace/certified-operators-vf8r7" Jan 27 17:25:19 crc kubenswrapper[4772]: I0127 17:25:19.833565 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46d7e227-3c40-4173-abf4-7c1ee4aee5d2-catalog-content\") pod \"certified-operators-vf8r7\" (UID: \"46d7e227-3c40-4173-abf4-7c1ee4aee5d2\") " pod="openshift-marketplace/certified-operators-vf8r7" Jan 27 17:25:19 crc kubenswrapper[4772]: I0127 17:25:19.935031 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46d7e227-3c40-4173-abf4-7c1ee4aee5d2-utilities\") pod \"certified-operators-vf8r7\" (UID: \"46d7e227-3c40-4173-abf4-7c1ee4aee5d2\") " pod="openshift-marketplace/certified-operators-vf8r7" Jan 27 17:25:19 crc kubenswrapper[4772]: I0127 17:25:19.935114 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46d7e227-3c40-4173-abf4-7c1ee4aee5d2-catalog-content\") pod \"certified-operators-vf8r7\" (UID: \"46d7e227-3c40-4173-abf4-7c1ee4aee5d2\") " pod="openshift-marketplace/certified-operators-vf8r7" Jan 27 17:25:19 crc kubenswrapper[4772]: I0127 17:25:19.935182 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjq7m\" (UniqueName: \"kubernetes.io/projected/46d7e227-3c40-4173-abf4-7c1ee4aee5d2-kube-api-access-zjq7m\") pod \"certified-operators-vf8r7\" (UID: \"46d7e227-3c40-4173-abf4-7c1ee4aee5d2\") " pod="openshift-marketplace/certified-operators-vf8r7" Jan 27 17:25:19 crc kubenswrapper[4772]: I0127 17:25:19.935635 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46d7e227-3c40-4173-abf4-7c1ee4aee5d2-utilities\") pod \"certified-operators-vf8r7\" (UID: \"46d7e227-3c40-4173-abf4-7c1ee4aee5d2\") " pod="openshift-marketplace/certified-operators-vf8r7" Jan 27 17:25:19 crc kubenswrapper[4772]: I0127 17:25:19.935737 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46d7e227-3c40-4173-abf4-7c1ee4aee5d2-catalog-content\") pod \"certified-operators-vf8r7\" (UID: \"46d7e227-3c40-4173-abf4-7c1ee4aee5d2\") " pod="openshift-marketplace/certified-operators-vf8r7" Jan 27 17:25:19 crc kubenswrapper[4772]: I0127 17:25:19.970996 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjq7m\" (UniqueName: \"kubernetes.io/projected/46d7e227-3c40-4173-abf4-7c1ee4aee5d2-kube-api-access-zjq7m\") pod \"certified-operators-vf8r7\" (UID: \"46d7e227-3c40-4173-abf4-7c1ee4aee5d2\") " pod="openshift-marketplace/certified-operators-vf8r7" Jan 27 17:25:20 crc kubenswrapper[4772]: I0127 17:25:20.107961 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vf8r7" Jan 27 17:25:20 crc kubenswrapper[4772]: I0127 17:25:20.689655 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vf8r7"] Jan 27 17:25:20 crc kubenswrapper[4772]: W0127 17:25:20.704360 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46d7e227_3c40_4173_abf4_7c1ee4aee5d2.slice/crio-a113e824463a2b79f76c2d98224ce0ade7752003b9acd3f5ff9e304b1ff68a8a WatchSource:0}: Error finding container a113e824463a2b79f76c2d98224ce0ade7752003b9acd3f5ff9e304b1ff68a8a: Status 404 returned error can't find the container with id a113e824463a2b79f76c2d98224ce0ade7752003b9acd3f5ff9e304b1ff68a8a Jan 27 17:25:20 crc kubenswrapper[4772]: I0127 17:25:20.880345 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vf8r7" event={"ID":"46d7e227-3c40-4173-abf4-7c1ee4aee5d2","Type":"ContainerStarted","Data":"a113e824463a2b79f76c2d98224ce0ade7752003b9acd3f5ff9e304b1ff68a8a"} Jan 27 17:25:21 crc kubenswrapper[4772]: I0127 17:25:21.892553 4772 generic.go:334] "Generic (PLEG): container finished" podID="46d7e227-3c40-4173-abf4-7c1ee4aee5d2" containerID="aad0b49b76f8d36172eb452dff4c5d1765e0da461be5445c32ece6a3405d8b20" exitCode=0 Jan 27 17:25:21 crc kubenswrapper[4772]: I0127 17:25:21.892644 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vf8r7" event={"ID":"46d7e227-3c40-4173-abf4-7c1ee4aee5d2","Type":"ContainerDied","Data":"aad0b49b76f8d36172eb452dff4c5d1765e0da461be5445c32ece6a3405d8b20"} Jan 27 17:25:21 crc kubenswrapper[4772]: I0127 17:25:21.895303 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 17:25:22 crc kubenswrapper[4772]: I0127 17:25:22.904329 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vf8r7" event={"ID":"46d7e227-3c40-4173-abf4-7c1ee4aee5d2","Type":"ContainerStarted","Data":"5f5443e55691e58714f2473c9100f2320f98b4ed0e3962a5ebdf967c7fe05757"} Jan 27 17:25:23 crc kubenswrapper[4772]: I0127 17:25:23.919259 4772 generic.go:334] "Generic (PLEG): container finished" podID="46d7e227-3c40-4173-abf4-7c1ee4aee5d2" containerID="5f5443e55691e58714f2473c9100f2320f98b4ed0e3962a5ebdf967c7fe05757" exitCode=0 Jan 27 17:25:23 crc kubenswrapper[4772]: I0127 17:25:23.919357 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vf8r7" event={"ID":"46d7e227-3c40-4173-abf4-7c1ee4aee5d2","Type":"ContainerDied","Data":"5f5443e55691e58714f2473c9100f2320f98b4ed0e3962a5ebdf967c7fe05757"} Jan 27 17:25:24 crc kubenswrapper[4772]: I0127 17:25:24.929407 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vf8r7" event={"ID":"46d7e227-3c40-4173-abf4-7c1ee4aee5d2","Type":"ContainerStarted","Data":"3b09782714157f33427572d6bbbe2fb46604581cbefffd50d34df0f457248ad1"} Jan 27 17:25:30 crc kubenswrapper[4772]: I0127 17:25:30.108794 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vf8r7" Jan 27 17:25:30 crc kubenswrapper[4772]: I0127 17:25:30.109632 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vf8r7" Jan 27 17:25:30 crc kubenswrapper[4772]: I0127 17:25:30.169259 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vf8r7" Jan 27 17:25:30 crc kubenswrapper[4772]: I0127 17:25:30.193250 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vf8r7" podStartSLOduration=8.720730494 podStartE2EDuration="11.193234716s" podCreationTimestamp="2026-01-27 17:25:19 +0000 UTC" firstStartedPulling="2026-01-27 17:25:21.895058081 +0000 UTC m=+8307.875667179" lastFinishedPulling="2026-01-27 17:25:24.367562273 +0000 UTC m=+8310.348171401" observedRunningTime="2026-01-27 17:25:24.956940442 +0000 UTC m=+8310.937549580" watchObservedRunningTime="2026-01-27 17:25:30.193234716 +0000 UTC m=+8316.173843814" Jan 27 17:25:31 crc kubenswrapper[4772]: I0127 17:25:31.052283 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vf8r7" Jan 27 17:25:31 crc kubenswrapper[4772]: I0127 17:25:31.104637 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vf8r7"] Jan 27 17:25:33 crc kubenswrapper[4772]: I0127 17:25:33.012474 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vf8r7" podUID="46d7e227-3c40-4173-abf4-7c1ee4aee5d2" containerName="registry-server" containerID="cri-o://3b09782714157f33427572d6bbbe2fb46604581cbefffd50d34df0f457248ad1" gracePeriod=2 Jan 27 17:25:33 crc kubenswrapper[4772]: I0127 17:25:33.580406 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vf8r7" Jan 27 17:25:33 crc kubenswrapper[4772]: I0127 17:25:33.617123 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46d7e227-3c40-4173-abf4-7c1ee4aee5d2-catalog-content\") pod \"46d7e227-3c40-4173-abf4-7c1ee4aee5d2\" (UID: \"46d7e227-3c40-4173-abf4-7c1ee4aee5d2\") " Jan 27 17:25:33 crc kubenswrapper[4772]: I0127 17:25:33.617400 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46d7e227-3c40-4173-abf4-7c1ee4aee5d2-utilities\") pod \"46d7e227-3c40-4173-abf4-7c1ee4aee5d2\" (UID: \"46d7e227-3c40-4173-abf4-7c1ee4aee5d2\") " Jan 27 17:25:33 crc kubenswrapper[4772]: I0127 17:25:33.617579 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjq7m\" (UniqueName: \"kubernetes.io/projected/46d7e227-3c40-4173-abf4-7c1ee4aee5d2-kube-api-access-zjq7m\") pod \"46d7e227-3c40-4173-abf4-7c1ee4aee5d2\" (UID: \"46d7e227-3c40-4173-abf4-7c1ee4aee5d2\") " Jan 27 17:25:33 crc kubenswrapper[4772]: I0127 17:25:33.618344 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46d7e227-3c40-4173-abf4-7c1ee4aee5d2-utilities" (OuterVolumeSpecName: "utilities") pod "46d7e227-3c40-4173-abf4-7c1ee4aee5d2" (UID: "46d7e227-3c40-4173-abf4-7c1ee4aee5d2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:25:33 crc kubenswrapper[4772]: I0127 17:25:33.623508 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46d7e227-3c40-4173-abf4-7c1ee4aee5d2-kube-api-access-zjq7m" (OuterVolumeSpecName: "kube-api-access-zjq7m") pod "46d7e227-3c40-4173-abf4-7c1ee4aee5d2" (UID: "46d7e227-3c40-4173-abf4-7c1ee4aee5d2"). InnerVolumeSpecName "kube-api-access-zjq7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:25:33 crc kubenswrapper[4772]: I0127 17:25:33.669113 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46d7e227-3c40-4173-abf4-7c1ee4aee5d2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "46d7e227-3c40-4173-abf4-7c1ee4aee5d2" (UID: "46d7e227-3c40-4173-abf4-7c1ee4aee5d2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:25:33 crc kubenswrapper[4772]: I0127 17:25:33.719467 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46d7e227-3c40-4173-abf4-7c1ee4aee5d2-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 17:25:33 crc kubenswrapper[4772]: I0127 17:25:33.719501 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjq7m\" (UniqueName: \"kubernetes.io/projected/46d7e227-3c40-4173-abf4-7c1ee4aee5d2-kube-api-access-zjq7m\") on node \"crc\" DevicePath \"\"" Jan 27 17:25:33 crc kubenswrapper[4772]: I0127 17:25:33.719512 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46d7e227-3c40-4173-abf4-7c1ee4aee5d2-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 17:25:34 crc kubenswrapper[4772]: I0127 17:25:34.025126 4772 generic.go:334] "Generic (PLEG): container finished" podID="46d7e227-3c40-4173-abf4-7c1ee4aee5d2" containerID="3b09782714157f33427572d6bbbe2fb46604581cbefffd50d34df0f457248ad1" exitCode=0 Jan 27 17:25:34 crc kubenswrapper[4772]: I0127 17:25:34.025194 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vf8r7" event={"ID":"46d7e227-3c40-4173-abf4-7c1ee4aee5d2","Type":"ContainerDied","Data":"3b09782714157f33427572d6bbbe2fb46604581cbefffd50d34df0f457248ad1"} Jan 27 17:25:34 crc kubenswrapper[4772]: I0127 17:25:34.025241 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vf8r7" event={"ID":"46d7e227-3c40-4173-abf4-7c1ee4aee5d2","Type":"ContainerDied","Data":"a113e824463a2b79f76c2d98224ce0ade7752003b9acd3f5ff9e304b1ff68a8a"} Jan 27 17:25:34 crc kubenswrapper[4772]: I0127 17:25:34.025262 4772 scope.go:117] "RemoveContainer" containerID="3b09782714157f33427572d6bbbe2fb46604581cbefffd50d34df0f457248ad1" Jan 27 17:25:34 crc kubenswrapper[4772]: I0127 17:25:34.025309 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vf8r7" Jan 27 17:25:34 crc kubenswrapper[4772]: I0127 17:25:34.046483 4772 scope.go:117] "RemoveContainer" containerID="5f5443e55691e58714f2473c9100f2320f98b4ed0e3962a5ebdf967c7fe05757" Jan 27 17:25:34 crc kubenswrapper[4772]: I0127 17:25:34.079458 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vf8r7"] Jan 27 17:25:34 crc kubenswrapper[4772]: I0127 17:25:34.091238 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vf8r7"] Jan 27 17:25:34 crc kubenswrapper[4772]: I0127 17:25:34.098152 4772 scope.go:117] "RemoveContainer" containerID="aad0b49b76f8d36172eb452dff4c5d1765e0da461be5445c32ece6a3405d8b20" Jan 27 17:25:34 crc kubenswrapper[4772]: I0127 17:25:34.158550 4772 scope.go:117] "RemoveContainer" containerID="3b09782714157f33427572d6bbbe2fb46604581cbefffd50d34df0f457248ad1" Jan 27 17:25:34 crc kubenswrapper[4772]: E0127 17:25:34.159143 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b09782714157f33427572d6bbbe2fb46604581cbefffd50d34df0f457248ad1\": container with ID starting with 3b09782714157f33427572d6bbbe2fb46604581cbefffd50d34df0f457248ad1 not found: ID does not exist" containerID="3b09782714157f33427572d6bbbe2fb46604581cbefffd50d34df0f457248ad1" Jan 27 17:25:34 crc kubenswrapper[4772]: I0127 17:25:34.159219 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b09782714157f33427572d6bbbe2fb46604581cbefffd50d34df0f457248ad1"} err="failed to get container status \"3b09782714157f33427572d6bbbe2fb46604581cbefffd50d34df0f457248ad1\": rpc error: code = NotFound desc = could not find container \"3b09782714157f33427572d6bbbe2fb46604581cbefffd50d34df0f457248ad1\": container with ID starting with 3b09782714157f33427572d6bbbe2fb46604581cbefffd50d34df0f457248ad1 not found: ID does not exist" Jan 27 17:25:34 crc kubenswrapper[4772]: I0127 17:25:34.159253 4772 scope.go:117] "RemoveContainer" containerID="5f5443e55691e58714f2473c9100f2320f98b4ed0e3962a5ebdf967c7fe05757" Jan 27 17:25:34 crc kubenswrapper[4772]: E0127 17:25:34.159766 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f5443e55691e58714f2473c9100f2320f98b4ed0e3962a5ebdf967c7fe05757\": container with ID starting with 5f5443e55691e58714f2473c9100f2320f98b4ed0e3962a5ebdf967c7fe05757 not found: ID does not exist" containerID="5f5443e55691e58714f2473c9100f2320f98b4ed0e3962a5ebdf967c7fe05757" Jan 27 17:25:34 crc kubenswrapper[4772]: I0127 17:25:34.159831 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f5443e55691e58714f2473c9100f2320f98b4ed0e3962a5ebdf967c7fe05757"} err="failed to get container status \"5f5443e55691e58714f2473c9100f2320f98b4ed0e3962a5ebdf967c7fe05757\": rpc error: code = NotFound desc = could not find container \"5f5443e55691e58714f2473c9100f2320f98b4ed0e3962a5ebdf967c7fe05757\": container with ID starting with 5f5443e55691e58714f2473c9100f2320f98b4ed0e3962a5ebdf967c7fe05757 not found: ID does not exist" Jan 27 17:25:34 crc kubenswrapper[4772]: I0127 17:25:34.159873 4772 scope.go:117] "RemoveContainer" containerID="aad0b49b76f8d36172eb452dff4c5d1765e0da461be5445c32ece6a3405d8b20" Jan 27 17:25:34 crc kubenswrapper[4772]: E0127 17:25:34.160609 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aad0b49b76f8d36172eb452dff4c5d1765e0da461be5445c32ece6a3405d8b20\": container with ID starting with aad0b49b76f8d36172eb452dff4c5d1765e0da461be5445c32ece6a3405d8b20 not found: ID does not exist" containerID="aad0b49b76f8d36172eb452dff4c5d1765e0da461be5445c32ece6a3405d8b20" Jan 27 17:25:34 crc kubenswrapper[4772]: I0127 17:25:34.160651 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aad0b49b76f8d36172eb452dff4c5d1765e0da461be5445c32ece6a3405d8b20"} err="failed to get container status \"aad0b49b76f8d36172eb452dff4c5d1765e0da461be5445c32ece6a3405d8b20\": rpc error: code = NotFound desc = could not find container \"aad0b49b76f8d36172eb452dff4c5d1765e0da461be5445c32ece6a3405d8b20\": container with ID starting with aad0b49b76f8d36172eb452dff4c5d1765e0da461be5445c32ece6a3405d8b20 not found: ID does not exist" Jan 27 17:25:34 crc kubenswrapper[4772]: I0127 17:25:34.681985 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46d7e227-3c40-4173-abf4-7c1ee4aee5d2" path="/var/lib/kubelet/pods/46d7e227-3c40-4173-abf4-7c1ee4aee5d2/volumes" Jan 27 17:26:42 crc kubenswrapper[4772]: I0127 17:26:42.059356 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:26:42 crc kubenswrapper[4772]: I0127 17:26:42.060369 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:27:12 crc kubenswrapper[4772]: I0127 17:27:12.058499 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:27:12 crc kubenswrapper[4772]: I0127 17:27:12.059152 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:27:42 crc kubenswrapper[4772]: I0127 17:27:42.058788 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:27:42 crc kubenswrapper[4772]: I0127 17:27:42.059656 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:27:42 crc kubenswrapper[4772]: I0127 17:27:42.059740 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 17:27:42 crc kubenswrapper[4772]: I0127 17:27:42.061101 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 17:27:42 crc kubenswrapper[4772]: I0127 17:27:42.061255 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" gracePeriod=600 Jan 27 17:27:42 crc kubenswrapper[4772]: E0127 17:27:42.198427 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:27:42 crc kubenswrapper[4772]: I0127 17:27:42.415531 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" exitCode=0 Jan 27 17:27:42 crc kubenswrapper[4772]: I0127 17:27:42.415581 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284"} Jan 27 17:27:42 crc kubenswrapper[4772]: I0127 17:27:42.415937 4772 scope.go:117] "RemoveContainer" containerID="a0596c152e2f2b0802267f33c0f2d3224ef4bcfda5d22940bb1eb3256403bf5f" Jan 27 17:27:42 crc kubenswrapper[4772]: I0127 17:27:42.416522 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:27:42 crc kubenswrapper[4772]: E0127 17:27:42.416845 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:27:57 crc kubenswrapper[4772]: I0127 17:27:57.663811 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:27:57 crc kubenswrapper[4772]: E0127 17:27:57.665128 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:28:10 crc kubenswrapper[4772]: I0127 17:28:10.663501 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:28:10 crc kubenswrapper[4772]: E0127 17:28:10.664850 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:28:25 crc kubenswrapper[4772]: I0127 17:28:25.663626 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:28:25 crc kubenswrapper[4772]: E0127 17:28:25.665033 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:28:40 crc kubenswrapper[4772]: I0127 17:28:40.663877 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:28:40 crc kubenswrapper[4772]: E0127 17:28:40.665254 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:28:53 crc kubenswrapper[4772]: I0127 17:28:53.664715 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:28:53 crc kubenswrapper[4772]: E0127 17:28:53.665869 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:29:04 crc kubenswrapper[4772]: I0127 17:29:04.675951 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:29:04 crc kubenswrapper[4772]: E0127 17:29:04.676858 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:29:18 crc kubenswrapper[4772]: I0127 17:29:18.664283 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:29:18 crc kubenswrapper[4772]: E0127 17:29:18.665623 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:29:29 crc kubenswrapper[4772]: I0127 17:29:29.040550 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-r5pg8"] Jan 27 17:29:29 crc kubenswrapper[4772]: E0127 17:29:29.041754 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46d7e227-3c40-4173-abf4-7c1ee4aee5d2" containerName="extract-content" Jan 27 17:29:29 crc kubenswrapper[4772]: I0127 17:29:29.041774 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="46d7e227-3c40-4173-abf4-7c1ee4aee5d2" containerName="extract-content" Jan 27 17:29:29 crc kubenswrapper[4772]: E0127 17:29:29.041828 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46d7e227-3c40-4173-abf4-7c1ee4aee5d2" containerName="registry-server" Jan 27 17:29:29 crc kubenswrapper[4772]: I0127 17:29:29.041841 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="46d7e227-3c40-4173-abf4-7c1ee4aee5d2" containerName="registry-server" Jan 27 17:29:29 crc kubenswrapper[4772]: E0127 17:29:29.041861 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46d7e227-3c40-4173-abf4-7c1ee4aee5d2" containerName="extract-utilities" Jan 27 17:29:29 crc kubenswrapper[4772]: I0127 17:29:29.041873 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="46d7e227-3c40-4173-abf4-7c1ee4aee5d2" containerName="extract-utilities" Jan 27 17:29:29 crc kubenswrapper[4772]: I0127 17:29:29.042142 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="46d7e227-3c40-4173-abf4-7c1ee4aee5d2" containerName="registry-server" Jan 27 17:29:29 crc kubenswrapper[4772]: I0127 17:29:29.044435 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r5pg8" Jan 27 17:29:29 crc kubenswrapper[4772]: I0127 17:29:29.059476 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r5pg8"] Jan 27 17:29:29 crc kubenswrapper[4772]: I0127 17:29:29.187369 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff1acde3-5e9f-4659-a5b5-88497d0ce4c5-utilities\") pod \"redhat-operators-r5pg8\" (UID: \"ff1acde3-5e9f-4659-a5b5-88497d0ce4c5\") " pod="openshift-marketplace/redhat-operators-r5pg8" Jan 27 17:29:29 crc kubenswrapper[4772]: I0127 17:29:29.187440 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45llf\" (UniqueName: \"kubernetes.io/projected/ff1acde3-5e9f-4659-a5b5-88497d0ce4c5-kube-api-access-45llf\") pod \"redhat-operators-r5pg8\" (UID: \"ff1acde3-5e9f-4659-a5b5-88497d0ce4c5\") " pod="openshift-marketplace/redhat-operators-r5pg8" Jan 27 17:29:29 crc kubenswrapper[4772]: I0127 17:29:29.187460 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff1acde3-5e9f-4659-a5b5-88497d0ce4c5-catalog-content\") pod \"redhat-operators-r5pg8\" (UID: \"ff1acde3-5e9f-4659-a5b5-88497d0ce4c5\") " pod="openshift-marketplace/redhat-operators-r5pg8" Jan 27 17:29:29 crc kubenswrapper[4772]: I0127 17:29:29.289834 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff1acde3-5e9f-4659-a5b5-88497d0ce4c5-utilities\") pod \"redhat-operators-r5pg8\" (UID: \"ff1acde3-5e9f-4659-a5b5-88497d0ce4c5\") " pod="openshift-marketplace/redhat-operators-r5pg8" Jan 27 17:29:29 crc kubenswrapper[4772]: I0127 17:29:29.289930 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45llf\" (UniqueName: \"kubernetes.io/projected/ff1acde3-5e9f-4659-a5b5-88497d0ce4c5-kube-api-access-45llf\") pod \"redhat-operators-r5pg8\" (UID: \"ff1acde3-5e9f-4659-a5b5-88497d0ce4c5\") " pod="openshift-marketplace/redhat-operators-r5pg8" Jan 27 17:29:29 crc kubenswrapper[4772]: I0127 17:29:29.289956 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff1acde3-5e9f-4659-a5b5-88497d0ce4c5-catalog-content\") pod \"redhat-operators-r5pg8\" (UID: \"ff1acde3-5e9f-4659-a5b5-88497d0ce4c5\") " pod="openshift-marketplace/redhat-operators-r5pg8" Jan 27 17:29:29 crc kubenswrapper[4772]: I0127 17:29:29.290467 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff1acde3-5e9f-4659-a5b5-88497d0ce4c5-catalog-content\") pod \"redhat-operators-r5pg8\" (UID: \"ff1acde3-5e9f-4659-a5b5-88497d0ce4c5\") " pod="openshift-marketplace/redhat-operators-r5pg8" Jan 27 17:29:29 crc kubenswrapper[4772]: I0127 17:29:29.290731 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff1acde3-5e9f-4659-a5b5-88497d0ce4c5-utilities\") pod \"redhat-operators-r5pg8\" (UID: \"ff1acde3-5e9f-4659-a5b5-88497d0ce4c5\") " pod="openshift-marketplace/redhat-operators-r5pg8" Jan 27 17:29:29 crc kubenswrapper[4772]: I0127 17:29:29.323718 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45llf\" (UniqueName: \"kubernetes.io/projected/ff1acde3-5e9f-4659-a5b5-88497d0ce4c5-kube-api-access-45llf\") pod \"redhat-operators-r5pg8\" (UID: \"ff1acde3-5e9f-4659-a5b5-88497d0ce4c5\") " pod="openshift-marketplace/redhat-operators-r5pg8" Jan 27 17:29:29 crc kubenswrapper[4772]: I0127 17:29:29.373765 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r5pg8" Jan 27 17:29:29 crc kubenswrapper[4772]: I0127 17:29:29.835496 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r5pg8"] Jan 27 17:29:30 crc kubenswrapper[4772]: I0127 17:29:30.611830 4772 generic.go:334] "Generic (PLEG): container finished" podID="ff1acde3-5e9f-4659-a5b5-88497d0ce4c5" containerID="726824e909b709ea955048da1b47107c917a9ead2621a9e65cc30363053366cb" exitCode=0 Jan 27 17:29:30 crc kubenswrapper[4772]: I0127 17:29:30.611937 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5pg8" event={"ID":"ff1acde3-5e9f-4659-a5b5-88497d0ce4c5","Type":"ContainerDied","Data":"726824e909b709ea955048da1b47107c917a9ead2621a9e65cc30363053366cb"} Jan 27 17:29:30 crc kubenswrapper[4772]: I0127 17:29:30.612144 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5pg8" event={"ID":"ff1acde3-5e9f-4659-a5b5-88497d0ce4c5","Type":"ContainerStarted","Data":"4880d61008d58f60627606eb86b9a7ff700103c4a3a917606afb4f4e5637a700"} Jan 27 17:29:31 crc kubenswrapper[4772]: I0127 17:29:31.631004 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5pg8" event={"ID":"ff1acde3-5e9f-4659-a5b5-88497d0ce4c5","Type":"ContainerStarted","Data":"be4fbbd1741965f3fcecf5b9a29d8a1a7d7296f88ab213fffd113439cfaa68b6"} Jan 27 17:29:32 crc kubenswrapper[4772]: I0127 17:29:32.663259 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:29:32 crc kubenswrapper[4772]: E0127 17:29:32.664042 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:29:33 crc kubenswrapper[4772]: I0127 17:29:33.659371 4772 generic.go:334] "Generic (PLEG): container finished" podID="ff1acde3-5e9f-4659-a5b5-88497d0ce4c5" containerID="be4fbbd1741965f3fcecf5b9a29d8a1a7d7296f88ab213fffd113439cfaa68b6" exitCode=0 Jan 27 17:29:33 crc kubenswrapper[4772]: I0127 17:29:33.659468 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5pg8" event={"ID":"ff1acde3-5e9f-4659-a5b5-88497d0ce4c5","Type":"ContainerDied","Data":"be4fbbd1741965f3fcecf5b9a29d8a1a7d7296f88ab213fffd113439cfaa68b6"} Jan 27 17:29:35 crc kubenswrapper[4772]: I0127 17:29:35.683556 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5pg8" event={"ID":"ff1acde3-5e9f-4659-a5b5-88497d0ce4c5","Type":"ContainerStarted","Data":"024773fe5b0145280bf522bfd231862b43616d6d3e7c272e497743d5b82823f5"} Jan 27 17:29:35 crc kubenswrapper[4772]: I0127 17:29:35.726123 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-r5pg8" podStartSLOduration=4.134854132 podStartE2EDuration="7.726093374s" podCreationTimestamp="2026-01-27 17:29:28 +0000 UTC" firstStartedPulling="2026-01-27 17:29:30.614228988 +0000 UTC m=+8556.594838086" lastFinishedPulling="2026-01-27 17:29:34.20546819 +0000 UTC m=+8560.186077328" observedRunningTime="2026-01-27 17:29:35.713882944 +0000 UTC m=+8561.694492072" watchObservedRunningTime="2026-01-27 17:29:35.726093374 +0000 UTC m=+8561.706702512" Jan 27 17:29:39 crc kubenswrapper[4772]: I0127 17:29:39.374513 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-r5pg8" Jan 27 17:29:39 crc kubenswrapper[4772]: I0127 17:29:39.374985 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-r5pg8" Jan 27 17:29:40 crc kubenswrapper[4772]: I0127 17:29:40.427517 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-r5pg8" podUID="ff1acde3-5e9f-4659-a5b5-88497d0ce4c5" containerName="registry-server" probeResult="failure" output=< Jan 27 17:29:40 crc kubenswrapper[4772]: timeout: failed to connect service ":50051" within 1s Jan 27 17:29:40 crc kubenswrapper[4772]: > Jan 27 17:29:45 crc kubenswrapper[4772]: I0127 17:29:45.664228 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:29:45 crc kubenswrapper[4772]: E0127 17:29:45.665588 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:29:49 crc kubenswrapper[4772]: I0127 17:29:49.461186 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-r5pg8" Jan 27 17:29:49 crc kubenswrapper[4772]: I0127 17:29:49.547888 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-r5pg8" Jan 27 17:29:49 crc kubenswrapper[4772]: I0127 17:29:49.707600 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r5pg8"] Jan 27 17:29:50 crc kubenswrapper[4772]: I0127 17:29:50.840796 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-r5pg8" podUID="ff1acde3-5e9f-4659-a5b5-88497d0ce4c5" containerName="registry-server" containerID="cri-o://024773fe5b0145280bf522bfd231862b43616d6d3e7c272e497743d5b82823f5" gracePeriod=2 Jan 27 17:29:51 crc kubenswrapper[4772]: I0127 17:29:51.852970 4772 generic.go:334] "Generic (PLEG): container finished" podID="ff1acde3-5e9f-4659-a5b5-88497d0ce4c5" containerID="024773fe5b0145280bf522bfd231862b43616d6d3e7c272e497743d5b82823f5" exitCode=0 Jan 27 17:29:51 crc kubenswrapper[4772]: I0127 17:29:51.853024 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5pg8" event={"ID":"ff1acde3-5e9f-4659-a5b5-88497d0ce4c5","Type":"ContainerDied","Data":"024773fe5b0145280bf522bfd231862b43616d6d3e7c272e497743d5b82823f5"} Jan 27 17:29:52 crc kubenswrapper[4772]: I0127 17:29:52.286235 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r5pg8" Jan 27 17:29:52 crc kubenswrapper[4772]: I0127 17:29:52.341794 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff1acde3-5e9f-4659-a5b5-88497d0ce4c5-catalog-content\") pod \"ff1acde3-5e9f-4659-a5b5-88497d0ce4c5\" (UID: \"ff1acde3-5e9f-4659-a5b5-88497d0ce4c5\") " Jan 27 17:29:52 crc kubenswrapper[4772]: I0127 17:29:52.341854 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff1acde3-5e9f-4659-a5b5-88497d0ce4c5-utilities\") pod \"ff1acde3-5e9f-4659-a5b5-88497d0ce4c5\" (UID: \"ff1acde3-5e9f-4659-a5b5-88497d0ce4c5\") " Jan 27 17:29:52 crc kubenswrapper[4772]: I0127 17:29:52.341929 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45llf\" (UniqueName: \"kubernetes.io/projected/ff1acde3-5e9f-4659-a5b5-88497d0ce4c5-kube-api-access-45llf\") pod \"ff1acde3-5e9f-4659-a5b5-88497d0ce4c5\" (UID: \"ff1acde3-5e9f-4659-a5b5-88497d0ce4c5\") " Jan 27 17:29:52 crc kubenswrapper[4772]: I0127 17:29:52.343251 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff1acde3-5e9f-4659-a5b5-88497d0ce4c5-utilities" (OuterVolumeSpecName: "utilities") pod "ff1acde3-5e9f-4659-a5b5-88497d0ce4c5" (UID: "ff1acde3-5e9f-4659-a5b5-88497d0ce4c5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:29:52 crc kubenswrapper[4772]: I0127 17:29:52.352546 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff1acde3-5e9f-4659-a5b5-88497d0ce4c5-kube-api-access-45llf" (OuterVolumeSpecName: "kube-api-access-45llf") pod "ff1acde3-5e9f-4659-a5b5-88497d0ce4c5" (UID: "ff1acde3-5e9f-4659-a5b5-88497d0ce4c5"). InnerVolumeSpecName "kube-api-access-45llf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:29:52 crc kubenswrapper[4772]: I0127 17:29:52.444601 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff1acde3-5e9f-4659-a5b5-88497d0ce4c5-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 17:29:52 crc kubenswrapper[4772]: I0127 17:29:52.444646 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45llf\" (UniqueName: \"kubernetes.io/projected/ff1acde3-5e9f-4659-a5b5-88497d0ce4c5-kube-api-access-45llf\") on node \"crc\" DevicePath \"\"" Jan 27 17:29:52 crc kubenswrapper[4772]: I0127 17:29:52.461109 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff1acde3-5e9f-4659-a5b5-88497d0ce4c5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff1acde3-5e9f-4659-a5b5-88497d0ce4c5" (UID: "ff1acde3-5e9f-4659-a5b5-88497d0ce4c5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:29:52 crc kubenswrapper[4772]: I0127 17:29:52.546650 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff1acde3-5e9f-4659-a5b5-88497d0ce4c5-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 17:29:52 crc kubenswrapper[4772]: I0127 17:29:52.869566 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5pg8" event={"ID":"ff1acde3-5e9f-4659-a5b5-88497d0ce4c5","Type":"ContainerDied","Data":"4880d61008d58f60627606eb86b9a7ff700103c4a3a917606afb4f4e5637a700"} Jan 27 17:29:52 crc kubenswrapper[4772]: I0127 17:29:52.870416 4772 scope.go:117] "RemoveContainer" containerID="024773fe5b0145280bf522bfd231862b43616d6d3e7c272e497743d5b82823f5" Jan 27 17:29:52 crc kubenswrapper[4772]: I0127 17:29:52.869677 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r5pg8" Jan 27 17:29:52 crc kubenswrapper[4772]: I0127 17:29:52.927254 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r5pg8"] Jan 27 17:29:52 crc kubenswrapper[4772]: I0127 17:29:52.938094 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-r5pg8"] Jan 27 17:29:52 crc kubenswrapper[4772]: I0127 17:29:52.943873 4772 scope.go:117] "RemoveContainer" containerID="be4fbbd1741965f3fcecf5b9a29d8a1a7d7296f88ab213fffd113439cfaa68b6" Jan 27 17:29:52 crc kubenswrapper[4772]: I0127 17:29:52.978577 4772 scope.go:117] "RemoveContainer" containerID="726824e909b709ea955048da1b47107c917a9ead2621a9e65cc30363053366cb" Jan 27 17:29:54 crc kubenswrapper[4772]: I0127 17:29:54.680538 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff1acde3-5e9f-4659-a5b5-88497d0ce4c5" path="/var/lib/kubelet/pods/ff1acde3-5e9f-4659-a5b5-88497d0ce4c5/volumes" Jan 27 17:29:59 crc kubenswrapper[4772]: I0127 17:29:59.664074 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:29:59 crc kubenswrapper[4772]: E0127 17:29:59.664838 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:30:00 crc kubenswrapper[4772]: I0127 17:30:00.156334 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492250-lgh54"] Jan 27 17:30:00 crc kubenswrapper[4772]: E0127 17:30:00.157507 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff1acde3-5e9f-4659-a5b5-88497d0ce4c5" containerName="extract-utilities" Jan 27 17:30:00 crc kubenswrapper[4772]: I0127 17:30:00.157551 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff1acde3-5e9f-4659-a5b5-88497d0ce4c5" containerName="extract-utilities" Jan 27 17:30:00 crc kubenswrapper[4772]: E0127 17:30:00.157573 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff1acde3-5e9f-4659-a5b5-88497d0ce4c5" containerName="extract-content" Jan 27 17:30:00 crc kubenswrapper[4772]: I0127 17:30:00.157589 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff1acde3-5e9f-4659-a5b5-88497d0ce4c5" containerName="extract-content" Jan 27 17:30:00 crc kubenswrapper[4772]: E0127 17:30:00.157611 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff1acde3-5e9f-4659-a5b5-88497d0ce4c5" containerName="registry-server" Jan 27 17:30:00 crc kubenswrapper[4772]: I0127 17:30:00.157624 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff1acde3-5e9f-4659-a5b5-88497d0ce4c5" containerName="registry-server" Jan 27 17:30:00 crc kubenswrapper[4772]: I0127 17:30:00.158028 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff1acde3-5e9f-4659-a5b5-88497d0ce4c5" containerName="registry-server" Jan 27 17:30:00 crc kubenswrapper[4772]: I0127 17:30:00.159084 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492250-lgh54" Jan 27 17:30:00 crc kubenswrapper[4772]: I0127 17:30:00.161075 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 27 17:30:00 crc kubenswrapper[4772]: I0127 17:30:00.161450 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 27 17:30:00 crc kubenswrapper[4772]: I0127 17:30:00.190659 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492250-lgh54"] Jan 27 17:30:00 crc kubenswrapper[4772]: I0127 17:30:00.318971 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4v8d\" (UniqueName: \"kubernetes.io/projected/c02097b9-7eb9-4e2c-ad5e-5ec0983cb874-kube-api-access-c4v8d\") pod \"collect-profiles-29492250-lgh54\" (UID: \"c02097b9-7eb9-4e2c-ad5e-5ec0983cb874\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492250-lgh54" Jan 27 17:30:00 crc kubenswrapper[4772]: I0127 17:30:00.319239 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c02097b9-7eb9-4e2c-ad5e-5ec0983cb874-config-volume\") pod \"collect-profiles-29492250-lgh54\" (UID: \"c02097b9-7eb9-4e2c-ad5e-5ec0983cb874\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492250-lgh54" Jan 27 17:30:00 crc kubenswrapper[4772]: I0127 17:30:00.319433 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c02097b9-7eb9-4e2c-ad5e-5ec0983cb874-secret-volume\") pod \"collect-profiles-29492250-lgh54\" (UID: \"c02097b9-7eb9-4e2c-ad5e-5ec0983cb874\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492250-lgh54" Jan 27 17:30:00 crc kubenswrapper[4772]: I0127 17:30:00.422152 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4v8d\" (UniqueName: \"kubernetes.io/projected/c02097b9-7eb9-4e2c-ad5e-5ec0983cb874-kube-api-access-c4v8d\") pod \"collect-profiles-29492250-lgh54\" (UID: \"c02097b9-7eb9-4e2c-ad5e-5ec0983cb874\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492250-lgh54" Jan 27 17:30:00 crc kubenswrapper[4772]: I0127 17:30:00.422321 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c02097b9-7eb9-4e2c-ad5e-5ec0983cb874-config-volume\") pod \"collect-profiles-29492250-lgh54\" (UID: \"c02097b9-7eb9-4e2c-ad5e-5ec0983cb874\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492250-lgh54" Jan 27 17:30:00 crc kubenswrapper[4772]: I0127 17:30:00.422394 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c02097b9-7eb9-4e2c-ad5e-5ec0983cb874-secret-volume\") pod \"collect-profiles-29492250-lgh54\" (UID: \"c02097b9-7eb9-4e2c-ad5e-5ec0983cb874\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492250-lgh54" Jan 27 17:30:00 crc kubenswrapper[4772]: I0127 17:30:00.424007 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c02097b9-7eb9-4e2c-ad5e-5ec0983cb874-config-volume\") pod \"collect-profiles-29492250-lgh54\" (UID: \"c02097b9-7eb9-4e2c-ad5e-5ec0983cb874\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492250-lgh54" Jan 27 17:30:00 crc kubenswrapper[4772]: I0127 17:30:00.885069 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c02097b9-7eb9-4e2c-ad5e-5ec0983cb874-secret-volume\") pod \"collect-profiles-29492250-lgh54\" (UID: \"c02097b9-7eb9-4e2c-ad5e-5ec0983cb874\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492250-lgh54" Jan 27 17:30:00 crc kubenswrapper[4772]: I0127 17:30:00.886289 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4v8d\" (UniqueName: \"kubernetes.io/projected/c02097b9-7eb9-4e2c-ad5e-5ec0983cb874-kube-api-access-c4v8d\") pod \"collect-profiles-29492250-lgh54\" (UID: \"c02097b9-7eb9-4e2c-ad5e-5ec0983cb874\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492250-lgh54" Jan 27 17:30:01 crc kubenswrapper[4772]: I0127 17:30:01.095578 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492250-lgh54" Jan 27 17:30:01 crc kubenswrapper[4772]: I0127 17:30:01.611767 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492250-lgh54"] Jan 27 17:30:01 crc kubenswrapper[4772]: W0127 17:30:01.627003 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc02097b9_7eb9_4e2c_ad5e_5ec0983cb874.slice/crio-e0a9f131bb0864deac9499968e098890c0ec60bcb62ff0e22a1c29d01928a81e WatchSource:0}: Error finding container e0a9f131bb0864deac9499968e098890c0ec60bcb62ff0e22a1c29d01928a81e: Status 404 returned error can't find the container with id e0a9f131bb0864deac9499968e098890c0ec60bcb62ff0e22a1c29d01928a81e Jan 27 17:30:01 crc kubenswrapper[4772]: I0127 17:30:01.952719 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492250-lgh54" event={"ID":"c02097b9-7eb9-4e2c-ad5e-5ec0983cb874","Type":"ContainerStarted","Data":"2c0ec74fa28e0e4c9ec0bad9a7085dcc6ba408566f72f39040535b3c2d917984"} Jan 27 17:30:01 crc kubenswrapper[4772]: I0127 17:30:01.952765 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492250-lgh54" event={"ID":"c02097b9-7eb9-4e2c-ad5e-5ec0983cb874","Type":"ContainerStarted","Data":"e0a9f131bb0864deac9499968e098890c0ec60bcb62ff0e22a1c29d01928a81e"} Jan 27 17:30:01 crc kubenswrapper[4772]: I0127 17:30:01.974314 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29492250-lgh54" podStartSLOduration=1.974299113 podStartE2EDuration="1.974299113s" podCreationTimestamp="2026-01-27 17:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 17:30:01.974280782 +0000 UTC m=+8587.954889900" watchObservedRunningTime="2026-01-27 17:30:01.974299113 +0000 UTC m=+8587.954908211" Jan 27 17:30:02 crc kubenswrapper[4772]: I0127 17:30:02.970647 4772 generic.go:334] "Generic (PLEG): container finished" podID="c02097b9-7eb9-4e2c-ad5e-5ec0983cb874" containerID="2c0ec74fa28e0e4c9ec0bad9a7085dcc6ba408566f72f39040535b3c2d917984" exitCode=0 Jan 27 17:30:02 crc kubenswrapper[4772]: I0127 17:30:02.970764 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492250-lgh54" event={"ID":"c02097b9-7eb9-4e2c-ad5e-5ec0983cb874","Type":"ContainerDied","Data":"2c0ec74fa28e0e4c9ec0bad9a7085dcc6ba408566f72f39040535b3c2d917984"} Jan 27 17:30:04 crc kubenswrapper[4772]: I0127 17:30:04.334041 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492250-lgh54" Jan 27 17:30:04 crc kubenswrapper[4772]: I0127 17:30:04.401316 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c02097b9-7eb9-4e2c-ad5e-5ec0983cb874-secret-volume\") pod \"c02097b9-7eb9-4e2c-ad5e-5ec0983cb874\" (UID: \"c02097b9-7eb9-4e2c-ad5e-5ec0983cb874\") " Jan 27 17:30:04 crc kubenswrapper[4772]: I0127 17:30:04.401684 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c02097b9-7eb9-4e2c-ad5e-5ec0983cb874-config-volume\") pod \"c02097b9-7eb9-4e2c-ad5e-5ec0983cb874\" (UID: \"c02097b9-7eb9-4e2c-ad5e-5ec0983cb874\") " Jan 27 17:30:04 crc kubenswrapper[4772]: I0127 17:30:04.401945 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4v8d\" (UniqueName: \"kubernetes.io/projected/c02097b9-7eb9-4e2c-ad5e-5ec0983cb874-kube-api-access-c4v8d\") pod \"c02097b9-7eb9-4e2c-ad5e-5ec0983cb874\" (UID: \"c02097b9-7eb9-4e2c-ad5e-5ec0983cb874\") " Jan 27 17:30:04 crc kubenswrapper[4772]: I0127 17:30:04.402672 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c02097b9-7eb9-4e2c-ad5e-5ec0983cb874-config-volume" (OuterVolumeSpecName: "config-volume") pod "c02097b9-7eb9-4e2c-ad5e-5ec0983cb874" (UID: "c02097b9-7eb9-4e2c-ad5e-5ec0983cb874"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 17:30:04 crc kubenswrapper[4772]: I0127 17:30:04.403539 4772 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c02097b9-7eb9-4e2c-ad5e-5ec0983cb874-config-volume\") on node \"crc\" DevicePath \"\"" Jan 27 17:30:04 crc kubenswrapper[4772]: I0127 17:30:04.409900 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c02097b9-7eb9-4e2c-ad5e-5ec0983cb874-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c02097b9-7eb9-4e2c-ad5e-5ec0983cb874" (UID: "c02097b9-7eb9-4e2c-ad5e-5ec0983cb874"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 17:30:04 crc kubenswrapper[4772]: I0127 17:30:04.412497 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c02097b9-7eb9-4e2c-ad5e-5ec0983cb874-kube-api-access-c4v8d" (OuterVolumeSpecName: "kube-api-access-c4v8d") pod "c02097b9-7eb9-4e2c-ad5e-5ec0983cb874" (UID: "c02097b9-7eb9-4e2c-ad5e-5ec0983cb874"). InnerVolumeSpecName "kube-api-access-c4v8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:30:04 crc kubenswrapper[4772]: I0127 17:30:04.505606 4772 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c02097b9-7eb9-4e2c-ad5e-5ec0983cb874-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 27 17:30:04 crc kubenswrapper[4772]: I0127 17:30:04.505681 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4v8d\" (UniqueName: \"kubernetes.io/projected/c02097b9-7eb9-4e2c-ad5e-5ec0983cb874-kube-api-access-c4v8d\") on node \"crc\" DevicePath \"\"" Jan 27 17:30:04 crc kubenswrapper[4772]: I0127 17:30:04.687790 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492205-9hv9l"] Jan 27 17:30:04 crc kubenswrapper[4772]: I0127 17:30:04.694455 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492205-9hv9l"] Jan 27 17:30:04 crc kubenswrapper[4772]: I0127 17:30:04.996745 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492250-lgh54" event={"ID":"c02097b9-7eb9-4e2c-ad5e-5ec0983cb874","Type":"ContainerDied","Data":"e0a9f131bb0864deac9499968e098890c0ec60bcb62ff0e22a1c29d01928a81e"} Jan 27 17:30:04 crc kubenswrapper[4772]: I0127 17:30:04.997127 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0a9f131bb0864deac9499968e098890c0ec60bcb62ff0e22a1c29d01928a81e" Jan 27 17:30:04 crc kubenswrapper[4772]: I0127 17:30:04.996816 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492250-lgh54" Jan 27 17:30:06 crc kubenswrapper[4772]: I0127 17:30:06.678640 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb399a66-1690-4026-9b2e-9e399d3270d2" path="/var/lib/kubelet/pods/eb399a66-1690-4026-9b2e-9e399d3270d2/volumes" Jan 27 17:30:11 crc kubenswrapper[4772]: I0127 17:30:11.158891 4772 scope.go:117] "RemoveContainer" containerID="3b5c3dfd99ca4b5982c3131c3d5ce465e41cbe4ff9774e156e9f425715410ede" Jan 27 17:30:11 crc kubenswrapper[4772]: I0127 17:30:11.663555 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:30:11 crc kubenswrapper[4772]: E0127 17:30:11.663900 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:30:22 crc kubenswrapper[4772]: I0127 17:30:22.663300 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:30:22 crc kubenswrapper[4772]: E0127 17:30:22.664218 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:30:34 crc kubenswrapper[4772]: I0127 17:30:34.678710 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:30:34 crc kubenswrapper[4772]: E0127 17:30:34.680161 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:30:48 crc kubenswrapper[4772]: I0127 17:30:48.663730 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:30:48 crc kubenswrapper[4772]: E0127 17:30:48.665052 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:31:00 crc kubenswrapper[4772]: I0127 17:31:00.664669 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:31:00 crc kubenswrapper[4772]: E0127 17:31:00.666013 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:31:15 crc kubenswrapper[4772]: I0127 17:31:15.662626 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:31:15 crc kubenswrapper[4772]: E0127 17:31:15.663485 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:31:26 crc kubenswrapper[4772]: I0127 17:31:26.663682 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:31:26 crc kubenswrapper[4772]: E0127 17:31:26.665636 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:31:40 crc kubenswrapper[4772]: I0127 17:31:40.665440 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:31:40 crc kubenswrapper[4772]: E0127 17:31:40.668057 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:31:55 crc kubenswrapper[4772]: I0127 17:31:55.663292 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:31:55 crc kubenswrapper[4772]: E0127 17:31:55.663981 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:32:10 crc kubenswrapper[4772]: I0127 17:32:10.663845 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:32:10 crc kubenswrapper[4772]: E0127 17:32:10.664835 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:32:23 crc kubenswrapper[4772]: I0127 17:32:23.663654 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:32:23 crc kubenswrapper[4772]: E0127 17:32:23.665468 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:32:34 crc kubenswrapper[4772]: I0127 17:32:34.684645 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:32:34 crc kubenswrapper[4772]: E0127 17:32:34.686202 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:32:47 crc kubenswrapper[4772]: I0127 17:32:47.678342 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:32:49 crc kubenswrapper[4772]: I0127 17:32:49.672459 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"7028bdbe5a39e55f1ad82b72de13a00af4f685c12d26eb6560d24b432826149d"} Jan 27 17:33:13 crc kubenswrapper[4772]: I0127 17:33:13.627047 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-x5jtw"] Jan 27 17:33:13 crc kubenswrapper[4772]: E0127 17:33:13.628346 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c02097b9-7eb9-4e2c-ad5e-5ec0983cb874" containerName="collect-profiles" Jan 27 17:33:13 crc kubenswrapper[4772]: I0127 17:33:13.628364 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="c02097b9-7eb9-4e2c-ad5e-5ec0983cb874" containerName="collect-profiles" Jan 27 17:33:13 crc kubenswrapper[4772]: I0127 17:33:13.628651 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="c02097b9-7eb9-4e2c-ad5e-5ec0983cb874" containerName="collect-profiles" Jan 27 17:33:13 crc kubenswrapper[4772]: I0127 17:33:13.630993 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x5jtw" Jan 27 17:33:13 crc kubenswrapper[4772]: I0127 17:33:13.637719 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x5jtw"] Jan 27 17:33:13 crc kubenswrapper[4772]: I0127 17:33:13.769828 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0effe0d1-557e-4294-b4c7-71060dec32e0-catalog-content\") pod \"community-operators-x5jtw\" (UID: \"0effe0d1-557e-4294-b4c7-71060dec32e0\") " pod="openshift-marketplace/community-operators-x5jtw" Jan 27 17:33:13 crc kubenswrapper[4772]: I0127 17:33:13.769942 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55llv\" (UniqueName: \"kubernetes.io/projected/0effe0d1-557e-4294-b4c7-71060dec32e0-kube-api-access-55llv\") pod \"community-operators-x5jtw\" (UID: \"0effe0d1-557e-4294-b4c7-71060dec32e0\") " pod="openshift-marketplace/community-operators-x5jtw" Jan 27 17:33:13 crc kubenswrapper[4772]: I0127 17:33:13.769994 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0effe0d1-557e-4294-b4c7-71060dec32e0-utilities\") pod \"community-operators-x5jtw\" (UID: \"0effe0d1-557e-4294-b4c7-71060dec32e0\") " pod="openshift-marketplace/community-operators-x5jtw" Jan 27 17:33:13 crc kubenswrapper[4772]: I0127 17:33:13.871710 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0effe0d1-557e-4294-b4c7-71060dec32e0-catalog-content\") pod \"community-operators-x5jtw\" (UID: \"0effe0d1-557e-4294-b4c7-71060dec32e0\") " pod="openshift-marketplace/community-operators-x5jtw" Jan 27 17:33:13 crc kubenswrapper[4772]: I0127 17:33:13.872046 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55llv\" (UniqueName: \"kubernetes.io/projected/0effe0d1-557e-4294-b4c7-71060dec32e0-kube-api-access-55llv\") pod \"community-operators-x5jtw\" (UID: \"0effe0d1-557e-4294-b4c7-71060dec32e0\") " pod="openshift-marketplace/community-operators-x5jtw" Jan 27 17:33:13 crc kubenswrapper[4772]: I0127 17:33:13.872194 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0effe0d1-557e-4294-b4c7-71060dec32e0-catalog-content\") pod \"community-operators-x5jtw\" (UID: \"0effe0d1-557e-4294-b4c7-71060dec32e0\") " pod="openshift-marketplace/community-operators-x5jtw" Jan 27 17:33:13 crc kubenswrapper[4772]: I0127 17:33:13.872316 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0effe0d1-557e-4294-b4c7-71060dec32e0-utilities\") pod \"community-operators-x5jtw\" (UID: \"0effe0d1-557e-4294-b4c7-71060dec32e0\") " pod="openshift-marketplace/community-operators-x5jtw" Jan 27 17:33:13 crc kubenswrapper[4772]: I0127 17:33:13.872545 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0effe0d1-557e-4294-b4c7-71060dec32e0-utilities\") pod \"community-operators-x5jtw\" (UID: \"0effe0d1-557e-4294-b4c7-71060dec32e0\") " pod="openshift-marketplace/community-operators-x5jtw" Jan 27 17:33:14 crc kubenswrapper[4772]: I0127 17:33:14.286147 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55llv\" (UniqueName: \"kubernetes.io/projected/0effe0d1-557e-4294-b4c7-71060dec32e0-kube-api-access-55llv\") pod \"community-operators-x5jtw\" (UID: \"0effe0d1-557e-4294-b4c7-71060dec32e0\") " pod="openshift-marketplace/community-operators-x5jtw" Jan 27 17:33:14 crc kubenswrapper[4772]: I0127 17:33:14.565664 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x5jtw" Jan 27 17:33:15 crc kubenswrapper[4772]: I0127 17:33:15.139888 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x5jtw"] Jan 27 17:33:15 crc kubenswrapper[4772]: I0127 17:33:15.971751 4772 generic.go:334] "Generic (PLEG): container finished" podID="0effe0d1-557e-4294-b4c7-71060dec32e0" containerID="d553db1d54a8e5c7fa11aff0982e1c195dfbd54e62c92c4bd24f05551d70a976" exitCode=0 Jan 27 17:33:15 crc kubenswrapper[4772]: I0127 17:33:15.971968 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x5jtw" event={"ID":"0effe0d1-557e-4294-b4c7-71060dec32e0","Type":"ContainerDied","Data":"d553db1d54a8e5c7fa11aff0982e1c195dfbd54e62c92c4bd24f05551d70a976"} Jan 27 17:33:15 crc kubenswrapper[4772]: I0127 17:33:15.972457 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x5jtw" event={"ID":"0effe0d1-557e-4294-b4c7-71060dec32e0","Type":"ContainerStarted","Data":"d2736ac9699bef9db0c5e6e51ca8cd5f0baa3e5f558e7d1f0fcced8539a9c015"} Jan 27 17:33:15 crc kubenswrapper[4772]: I0127 17:33:15.974895 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 17:33:18 crc kubenswrapper[4772]: I0127 17:33:18.005143 4772 generic.go:334] "Generic (PLEG): container finished" podID="0effe0d1-557e-4294-b4c7-71060dec32e0" containerID="e03d3a2bd2216c5cd5a13cc3d14af89ab95fccb13a8daba4c24389313380f97e" exitCode=0 Jan 27 17:33:18 crc kubenswrapper[4772]: I0127 17:33:18.005213 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x5jtw" event={"ID":"0effe0d1-557e-4294-b4c7-71060dec32e0","Type":"ContainerDied","Data":"e03d3a2bd2216c5cd5a13cc3d14af89ab95fccb13a8daba4c24389313380f97e"} Jan 27 17:33:19 crc kubenswrapper[4772]: I0127 17:33:19.022500 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x5jtw" event={"ID":"0effe0d1-557e-4294-b4c7-71060dec32e0","Type":"ContainerStarted","Data":"bf765e64a424bbf22c0d9aa03a33fb814fb2c400906dcca8835df8f59a67eddc"} Jan 27 17:33:19 crc kubenswrapper[4772]: I0127 17:33:19.055721 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-x5jtw" podStartSLOduration=3.5990815830000003 podStartE2EDuration="6.055703774s" podCreationTimestamp="2026-01-27 17:33:13 +0000 UTC" firstStartedPulling="2026-01-27 17:33:15.974334131 +0000 UTC m=+8781.954943259" lastFinishedPulling="2026-01-27 17:33:18.430956332 +0000 UTC m=+8784.411565450" observedRunningTime="2026-01-27 17:33:19.053473392 +0000 UTC m=+8785.034082490" watchObservedRunningTime="2026-01-27 17:33:19.055703774 +0000 UTC m=+8785.036312872" Jan 27 17:33:24 crc kubenswrapper[4772]: I0127 17:33:24.565858 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-x5jtw" Jan 27 17:33:24 crc kubenswrapper[4772]: I0127 17:33:24.566608 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-x5jtw" Jan 27 17:33:24 crc kubenswrapper[4772]: I0127 17:33:24.639837 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-x5jtw" Jan 27 17:33:25 crc kubenswrapper[4772]: I0127 17:33:25.167963 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-x5jtw" Jan 27 17:33:25 crc kubenswrapper[4772]: I0127 17:33:25.248049 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x5jtw"] Jan 27 17:33:27 crc kubenswrapper[4772]: I0127 17:33:27.114050 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-x5jtw" podUID="0effe0d1-557e-4294-b4c7-71060dec32e0" containerName="registry-server" containerID="cri-o://bf765e64a424bbf22c0d9aa03a33fb814fb2c400906dcca8835df8f59a67eddc" gracePeriod=2 Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.025738 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x5jtw" Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.123957 4772 generic.go:334] "Generic (PLEG): container finished" podID="0effe0d1-557e-4294-b4c7-71060dec32e0" containerID="bf765e64a424bbf22c0d9aa03a33fb814fb2c400906dcca8835df8f59a67eddc" exitCode=0 Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.123993 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x5jtw" event={"ID":"0effe0d1-557e-4294-b4c7-71060dec32e0","Type":"ContainerDied","Data":"bf765e64a424bbf22c0d9aa03a33fb814fb2c400906dcca8835df8f59a67eddc"} Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.124021 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x5jtw" event={"ID":"0effe0d1-557e-4294-b4c7-71060dec32e0","Type":"ContainerDied","Data":"d2736ac9699bef9db0c5e6e51ca8cd5f0baa3e5f558e7d1f0fcced8539a9c015"} Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.124036 4772 scope.go:117] "RemoveContainer" containerID="bf765e64a424bbf22c0d9aa03a33fb814fb2c400906dcca8835df8f59a67eddc" Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.124059 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x5jtw" Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.144012 4772 scope.go:117] "RemoveContainer" containerID="e03d3a2bd2216c5cd5a13cc3d14af89ab95fccb13a8daba4c24389313380f97e" Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.164217 4772 scope.go:117] "RemoveContainer" containerID="d553db1d54a8e5c7fa11aff0982e1c195dfbd54e62c92c4bd24f05551d70a976" Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.193826 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0effe0d1-557e-4294-b4c7-71060dec32e0-utilities\") pod \"0effe0d1-557e-4294-b4c7-71060dec32e0\" (UID: \"0effe0d1-557e-4294-b4c7-71060dec32e0\") " Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.193891 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55llv\" (UniqueName: \"kubernetes.io/projected/0effe0d1-557e-4294-b4c7-71060dec32e0-kube-api-access-55llv\") pod \"0effe0d1-557e-4294-b4c7-71060dec32e0\" (UID: \"0effe0d1-557e-4294-b4c7-71060dec32e0\") " Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.194064 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0effe0d1-557e-4294-b4c7-71060dec32e0-catalog-content\") pod \"0effe0d1-557e-4294-b4c7-71060dec32e0\" (UID: \"0effe0d1-557e-4294-b4c7-71060dec32e0\") " Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.194975 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0effe0d1-557e-4294-b4c7-71060dec32e0-utilities" (OuterVolumeSpecName: "utilities") pod "0effe0d1-557e-4294-b4c7-71060dec32e0" (UID: "0effe0d1-557e-4294-b4c7-71060dec32e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.200241 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0effe0d1-557e-4294-b4c7-71060dec32e0-kube-api-access-55llv" (OuterVolumeSpecName: "kube-api-access-55llv") pod "0effe0d1-557e-4294-b4c7-71060dec32e0" (UID: "0effe0d1-557e-4294-b4c7-71060dec32e0"). InnerVolumeSpecName "kube-api-access-55llv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.268382 4772 scope.go:117] "RemoveContainer" containerID="bf765e64a424bbf22c0d9aa03a33fb814fb2c400906dcca8835df8f59a67eddc" Jan 27 17:33:28 crc kubenswrapper[4772]: E0127 17:33:28.269211 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf765e64a424bbf22c0d9aa03a33fb814fb2c400906dcca8835df8f59a67eddc\": container with ID starting with bf765e64a424bbf22c0d9aa03a33fb814fb2c400906dcca8835df8f59a67eddc not found: ID does not exist" containerID="bf765e64a424bbf22c0d9aa03a33fb814fb2c400906dcca8835df8f59a67eddc" Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.269281 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf765e64a424bbf22c0d9aa03a33fb814fb2c400906dcca8835df8f59a67eddc"} err="failed to get container status \"bf765e64a424bbf22c0d9aa03a33fb814fb2c400906dcca8835df8f59a67eddc\": rpc error: code = NotFound desc = could not find container \"bf765e64a424bbf22c0d9aa03a33fb814fb2c400906dcca8835df8f59a67eddc\": container with ID starting with bf765e64a424bbf22c0d9aa03a33fb814fb2c400906dcca8835df8f59a67eddc not found: ID does not exist" Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.269364 4772 scope.go:117] "RemoveContainer" containerID="e03d3a2bd2216c5cd5a13cc3d14af89ab95fccb13a8daba4c24389313380f97e" Jan 27 17:33:28 crc kubenswrapper[4772]: E0127 17:33:28.270323 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e03d3a2bd2216c5cd5a13cc3d14af89ab95fccb13a8daba4c24389313380f97e\": container with ID starting with e03d3a2bd2216c5cd5a13cc3d14af89ab95fccb13a8daba4c24389313380f97e not found: ID does not exist" containerID="e03d3a2bd2216c5cd5a13cc3d14af89ab95fccb13a8daba4c24389313380f97e" Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.270385 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e03d3a2bd2216c5cd5a13cc3d14af89ab95fccb13a8daba4c24389313380f97e"} err="failed to get container status \"e03d3a2bd2216c5cd5a13cc3d14af89ab95fccb13a8daba4c24389313380f97e\": rpc error: code = NotFound desc = could not find container \"e03d3a2bd2216c5cd5a13cc3d14af89ab95fccb13a8daba4c24389313380f97e\": container with ID starting with e03d3a2bd2216c5cd5a13cc3d14af89ab95fccb13a8daba4c24389313380f97e not found: ID does not exist" Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.270416 4772 scope.go:117] "RemoveContainer" containerID="d553db1d54a8e5c7fa11aff0982e1c195dfbd54e62c92c4bd24f05551d70a976" Jan 27 17:33:28 crc kubenswrapper[4772]: E0127 17:33:28.271144 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d553db1d54a8e5c7fa11aff0982e1c195dfbd54e62c92c4bd24f05551d70a976\": container with ID starting with d553db1d54a8e5c7fa11aff0982e1c195dfbd54e62c92c4bd24f05551d70a976 not found: ID does not exist" containerID="d553db1d54a8e5c7fa11aff0982e1c195dfbd54e62c92c4bd24f05551d70a976" Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.271219 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d553db1d54a8e5c7fa11aff0982e1c195dfbd54e62c92c4bd24f05551d70a976"} err="failed to get container status \"d553db1d54a8e5c7fa11aff0982e1c195dfbd54e62c92c4bd24f05551d70a976\": rpc error: code = NotFound desc = could not find container \"d553db1d54a8e5c7fa11aff0982e1c195dfbd54e62c92c4bd24f05551d70a976\": container with ID starting with d553db1d54a8e5c7fa11aff0982e1c195dfbd54e62c92c4bd24f05551d70a976 not found: ID does not exist" Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.296469 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0effe0d1-557e-4294-b4c7-71060dec32e0-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.296506 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55llv\" (UniqueName: \"kubernetes.io/projected/0effe0d1-557e-4294-b4c7-71060dec32e0-kube-api-access-55llv\") on node \"crc\" DevicePath \"\"" Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.521004 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0effe0d1-557e-4294-b4c7-71060dec32e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0effe0d1-557e-4294-b4c7-71060dec32e0" (UID: "0effe0d1-557e-4294-b4c7-71060dec32e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.601439 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0effe0d1-557e-4294-b4c7-71060dec32e0-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.757701 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x5jtw"] Jan 27 17:33:28 crc kubenswrapper[4772]: I0127 17:33:28.774372 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-x5jtw"] Jan 27 17:33:30 crc kubenswrapper[4772]: I0127 17:33:30.675219 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0effe0d1-557e-4294-b4c7-71060dec32e0" path="/var/lib/kubelet/pods/0effe0d1-557e-4294-b4c7-71060dec32e0/volumes" Jan 27 17:33:35 crc kubenswrapper[4772]: I0127 17:33:35.843387 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mnwvb"] Jan 27 17:33:35 crc kubenswrapper[4772]: E0127 17:33:35.846706 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0effe0d1-557e-4294-b4c7-71060dec32e0" containerName="registry-server" Jan 27 17:33:35 crc kubenswrapper[4772]: I0127 17:33:35.846781 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0effe0d1-557e-4294-b4c7-71060dec32e0" containerName="registry-server" Jan 27 17:33:35 crc kubenswrapper[4772]: E0127 17:33:35.846803 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0effe0d1-557e-4294-b4c7-71060dec32e0" containerName="extract-utilities" Jan 27 17:33:35 crc kubenswrapper[4772]: I0127 17:33:35.846816 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0effe0d1-557e-4294-b4c7-71060dec32e0" containerName="extract-utilities" Jan 27 17:33:35 crc kubenswrapper[4772]: E0127 17:33:35.847017 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0effe0d1-557e-4294-b4c7-71060dec32e0" containerName="extract-content" Jan 27 17:33:35 crc kubenswrapper[4772]: I0127 17:33:35.847035 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0effe0d1-557e-4294-b4c7-71060dec32e0" containerName="extract-content" Jan 27 17:33:35 crc kubenswrapper[4772]: I0127 17:33:35.847485 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="0effe0d1-557e-4294-b4c7-71060dec32e0" containerName="registry-server" Jan 27 17:33:35 crc kubenswrapper[4772]: I0127 17:33:35.850776 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mnwvb" Jan 27 17:33:35 crc kubenswrapper[4772]: I0127 17:33:35.874734 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mnwvb"] Jan 27 17:33:36 crc kubenswrapper[4772]: I0127 17:33:36.003805 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zpgm\" (UniqueName: \"kubernetes.io/projected/480c784a-4028-480a-9400-d0c5a26072aa-kube-api-access-4zpgm\") pod \"redhat-marketplace-mnwvb\" (UID: \"480c784a-4028-480a-9400-d0c5a26072aa\") " pod="openshift-marketplace/redhat-marketplace-mnwvb" Jan 27 17:33:36 crc kubenswrapper[4772]: I0127 17:33:36.003869 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/480c784a-4028-480a-9400-d0c5a26072aa-catalog-content\") pod \"redhat-marketplace-mnwvb\" (UID: \"480c784a-4028-480a-9400-d0c5a26072aa\") " pod="openshift-marketplace/redhat-marketplace-mnwvb" Jan 27 17:33:36 crc kubenswrapper[4772]: I0127 17:33:36.004062 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/480c784a-4028-480a-9400-d0c5a26072aa-utilities\") pod \"redhat-marketplace-mnwvb\" (UID: \"480c784a-4028-480a-9400-d0c5a26072aa\") " pod="openshift-marketplace/redhat-marketplace-mnwvb" Jan 27 17:33:36 crc kubenswrapper[4772]: I0127 17:33:36.105590 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zpgm\" (UniqueName: \"kubernetes.io/projected/480c784a-4028-480a-9400-d0c5a26072aa-kube-api-access-4zpgm\") pod \"redhat-marketplace-mnwvb\" (UID: \"480c784a-4028-480a-9400-d0c5a26072aa\") " pod="openshift-marketplace/redhat-marketplace-mnwvb" Jan 27 17:33:36 crc kubenswrapper[4772]: I0127 17:33:36.105690 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/480c784a-4028-480a-9400-d0c5a26072aa-catalog-content\") pod \"redhat-marketplace-mnwvb\" (UID: \"480c784a-4028-480a-9400-d0c5a26072aa\") " pod="openshift-marketplace/redhat-marketplace-mnwvb" Jan 27 17:33:36 crc kubenswrapper[4772]: I0127 17:33:36.105759 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/480c784a-4028-480a-9400-d0c5a26072aa-utilities\") pod \"redhat-marketplace-mnwvb\" (UID: \"480c784a-4028-480a-9400-d0c5a26072aa\") " pod="openshift-marketplace/redhat-marketplace-mnwvb" Jan 27 17:33:36 crc kubenswrapper[4772]: I0127 17:33:36.106287 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/480c784a-4028-480a-9400-d0c5a26072aa-utilities\") pod \"redhat-marketplace-mnwvb\" (UID: \"480c784a-4028-480a-9400-d0c5a26072aa\") " pod="openshift-marketplace/redhat-marketplace-mnwvb" Jan 27 17:33:36 crc kubenswrapper[4772]: I0127 17:33:36.106557 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/480c784a-4028-480a-9400-d0c5a26072aa-catalog-content\") pod \"redhat-marketplace-mnwvb\" (UID: \"480c784a-4028-480a-9400-d0c5a26072aa\") " pod="openshift-marketplace/redhat-marketplace-mnwvb" Jan 27 17:33:36 crc kubenswrapper[4772]: I0127 17:33:36.123850 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zpgm\" (UniqueName: \"kubernetes.io/projected/480c784a-4028-480a-9400-d0c5a26072aa-kube-api-access-4zpgm\") pod \"redhat-marketplace-mnwvb\" (UID: \"480c784a-4028-480a-9400-d0c5a26072aa\") " pod="openshift-marketplace/redhat-marketplace-mnwvb" Jan 27 17:33:36 crc kubenswrapper[4772]: I0127 17:33:36.178001 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mnwvb" Jan 27 17:33:36 crc kubenswrapper[4772]: I0127 17:33:36.716893 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mnwvb"] Jan 27 17:33:37 crc kubenswrapper[4772]: I0127 17:33:37.216609 4772 generic.go:334] "Generic (PLEG): container finished" podID="480c784a-4028-480a-9400-d0c5a26072aa" containerID="49101020f0d6ef70ba5d6fc50a8fcd66dcfad489ffd620338fb72a9aa2b0bab6" exitCode=0 Jan 27 17:33:37 crc kubenswrapper[4772]: I0127 17:33:37.216651 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mnwvb" event={"ID":"480c784a-4028-480a-9400-d0c5a26072aa","Type":"ContainerDied","Data":"49101020f0d6ef70ba5d6fc50a8fcd66dcfad489ffd620338fb72a9aa2b0bab6"} Jan 27 17:33:37 crc kubenswrapper[4772]: I0127 17:33:37.216677 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mnwvb" event={"ID":"480c784a-4028-480a-9400-d0c5a26072aa","Type":"ContainerStarted","Data":"64fe0e4178b2f7c828957a4dc7ec17e010d253740ac714a22f2b3416b207e566"} Jan 27 17:33:39 crc kubenswrapper[4772]: I0127 17:33:39.242262 4772 generic.go:334] "Generic (PLEG): container finished" podID="480c784a-4028-480a-9400-d0c5a26072aa" containerID="c0832a14e25fd5e95433f847268a64854b319bd9eea88f84e16396ac7edd5e8d" exitCode=0 Jan 27 17:33:39 crc kubenswrapper[4772]: I0127 17:33:39.242350 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mnwvb" event={"ID":"480c784a-4028-480a-9400-d0c5a26072aa","Type":"ContainerDied","Data":"c0832a14e25fd5e95433f847268a64854b319bd9eea88f84e16396ac7edd5e8d"} Jan 27 17:33:40 crc kubenswrapper[4772]: I0127 17:33:40.254867 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mnwvb" event={"ID":"480c784a-4028-480a-9400-d0c5a26072aa","Type":"ContainerStarted","Data":"39dbf45dc1eeb8a5861376f800ef838077a260ad19d02ba9cc20d0c42f53df72"} Jan 27 17:33:40 crc kubenswrapper[4772]: I0127 17:33:40.284002 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mnwvb" podStartSLOduration=2.833658369 podStartE2EDuration="5.283983716s" podCreationTimestamp="2026-01-27 17:33:35 +0000 UTC" firstStartedPulling="2026-01-27 17:33:37.219674427 +0000 UTC m=+8803.200283565" lastFinishedPulling="2026-01-27 17:33:39.669999794 +0000 UTC m=+8805.650608912" observedRunningTime="2026-01-27 17:33:40.278338029 +0000 UTC m=+8806.258947127" watchObservedRunningTime="2026-01-27 17:33:40.283983716 +0000 UTC m=+8806.264592814" Jan 27 17:33:46 crc kubenswrapper[4772]: I0127 17:33:46.178829 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mnwvb" Jan 27 17:33:46 crc kubenswrapper[4772]: I0127 17:33:46.179671 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mnwvb" Jan 27 17:33:46 crc kubenswrapper[4772]: I0127 17:33:46.253951 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mnwvb" Jan 27 17:33:46 crc kubenswrapper[4772]: I0127 17:33:46.370924 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mnwvb" Jan 27 17:33:46 crc kubenswrapper[4772]: I0127 17:33:46.505190 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mnwvb"] Jan 27 17:33:48 crc kubenswrapper[4772]: I0127 17:33:48.332164 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mnwvb" podUID="480c784a-4028-480a-9400-d0c5a26072aa" containerName="registry-server" containerID="cri-o://39dbf45dc1eeb8a5861376f800ef838077a260ad19d02ba9cc20d0c42f53df72" gracePeriod=2 Jan 27 17:33:48 crc kubenswrapper[4772]: I0127 17:33:48.841828 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mnwvb" Jan 27 17:33:48 crc kubenswrapper[4772]: I0127 17:33:48.927266 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/480c784a-4028-480a-9400-d0c5a26072aa-utilities\") pod \"480c784a-4028-480a-9400-d0c5a26072aa\" (UID: \"480c784a-4028-480a-9400-d0c5a26072aa\") " Jan 27 17:33:48 crc kubenswrapper[4772]: I0127 17:33:48.927334 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/480c784a-4028-480a-9400-d0c5a26072aa-catalog-content\") pod \"480c784a-4028-480a-9400-d0c5a26072aa\" (UID: \"480c784a-4028-480a-9400-d0c5a26072aa\") " Jan 27 17:33:48 crc kubenswrapper[4772]: I0127 17:33:48.927536 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zpgm\" (UniqueName: \"kubernetes.io/projected/480c784a-4028-480a-9400-d0c5a26072aa-kube-api-access-4zpgm\") pod \"480c784a-4028-480a-9400-d0c5a26072aa\" (UID: \"480c784a-4028-480a-9400-d0c5a26072aa\") " Jan 27 17:33:48 crc kubenswrapper[4772]: I0127 17:33:48.928995 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/480c784a-4028-480a-9400-d0c5a26072aa-utilities" (OuterVolumeSpecName: "utilities") pod "480c784a-4028-480a-9400-d0c5a26072aa" (UID: "480c784a-4028-480a-9400-d0c5a26072aa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:33:48 crc kubenswrapper[4772]: I0127 17:33:48.933646 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/480c784a-4028-480a-9400-d0c5a26072aa-kube-api-access-4zpgm" (OuterVolumeSpecName: "kube-api-access-4zpgm") pod "480c784a-4028-480a-9400-d0c5a26072aa" (UID: "480c784a-4028-480a-9400-d0c5a26072aa"). InnerVolumeSpecName "kube-api-access-4zpgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:33:48 crc kubenswrapper[4772]: I0127 17:33:48.951234 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/480c784a-4028-480a-9400-d0c5a26072aa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "480c784a-4028-480a-9400-d0c5a26072aa" (UID: "480c784a-4028-480a-9400-d0c5a26072aa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:33:49 crc kubenswrapper[4772]: I0127 17:33:49.029588 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/480c784a-4028-480a-9400-d0c5a26072aa-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 17:33:49 crc kubenswrapper[4772]: I0127 17:33:49.029630 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/480c784a-4028-480a-9400-d0c5a26072aa-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 17:33:49 crc kubenswrapper[4772]: I0127 17:33:49.029646 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zpgm\" (UniqueName: \"kubernetes.io/projected/480c784a-4028-480a-9400-d0c5a26072aa-kube-api-access-4zpgm\") on node \"crc\" DevicePath \"\"" Jan 27 17:33:49 crc kubenswrapper[4772]: I0127 17:33:49.347242 4772 generic.go:334] "Generic (PLEG): container finished" podID="480c784a-4028-480a-9400-d0c5a26072aa" containerID="39dbf45dc1eeb8a5861376f800ef838077a260ad19d02ba9cc20d0c42f53df72" exitCode=0 Jan 27 17:33:49 crc kubenswrapper[4772]: I0127 17:33:49.347396 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mnwvb" Jan 27 17:33:49 crc kubenswrapper[4772]: I0127 17:33:49.348495 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mnwvb" event={"ID":"480c784a-4028-480a-9400-d0c5a26072aa","Type":"ContainerDied","Data":"39dbf45dc1eeb8a5861376f800ef838077a260ad19d02ba9cc20d0c42f53df72"} Jan 27 17:33:49 crc kubenswrapper[4772]: I0127 17:33:49.348624 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mnwvb" event={"ID":"480c784a-4028-480a-9400-d0c5a26072aa","Type":"ContainerDied","Data":"64fe0e4178b2f7c828957a4dc7ec17e010d253740ac714a22f2b3416b207e566"} Jan 27 17:33:49 crc kubenswrapper[4772]: I0127 17:33:49.348720 4772 scope.go:117] "RemoveContainer" containerID="39dbf45dc1eeb8a5861376f800ef838077a260ad19d02ba9cc20d0c42f53df72" Jan 27 17:33:49 crc kubenswrapper[4772]: I0127 17:33:49.389030 4772 scope.go:117] "RemoveContainer" containerID="c0832a14e25fd5e95433f847268a64854b319bd9eea88f84e16396ac7edd5e8d" Jan 27 17:33:49 crc kubenswrapper[4772]: I0127 17:33:49.406049 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mnwvb"] Jan 27 17:33:49 crc kubenswrapper[4772]: I0127 17:33:49.429927 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mnwvb"] Jan 27 17:33:49 crc kubenswrapper[4772]: I0127 17:33:49.434837 4772 scope.go:117] "RemoveContainer" containerID="49101020f0d6ef70ba5d6fc50a8fcd66dcfad489ffd620338fb72a9aa2b0bab6" Jan 27 17:33:49 crc kubenswrapper[4772]: I0127 17:33:49.497135 4772 scope.go:117] "RemoveContainer" containerID="39dbf45dc1eeb8a5861376f800ef838077a260ad19d02ba9cc20d0c42f53df72" Jan 27 17:33:49 crc kubenswrapper[4772]: E0127 17:33:49.497677 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39dbf45dc1eeb8a5861376f800ef838077a260ad19d02ba9cc20d0c42f53df72\": container with ID starting with 39dbf45dc1eeb8a5861376f800ef838077a260ad19d02ba9cc20d0c42f53df72 not found: ID does not exist" containerID="39dbf45dc1eeb8a5861376f800ef838077a260ad19d02ba9cc20d0c42f53df72" Jan 27 17:33:49 crc kubenswrapper[4772]: I0127 17:33:49.497718 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39dbf45dc1eeb8a5861376f800ef838077a260ad19d02ba9cc20d0c42f53df72"} err="failed to get container status \"39dbf45dc1eeb8a5861376f800ef838077a260ad19d02ba9cc20d0c42f53df72\": rpc error: code = NotFound desc = could not find container \"39dbf45dc1eeb8a5861376f800ef838077a260ad19d02ba9cc20d0c42f53df72\": container with ID starting with 39dbf45dc1eeb8a5861376f800ef838077a260ad19d02ba9cc20d0c42f53df72 not found: ID does not exist" Jan 27 17:33:49 crc kubenswrapper[4772]: I0127 17:33:49.497744 4772 scope.go:117] "RemoveContainer" containerID="c0832a14e25fd5e95433f847268a64854b319bd9eea88f84e16396ac7edd5e8d" Jan 27 17:33:49 crc kubenswrapper[4772]: E0127 17:33:49.498117 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0832a14e25fd5e95433f847268a64854b319bd9eea88f84e16396ac7edd5e8d\": container with ID starting with c0832a14e25fd5e95433f847268a64854b319bd9eea88f84e16396ac7edd5e8d not found: ID does not exist" containerID="c0832a14e25fd5e95433f847268a64854b319bd9eea88f84e16396ac7edd5e8d" Jan 27 17:33:49 crc kubenswrapper[4772]: I0127 17:33:49.498153 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0832a14e25fd5e95433f847268a64854b319bd9eea88f84e16396ac7edd5e8d"} err="failed to get container status \"c0832a14e25fd5e95433f847268a64854b319bd9eea88f84e16396ac7edd5e8d\": rpc error: code = NotFound desc = could not find container \"c0832a14e25fd5e95433f847268a64854b319bd9eea88f84e16396ac7edd5e8d\": container with ID starting with c0832a14e25fd5e95433f847268a64854b319bd9eea88f84e16396ac7edd5e8d not found: ID does not exist" Jan 27 17:33:49 crc kubenswrapper[4772]: I0127 17:33:49.498207 4772 scope.go:117] "RemoveContainer" containerID="49101020f0d6ef70ba5d6fc50a8fcd66dcfad489ffd620338fb72a9aa2b0bab6" Jan 27 17:33:49 crc kubenswrapper[4772]: E0127 17:33:49.498680 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49101020f0d6ef70ba5d6fc50a8fcd66dcfad489ffd620338fb72a9aa2b0bab6\": container with ID starting with 49101020f0d6ef70ba5d6fc50a8fcd66dcfad489ffd620338fb72a9aa2b0bab6 not found: ID does not exist" containerID="49101020f0d6ef70ba5d6fc50a8fcd66dcfad489ffd620338fb72a9aa2b0bab6" Jan 27 17:33:49 crc kubenswrapper[4772]: I0127 17:33:49.498707 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49101020f0d6ef70ba5d6fc50a8fcd66dcfad489ffd620338fb72a9aa2b0bab6"} err="failed to get container status \"49101020f0d6ef70ba5d6fc50a8fcd66dcfad489ffd620338fb72a9aa2b0bab6\": rpc error: code = NotFound desc = could not find container \"49101020f0d6ef70ba5d6fc50a8fcd66dcfad489ffd620338fb72a9aa2b0bab6\": container with ID starting with 49101020f0d6ef70ba5d6fc50a8fcd66dcfad489ffd620338fb72a9aa2b0bab6 not found: ID does not exist" Jan 27 17:33:50 crc kubenswrapper[4772]: I0127 17:33:50.672822 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="480c784a-4028-480a-9400-d0c5a26072aa" path="/var/lib/kubelet/pods/480c784a-4028-480a-9400-d0c5a26072aa/volumes" Jan 27 17:35:12 crc kubenswrapper[4772]: I0127 17:35:12.058437 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:35:12 crc kubenswrapper[4772]: I0127 17:35:12.059694 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:35:42 crc kubenswrapper[4772]: I0127 17:35:42.059005 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:35:42 crc kubenswrapper[4772]: I0127 17:35:42.059657 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:35:51 crc kubenswrapper[4772]: I0127 17:35:51.630735 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nz2jj"] Jan 27 17:35:51 crc kubenswrapper[4772]: E0127 17:35:51.634006 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="480c784a-4028-480a-9400-d0c5a26072aa" containerName="extract-content" Jan 27 17:35:51 crc kubenswrapper[4772]: I0127 17:35:51.634148 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="480c784a-4028-480a-9400-d0c5a26072aa" containerName="extract-content" Jan 27 17:35:51 crc kubenswrapper[4772]: E0127 17:35:51.634282 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="480c784a-4028-480a-9400-d0c5a26072aa" containerName="extract-utilities" Jan 27 17:35:51 crc kubenswrapper[4772]: I0127 17:35:51.634378 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="480c784a-4028-480a-9400-d0c5a26072aa" containerName="extract-utilities" Jan 27 17:35:51 crc kubenswrapper[4772]: E0127 17:35:51.634528 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="480c784a-4028-480a-9400-d0c5a26072aa" containerName="registry-server" Jan 27 17:35:51 crc kubenswrapper[4772]: I0127 17:35:51.634612 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="480c784a-4028-480a-9400-d0c5a26072aa" containerName="registry-server" Jan 27 17:35:51 crc kubenswrapper[4772]: I0127 17:35:51.635125 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="480c784a-4028-480a-9400-d0c5a26072aa" containerName="registry-server" Jan 27 17:35:51 crc kubenswrapper[4772]: I0127 17:35:51.637515 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nz2jj" Jan 27 17:35:51 crc kubenswrapper[4772]: I0127 17:35:51.658542 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nz2jj"] Jan 27 17:35:51 crc kubenswrapper[4772]: I0127 17:35:51.740321 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3da4f7ee-931f-4606-b652-ca6bb3b36bcc-catalog-content\") pod \"certified-operators-nz2jj\" (UID: \"3da4f7ee-931f-4606-b652-ca6bb3b36bcc\") " pod="openshift-marketplace/certified-operators-nz2jj" Jan 27 17:35:51 crc kubenswrapper[4772]: I0127 17:35:51.740387 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmtv8\" (UniqueName: \"kubernetes.io/projected/3da4f7ee-931f-4606-b652-ca6bb3b36bcc-kube-api-access-wmtv8\") pod \"certified-operators-nz2jj\" (UID: \"3da4f7ee-931f-4606-b652-ca6bb3b36bcc\") " pod="openshift-marketplace/certified-operators-nz2jj" Jan 27 17:35:51 crc kubenswrapper[4772]: I0127 17:35:51.741405 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3da4f7ee-931f-4606-b652-ca6bb3b36bcc-utilities\") pod \"certified-operators-nz2jj\" (UID: \"3da4f7ee-931f-4606-b652-ca6bb3b36bcc\") " pod="openshift-marketplace/certified-operators-nz2jj" Jan 27 17:35:51 crc kubenswrapper[4772]: I0127 17:35:51.843016 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3da4f7ee-931f-4606-b652-ca6bb3b36bcc-utilities\") pod \"certified-operators-nz2jj\" (UID: \"3da4f7ee-931f-4606-b652-ca6bb3b36bcc\") " pod="openshift-marketplace/certified-operators-nz2jj" Jan 27 17:35:51 crc kubenswrapper[4772]: I0127 17:35:51.843189 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3da4f7ee-931f-4606-b652-ca6bb3b36bcc-catalog-content\") pod \"certified-operators-nz2jj\" (UID: \"3da4f7ee-931f-4606-b652-ca6bb3b36bcc\") " pod="openshift-marketplace/certified-operators-nz2jj" Jan 27 17:35:51 crc kubenswrapper[4772]: I0127 17:35:51.843216 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmtv8\" (UniqueName: \"kubernetes.io/projected/3da4f7ee-931f-4606-b652-ca6bb3b36bcc-kube-api-access-wmtv8\") pod \"certified-operators-nz2jj\" (UID: \"3da4f7ee-931f-4606-b652-ca6bb3b36bcc\") " pod="openshift-marketplace/certified-operators-nz2jj" Jan 27 17:35:51 crc kubenswrapper[4772]: I0127 17:35:51.843582 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3da4f7ee-931f-4606-b652-ca6bb3b36bcc-utilities\") pod \"certified-operators-nz2jj\" (UID: \"3da4f7ee-931f-4606-b652-ca6bb3b36bcc\") " pod="openshift-marketplace/certified-operators-nz2jj" Jan 27 17:35:51 crc kubenswrapper[4772]: I0127 17:35:51.844088 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3da4f7ee-931f-4606-b652-ca6bb3b36bcc-catalog-content\") pod \"certified-operators-nz2jj\" (UID: \"3da4f7ee-931f-4606-b652-ca6bb3b36bcc\") " pod="openshift-marketplace/certified-operators-nz2jj" Jan 27 17:35:51 crc kubenswrapper[4772]: I0127 17:35:51.869120 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmtv8\" (UniqueName: \"kubernetes.io/projected/3da4f7ee-931f-4606-b652-ca6bb3b36bcc-kube-api-access-wmtv8\") pod \"certified-operators-nz2jj\" (UID: \"3da4f7ee-931f-4606-b652-ca6bb3b36bcc\") " pod="openshift-marketplace/certified-operators-nz2jj" Jan 27 17:35:51 crc kubenswrapper[4772]: I0127 17:35:51.999452 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nz2jj" Jan 27 17:35:52 crc kubenswrapper[4772]: I0127 17:35:52.508840 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nz2jj"] Jan 27 17:35:52 crc kubenswrapper[4772]: I0127 17:35:52.560537 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nz2jj" event={"ID":"3da4f7ee-931f-4606-b652-ca6bb3b36bcc","Type":"ContainerStarted","Data":"b49ab6bff4c5c8f14d6eb5920a5654de880c1aee811c325479fc5b2f59a6b28c"} Jan 27 17:35:53 crc kubenswrapper[4772]: I0127 17:35:53.576819 4772 generic.go:334] "Generic (PLEG): container finished" podID="3da4f7ee-931f-4606-b652-ca6bb3b36bcc" containerID="446754c22019eb637f680b802348c7e55f93e4b8311ed203bf99915655bc42a7" exitCode=0 Jan 27 17:35:53 crc kubenswrapper[4772]: I0127 17:35:53.576924 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nz2jj" event={"ID":"3da4f7ee-931f-4606-b652-ca6bb3b36bcc","Type":"ContainerDied","Data":"446754c22019eb637f680b802348c7e55f93e4b8311ed203bf99915655bc42a7"} Jan 27 17:35:54 crc kubenswrapper[4772]: I0127 17:35:54.599708 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nz2jj" event={"ID":"3da4f7ee-931f-4606-b652-ca6bb3b36bcc","Type":"ContainerStarted","Data":"96788f8c0400100015b3d653c4c77506acec40f39e8f9a95a39c10d64447317d"} Jan 27 17:35:55 crc kubenswrapper[4772]: I0127 17:35:55.613070 4772 generic.go:334] "Generic (PLEG): container finished" podID="3da4f7ee-931f-4606-b652-ca6bb3b36bcc" containerID="96788f8c0400100015b3d653c4c77506acec40f39e8f9a95a39c10d64447317d" exitCode=0 Jan 27 17:35:55 crc kubenswrapper[4772]: I0127 17:35:55.613110 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nz2jj" event={"ID":"3da4f7ee-931f-4606-b652-ca6bb3b36bcc","Type":"ContainerDied","Data":"96788f8c0400100015b3d653c4c77506acec40f39e8f9a95a39c10d64447317d"} Jan 27 17:35:56 crc kubenswrapper[4772]: I0127 17:35:56.623797 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nz2jj" event={"ID":"3da4f7ee-931f-4606-b652-ca6bb3b36bcc","Type":"ContainerStarted","Data":"1b92529d945a8ea40be9f2fcb29ba3d51fc168d67d9414057934202d95b840dd"} Jan 27 17:35:56 crc kubenswrapper[4772]: I0127 17:35:56.652979 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nz2jj" podStartSLOduration=3.221901419 podStartE2EDuration="5.65295613s" podCreationTimestamp="2026-01-27 17:35:51 +0000 UTC" firstStartedPulling="2026-01-27 17:35:53.585809672 +0000 UTC m=+8939.566418770" lastFinishedPulling="2026-01-27 17:35:56.016864363 +0000 UTC m=+8941.997473481" observedRunningTime="2026-01-27 17:35:56.645337818 +0000 UTC m=+8942.625946946" watchObservedRunningTime="2026-01-27 17:35:56.65295613 +0000 UTC m=+8942.633565268" Jan 27 17:36:02 crc kubenswrapper[4772]: I0127 17:36:01.999633 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nz2jj" Jan 27 17:36:02 crc kubenswrapper[4772]: I0127 17:36:02.001801 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nz2jj" Jan 27 17:36:02 crc kubenswrapper[4772]: I0127 17:36:02.087301 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nz2jj" Jan 27 17:36:02 crc kubenswrapper[4772]: I0127 17:36:02.762034 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nz2jj" Jan 27 17:36:02 crc kubenswrapper[4772]: I0127 17:36:02.821351 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nz2jj"] Jan 27 17:36:04 crc kubenswrapper[4772]: I0127 17:36:04.721944 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nz2jj" podUID="3da4f7ee-931f-4606-b652-ca6bb3b36bcc" containerName="registry-server" containerID="cri-o://1b92529d945a8ea40be9f2fcb29ba3d51fc168d67d9414057934202d95b840dd" gracePeriod=2 Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.195218 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nz2jj" Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.225638 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3da4f7ee-931f-4606-b652-ca6bb3b36bcc-catalog-content\") pod \"3da4f7ee-931f-4606-b652-ca6bb3b36bcc\" (UID: \"3da4f7ee-931f-4606-b652-ca6bb3b36bcc\") " Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.225725 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3da4f7ee-931f-4606-b652-ca6bb3b36bcc-utilities\") pod \"3da4f7ee-931f-4606-b652-ca6bb3b36bcc\" (UID: \"3da4f7ee-931f-4606-b652-ca6bb3b36bcc\") " Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.225876 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmtv8\" (UniqueName: \"kubernetes.io/projected/3da4f7ee-931f-4606-b652-ca6bb3b36bcc-kube-api-access-wmtv8\") pod \"3da4f7ee-931f-4606-b652-ca6bb3b36bcc\" (UID: \"3da4f7ee-931f-4606-b652-ca6bb3b36bcc\") " Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.228382 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3da4f7ee-931f-4606-b652-ca6bb3b36bcc-utilities" (OuterVolumeSpecName: "utilities") pod "3da4f7ee-931f-4606-b652-ca6bb3b36bcc" (UID: "3da4f7ee-931f-4606-b652-ca6bb3b36bcc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.243088 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3da4f7ee-931f-4606-b652-ca6bb3b36bcc-kube-api-access-wmtv8" (OuterVolumeSpecName: "kube-api-access-wmtv8") pod "3da4f7ee-931f-4606-b652-ca6bb3b36bcc" (UID: "3da4f7ee-931f-4606-b652-ca6bb3b36bcc"). InnerVolumeSpecName "kube-api-access-wmtv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.332751 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3da4f7ee-931f-4606-b652-ca6bb3b36bcc-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.332804 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmtv8\" (UniqueName: \"kubernetes.io/projected/3da4f7ee-931f-4606-b652-ca6bb3b36bcc-kube-api-access-wmtv8\") on node \"crc\" DevicePath \"\"" Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.374419 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3da4f7ee-931f-4606-b652-ca6bb3b36bcc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3da4f7ee-931f-4606-b652-ca6bb3b36bcc" (UID: "3da4f7ee-931f-4606-b652-ca6bb3b36bcc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.435102 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3da4f7ee-931f-4606-b652-ca6bb3b36bcc-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.734897 4772 generic.go:334] "Generic (PLEG): container finished" podID="3da4f7ee-931f-4606-b652-ca6bb3b36bcc" containerID="1b92529d945a8ea40be9f2fcb29ba3d51fc168d67d9414057934202d95b840dd" exitCode=0 Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.734967 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nz2jj" Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.735001 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nz2jj" event={"ID":"3da4f7ee-931f-4606-b652-ca6bb3b36bcc","Type":"ContainerDied","Data":"1b92529d945a8ea40be9f2fcb29ba3d51fc168d67d9414057934202d95b840dd"} Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.735374 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nz2jj" event={"ID":"3da4f7ee-931f-4606-b652-ca6bb3b36bcc","Type":"ContainerDied","Data":"b49ab6bff4c5c8f14d6eb5920a5654de880c1aee811c325479fc5b2f59a6b28c"} Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.735429 4772 scope.go:117] "RemoveContainer" containerID="1b92529d945a8ea40be9f2fcb29ba3d51fc168d67d9414057934202d95b840dd" Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.767767 4772 scope.go:117] "RemoveContainer" containerID="96788f8c0400100015b3d653c4c77506acec40f39e8f9a95a39c10d64447317d" Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.807446 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nz2jj"] Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.815359 4772 scope.go:117] "RemoveContainer" containerID="446754c22019eb637f680b802348c7e55f93e4b8311ed203bf99915655bc42a7" Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.824157 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nz2jj"] Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.868491 4772 scope.go:117] "RemoveContainer" containerID="1b92529d945a8ea40be9f2fcb29ba3d51fc168d67d9414057934202d95b840dd" Jan 27 17:36:05 crc kubenswrapper[4772]: E0127 17:36:05.869125 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b92529d945a8ea40be9f2fcb29ba3d51fc168d67d9414057934202d95b840dd\": container with ID starting with 1b92529d945a8ea40be9f2fcb29ba3d51fc168d67d9414057934202d95b840dd not found: ID does not exist" containerID="1b92529d945a8ea40be9f2fcb29ba3d51fc168d67d9414057934202d95b840dd" Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.869268 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b92529d945a8ea40be9f2fcb29ba3d51fc168d67d9414057934202d95b840dd"} err="failed to get container status \"1b92529d945a8ea40be9f2fcb29ba3d51fc168d67d9414057934202d95b840dd\": rpc error: code = NotFound desc = could not find container \"1b92529d945a8ea40be9f2fcb29ba3d51fc168d67d9414057934202d95b840dd\": container with ID starting with 1b92529d945a8ea40be9f2fcb29ba3d51fc168d67d9414057934202d95b840dd not found: ID does not exist" Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.869314 4772 scope.go:117] "RemoveContainer" containerID="96788f8c0400100015b3d653c4c77506acec40f39e8f9a95a39c10d64447317d" Jan 27 17:36:05 crc kubenswrapper[4772]: E0127 17:36:05.869726 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96788f8c0400100015b3d653c4c77506acec40f39e8f9a95a39c10d64447317d\": container with ID starting with 96788f8c0400100015b3d653c4c77506acec40f39e8f9a95a39c10d64447317d not found: ID does not exist" containerID="96788f8c0400100015b3d653c4c77506acec40f39e8f9a95a39c10d64447317d" Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.869771 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96788f8c0400100015b3d653c4c77506acec40f39e8f9a95a39c10d64447317d"} err="failed to get container status \"96788f8c0400100015b3d653c4c77506acec40f39e8f9a95a39c10d64447317d\": rpc error: code = NotFound desc = could not find container \"96788f8c0400100015b3d653c4c77506acec40f39e8f9a95a39c10d64447317d\": container with ID starting with 96788f8c0400100015b3d653c4c77506acec40f39e8f9a95a39c10d64447317d not found: ID does not exist" Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.869801 4772 scope.go:117] "RemoveContainer" containerID="446754c22019eb637f680b802348c7e55f93e4b8311ed203bf99915655bc42a7" Jan 27 17:36:05 crc kubenswrapper[4772]: E0127 17:36:05.870513 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"446754c22019eb637f680b802348c7e55f93e4b8311ed203bf99915655bc42a7\": container with ID starting with 446754c22019eb637f680b802348c7e55f93e4b8311ed203bf99915655bc42a7 not found: ID does not exist" containerID="446754c22019eb637f680b802348c7e55f93e4b8311ed203bf99915655bc42a7" Jan 27 17:36:05 crc kubenswrapper[4772]: I0127 17:36:05.871197 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"446754c22019eb637f680b802348c7e55f93e4b8311ed203bf99915655bc42a7"} err="failed to get container status \"446754c22019eb637f680b802348c7e55f93e4b8311ed203bf99915655bc42a7\": rpc error: code = NotFound desc = could not find container \"446754c22019eb637f680b802348c7e55f93e4b8311ed203bf99915655bc42a7\": container with ID starting with 446754c22019eb637f680b802348c7e55f93e4b8311ed203bf99915655bc42a7 not found: ID does not exist" Jan 27 17:36:06 crc kubenswrapper[4772]: I0127 17:36:06.685356 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3da4f7ee-931f-4606-b652-ca6bb3b36bcc" path="/var/lib/kubelet/pods/3da4f7ee-931f-4606-b652-ca6bb3b36bcc/volumes" Jan 27 17:36:12 crc kubenswrapper[4772]: I0127 17:36:12.058092 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:36:12 crc kubenswrapper[4772]: I0127 17:36:12.058992 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:36:12 crc kubenswrapper[4772]: I0127 17:36:12.059077 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 17:36:12 crc kubenswrapper[4772]: I0127 17:36:12.060407 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7028bdbe5a39e55f1ad82b72de13a00af4f685c12d26eb6560d24b432826149d"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 17:36:12 crc kubenswrapper[4772]: I0127 17:36:12.060524 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://7028bdbe5a39e55f1ad82b72de13a00af4f685c12d26eb6560d24b432826149d" gracePeriod=600 Jan 27 17:36:12 crc kubenswrapper[4772]: I0127 17:36:12.834714 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="7028bdbe5a39e55f1ad82b72de13a00af4f685c12d26eb6560d24b432826149d" exitCode=0 Jan 27 17:36:12 crc kubenswrapper[4772]: I0127 17:36:12.834943 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"7028bdbe5a39e55f1ad82b72de13a00af4f685c12d26eb6560d24b432826149d"} Jan 27 17:36:12 crc kubenswrapper[4772]: I0127 17:36:12.835273 4772 scope.go:117] "RemoveContainer" containerID="af2493a7ca3ac75fd4192b599bed251fe8ce4cca24c4715f3493ee05b00e8284" Jan 27 17:36:13 crc kubenswrapper[4772]: I0127 17:36:13.847909 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d"} Jan 27 17:38:42 crc kubenswrapper[4772]: I0127 17:38:42.058056 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:38:42 crc kubenswrapper[4772]: I0127 17:38:42.058578 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:39:12 crc kubenswrapper[4772]: I0127 17:39:12.058693 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:39:12 crc kubenswrapper[4772]: I0127 17:39:12.059564 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:39:42 crc kubenswrapper[4772]: I0127 17:39:42.058639 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:39:42 crc kubenswrapper[4772]: I0127 17:39:42.059332 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:39:42 crc kubenswrapper[4772]: I0127 17:39:42.059392 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 17:39:42 crc kubenswrapper[4772]: I0127 17:39:42.060443 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 17:39:42 crc kubenswrapper[4772]: I0127 17:39:42.060572 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" gracePeriod=600 Jan 27 17:39:42 crc kubenswrapper[4772]: E0127 17:39:42.188744 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:39:42 crc kubenswrapper[4772]: I0127 17:39:42.355943 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" exitCode=0 Jan 27 17:39:42 crc kubenswrapper[4772]: I0127 17:39:42.356044 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d"} Jan 27 17:39:42 crc kubenswrapper[4772]: I0127 17:39:42.356307 4772 scope.go:117] "RemoveContainer" containerID="7028bdbe5a39e55f1ad82b72de13a00af4f685c12d26eb6560d24b432826149d" Jan 27 17:39:42 crc kubenswrapper[4772]: I0127 17:39:42.357220 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:39:42 crc kubenswrapper[4772]: E0127 17:39:42.357714 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:39:53 crc kubenswrapper[4772]: I0127 17:39:53.663742 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:39:53 crc kubenswrapper[4772]: E0127 17:39:53.664816 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:40:04 crc kubenswrapper[4772]: I0127 17:40:04.676956 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:40:04 crc kubenswrapper[4772]: E0127 17:40:04.678277 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:40:16 crc kubenswrapper[4772]: I0127 17:40:16.697641 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6f7k5"] Jan 27 17:40:16 crc kubenswrapper[4772]: E0127 17:40:16.698948 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3da4f7ee-931f-4606-b652-ca6bb3b36bcc" containerName="extract-utilities" Jan 27 17:40:16 crc kubenswrapper[4772]: I0127 17:40:16.698974 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3da4f7ee-931f-4606-b652-ca6bb3b36bcc" containerName="extract-utilities" Jan 27 17:40:16 crc kubenswrapper[4772]: E0127 17:40:16.698991 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3da4f7ee-931f-4606-b652-ca6bb3b36bcc" containerName="extract-content" Jan 27 17:40:16 crc kubenswrapper[4772]: I0127 17:40:16.699002 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3da4f7ee-931f-4606-b652-ca6bb3b36bcc" containerName="extract-content" Jan 27 17:40:16 crc kubenswrapper[4772]: E0127 17:40:16.699016 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3da4f7ee-931f-4606-b652-ca6bb3b36bcc" containerName="registry-server" Jan 27 17:40:16 crc kubenswrapper[4772]: I0127 17:40:16.699026 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3da4f7ee-931f-4606-b652-ca6bb3b36bcc" containerName="registry-server" Jan 27 17:40:16 crc kubenswrapper[4772]: I0127 17:40:16.699351 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3da4f7ee-931f-4606-b652-ca6bb3b36bcc" containerName="registry-server" Jan 27 17:40:16 crc kubenswrapper[4772]: I0127 17:40:16.701334 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6f7k5" Jan 27 17:40:16 crc kubenswrapper[4772]: I0127 17:40:16.710385 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6f7k5"] Jan 27 17:40:16 crc kubenswrapper[4772]: I0127 17:40:16.790676 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fjnv\" (UniqueName: \"kubernetes.io/projected/a3b36acb-1d5b-4384-9090-ce95e3d89a21-kube-api-access-7fjnv\") pod \"redhat-operators-6f7k5\" (UID: \"a3b36acb-1d5b-4384-9090-ce95e3d89a21\") " pod="openshift-marketplace/redhat-operators-6f7k5" Jan 27 17:40:16 crc kubenswrapper[4772]: I0127 17:40:16.790772 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3b36acb-1d5b-4384-9090-ce95e3d89a21-catalog-content\") pod \"redhat-operators-6f7k5\" (UID: \"a3b36acb-1d5b-4384-9090-ce95e3d89a21\") " pod="openshift-marketplace/redhat-operators-6f7k5" Jan 27 17:40:16 crc kubenswrapper[4772]: I0127 17:40:16.791476 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3b36acb-1d5b-4384-9090-ce95e3d89a21-utilities\") pod \"redhat-operators-6f7k5\" (UID: \"a3b36acb-1d5b-4384-9090-ce95e3d89a21\") " pod="openshift-marketplace/redhat-operators-6f7k5" Jan 27 17:40:16 crc kubenswrapper[4772]: I0127 17:40:16.893898 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3b36acb-1d5b-4384-9090-ce95e3d89a21-utilities\") pod \"redhat-operators-6f7k5\" (UID: \"a3b36acb-1d5b-4384-9090-ce95e3d89a21\") " pod="openshift-marketplace/redhat-operators-6f7k5" Jan 27 17:40:16 crc kubenswrapper[4772]: I0127 17:40:16.894211 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fjnv\" (UniqueName: \"kubernetes.io/projected/a3b36acb-1d5b-4384-9090-ce95e3d89a21-kube-api-access-7fjnv\") pod \"redhat-operators-6f7k5\" (UID: \"a3b36acb-1d5b-4384-9090-ce95e3d89a21\") " pod="openshift-marketplace/redhat-operators-6f7k5" Jan 27 17:40:16 crc kubenswrapper[4772]: I0127 17:40:16.894393 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3b36acb-1d5b-4384-9090-ce95e3d89a21-catalog-content\") pod \"redhat-operators-6f7k5\" (UID: \"a3b36acb-1d5b-4384-9090-ce95e3d89a21\") " pod="openshift-marketplace/redhat-operators-6f7k5" Jan 27 17:40:16 crc kubenswrapper[4772]: I0127 17:40:16.894703 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3b36acb-1d5b-4384-9090-ce95e3d89a21-utilities\") pod \"redhat-operators-6f7k5\" (UID: \"a3b36acb-1d5b-4384-9090-ce95e3d89a21\") " pod="openshift-marketplace/redhat-operators-6f7k5" Jan 27 17:40:16 crc kubenswrapper[4772]: I0127 17:40:16.894909 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3b36acb-1d5b-4384-9090-ce95e3d89a21-catalog-content\") pod \"redhat-operators-6f7k5\" (UID: \"a3b36acb-1d5b-4384-9090-ce95e3d89a21\") " pod="openshift-marketplace/redhat-operators-6f7k5" Jan 27 17:40:16 crc kubenswrapper[4772]: I0127 17:40:16.991064 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fjnv\" (UniqueName: \"kubernetes.io/projected/a3b36acb-1d5b-4384-9090-ce95e3d89a21-kube-api-access-7fjnv\") pod \"redhat-operators-6f7k5\" (UID: \"a3b36acb-1d5b-4384-9090-ce95e3d89a21\") " pod="openshift-marketplace/redhat-operators-6f7k5" Jan 27 17:40:17 crc kubenswrapper[4772]: I0127 17:40:17.044895 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6f7k5" Jan 27 17:40:17 crc kubenswrapper[4772]: I0127 17:40:17.500637 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6f7k5"] Jan 27 17:40:17 crc kubenswrapper[4772]: I0127 17:40:17.741242 4772 generic.go:334] "Generic (PLEG): container finished" podID="a3b36acb-1d5b-4384-9090-ce95e3d89a21" containerID="b09d5a873a7a863058f1edfb234292bbee9181b432c9d473d092af477ae78419" exitCode=0 Jan 27 17:40:17 crc kubenswrapper[4772]: I0127 17:40:17.741334 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6f7k5" event={"ID":"a3b36acb-1d5b-4384-9090-ce95e3d89a21","Type":"ContainerDied","Data":"b09d5a873a7a863058f1edfb234292bbee9181b432c9d473d092af477ae78419"} Jan 27 17:40:17 crc kubenswrapper[4772]: I0127 17:40:17.741579 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6f7k5" event={"ID":"a3b36acb-1d5b-4384-9090-ce95e3d89a21","Type":"ContainerStarted","Data":"ba22c7075b7ea96ca1087e7fff4b1cec7517736240e86b21b900c9d7c35d0e36"} Jan 27 17:40:17 crc kubenswrapper[4772]: I0127 17:40:17.742934 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 17:40:18 crc kubenswrapper[4772]: I0127 17:40:18.665609 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:40:18 crc kubenswrapper[4772]: E0127 17:40:18.666753 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:40:18 crc kubenswrapper[4772]: I0127 17:40:18.756903 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6f7k5" event={"ID":"a3b36acb-1d5b-4384-9090-ce95e3d89a21","Type":"ContainerStarted","Data":"5aea5fa5756b456fa46058d90e0286833506c245fcf9cb1a7c973fe4db57abeb"} Jan 27 17:40:21 crc kubenswrapper[4772]: I0127 17:40:21.792555 4772 generic.go:334] "Generic (PLEG): container finished" podID="a3b36acb-1d5b-4384-9090-ce95e3d89a21" containerID="5aea5fa5756b456fa46058d90e0286833506c245fcf9cb1a7c973fe4db57abeb" exitCode=0 Jan 27 17:40:21 crc kubenswrapper[4772]: I0127 17:40:21.792651 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6f7k5" event={"ID":"a3b36acb-1d5b-4384-9090-ce95e3d89a21","Type":"ContainerDied","Data":"5aea5fa5756b456fa46058d90e0286833506c245fcf9cb1a7c973fe4db57abeb"} Jan 27 17:40:22 crc kubenswrapper[4772]: I0127 17:40:22.803295 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6f7k5" event={"ID":"a3b36acb-1d5b-4384-9090-ce95e3d89a21","Type":"ContainerStarted","Data":"593d19ce0cdbbf6fdbd0b96cef5cb458224a327825a61f25f1b4182c27097b38"} Jan 27 17:40:22 crc kubenswrapper[4772]: I0127 17:40:22.829099 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6f7k5" podStartSLOduration=2.30730743 podStartE2EDuration="6.829077118s" podCreationTimestamp="2026-01-27 17:40:16 +0000 UTC" firstStartedPulling="2026-01-27 17:40:17.742727833 +0000 UTC m=+9203.723336931" lastFinishedPulling="2026-01-27 17:40:22.264497521 +0000 UTC m=+9208.245106619" observedRunningTime="2026-01-27 17:40:22.825823005 +0000 UTC m=+9208.806432113" watchObservedRunningTime="2026-01-27 17:40:22.829077118 +0000 UTC m=+9208.809686236" Jan 27 17:40:27 crc kubenswrapper[4772]: I0127 17:40:27.046144 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6f7k5" Jan 27 17:40:27 crc kubenswrapper[4772]: I0127 17:40:27.047094 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6f7k5" Jan 27 17:40:28 crc kubenswrapper[4772]: I0127 17:40:28.123487 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6f7k5" podUID="a3b36acb-1d5b-4384-9090-ce95e3d89a21" containerName="registry-server" probeResult="failure" output=< Jan 27 17:40:28 crc kubenswrapper[4772]: timeout: failed to connect service ":50051" within 1s Jan 27 17:40:28 crc kubenswrapper[4772]: > Jan 27 17:40:30 crc kubenswrapper[4772]: I0127 17:40:30.664005 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:40:30 crc kubenswrapper[4772]: E0127 17:40:30.665036 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:40:37 crc kubenswrapper[4772]: I0127 17:40:37.122129 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6f7k5" Jan 27 17:40:37 crc kubenswrapper[4772]: I0127 17:40:37.208639 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6f7k5" Jan 27 17:40:37 crc kubenswrapper[4772]: I0127 17:40:37.380069 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6f7k5"] Jan 27 17:40:38 crc kubenswrapper[4772]: I0127 17:40:38.979950 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6f7k5" podUID="a3b36acb-1d5b-4384-9090-ce95e3d89a21" containerName="registry-server" containerID="cri-o://593d19ce0cdbbf6fdbd0b96cef5cb458224a327825a61f25f1b4182c27097b38" gracePeriod=2 Jan 27 17:40:39 crc kubenswrapper[4772]: I0127 17:40:39.443959 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6f7k5" Jan 27 17:40:39 crc kubenswrapper[4772]: I0127 17:40:39.473606 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3b36acb-1d5b-4384-9090-ce95e3d89a21-utilities\") pod \"a3b36acb-1d5b-4384-9090-ce95e3d89a21\" (UID: \"a3b36acb-1d5b-4384-9090-ce95e3d89a21\") " Jan 27 17:40:39 crc kubenswrapper[4772]: I0127 17:40:39.473815 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3b36acb-1d5b-4384-9090-ce95e3d89a21-catalog-content\") pod \"a3b36acb-1d5b-4384-9090-ce95e3d89a21\" (UID: \"a3b36acb-1d5b-4384-9090-ce95e3d89a21\") " Jan 27 17:40:39 crc kubenswrapper[4772]: I0127 17:40:39.474011 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fjnv\" (UniqueName: \"kubernetes.io/projected/a3b36acb-1d5b-4384-9090-ce95e3d89a21-kube-api-access-7fjnv\") pod \"a3b36acb-1d5b-4384-9090-ce95e3d89a21\" (UID: \"a3b36acb-1d5b-4384-9090-ce95e3d89a21\") " Jan 27 17:40:39 crc kubenswrapper[4772]: I0127 17:40:39.474557 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3b36acb-1d5b-4384-9090-ce95e3d89a21-utilities" (OuterVolumeSpecName: "utilities") pod "a3b36acb-1d5b-4384-9090-ce95e3d89a21" (UID: "a3b36acb-1d5b-4384-9090-ce95e3d89a21"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:40:39 crc kubenswrapper[4772]: I0127 17:40:39.475626 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3b36acb-1d5b-4384-9090-ce95e3d89a21-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 17:40:39 crc kubenswrapper[4772]: I0127 17:40:39.482084 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3b36acb-1d5b-4384-9090-ce95e3d89a21-kube-api-access-7fjnv" (OuterVolumeSpecName: "kube-api-access-7fjnv") pod "a3b36acb-1d5b-4384-9090-ce95e3d89a21" (UID: "a3b36acb-1d5b-4384-9090-ce95e3d89a21"). InnerVolumeSpecName "kube-api-access-7fjnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:40:39 crc kubenswrapper[4772]: I0127 17:40:39.578321 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fjnv\" (UniqueName: \"kubernetes.io/projected/a3b36acb-1d5b-4384-9090-ce95e3d89a21-kube-api-access-7fjnv\") on node \"crc\" DevicePath \"\"" Jan 27 17:40:39 crc kubenswrapper[4772]: I0127 17:40:39.618042 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3b36acb-1d5b-4384-9090-ce95e3d89a21-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a3b36acb-1d5b-4384-9090-ce95e3d89a21" (UID: "a3b36acb-1d5b-4384-9090-ce95e3d89a21"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:40:39 crc kubenswrapper[4772]: I0127 17:40:39.680110 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3b36acb-1d5b-4384-9090-ce95e3d89a21-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 17:40:39 crc kubenswrapper[4772]: I0127 17:40:39.991334 4772 generic.go:334] "Generic (PLEG): container finished" podID="a3b36acb-1d5b-4384-9090-ce95e3d89a21" containerID="593d19ce0cdbbf6fdbd0b96cef5cb458224a327825a61f25f1b4182c27097b38" exitCode=0 Jan 27 17:40:39 crc kubenswrapper[4772]: I0127 17:40:39.991388 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6f7k5" event={"ID":"a3b36acb-1d5b-4384-9090-ce95e3d89a21","Type":"ContainerDied","Data":"593d19ce0cdbbf6fdbd0b96cef5cb458224a327825a61f25f1b4182c27097b38"} Jan 27 17:40:39 crc kubenswrapper[4772]: I0127 17:40:39.991444 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6f7k5" event={"ID":"a3b36acb-1d5b-4384-9090-ce95e3d89a21","Type":"ContainerDied","Data":"ba22c7075b7ea96ca1087e7fff4b1cec7517736240e86b21b900c9d7c35d0e36"} Jan 27 17:40:39 crc kubenswrapper[4772]: I0127 17:40:39.991449 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6f7k5" Jan 27 17:40:39 crc kubenswrapper[4772]: I0127 17:40:39.991467 4772 scope.go:117] "RemoveContainer" containerID="593d19ce0cdbbf6fdbd0b96cef5cb458224a327825a61f25f1b4182c27097b38" Jan 27 17:40:40 crc kubenswrapper[4772]: I0127 17:40:40.013314 4772 scope.go:117] "RemoveContainer" containerID="5aea5fa5756b456fa46058d90e0286833506c245fcf9cb1a7c973fe4db57abeb" Jan 27 17:40:40 crc kubenswrapper[4772]: I0127 17:40:40.032349 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6f7k5"] Jan 27 17:40:40 crc kubenswrapper[4772]: I0127 17:40:40.042130 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6f7k5"] Jan 27 17:40:40 crc kubenswrapper[4772]: I0127 17:40:40.051867 4772 scope.go:117] "RemoveContainer" containerID="b09d5a873a7a863058f1edfb234292bbee9181b432c9d473d092af477ae78419" Jan 27 17:40:40 crc kubenswrapper[4772]: I0127 17:40:40.078152 4772 scope.go:117] "RemoveContainer" containerID="593d19ce0cdbbf6fdbd0b96cef5cb458224a327825a61f25f1b4182c27097b38" Jan 27 17:40:40 crc kubenswrapper[4772]: E0127 17:40:40.078884 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"593d19ce0cdbbf6fdbd0b96cef5cb458224a327825a61f25f1b4182c27097b38\": container with ID starting with 593d19ce0cdbbf6fdbd0b96cef5cb458224a327825a61f25f1b4182c27097b38 not found: ID does not exist" containerID="593d19ce0cdbbf6fdbd0b96cef5cb458224a327825a61f25f1b4182c27097b38" Jan 27 17:40:40 crc kubenswrapper[4772]: I0127 17:40:40.078950 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"593d19ce0cdbbf6fdbd0b96cef5cb458224a327825a61f25f1b4182c27097b38"} err="failed to get container status \"593d19ce0cdbbf6fdbd0b96cef5cb458224a327825a61f25f1b4182c27097b38\": rpc error: code = NotFound desc = could not find container \"593d19ce0cdbbf6fdbd0b96cef5cb458224a327825a61f25f1b4182c27097b38\": container with ID starting with 593d19ce0cdbbf6fdbd0b96cef5cb458224a327825a61f25f1b4182c27097b38 not found: ID does not exist" Jan 27 17:40:40 crc kubenswrapper[4772]: I0127 17:40:40.078994 4772 scope.go:117] "RemoveContainer" containerID="5aea5fa5756b456fa46058d90e0286833506c245fcf9cb1a7c973fe4db57abeb" Jan 27 17:40:40 crc kubenswrapper[4772]: E0127 17:40:40.079446 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5aea5fa5756b456fa46058d90e0286833506c245fcf9cb1a7c973fe4db57abeb\": container with ID starting with 5aea5fa5756b456fa46058d90e0286833506c245fcf9cb1a7c973fe4db57abeb not found: ID does not exist" containerID="5aea5fa5756b456fa46058d90e0286833506c245fcf9cb1a7c973fe4db57abeb" Jan 27 17:40:40 crc kubenswrapper[4772]: I0127 17:40:40.079509 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5aea5fa5756b456fa46058d90e0286833506c245fcf9cb1a7c973fe4db57abeb"} err="failed to get container status \"5aea5fa5756b456fa46058d90e0286833506c245fcf9cb1a7c973fe4db57abeb\": rpc error: code = NotFound desc = could not find container \"5aea5fa5756b456fa46058d90e0286833506c245fcf9cb1a7c973fe4db57abeb\": container with ID starting with 5aea5fa5756b456fa46058d90e0286833506c245fcf9cb1a7c973fe4db57abeb not found: ID does not exist" Jan 27 17:40:40 crc kubenswrapper[4772]: I0127 17:40:40.079549 4772 scope.go:117] "RemoveContainer" containerID="b09d5a873a7a863058f1edfb234292bbee9181b432c9d473d092af477ae78419" Jan 27 17:40:40 crc kubenswrapper[4772]: E0127 17:40:40.079895 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b09d5a873a7a863058f1edfb234292bbee9181b432c9d473d092af477ae78419\": container with ID starting with b09d5a873a7a863058f1edfb234292bbee9181b432c9d473d092af477ae78419 not found: ID does not exist" containerID="b09d5a873a7a863058f1edfb234292bbee9181b432c9d473d092af477ae78419" Jan 27 17:40:40 crc kubenswrapper[4772]: I0127 17:40:40.079944 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b09d5a873a7a863058f1edfb234292bbee9181b432c9d473d092af477ae78419"} err="failed to get container status \"b09d5a873a7a863058f1edfb234292bbee9181b432c9d473d092af477ae78419\": rpc error: code = NotFound desc = could not find container \"b09d5a873a7a863058f1edfb234292bbee9181b432c9d473d092af477ae78419\": container with ID starting with b09d5a873a7a863058f1edfb234292bbee9181b432c9d473d092af477ae78419 not found: ID does not exist" Jan 27 17:40:40 crc kubenswrapper[4772]: I0127 17:40:40.678012 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3b36acb-1d5b-4384-9090-ce95e3d89a21" path="/var/lib/kubelet/pods/a3b36acb-1d5b-4384-9090-ce95e3d89a21/volumes" Jan 27 17:40:41 crc kubenswrapper[4772]: I0127 17:40:41.663215 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:40:41 crc kubenswrapper[4772]: E0127 17:40:41.663591 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:40:56 crc kubenswrapper[4772]: I0127 17:40:56.663301 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:40:56 crc kubenswrapper[4772]: E0127 17:40:56.664722 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:41:11 crc kubenswrapper[4772]: I0127 17:41:11.664146 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:41:11 crc kubenswrapper[4772]: E0127 17:41:11.665273 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:41:23 crc kubenswrapper[4772]: I0127 17:41:23.664389 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:41:23 crc kubenswrapper[4772]: E0127 17:41:23.665290 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:41:37 crc kubenswrapper[4772]: I0127 17:41:37.668790 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:41:37 crc kubenswrapper[4772]: E0127 17:41:37.670323 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:41:50 crc kubenswrapper[4772]: I0127 17:41:50.662881 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:41:50 crc kubenswrapper[4772]: E0127 17:41:50.663790 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:42:04 crc kubenswrapper[4772]: I0127 17:42:04.671320 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:42:04 crc kubenswrapper[4772]: E0127 17:42:04.672107 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:42:18 crc kubenswrapper[4772]: I0127 17:42:18.663489 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:42:18 crc kubenswrapper[4772]: E0127 17:42:18.664607 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:42:29 crc kubenswrapper[4772]: I0127 17:42:29.663530 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:42:29 crc kubenswrapper[4772]: E0127 17:42:29.664287 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:42:41 crc kubenswrapper[4772]: I0127 17:42:41.662865 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:42:41 crc kubenswrapper[4772]: E0127 17:42:41.663706 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:42:52 crc kubenswrapper[4772]: I0127 17:42:52.663919 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:42:52 crc kubenswrapper[4772]: E0127 17:42:52.664813 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:43:04 crc kubenswrapper[4772]: I0127 17:43:04.676030 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:43:04 crc kubenswrapper[4772]: E0127 17:43:04.677673 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:43:16 crc kubenswrapper[4772]: I0127 17:43:16.664549 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:43:16 crc kubenswrapper[4772]: E0127 17:43:16.665814 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:43:17 crc kubenswrapper[4772]: I0127 17:43:17.488796 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5kdsn"] Jan 27 17:43:17 crc kubenswrapper[4772]: E0127 17:43:17.489608 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3b36acb-1d5b-4384-9090-ce95e3d89a21" containerName="extract-content" Jan 27 17:43:17 crc kubenswrapper[4772]: I0127 17:43:17.489636 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3b36acb-1d5b-4384-9090-ce95e3d89a21" containerName="extract-content" Jan 27 17:43:17 crc kubenswrapper[4772]: E0127 17:43:17.489656 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3b36acb-1d5b-4384-9090-ce95e3d89a21" containerName="registry-server" Jan 27 17:43:17 crc kubenswrapper[4772]: I0127 17:43:17.489664 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3b36acb-1d5b-4384-9090-ce95e3d89a21" containerName="registry-server" Jan 27 17:43:17 crc kubenswrapper[4772]: E0127 17:43:17.489701 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3b36acb-1d5b-4384-9090-ce95e3d89a21" containerName="extract-utilities" Jan 27 17:43:17 crc kubenswrapper[4772]: I0127 17:43:17.489710 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3b36acb-1d5b-4384-9090-ce95e3d89a21" containerName="extract-utilities" Jan 27 17:43:17 crc kubenswrapper[4772]: I0127 17:43:17.489947 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3b36acb-1d5b-4384-9090-ce95e3d89a21" containerName="registry-server" Jan 27 17:43:17 crc kubenswrapper[4772]: I0127 17:43:17.491532 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5kdsn" Jan 27 17:43:17 crc kubenswrapper[4772]: I0127 17:43:17.510540 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5kdsn"] Jan 27 17:43:17 crc kubenswrapper[4772]: I0127 17:43:17.647343 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c6ade8d-638e-415e-8363-af0aa99994b2-catalog-content\") pod \"community-operators-5kdsn\" (UID: \"0c6ade8d-638e-415e-8363-af0aa99994b2\") " pod="openshift-marketplace/community-operators-5kdsn" Jan 27 17:43:17 crc kubenswrapper[4772]: I0127 17:43:17.647487 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqn9m\" (UniqueName: \"kubernetes.io/projected/0c6ade8d-638e-415e-8363-af0aa99994b2-kube-api-access-kqn9m\") pod \"community-operators-5kdsn\" (UID: \"0c6ade8d-638e-415e-8363-af0aa99994b2\") " pod="openshift-marketplace/community-operators-5kdsn" Jan 27 17:43:17 crc kubenswrapper[4772]: I0127 17:43:17.649274 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c6ade8d-638e-415e-8363-af0aa99994b2-utilities\") pod \"community-operators-5kdsn\" (UID: \"0c6ade8d-638e-415e-8363-af0aa99994b2\") " pod="openshift-marketplace/community-operators-5kdsn" Jan 27 17:43:17 crc kubenswrapper[4772]: I0127 17:43:17.751455 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c6ade8d-638e-415e-8363-af0aa99994b2-utilities\") pod \"community-operators-5kdsn\" (UID: \"0c6ade8d-638e-415e-8363-af0aa99994b2\") " pod="openshift-marketplace/community-operators-5kdsn" Jan 27 17:43:17 crc kubenswrapper[4772]: I0127 17:43:17.751551 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c6ade8d-638e-415e-8363-af0aa99994b2-catalog-content\") pod \"community-operators-5kdsn\" (UID: \"0c6ade8d-638e-415e-8363-af0aa99994b2\") " pod="openshift-marketplace/community-operators-5kdsn" Jan 27 17:43:17 crc kubenswrapper[4772]: I0127 17:43:17.751579 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqn9m\" (UniqueName: \"kubernetes.io/projected/0c6ade8d-638e-415e-8363-af0aa99994b2-kube-api-access-kqn9m\") pod \"community-operators-5kdsn\" (UID: \"0c6ade8d-638e-415e-8363-af0aa99994b2\") " pod="openshift-marketplace/community-operators-5kdsn" Jan 27 17:43:17 crc kubenswrapper[4772]: I0127 17:43:17.752146 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c6ade8d-638e-415e-8363-af0aa99994b2-utilities\") pod \"community-operators-5kdsn\" (UID: \"0c6ade8d-638e-415e-8363-af0aa99994b2\") " pod="openshift-marketplace/community-operators-5kdsn" Jan 27 17:43:17 crc kubenswrapper[4772]: I0127 17:43:17.752220 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c6ade8d-638e-415e-8363-af0aa99994b2-catalog-content\") pod \"community-operators-5kdsn\" (UID: \"0c6ade8d-638e-415e-8363-af0aa99994b2\") " pod="openshift-marketplace/community-operators-5kdsn" Jan 27 17:43:18 crc kubenswrapper[4772]: I0127 17:43:18.191877 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqn9m\" (UniqueName: \"kubernetes.io/projected/0c6ade8d-638e-415e-8363-af0aa99994b2-kube-api-access-kqn9m\") pod \"community-operators-5kdsn\" (UID: \"0c6ade8d-638e-415e-8363-af0aa99994b2\") " pod="openshift-marketplace/community-operators-5kdsn" Jan 27 17:43:18 crc kubenswrapper[4772]: I0127 17:43:18.416056 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5kdsn" Jan 27 17:43:18 crc kubenswrapper[4772]: I0127 17:43:18.862457 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5kdsn"] Jan 27 17:43:19 crc kubenswrapper[4772]: I0127 17:43:19.629057 4772 generic.go:334] "Generic (PLEG): container finished" podID="0c6ade8d-638e-415e-8363-af0aa99994b2" containerID="1de1923bc262c5aaefa8cf7b56877449ea107eda643f14cf3ab05e1b3236c906" exitCode=0 Jan 27 17:43:19 crc kubenswrapper[4772]: I0127 17:43:19.629357 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kdsn" event={"ID":"0c6ade8d-638e-415e-8363-af0aa99994b2","Type":"ContainerDied","Data":"1de1923bc262c5aaefa8cf7b56877449ea107eda643f14cf3ab05e1b3236c906"} Jan 27 17:43:19 crc kubenswrapper[4772]: I0127 17:43:19.629387 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kdsn" event={"ID":"0c6ade8d-638e-415e-8363-af0aa99994b2","Type":"ContainerStarted","Data":"bd47383e58f847ae39fb7fb3c9b02b7da360d7ebf0041c54c2a55ffc5d783006"} Jan 27 17:43:21 crc kubenswrapper[4772]: I0127 17:43:21.665417 4772 generic.go:334] "Generic (PLEG): container finished" podID="0c6ade8d-638e-415e-8363-af0aa99994b2" containerID="d61e1ffdf7b107647ae33c5ea2c7619bc98bdd510e69bf094fc6877e8690c610" exitCode=0 Jan 27 17:43:21 crc kubenswrapper[4772]: I0127 17:43:21.665506 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kdsn" event={"ID":"0c6ade8d-638e-415e-8363-af0aa99994b2","Type":"ContainerDied","Data":"d61e1ffdf7b107647ae33c5ea2c7619bc98bdd510e69bf094fc6877e8690c610"} Jan 27 17:43:22 crc kubenswrapper[4772]: I0127 17:43:22.696870 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kdsn" event={"ID":"0c6ade8d-638e-415e-8363-af0aa99994b2","Type":"ContainerStarted","Data":"44933a5bde23ef59092f591e6e7a85ecc728ca650a6bece714d8a8b7d09229e6"} Jan 27 17:43:22 crc kubenswrapper[4772]: I0127 17:43:22.717435 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5kdsn" podStartSLOduration=3.276583875 podStartE2EDuration="5.717418478s" podCreationTimestamp="2026-01-27 17:43:17 +0000 UTC" firstStartedPulling="2026-01-27 17:43:19.630918082 +0000 UTC m=+9385.611527190" lastFinishedPulling="2026-01-27 17:43:22.071752675 +0000 UTC m=+9388.052361793" observedRunningTime="2026-01-27 17:43:22.717034197 +0000 UTC m=+9388.697643315" watchObservedRunningTime="2026-01-27 17:43:22.717418478 +0000 UTC m=+9388.698027576" Jan 27 17:43:28 crc kubenswrapper[4772]: I0127 17:43:28.416391 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5kdsn" Jan 27 17:43:28 crc kubenswrapper[4772]: I0127 17:43:28.416969 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5kdsn" Jan 27 17:43:28 crc kubenswrapper[4772]: I0127 17:43:28.471118 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5kdsn" Jan 27 17:43:28 crc kubenswrapper[4772]: I0127 17:43:28.663607 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:43:28 crc kubenswrapper[4772]: E0127 17:43:28.664217 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:43:28 crc kubenswrapper[4772]: I0127 17:43:28.880796 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5kdsn" Jan 27 17:43:28 crc kubenswrapper[4772]: I0127 17:43:28.948536 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5kdsn"] Jan 27 17:43:30 crc kubenswrapper[4772]: I0127 17:43:30.816641 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5kdsn" podUID="0c6ade8d-638e-415e-8363-af0aa99994b2" containerName="registry-server" containerID="cri-o://44933a5bde23ef59092f591e6e7a85ecc728ca650a6bece714d8a8b7d09229e6" gracePeriod=2 Jan 27 17:43:31 crc kubenswrapper[4772]: I0127 17:43:31.829138 4772 generic.go:334] "Generic (PLEG): container finished" podID="0c6ade8d-638e-415e-8363-af0aa99994b2" containerID="44933a5bde23ef59092f591e6e7a85ecc728ca650a6bece714d8a8b7d09229e6" exitCode=0 Jan 27 17:43:31 crc kubenswrapper[4772]: I0127 17:43:31.829215 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kdsn" event={"ID":"0c6ade8d-638e-415e-8363-af0aa99994b2","Type":"ContainerDied","Data":"44933a5bde23ef59092f591e6e7a85ecc728ca650a6bece714d8a8b7d09229e6"} Jan 27 17:43:32 crc kubenswrapper[4772]: I0127 17:43:32.202804 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5kdsn" Jan 27 17:43:32 crc kubenswrapper[4772]: I0127 17:43:32.354108 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c6ade8d-638e-415e-8363-af0aa99994b2-catalog-content\") pod \"0c6ade8d-638e-415e-8363-af0aa99994b2\" (UID: \"0c6ade8d-638e-415e-8363-af0aa99994b2\") " Jan 27 17:43:32 crc kubenswrapper[4772]: I0127 17:43:32.354212 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqn9m\" (UniqueName: \"kubernetes.io/projected/0c6ade8d-638e-415e-8363-af0aa99994b2-kube-api-access-kqn9m\") pod \"0c6ade8d-638e-415e-8363-af0aa99994b2\" (UID: \"0c6ade8d-638e-415e-8363-af0aa99994b2\") " Jan 27 17:43:32 crc kubenswrapper[4772]: I0127 17:43:32.354597 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c6ade8d-638e-415e-8363-af0aa99994b2-utilities\") pod \"0c6ade8d-638e-415e-8363-af0aa99994b2\" (UID: \"0c6ade8d-638e-415e-8363-af0aa99994b2\") " Jan 27 17:43:32 crc kubenswrapper[4772]: I0127 17:43:32.356544 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c6ade8d-638e-415e-8363-af0aa99994b2-utilities" (OuterVolumeSpecName: "utilities") pod "0c6ade8d-638e-415e-8363-af0aa99994b2" (UID: "0c6ade8d-638e-415e-8363-af0aa99994b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:43:32 crc kubenswrapper[4772]: I0127 17:43:32.361850 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c6ade8d-638e-415e-8363-af0aa99994b2-kube-api-access-kqn9m" (OuterVolumeSpecName: "kube-api-access-kqn9m") pod "0c6ade8d-638e-415e-8363-af0aa99994b2" (UID: "0c6ade8d-638e-415e-8363-af0aa99994b2"). InnerVolumeSpecName "kube-api-access-kqn9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:43:32 crc kubenswrapper[4772]: I0127 17:43:32.402031 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c6ade8d-638e-415e-8363-af0aa99994b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c6ade8d-638e-415e-8363-af0aa99994b2" (UID: "0c6ade8d-638e-415e-8363-af0aa99994b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:43:32 crc kubenswrapper[4772]: I0127 17:43:32.456750 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c6ade8d-638e-415e-8363-af0aa99994b2-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 17:43:32 crc kubenswrapper[4772]: I0127 17:43:32.456797 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqn9m\" (UniqueName: \"kubernetes.io/projected/0c6ade8d-638e-415e-8363-af0aa99994b2-kube-api-access-kqn9m\") on node \"crc\" DevicePath \"\"" Jan 27 17:43:32 crc kubenswrapper[4772]: I0127 17:43:32.456812 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c6ade8d-638e-415e-8363-af0aa99994b2-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 17:43:32 crc kubenswrapper[4772]: I0127 17:43:32.844757 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5kdsn" event={"ID":"0c6ade8d-638e-415e-8363-af0aa99994b2","Type":"ContainerDied","Data":"bd47383e58f847ae39fb7fb3c9b02b7da360d7ebf0041c54c2a55ffc5d783006"} Jan 27 17:43:32 crc kubenswrapper[4772]: I0127 17:43:32.844905 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5kdsn" Jan 27 17:43:32 crc kubenswrapper[4772]: I0127 17:43:32.846251 4772 scope.go:117] "RemoveContainer" containerID="44933a5bde23ef59092f591e6e7a85ecc728ca650a6bece714d8a8b7d09229e6" Jan 27 17:43:32 crc kubenswrapper[4772]: I0127 17:43:32.885818 4772 scope.go:117] "RemoveContainer" containerID="d61e1ffdf7b107647ae33c5ea2c7619bc98bdd510e69bf094fc6877e8690c610" Jan 27 17:43:32 crc kubenswrapper[4772]: I0127 17:43:32.901576 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5kdsn"] Jan 27 17:43:32 crc kubenswrapper[4772]: I0127 17:43:32.915579 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5kdsn"] Jan 27 17:43:32 crc kubenswrapper[4772]: I0127 17:43:32.928593 4772 scope.go:117] "RemoveContainer" containerID="1de1923bc262c5aaefa8cf7b56877449ea107eda643f14cf3ab05e1b3236c906" Jan 27 17:43:34 crc kubenswrapper[4772]: I0127 17:43:34.682685 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c6ade8d-638e-415e-8363-af0aa99994b2" path="/var/lib/kubelet/pods/0c6ade8d-638e-415e-8363-af0aa99994b2/volumes" Jan 27 17:43:41 crc kubenswrapper[4772]: I0127 17:43:41.663261 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:43:41 crc kubenswrapper[4772]: E0127 17:43:41.664522 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:43:52 crc kubenswrapper[4772]: I0127 17:43:52.664204 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:43:52 crc kubenswrapper[4772]: E0127 17:43:52.665009 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:44:04 crc kubenswrapper[4772]: I0127 17:44:04.674700 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:44:04 crc kubenswrapper[4772]: E0127 17:44:04.675972 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:44:16 crc kubenswrapper[4772]: I0127 17:44:16.485025 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ztd7g"] Jan 27 17:44:16 crc kubenswrapper[4772]: E0127 17:44:16.486118 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c6ade8d-638e-415e-8363-af0aa99994b2" containerName="extract-utilities" Jan 27 17:44:16 crc kubenswrapper[4772]: I0127 17:44:16.486134 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c6ade8d-638e-415e-8363-af0aa99994b2" containerName="extract-utilities" Jan 27 17:44:16 crc kubenswrapper[4772]: E0127 17:44:16.486150 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c6ade8d-638e-415e-8363-af0aa99994b2" containerName="extract-content" Jan 27 17:44:16 crc kubenswrapper[4772]: I0127 17:44:16.486158 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c6ade8d-638e-415e-8363-af0aa99994b2" containerName="extract-content" Jan 27 17:44:16 crc kubenswrapper[4772]: E0127 17:44:16.486193 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c6ade8d-638e-415e-8363-af0aa99994b2" containerName="registry-server" Jan 27 17:44:16 crc kubenswrapper[4772]: I0127 17:44:16.486202 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c6ade8d-638e-415e-8363-af0aa99994b2" containerName="registry-server" Jan 27 17:44:16 crc kubenswrapper[4772]: I0127 17:44:16.486422 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c6ade8d-638e-415e-8363-af0aa99994b2" containerName="registry-server" Jan 27 17:44:16 crc kubenswrapper[4772]: I0127 17:44:16.488074 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ztd7g" Jan 27 17:44:16 crc kubenswrapper[4772]: I0127 17:44:16.505967 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ztd7g"] Jan 27 17:44:16 crc kubenswrapper[4772]: I0127 17:44:16.610180 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx5j7\" (UniqueName: \"kubernetes.io/projected/cb56c62f-1902-42bd-9f7a-3f0e20c01d88-kube-api-access-nx5j7\") pod \"redhat-marketplace-ztd7g\" (UID: \"cb56c62f-1902-42bd-9f7a-3f0e20c01d88\") " pod="openshift-marketplace/redhat-marketplace-ztd7g" Jan 27 17:44:16 crc kubenswrapper[4772]: I0127 17:44:16.610451 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb56c62f-1902-42bd-9f7a-3f0e20c01d88-utilities\") pod \"redhat-marketplace-ztd7g\" (UID: \"cb56c62f-1902-42bd-9f7a-3f0e20c01d88\") " pod="openshift-marketplace/redhat-marketplace-ztd7g" Jan 27 17:44:16 crc kubenswrapper[4772]: I0127 17:44:16.610673 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb56c62f-1902-42bd-9f7a-3f0e20c01d88-catalog-content\") pod \"redhat-marketplace-ztd7g\" (UID: \"cb56c62f-1902-42bd-9f7a-3f0e20c01d88\") " pod="openshift-marketplace/redhat-marketplace-ztd7g" Jan 27 17:44:16 crc kubenswrapper[4772]: I0127 17:44:16.712093 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx5j7\" (UniqueName: \"kubernetes.io/projected/cb56c62f-1902-42bd-9f7a-3f0e20c01d88-kube-api-access-nx5j7\") pod \"redhat-marketplace-ztd7g\" (UID: \"cb56c62f-1902-42bd-9f7a-3f0e20c01d88\") " pod="openshift-marketplace/redhat-marketplace-ztd7g" Jan 27 17:44:16 crc kubenswrapper[4772]: I0127 17:44:16.712143 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb56c62f-1902-42bd-9f7a-3f0e20c01d88-utilities\") pod \"redhat-marketplace-ztd7g\" (UID: \"cb56c62f-1902-42bd-9f7a-3f0e20c01d88\") " pod="openshift-marketplace/redhat-marketplace-ztd7g" Jan 27 17:44:16 crc kubenswrapper[4772]: I0127 17:44:16.712259 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb56c62f-1902-42bd-9f7a-3f0e20c01d88-catalog-content\") pod \"redhat-marketplace-ztd7g\" (UID: \"cb56c62f-1902-42bd-9f7a-3f0e20c01d88\") " pod="openshift-marketplace/redhat-marketplace-ztd7g" Jan 27 17:44:16 crc kubenswrapper[4772]: I0127 17:44:16.712844 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb56c62f-1902-42bd-9f7a-3f0e20c01d88-catalog-content\") pod \"redhat-marketplace-ztd7g\" (UID: \"cb56c62f-1902-42bd-9f7a-3f0e20c01d88\") " pod="openshift-marketplace/redhat-marketplace-ztd7g" Jan 27 17:44:16 crc kubenswrapper[4772]: I0127 17:44:16.712860 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb56c62f-1902-42bd-9f7a-3f0e20c01d88-utilities\") pod \"redhat-marketplace-ztd7g\" (UID: \"cb56c62f-1902-42bd-9f7a-3f0e20c01d88\") " pod="openshift-marketplace/redhat-marketplace-ztd7g" Jan 27 17:44:16 crc kubenswrapper[4772]: I0127 17:44:16.735125 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx5j7\" (UniqueName: \"kubernetes.io/projected/cb56c62f-1902-42bd-9f7a-3f0e20c01d88-kube-api-access-nx5j7\") pod \"redhat-marketplace-ztd7g\" (UID: \"cb56c62f-1902-42bd-9f7a-3f0e20c01d88\") " pod="openshift-marketplace/redhat-marketplace-ztd7g" Jan 27 17:44:16 crc kubenswrapper[4772]: I0127 17:44:16.877942 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ztd7g" Jan 27 17:44:17 crc kubenswrapper[4772]: I0127 17:44:17.355122 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ztd7g"] Jan 27 17:44:17 crc kubenswrapper[4772]: I0127 17:44:17.663060 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:44:17 crc kubenswrapper[4772]: E0127 17:44:17.663408 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:44:18 crc kubenswrapper[4772]: I0127 17:44:18.374826 4772 generic.go:334] "Generic (PLEG): container finished" podID="cb56c62f-1902-42bd-9f7a-3f0e20c01d88" containerID="b9cc131aafb6b8b3f0d0b0358d7d469d28588ef561bafb8cf718daf3bd6b115a" exitCode=0 Jan 27 17:44:18 crc kubenswrapper[4772]: I0127 17:44:18.374875 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ztd7g" event={"ID":"cb56c62f-1902-42bd-9f7a-3f0e20c01d88","Type":"ContainerDied","Data":"b9cc131aafb6b8b3f0d0b0358d7d469d28588ef561bafb8cf718daf3bd6b115a"} Jan 27 17:44:18 crc kubenswrapper[4772]: I0127 17:44:18.375153 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ztd7g" event={"ID":"cb56c62f-1902-42bd-9f7a-3f0e20c01d88","Type":"ContainerStarted","Data":"5dcb1a2066c81ec8078b97d12fda7371f769744554e102f8bcf84749d10bd1d9"} Jan 27 17:44:19 crc kubenswrapper[4772]: I0127 17:44:19.386992 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ztd7g" event={"ID":"cb56c62f-1902-42bd-9f7a-3f0e20c01d88","Type":"ContainerStarted","Data":"e25227c326a2c65eb0e09fe16b317148fadaabfa5b8ca0ec4137d03294784b1f"} Jan 27 17:44:20 crc kubenswrapper[4772]: I0127 17:44:20.402612 4772 generic.go:334] "Generic (PLEG): container finished" podID="cb56c62f-1902-42bd-9f7a-3f0e20c01d88" containerID="e25227c326a2c65eb0e09fe16b317148fadaabfa5b8ca0ec4137d03294784b1f" exitCode=0 Jan 27 17:44:20 crc kubenswrapper[4772]: I0127 17:44:20.402665 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ztd7g" event={"ID":"cb56c62f-1902-42bd-9f7a-3f0e20c01d88","Type":"ContainerDied","Data":"e25227c326a2c65eb0e09fe16b317148fadaabfa5b8ca0ec4137d03294784b1f"} Jan 27 17:44:21 crc kubenswrapper[4772]: I0127 17:44:21.412065 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ztd7g" event={"ID":"cb56c62f-1902-42bd-9f7a-3f0e20c01d88","Type":"ContainerStarted","Data":"4cbc758915cd523a83bffdd5d4476bbc8eb7f635dfad679b235f04fb5cbef507"} Jan 27 17:44:21 crc kubenswrapper[4772]: I0127 17:44:21.434330 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ztd7g" podStartSLOduration=3.00399019 podStartE2EDuration="5.434311503s" podCreationTimestamp="2026-01-27 17:44:16 +0000 UTC" firstStartedPulling="2026-01-27 17:44:18.37700394 +0000 UTC m=+9444.357613038" lastFinishedPulling="2026-01-27 17:44:20.807325263 +0000 UTC m=+9446.787934351" observedRunningTime="2026-01-27 17:44:21.427944581 +0000 UTC m=+9447.408553709" watchObservedRunningTime="2026-01-27 17:44:21.434311503 +0000 UTC m=+9447.414920611" Jan 27 17:44:26 crc kubenswrapper[4772]: I0127 17:44:26.878950 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ztd7g" Jan 27 17:44:26 crc kubenswrapper[4772]: I0127 17:44:26.879763 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ztd7g" Jan 27 17:44:26 crc kubenswrapper[4772]: I0127 17:44:26.941019 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ztd7g" Jan 27 17:44:27 crc kubenswrapper[4772]: I0127 17:44:27.539939 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ztd7g" Jan 27 17:44:27 crc kubenswrapper[4772]: I0127 17:44:27.609510 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ztd7g"] Jan 27 17:44:29 crc kubenswrapper[4772]: I0127 17:44:29.512043 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ztd7g" podUID="cb56c62f-1902-42bd-9f7a-3f0e20c01d88" containerName="registry-server" containerID="cri-o://4cbc758915cd523a83bffdd5d4476bbc8eb7f635dfad679b235f04fb5cbef507" gracePeriod=2 Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.012900 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ztd7g" Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.118416 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nx5j7\" (UniqueName: \"kubernetes.io/projected/cb56c62f-1902-42bd-9f7a-3f0e20c01d88-kube-api-access-nx5j7\") pod \"cb56c62f-1902-42bd-9f7a-3f0e20c01d88\" (UID: \"cb56c62f-1902-42bd-9f7a-3f0e20c01d88\") " Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.118984 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb56c62f-1902-42bd-9f7a-3f0e20c01d88-utilities\") pod \"cb56c62f-1902-42bd-9f7a-3f0e20c01d88\" (UID: \"cb56c62f-1902-42bd-9f7a-3f0e20c01d88\") " Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.119061 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb56c62f-1902-42bd-9f7a-3f0e20c01d88-catalog-content\") pod \"cb56c62f-1902-42bd-9f7a-3f0e20c01d88\" (UID: \"cb56c62f-1902-42bd-9f7a-3f0e20c01d88\") " Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.120033 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb56c62f-1902-42bd-9f7a-3f0e20c01d88-utilities" (OuterVolumeSpecName: "utilities") pod "cb56c62f-1902-42bd-9f7a-3f0e20c01d88" (UID: "cb56c62f-1902-42bd-9f7a-3f0e20c01d88"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.139273 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb56c62f-1902-42bd-9f7a-3f0e20c01d88-kube-api-access-nx5j7" (OuterVolumeSpecName: "kube-api-access-nx5j7") pod "cb56c62f-1902-42bd-9f7a-3f0e20c01d88" (UID: "cb56c62f-1902-42bd-9f7a-3f0e20c01d88"). InnerVolumeSpecName "kube-api-access-nx5j7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.144537 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb56c62f-1902-42bd-9f7a-3f0e20c01d88-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb56c62f-1902-42bd-9f7a-3f0e20c01d88" (UID: "cb56c62f-1902-42bd-9f7a-3f0e20c01d88"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.222618 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb56c62f-1902-42bd-9f7a-3f0e20c01d88-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.222705 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb56c62f-1902-42bd-9f7a-3f0e20c01d88-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.222738 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nx5j7\" (UniqueName: \"kubernetes.io/projected/cb56c62f-1902-42bd-9f7a-3f0e20c01d88-kube-api-access-nx5j7\") on node \"crc\" DevicePath \"\"" Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.527916 4772 generic.go:334] "Generic (PLEG): container finished" podID="cb56c62f-1902-42bd-9f7a-3f0e20c01d88" containerID="4cbc758915cd523a83bffdd5d4476bbc8eb7f635dfad679b235f04fb5cbef507" exitCode=0 Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.527983 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ztd7g" event={"ID":"cb56c62f-1902-42bd-9f7a-3f0e20c01d88","Type":"ContainerDied","Data":"4cbc758915cd523a83bffdd5d4476bbc8eb7f635dfad679b235f04fb5cbef507"} Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.528028 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ztd7g" event={"ID":"cb56c62f-1902-42bd-9f7a-3f0e20c01d88","Type":"ContainerDied","Data":"5dcb1a2066c81ec8078b97d12fda7371f769744554e102f8bcf84749d10bd1d9"} Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.528037 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ztd7g" Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.528079 4772 scope.go:117] "RemoveContainer" containerID="4cbc758915cd523a83bffdd5d4476bbc8eb7f635dfad679b235f04fb5cbef507" Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.571560 4772 scope.go:117] "RemoveContainer" containerID="e25227c326a2c65eb0e09fe16b317148fadaabfa5b8ca0ec4137d03294784b1f" Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.600050 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ztd7g"] Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.610332 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ztd7g"] Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.617946 4772 scope.go:117] "RemoveContainer" containerID="b9cc131aafb6b8b3f0d0b0358d7d469d28588ef561bafb8cf718daf3bd6b115a" Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.648912 4772 scope.go:117] "RemoveContainer" containerID="4cbc758915cd523a83bffdd5d4476bbc8eb7f635dfad679b235f04fb5cbef507" Jan 27 17:44:30 crc kubenswrapper[4772]: E0127 17:44:30.649584 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cbc758915cd523a83bffdd5d4476bbc8eb7f635dfad679b235f04fb5cbef507\": container with ID starting with 4cbc758915cd523a83bffdd5d4476bbc8eb7f635dfad679b235f04fb5cbef507 not found: ID does not exist" containerID="4cbc758915cd523a83bffdd5d4476bbc8eb7f635dfad679b235f04fb5cbef507" Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.649628 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cbc758915cd523a83bffdd5d4476bbc8eb7f635dfad679b235f04fb5cbef507"} err="failed to get container status \"4cbc758915cd523a83bffdd5d4476bbc8eb7f635dfad679b235f04fb5cbef507\": rpc error: code = NotFound desc = could not find container \"4cbc758915cd523a83bffdd5d4476bbc8eb7f635dfad679b235f04fb5cbef507\": container with ID starting with 4cbc758915cd523a83bffdd5d4476bbc8eb7f635dfad679b235f04fb5cbef507 not found: ID does not exist" Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.649653 4772 scope.go:117] "RemoveContainer" containerID="e25227c326a2c65eb0e09fe16b317148fadaabfa5b8ca0ec4137d03294784b1f" Jan 27 17:44:30 crc kubenswrapper[4772]: E0127 17:44:30.649973 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e25227c326a2c65eb0e09fe16b317148fadaabfa5b8ca0ec4137d03294784b1f\": container with ID starting with e25227c326a2c65eb0e09fe16b317148fadaabfa5b8ca0ec4137d03294784b1f not found: ID does not exist" containerID="e25227c326a2c65eb0e09fe16b317148fadaabfa5b8ca0ec4137d03294784b1f" Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.650000 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e25227c326a2c65eb0e09fe16b317148fadaabfa5b8ca0ec4137d03294784b1f"} err="failed to get container status \"e25227c326a2c65eb0e09fe16b317148fadaabfa5b8ca0ec4137d03294784b1f\": rpc error: code = NotFound desc = could not find container \"e25227c326a2c65eb0e09fe16b317148fadaabfa5b8ca0ec4137d03294784b1f\": container with ID starting with e25227c326a2c65eb0e09fe16b317148fadaabfa5b8ca0ec4137d03294784b1f not found: ID does not exist" Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.650020 4772 scope.go:117] "RemoveContainer" containerID="b9cc131aafb6b8b3f0d0b0358d7d469d28588ef561bafb8cf718daf3bd6b115a" Jan 27 17:44:30 crc kubenswrapper[4772]: E0127 17:44:30.650442 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9cc131aafb6b8b3f0d0b0358d7d469d28588ef561bafb8cf718daf3bd6b115a\": container with ID starting with b9cc131aafb6b8b3f0d0b0358d7d469d28588ef561bafb8cf718daf3bd6b115a not found: ID does not exist" containerID="b9cc131aafb6b8b3f0d0b0358d7d469d28588ef561bafb8cf718daf3bd6b115a" Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.650563 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9cc131aafb6b8b3f0d0b0358d7d469d28588ef561bafb8cf718daf3bd6b115a"} err="failed to get container status \"b9cc131aafb6b8b3f0d0b0358d7d469d28588ef561bafb8cf718daf3bd6b115a\": rpc error: code = NotFound desc = could not find container \"b9cc131aafb6b8b3f0d0b0358d7d469d28588ef561bafb8cf718daf3bd6b115a\": container with ID starting with b9cc131aafb6b8b3f0d0b0358d7d469d28588ef561bafb8cf718daf3bd6b115a not found: ID does not exist" Jan 27 17:44:30 crc kubenswrapper[4772]: I0127 17:44:30.675902 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb56c62f-1902-42bd-9f7a-3f0e20c01d88" path="/var/lib/kubelet/pods/cb56c62f-1902-42bd-9f7a-3f0e20c01d88/volumes" Jan 27 17:44:31 crc kubenswrapper[4772]: I0127 17:44:31.662820 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:44:31 crc kubenswrapper[4772]: E0127 17:44:31.663678 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:44:42 crc kubenswrapper[4772]: I0127 17:44:42.663697 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:44:43 crc kubenswrapper[4772]: I0127 17:44:43.678627 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"c95b967b4844ba74a00799daa1c360319d22d3be7e94ef067fa431b967e2d966"} Jan 27 17:45:00 crc kubenswrapper[4772]: I0127 17:45:00.146239 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492265-rzjpq"] Jan 27 17:45:00 crc kubenswrapper[4772]: E0127 17:45:00.147465 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb56c62f-1902-42bd-9f7a-3f0e20c01d88" containerName="extract-content" Jan 27 17:45:00 crc kubenswrapper[4772]: I0127 17:45:00.147491 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb56c62f-1902-42bd-9f7a-3f0e20c01d88" containerName="extract-content" Jan 27 17:45:00 crc kubenswrapper[4772]: E0127 17:45:00.147518 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb56c62f-1902-42bd-9f7a-3f0e20c01d88" containerName="extract-utilities" Jan 27 17:45:00 crc kubenswrapper[4772]: I0127 17:45:00.147529 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb56c62f-1902-42bd-9f7a-3f0e20c01d88" containerName="extract-utilities" Jan 27 17:45:00 crc kubenswrapper[4772]: E0127 17:45:00.147577 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb56c62f-1902-42bd-9f7a-3f0e20c01d88" containerName="registry-server" Jan 27 17:45:00 crc kubenswrapper[4772]: I0127 17:45:00.147589 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb56c62f-1902-42bd-9f7a-3f0e20c01d88" containerName="registry-server" Jan 27 17:45:00 crc kubenswrapper[4772]: I0127 17:45:00.147934 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb56c62f-1902-42bd-9f7a-3f0e20c01d88" containerName="registry-server" Jan 27 17:45:00 crc kubenswrapper[4772]: I0127 17:45:00.148999 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492265-rzjpq" Jan 27 17:45:00 crc kubenswrapper[4772]: I0127 17:45:00.152736 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 27 17:45:00 crc kubenswrapper[4772]: I0127 17:45:00.153041 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 27 17:45:00 crc kubenswrapper[4772]: I0127 17:45:00.158966 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492265-rzjpq"] Jan 27 17:45:00 crc kubenswrapper[4772]: I0127 17:45:00.232646 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c190402b-ed4d-48d2-984a-6ae0c4457911-config-volume\") pod \"collect-profiles-29492265-rzjpq\" (UID: \"c190402b-ed4d-48d2-984a-6ae0c4457911\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492265-rzjpq" Jan 27 17:45:00 crc kubenswrapper[4772]: I0127 17:45:00.232836 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzgdf\" (UniqueName: \"kubernetes.io/projected/c190402b-ed4d-48d2-984a-6ae0c4457911-kube-api-access-zzgdf\") pod \"collect-profiles-29492265-rzjpq\" (UID: \"c190402b-ed4d-48d2-984a-6ae0c4457911\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492265-rzjpq" Jan 27 17:45:00 crc kubenswrapper[4772]: I0127 17:45:00.232930 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c190402b-ed4d-48d2-984a-6ae0c4457911-secret-volume\") pod \"collect-profiles-29492265-rzjpq\" (UID: \"c190402b-ed4d-48d2-984a-6ae0c4457911\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492265-rzjpq" Jan 27 17:45:00 crc kubenswrapper[4772]: I0127 17:45:00.333853 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzgdf\" (UniqueName: \"kubernetes.io/projected/c190402b-ed4d-48d2-984a-6ae0c4457911-kube-api-access-zzgdf\") pod \"collect-profiles-29492265-rzjpq\" (UID: \"c190402b-ed4d-48d2-984a-6ae0c4457911\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492265-rzjpq" Jan 27 17:45:00 crc kubenswrapper[4772]: I0127 17:45:00.333950 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c190402b-ed4d-48d2-984a-6ae0c4457911-secret-volume\") pod \"collect-profiles-29492265-rzjpq\" (UID: \"c190402b-ed4d-48d2-984a-6ae0c4457911\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492265-rzjpq" Jan 27 17:45:00 crc kubenswrapper[4772]: I0127 17:45:00.334010 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c190402b-ed4d-48d2-984a-6ae0c4457911-config-volume\") pod \"collect-profiles-29492265-rzjpq\" (UID: \"c190402b-ed4d-48d2-984a-6ae0c4457911\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492265-rzjpq" Jan 27 17:45:00 crc kubenswrapper[4772]: I0127 17:45:00.334937 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c190402b-ed4d-48d2-984a-6ae0c4457911-config-volume\") pod \"collect-profiles-29492265-rzjpq\" (UID: \"c190402b-ed4d-48d2-984a-6ae0c4457911\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492265-rzjpq" Jan 27 17:45:00 crc kubenswrapper[4772]: I0127 17:45:00.340058 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c190402b-ed4d-48d2-984a-6ae0c4457911-secret-volume\") pod \"collect-profiles-29492265-rzjpq\" (UID: \"c190402b-ed4d-48d2-984a-6ae0c4457911\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492265-rzjpq" Jan 27 17:45:00 crc kubenswrapper[4772]: I0127 17:45:00.364645 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzgdf\" (UniqueName: \"kubernetes.io/projected/c190402b-ed4d-48d2-984a-6ae0c4457911-kube-api-access-zzgdf\") pod \"collect-profiles-29492265-rzjpq\" (UID: \"c190402b-ed4d-48d2-984a-6ae0c4457911\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29492265-rzjpq" Jan 27 17:45:00 crc kubenswrapper[4772]: I0127 17:45:00.472207 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492265-rzjpq" Jan 27 17:45:00 crc kubenswrapper[4772]: I0127 17:45:00.964510 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492265-rzjpq"] Jan 27 17:45:00 crc kubenswrapper[4772]: W0127 17:45:00.977339 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc190402b_ed4d_48d2_984a_6ae0c4457911.slice/crio-5afbed58df246672c1baa506912246529be637fdd0a5d30927af1bb8c0f2373f WatchSource:0}: Error finding container 5afbed58df246672c1baa506912246529be637fdd0a5d30927af1bb8c0f2373f: Status 404 returned error can't find the container with id 5afbed58df246672c1baa506912246529be637fdd0a5d30927af1bb8c0f2373f Jan 27 17:45:01 crc kubenswrapper[4772]: I0127 17:45:01.950949 4772 generic.go:334] "Generic (PLEG): container finished" podID="c190402b-ed4d-48d2-984a-6ae0c4457911" containerID="b9f31bd1c3fa9ae295f35e740124fa42856302304faaba875d6731b304dbd918" exitCode=0 Jan 27 17:45:01 crc kubenswrapper[4772]: I0127 17:45:01.951208 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492265-rzjpq" event={"ID":"c190402b-ed4d-48d2-984a-6ae0c4457911","Type":"ContainerDied","Data":"b9f31bd1c3fa9ae295f35e740124fa42856302304faaba875d6731b304dbd918"} Jan 27 17:45:01 crc kubenswrapper[4772]: I0127 17:45:01.951297 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492265-rzjpq" event={"ID":"c190402b-ed4d-48d2-984a-6ae0c4457911","Type":"ContainerStarted","Data":"5afbed58df246672c1baa506912246529be637fdd0a5d30927af1bb8c0f2373f"} Jan 27 17:45:03 crc kubenswrapper[4772]: I0127 17:45:03.329615 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492265-rzjpq" Jan 27 17:45:03 crc kubenswrapper[4772]: I0127 17:45:03.389968 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzgdf\" (UniqueName: \"kubernetes.io/projected/c190402b-ed4d-48d2-984a-6ae0c4457911-kube-api-access-zzgdf\") pod \"c190402b-ed4d-48d2-984a-6ae0c4457911\" (UID: \"c190402b-ed4d-48d2-984a-6ae0c4457911\") " Jan 27 17:45:03 crc kubenswrapper[4772]: I0127 17:45:03.390260 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c190402b-ed4d-48d2-984a-6ae0c4457911-secret-volume\") pod \"c190402b-ed4d-48d2-984a-6ae0c4457911\" (UID: \"c190402b-ed4d-48d2-984a-6ae0c4457911\") " Jan 27 17:45:03 crc kubenswrapper[4772]: I0127 17:45:03.390312 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c190402b-ed4d-48d2-984a-6ae0c4457911-config-volume\") pod \"c190402b-ed4d-48d2-984a-6ae0c4457911\" (UID: \"c190402b-ed4d-48d2-984a-6ae0c4457911\") " Jan 27 17:45:03 crc kubenswrapper[4772]: I0127 17:45:03.391089 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c190402b-ed4d-48d2-984a-6ae0c4457911-config-volume" (OuterVolumeSpecName: "config-volume") pod "c190402b-ed4d-48d2-984a-6ae0c4457911" (UID: "c190402b-ed4d-48d2-984a-6ae0c4457911"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 27 17:45:03 crc kubenswrapper[4772]: I0127 17:45:03.396901 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c190402b-ed4d-48d2-984a-6ae0c4457911-kube-api-access-zzgdf" (OuterVolumeSpecName: "kube-api-access-zzgdf") pod "c190402b-ed4d-48d2-984a-6ae0c4457911" (UID: "c190402b-ed4d-48d2-984a-6ae0c4457911"). InnerVolumeSpecName "kube-api-access-zzgdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:45:03 crc kubenswrapper[4772]: I0127 17:45:03.397473 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c190402b-ed4d-48d2-984a-6ae0c4457911-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c190402b-ed4d-48d2-984a-6ae0c4457911" (UID: "c190402b-ed4d-48d2-984a-6ae0c4457911"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 27 17:45:03 crc kubenswrapper[4772]: I0127 17:45:03.492515 4772 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c190402b-ed4d-48d2-984a-6ae0c4457911-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 27 17:45:03 crc kubenswrapper[4772]: I0127 17:45:03.492561 4772 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c190402b-ed4d-48d2-984a-6ae0c4457911-config-volume\") on node \"crc\" DevicePath \"\"" Jan 27 17:45:03 crc kubenswrapper[4772]: I0127 17:45:03.492578 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzgdf\" (UniqueName: \"kubernetes.io/projected/c190402b-ed4d-48d2-984a-6ae0c4457911-kube-api-access-zzgdf\") on node \"crc\" DevicePath \"\"" Jan 27 17:45:03 crc kubenswrapper[4772]: I0127 17:45:03.970881 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29492265-rzjpq" event={"ID":"c190402b-ed4d-48d2-984a-6ae0c4457911","Type":"ContainerDied","Data":"5afbed58df246672c1baa506912246529be637fdd0a5d30927af1bb8c0f2373f"} Jan 27 17:45:03 crc kubenswrapper[4772]: I0127 17:45:03.970924 4772 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5afbed58df246672c1baa506912246529be637fdd0a5d30927af1bb8c0f2373f" Jan 27 17:45:03 crc kubenswrapper[4772]: I0127 17:45:03.970989 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29492265-rzjpq" Jan 27 17:45:04 crc kubenswrapper[4772]: I0127 17:45:04.427131 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492220-sqfps"] Jan 27 17:45:04 crc kubenswrapper[4772]: I0127 17:45:04.434680 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29492220-sqfps"] Jan 27 17:45:04 crc kubenswrapper[4772]: I0127 17:45:04.683048 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73038a7f-6c26-47b7-ad06-bd235e268224" path="/var/lib/kubelet/pods/73038a7f-6c26-47b7-ad06-bd235e268224/volumes" Jan 27 17:45:11 crc kubenswrapper[4772]: I0127 17:45:11.648072 4772 scope.go:117] "RemoveContainer" containerID="21fbf772d614ea3a35cb7d6244635ba0574a7b4a610726fba6575e200d3d3209" Jan 27 17:46:42 crc kubenswrapper[4772]: I0127 17:46:42.058956 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:46:42 crc kubenswrapper[4772]: I0127 17:46:42.059568 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:46:50 crc kubenswrapper[4772]: I0127 17:46:50.345314 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gj8cc"] Jan 27 17:46:50 crc kubenswrapper[4772]: E0127 17:46:50.346833 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c190402b-ed4d-48d2-984a-6ae0c4457911" containerName="collect-profiles" Jan 27 17:46:50 crc kubenswrapper[4772]: I0127 17:46:50.346851 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="c190402b-ed4d-48d2-984a-6ae0c4457911" containerName="collect-profiles" Jan 27 17:46:50 crc kubenswrapper[4772]: I0127 17:46:50.347081 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="c190402b-ed4d-48d2-984a-6ae0c4457911" containerName="collect-profiles" Jan 27 17:46:50 crc kubenswrapper[4772]: I0127 17:46:50.348682 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gj8cc" Jan 27 17:46:50 crc kubenswrapper[4772]: I0127 17:46:50.353845 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gj8cc"] Jan 27 17:46:50 crc kubenswrapper[4772]: I0127 17:46:50.547033 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gljr4\" (UniqueName: \"kubernetes.io/projected/4d35ce00-8f21-48c4-ac86-b51879a0f1a0-kube-api-access-gljr4\") pod \"certified-operators-gj8cc\" (UID: \"4d35ce00-8f21-48c4-ac86-b51879a0f1a0\") " pod="openshift-marketplace/certified-operators-gj8cc" Jan 27 17:46:50 crc kubenswrapper[4772]: I0127 17:46:50.547759 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d35ce00-8f21-48c4-ac86-b51879a0f1a0-catalog-content\") pod \"certified-operators-gj8cc\" (UID: \"4d35ce00-8f21-48c4-ac86-b51879a0f1a0\") " pod="openshift-marketplace/certified-operators-gj8cc" Jan 27 17:46:50 crc kubenswrapper[4772]: I0127 17:46:50.548019 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d35ce00-8f21-48c4-ac86-b51879a0f1a0-utilities\") pod \"certified-operators-gj8cc\" (UID: \"4d35ce00-8f21-48c4-ac86-b51879a0f1a0\") " pod="openshift-marketplace/certified-operators-gj8cc" Jan 27 17:46:50 crc kubenswrapper[4772]: I0127 17:46:50.649775 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d35ce00-8f21-48c4-ac86-b51879a0f1a0-catalog-content\") pod \"certified-operators-gj8cc\" (UID: \"4d35ce00-8f21-48c4-ac86-b51879a0f1a0\") " pod="openshift-marketplace/certified-operators-gj8cc" Jan 27 17:46:50 crc kubenswrapper[4772]: I0127 17:46:50.649868 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d35ce00-8f21-48c4-ac86-b51879a0f1a0-utilities\") pod \"certified-operators-gj8cc\" (UID: \"4d35ce00-8f21-48c4-ac86-b51879a0f1a0\") " pod="openshift-marketplace/certified-operators-gj8cc" Jan 27 17:46:50 crc kubenswrapper[4772]: I0127 17:46:50.649938 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gljr4\" (UniqueName: \"kubernetes.io/projected/4d35ce00-8f21-48c4-ac86-b51879a0f1a0-kube-api-access-gljr4\") pod \"certified-operators-gj8cc\" (UID: \"4d35ce00-8f21-48c4-ac86-b51879a0f1a0\") " pod="openshift-marketplace/certified-operators-gj8cc" Jan 27 17:46:50 crc kubenswrapper[4772]: I0127 17:46:50.650395 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d35ce00-8f21-48c4-ac86-b51879a0f1a0-catalog-content\") pod \"certified-operators-gj8cc\" (UID: \"4d35ce00-8f21-48c4-ac86-b51879a0f1a0\") " pod="openshift-marketplace/certified-operators-gj8cc" Jan 27 17:46:50 crc kubenswrapper[4772]: I0127 17:46:50.650571 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d35ce00-8f21-48c4-ac86-b51879a0f1a0-utilities\") pod \"certified-operators-gj8cc\" (UID: \"4d35ce00-8f21-48c4-ac86-b51879a0f1a0\") " pod="openshift-marketplace/certified-operators-gj8cc" Jan 27 17:46:50 crc kubenswrapper[4772]: I0127 17:46:50.681140 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gljr4\" (UniqueName: \"kubernetes.io/projected/4d35ce00-8f21-48c4-ac86-b51879a0f1a0-kube-api-access-gljr4\") pod \"certified-operators-gj8cc\" (UID: \"4d35ce00-8f21-48c4-ac86-b51879a0f1a0\") " pod="openshift-marketplace/certified-operators-gj8cc" Jan 27 17:46:50 crc kubenswrapper[4772]: I0127 17:46:50.975423 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gj8cc" Jan 27 17:46:51 crc kubenswrapper[4772]: I0127 17:46:51.427725 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gj8cc"] Jan 27 17:46:52 crc kubenswrapper[4772]: I0127 17:46:52.065518 4772 generic.go:334] "Generic (PLEG): container finished" podID="4d35ce00-8f21-48c4-ac86-b51879a0f1a0" containerID="3563a42c4868e6463dd4fd0526fe31c9ba5bd18efc51c3bed319b5a0efe056aa" exitCode=0 Jan 27 17:46:52 crc kubenswrapper[4772]: I0127 17:46:52.065573 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gj8cc" event={"ID":"4d35ce00-8f21-48c4-ac86-b51879a0f1a0","Type":"ContainerDied","Data":"3563a42c4868e6463dd4fd0526fe31c9ba5bd18efc51c3bed319b5a0efe056aa"} Jan 27 17:46:52 crc kubenswrapper[4772]: I0127 17:46:52.065861 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gj8cc" event={"ID":"4d35ce00-8f21-48c4-ac86-b51879a0f1a0","Type":"ContainerStarted","Data":"27d4dc9c40f39e670cd2a0ed4833fa464d1c4b2f7c7516785daa6f3707fdf84e"} Jan 27 17:46:52 crc kubenswrapper[4772]: I0127 17:46:52.067561 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 17:46:53 crc kubenswrapper[4772]: I0127 17:46:53.077857 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gj8cc" event={"ID":"4d35ce00-8f21-48c4-ac86-b51879a0f1a0","Type":"ContainerStarted","Data":"943d9b9f50746b3fc3343e8866594a5e4a51e753c34e1d5555f821e76c7457c4"} Jan 27 17:46:54 crc kubenswrapper[4772]: I0127 17:46:54.093525 4772 generic.go:334] "Generic (PLEG): container finished" podID="4d35ce00-8f21-48c4-ac86-b51879a0f1a0" containerID="943d9b9f50746b3fc3343e8866594a5e4a51e753c34e1d5555f821e76c7457c4" exitCode=0 Jan 27 17:46:54 crc kubenswrapper[4772]: I0127 17:46:54.093858 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gj8cc" event={"ID":"4d35ce00-8f21-48c4-ac86-b51879a0f1a0","Type":"ContainerDied","Data":"943d9b9f50746b3fc3343e8866594a5e4a51e753c34e1d5555f821e76c7457c4"} Jan 27 17:46:55 crc kubenswrapper[4772]: I0127 17:46:55.105302 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gj8cc" event={"ID":"4d35ce00-8f21-48c4-ac86-b51879a0f1a0","Type":"ContainerStarted","Data":"6f1f87f9eb4015bb32f09732614123231dc5df6622454fbdaa7783dc972f7400"} Jan 27 17:46:55 crc kubenswrapper[4772]: I0127 17:46:55.130456 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gj8cc" podStartSLOduration=2.551489794 podStartE2EDuration="5.130441511s" podCreationTimestamp="2026-01-27 17:46:50 +0000 UTC" firstStartedPulling="2026-01-27 17:46:52.067362604 +0000 UTC m=+9598.047971702" lastFinishedPulling="2026-01-27 17:46:54.646314321 +0000 UTC m=+9600.626923419" observedRunningTime="2026-01-27 17:46:55.122053872 +0000 UTC m=+9601.102662990" watchObservedRunningTime="2026-01-27 17:46:55.130441511 +0000 UTC m=+9601.111050609" Jan 27 17:47:00 crc kubenswrapper[4772]: I0127 17:47:00.976457 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gj8cc" Jan 27 17:47:00 crc kubenswrapper[4772]: I0127 17:47:00.978269 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gj8cc" Jan 27 17:47:01 crc kubenswrapper[4772]: I0127 17:47:01.034722 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gj8cc" Jan 27 17:47:01 crc kubenswrapper[4772]: I0127 17:47:01.221075 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gj8cc" Jan 27 17:47:01 crc kubenswrapper[4772]: I0127 17:47:01.280830 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gj8cc"] Jan 27 17:47:03 crc kubenswrapper[4772]: I0127 17:47:03.188678 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gj8cc" podUID="4d35ce00-8f21-48c4-ac86-b51879a0f1a0" containerName="registry-server" containerID="cri-o://6f1f87f9eb4015bb32f09732614123231dc5df6622454fbdaa7783dc972f7400" gracePeriod=2 Jan 27 17:47:03 crc kubenswrapper[4772]: I0127 17:47:03.643562 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gj8cc" Jan 27 17:47:03 crc kubenswrapper[4772]: I0127 17:47:03.766663 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d35ce00-8f21-48c4-ac86-b51879a0f1a0-catalog-content\") pod \"4d35ce00-8f21-48c4-ac86-b51879a0f1a0\" (UID: \"4d35ce00-8f21-48c4-ac86-b51879a0f1a0\") " Jan 27 17:47:03 crc kubenswrapper[4772]: I0127 17:47:03.766762 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gljr4\" (UniqueName: \"kubernetes.io/projected/4d35ce00-8f21-48c4-ac86-b51879a0f1a0-kube-api-access-gljr4\") pod \"4d35ce00-8f21-48c4-ac86-b51879a0f1a0\" (UID: \"4d35ce00-8f21-48c4-ac86-b51879a0f1a0\") " Jan 27 17:47:03 crc kubenswrapper[4772]: I0127 17:47:03.766802 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d35ce00-8f21-48c4-ac86-b51879a0f1a0-utilities\") pod \"4d35ce00-8f21-48c4-ac86-b51879a0f1a0\" (UID: \"4d35ce00-8f21-48c4-ac86-b51879a0f1a0\") " Jan 27 17:47:03 crc kubenswrapper[4772]: I0127 17:47:03.769239 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d35ce00-8f21-48c4-ac86-b51879a0f1a0-utilities" (OuterVolumeSpecName: "utilities") pod "4d35ce00-8f21-48c4-ac86-b51879a0f1a0" (UID: "4d35ce00-8f21-48c4-ac86-b51879a0f1a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:47:03 crc kubenswrapper[4772]: I0127 17:47:03.777519 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d35ce00-8f21-48c4-ac86-b51879a0f1a0-kube-api-access-gljr4" (OuterVolumeSpecName: "kube-api-access-gljr4") pod "4d35ce00-8f21-48c4-ac86-b51879a0f1a0" (UID: "4d35ce00-8f21-48c4-ac86-b51879a0f1a0"). InnerVolumeSpecName "kube-api-access-gljr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:47:03 crc kubenswrapper[4772]: I0127 17:47:03.868746 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gljr4\" (UniqueName: \"kubernetes.io/projected/4d35ce00-8f21-48c4-ac86-b51879a0f1a0-kube-api-access-gljr4\") on node \"crc\" DevicePath \"\"" Jan 27 17:47:03 crc kubenswrapper[4772]: I0127 17:47:03.868967 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d35ce00-8f21-48c4-ac86-b51879a0f1a0-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 17:47:04 crc kubenswrapper[4772]: I0127 17:47:04.091570 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d35ce00-8f21-48c4-ac86-b51879a0f1a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4d35ce00-8f21-48c4-ac86-b51879a0f1a0" (UID: "4d35ce00-8f21-48c4-ac86-b51879a0f1a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:47:04 crc kubenswrapper[4772]: I0127 17:47:04.175205 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d35ce00-8f21-48c4-ac86-b51879a0f1a0-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 17:47:04 crc kubenswrapper[4772]: I0127 17:47:04.214836 4772 generic.go:334] "Generic (PLEG): container finished" podID="4d35ce00-8f21-48c4-ac86-b51879a0f1a0" containerID="6f1f87f9eb4015bb32f09732614123231dc5df6622454fbdaa7783dc972f7400" exitCode=0 Jan 27 17:47:04 crc kubenswrapper[4772]: I0127 17:47:04.214899 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gj8cc" event={"ID":"4d35ce00-8f21-48c4-ac86-b51879a0f1a0","Type":"ContainerDied","Data":"6f1f87f9eb4015bb32f09732614123231dc5df6622454fbdaa7783dc972f7400"} Jan 27 17:47:04 crc kubenswrapper[4772]: I0127 17:47:04.214916 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gj8cc" Jan 27 17:47:04 crc kubenswrapper[4772]: I0127 17:47:04.214939 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gj8cc" event={"ID":"4d35ce00-8f21-48c4-ac86-b51879a0f1a0","Type":"ContainerDied","Data":"27d4dc9c40f39e670cd2a0ed4833fa464d1c4b2f7c7516785daa6f3707fdf84e"} Jan 27 17:47:04 crc kubenswrapper[4772]: I0127 17:47:04.214957 4772 scope.go:117] "RemoveContainer" containerID="6f1f87f9eb4015bb32f09732614123231dc5df6622454fbdaa7783dc972f7400" Jan 27 17:47:04 crc kubenswrapper[4772]: I0127 17:47:04.265523 4772 scope.go:117] "RemoveContainer" containerID="943d9b9f50746b3fc3343e8866594a5e4a51e753c34e1d5555f821e76c7457c4" Jan 27 17:47:04 crc kubenswrapper[4772]: I0127 17:47:04.269191 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gj8cc"] Jan 27 17:47:04 crc kubenswrapper[4772]: I0127 17:47:04.279054 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gj8cc"] Jan 27 17:47:04 crc kubenswrapper[4772]: I0127 17:47:04.293867 4772 scope.go:117] "RemoveContainer" containerID="3563a42c4868e6463dd4fd0526fe31c9ba5bd18efc51c3bed319b5a0efe056aa" Jan 27 17:47:04 crc kubenswrapper[4772]: I0127 17:47:04.343269 4772 scope.go:117] "RemoveContainer" containerID="6f1f87f9eb4015bb32f09732614123231dc5df6622454fbdaa7783dc972f7400" Jan 27 17:47:04 crc kubenswrapper[4772]: E0127 17:47:04.343754 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f1f87f9eb4015bb32f09732614123231dc5df6622454fbdaa7783dc972f7400\": container with ID starting with 6f1f87f9eb4015bb32f09732614123231dc5df6622454fbdaa7783dc972f7400 not found: ID does not exist" containerID="6f1f87f9eb4015bb32f09732614123231dc5df6622454fbdaa7783dc972f7400" Jan 27 17:47:04 crc kubenswrapper[4772]: I0127 17:47:04.343808 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f1f87f9eb4015bb32f09732614123231dc5df6622454fbdaa7783dc972f7400"} err="failed to get container status \"6f1f87f9eb4015bb32f09732614123231dc5df6622454fbdaa7783dc972f7400\": rpc error: code = NotFound desc = could not find container \"6f1f87f9eb4015bb32f09732614123231dc5df6622454fbdaa7783dc972f7400\": container with ID starting with 6f1f87f9eb4015bb32f09732614123231dc5df6622454fbdaa7783dc972f7400 not found: ID does not exist" Jan 27 17:47:04 crc kubenswrapper[4772]: I0127 17:47:04.343840 4772 scope.go:117] "RemoveContainer" containerID="943d9b9f50746b3fc3343e8866594a5e4a51e753c34e1d5555f821e76c7457c4" Jan 27 17:47:04 crc kubenswrapper[4772]: E0127 17:47:04.344108 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"943d9b9f50746b3fc3343e8866594a5e4a51e753c34e1d5555f821e76c7457c4\": container with ID starting with 943d9b9f50746b3fc3343e8866594a5e4a51e753c34e1d5555f821e76c7457c4 not found: ID does not exist" containerID="943d9b9f50746b3fc3343e8866594a5e4a51e753c34e1d5555f821e76c7457c4" Jan 27 17:47:04 crc kubenswrapper[4772]: I0127 17:47:04.344141 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"943d9b9f50746b3fc3343e8866594a5e4a51e753c34e1d5555f821e76c7457c4"} err="failed to get container status \"943d9b9f50746b3fc3343e8866594a5e4a51e753c34e1d5555f821e76c7457c4\": rpc error: code = NotFound desc = could not find container \"943d9b9f50746b3fc3343e8866594a5e4a51e753c34e1d5555f821e76c7457c4\": container with ID starting with 943d9b9f50746b3fc3343e8866594a5e4a51e753c34e1d5555f821e76c7457c4 not found: ID does not exist" Jan 27 17:47:04 crc kubenswrapper[4772]: I0127 17:47:04.344159 4772 scope.go:117] "RemoveContainer" containerID="3563a42c4868e6463dd4fd0526fe31c9ba5bd18efc51c3bed319b5a0efe056aa" Jan 27 17:47:04 crc kubenswrapper[4772]: E0127 17:47:04.344707 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3563a42c4868e6463dd4fd0526fe31c9ba5bd18efc51c3bed319b5a0efe056aa\": container with ID starting with 3563a42c4868e6463dd4fd0526fe31c9ba5bd18efc51c3bed319b5a0efe056aa not found: ID does not exist" containerID="3563a42c4868e6463dd4fd0526fe31c9ba5bd18efc51c3bed319b5a0efe056aa" Jan 27 17:47:04 crc kubenswrapper[4772]: I0127 17:47:04.344794 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3563a42c4868e6463dd4fd0526fe31c9ba5bd18efc51c3bed319b5a0efe056aa"} err="failed to get container status \"3563a42c4868e6463dd4fd0526fe31c9ba5bd18efc51c3bed319b5a0efe056aa\": rpc error: code = NotFound desc = could not find container \"3563a42c4868e6463dd4fd0526fe31c9ba5bd18efc51c3bed319b5a0efe056aa\": container with ID starting with 3563a42c4868e6463dd4fd0526fe31c9ba5bd18efc51c3bed319b5a0efe056aa not found: ID does not exist" Jan 27 17:47:04 crc kubenswrapper[4772]: I0127 17:47:04.677109 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d35ce00-8f21-48c4-ac86-b51879a0f1a0" path="/var/lib/kubelet/pods/4d35ce00-8f21-48c4-ac86-b51879a0f1a0/volumes" Jan 27 17:47:12 crc kubenswrapper[4772]: I0127 17:47:12.058972 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:47:12 crc kubenswrapper[4772]: I0127 17:47:12.059685 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:47:42 crc kubenswrapper[4772]: I0127 17:47:42.058709 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:47:42 crc kubenswrapper[4772]: I0127 17:47:42.059575 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:47:42 crc kubenswrapper[4772]: I0127 17:47:42.060467 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 17:47:42 crc kubenswrapper[4772]: I0127 17:47:42.061775 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c95b967b4844ba74a00799daa1c360319d22d3be7e94ef067fa431b967e2d966"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 17:47:42 crc kubenswrapper[4772]: I0127 17:47:42.061912 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://c95b967b4844ba74a00799daa1c360319d22d3be7e94ef067fa431b967e2d966" gracePeriod=600 Jan 27 17:47:42 crc kubenswrapper[4772]: I0127 17:47:42.644351 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="c95b967b4844ba74a00799daa1c360319d22d3be7e94ef067fa431b967e2d966" exitCode=0 Jan 27 17:47:42 crc kubenswrapper[4772]: I0127 17:47:42.644450 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"c95b967b4844ba74a00799daa1c360319d22d3be7e94ef067fa431b967e2d966"} Jan 27 17:47:42 crc kubenswrapper[4772]: I0127 17:47:42.644713 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025"} Jan 27 17:47:42 crc kubenswrapper[4772]: I0127 17:47:42.644734 4772 scope.go:117] "RemoveContainer" containerID="900c1aed2e392c4c453d12a872ae4215d2e07d45524a30632831cba37945e88d" Jan 27 17:49:42 crc kubenswrapper[4772]: I0127 17:49:42.058963 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:49:42 crc kubenswrapper[4772]: I0127 17:49:42.059848 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:50:07 crc kubenswrapper[4772]: I0127 17:50:07.714221 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7dvs7/must-gather-vr6mx"] Jan 27 17:50:07 crc kubenswrapper[4772]: E0127 17:50:07.716254 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d35ce00-8f21-48c4-ac86-b51879a0f1a0" containerName="extract-content" Jan 27 17:50:07 crc kubenswrapper[4772]: I0127 17:50:07.716361 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d35ce00-8f21-48c4-ac86-b51879a0f1a0" containerName="extract-content" Jan 27 17:50:07 crc kubenswrapper[4772]: E0127 17:50:07.716455 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d35ce00-8f21-48c4-ac86-b51879a0f1a0" containerName="extract-utilities" Jan 27 17:50:07 crc kubenswrapper[4772]: I0127 17:50:07.716530 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d35ce00-8f21-48c4-ac86-b51879a0f1a0" containerName="extract-utilities" Jan 27 17:50:07 crc kubenswrapper[4772]: E0127 17:50:07.716608 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d35ce00-8f21-48c4-ac86-b51879a0f1a0" containerName="registry-server" Jan 27 17:50:07 crc kubenswrapper[4772]: I0127 17:50:07.716689 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d35ce00-8f21-48c4-ac86-b51879a0f1a0" containerName="registry-server" Jan 27 17:50:07 crc kubenswrapper[4772]: I0127 17:50:07.716999 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d35ce00-8f21-48c4-ac86-b51879a0f1a0" containerName="registry-server" Jan 27 17:50:07 crc kubenswrapper[4772]: I0127 17:50:07.718363 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7dvs7/must-gather-vr6mx" Jan 27 17:50:07 crc kubenswrapper[4772]: I0127 17:50:07.726215 4772 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-7dvs7"/"default-dockercfg-rfw2p" Jan 27 17:50:07 crc kubenswrapper[4772]: I0127 17:50:07.726588 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-7dvs7"/"kube-root-ca.crt" Jan 27 17:50:07 crc kubenswrapper[4772]: I0127 17:50:07.726756 4772 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-7dvs7"/"openshift-service-ca.crt" Jan 27 17:50:07 crc kubenswrapper[4772]: I0127 17:50:07.732331 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-7dvs7/must-gather-vr6mx"] Jan 27 17:50:07 crc kubenswrapper[4772]: I0127 17:50:07.801229 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/62afdf43-7cc5-4d53-aff8-2fd18fbfd493-must-gather-output\") pod \"must-gather-vr6mx\" (UID: \"62afdf43-7cc5-4d53-aff8-2fd18fbfd493\") " pod="openshift-must-gather-7dvs7/must-gather-vr6mx" Jan 27 17:50:07 crc kubenswrapper[4772]: I0127 17:50:07.801316 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwwmf\" (UniqueName: \"kubernetes.io/projected/62afdf43-7cc5-4d53-aff8-2fd18fbfd493-kube-api-access-vwwmf\") pod \"must-gather-vr6mx\" (UID: \"62afdf43-7cc5-4d53-aff8-2fd18fbfd493\") " pod="openshift-must-gather-7dvs7/must-gather-vr6mx" Jan 27 17:50:07 crc kubenswrapper[4772]: I0127 17:50:07.903635 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/62afdf43-7cc5-4d53-aff8-2fd18fbfd493-must-gather-output\") pod \"must-gather-vr6mx\" (UID: \"62afdf43-7cc5-4d53-aff8-2fd18fbfd493\") " pod="openshift-must-gather-7dvs7/must-gather-vr6mx" Jan 27 17:50:07 crc kubenswrapper[4772]: I0127 17:50:07.903798 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwwmf\" (UniqueName: \"kubernetes.io/projected/62afdf43-7cc5-4d53-aff8-2fd18fbfd493-kube-api-access-vwwmf\") pod \"must-gather-vr6mx\" (UID: \"62afdf43-7cc5-4d53-aff8-2fd18fbfd493\") " pod="openshift-must-gather-7dvs7/must-gather-vr6mx" Jan 27 17:50:07 crc kubenswrapper[4772]: I0127 17:50:07.904069 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/62afdf43-7cc5-4d53-aff8-2fd18fbfd493-must-gather-output\") pod \"must-gather-vr6mx\" (UID: \"62afdf43-7cc5-4d53-aff8-2fd18fbfd493\") " pod="openshift-must-gather-7dvs7/must-gather-vr6mx" Jan 27 17:50:07 crc kubenswrapper[4772]: I0127 17:50:07.936426 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwwmf\" (UniqueName: \"kubernetes.io/projected/62afdf43-7cc5-4d53-aff8-2fd18fbfd493-kube-api-access-vwwmf\") pod \"must-gather-vr6mx\" (UID: \"62afdf43-7cc5-4d53-aff8-2fd18fbfd493\") " pod="openshift-must-gather-7dvs7/must-gather-vr6mx" Jan 27 17:50:08 crc kubenswrapper[4772]: I0127 17:50:08.052821 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7dvs7/must-gather-vr6mx" Jan 27 17:50:08 crc kubenswrapper[4772]: I0127 17:50:08.525668 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-7dvs7/must-gather-vr6mx"] Jan 27 17:50:09 crc kubenswrapper[4772]: I0127 17:50:09.110032 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7dvs7/must-gather-vr6mx" event={"ID":"62afdf43-7cc5-4d53-aff8-2fd18fbfd493","Type":"ContainerStarted","Data":"73a87bca576775c52e6251c92bef7e518f388fa9a11be3bd3902b9b4c38f77b6"} Jan 27 17:50:12 crc kubenswrapper[4772]: I0127 17:50:12.059097 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:50:12 crc kubenswrapper[4772]: I0127 17:50:12.059910 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:50:16 crc kubenswrapper[4772]: I0127 17:50:16.182097 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7dvs7/must-gather-vr6mx" event={"ID":"62afdf43-7cc5-4d53-aff8-2fd18fbfd493","Type":"ContainerStarted","Data":"7d2e1870707eb567a3ae7d6fc16a9352a0ad811fd3af2c67ad9db194e0056b53"} Jan 27 17:50:16 crc kubenswrapper[4772]: I0127 17:50:16.182717 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7dvs7/must-gather-vr6mx" event={"ID":"62afdf43-7cc5-4d53-aff8-2fd18fbfd493","Type":"ContainerStarted","Data":"daea92a47f41c2b8eac7ad3a9eb9829829dd28a915856ca1714eacc250cb2602"} Jan 27 17:50:16 crc kubenswrapper[4772]: I0127 17:50:16.217729 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-7dvs7/must-gather-vr6mx" podStartSLOduration=2.695294125 podStartE2EDuration="9.217701194s" podCreationTimestamp="2026-01-27 17:50:07 +0000 UTC" firstStartedPulling="2026-01-27 17:50:08.534806985 +0000 UTC m=+9794.515416093" lastFinishedPulling="2026-01-27 17:50:15.057214064 +0000 UTC m=+9801.037823162" observedRunningTime="2026-01-27 17:50:16.207139393 +0000 UTC m=+9802.187748491" watchObservedRunningTime="2026-01-27 17:50:16.217701194 +0000 UTC m=+9802.198310312" Jan 27 17:50:20 crc kubenswrapper[4772]: I0127 17:50:20.302019 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7dvs7/crc-debug-fch7g"] Jan 27 17:50:20 crc kubenswrapper[4772]: I0127 17:50:20.306278 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7dvs7/crc-debug-fch7g" Jan 27 17:50:20 crc kubenswrapper[4772]: I0127 17:50:20.360845 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3b7e9e60-6e19-49be-866a-fba6fd4b0780-host\") pod \"crc-debug-fch7g\" (UID: \"3b7e9e60-6e19-49be-866a-fba6fd4b0780\") " pod="openshift-must-gather-7dvs7/crc-debug-fch7g" Jan 27 17:50:20 crc kubenswrapper[4772]: I0127 17:50:20.360981 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpdvc\" (UniqueName: \"kubernetes.io/projected/3b7e9e60-6e19-49be-866a-fba6fd4b0780-kube-api-access-xpdvc\") pod \"crc-debug-fch7g\" (UID: \"3b7e9e60-6e19-49be-866a-fba6fd4b0780\") " pod="openshift-must-gather-7dvs7/crc-debug-fch7g" Jan 27 17:50:20 crc kubenswrapper[4772]: I0127 17:50:20.462626 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpdvc\" (UniqueName: \"kubernetes.io/projected/3b7e9e60-6e19-49be-866a-fba6fd4b0780-kube-api-access-xpdvc\") pod \"crc-debug-fch7g\" (UID: \"3b7e9e60-6e19-49be-866a-fba6fd4b0780\") " pod="openshift-must-gather-7dvs7/crc-debug-fch7g" Jan 27 17:50:20 crc kubenswrapper[4772]: I0127 17:50:20.462790 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3b7e9e60-6e19-49be-866a-fba6fd4b0780-host\") pod \"crc-debug-fch7g\" (UID: \"3b7e9e60-6e19-49be-866a-fba6fd4b0780\") " pod="openshift-must-gather-7dvs7/crc-debug-fch7g" Jan 27 17:50:20 crc kubenswrapper[4772]: I0127 17:50:20.462931 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3b7e9e60-6e19-49be-866a-fba6fd4b0780-host\") pod \"crc-debug-fch7g\" (UID: \"3b7e9e60-6e19-49be-866a-fba6fd4b0780\") " pod="openshift-must-gather-7dvs7/crc-debug-fch7g" Jan 27 17:50:20 crc kubenswrapper[4772]: I0127 17:50:20.488942 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpdvc\" (UniqueName: \"kubernetes.io/projected/3b7e9e60-6e19-49be-866a-fba6fd4b0780-kube-api-access-xpdvc\") pod \"crc-debug-fch7g\" (UID: \"3b7e9e60-6e19-49be-866a-fba6fd4b0780\") " pod="openshift-must-gather-7dvs7/crc-debug-fch7g" Jan 27 17:50:20 crc kubenswrapper[4772]: I0127 17:50:20.647365 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7dvs7/crc-debug-fch7g" Jan 27 17:50:21 crc kubenswrapper[4772]: I0127 17:50:21.224730 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7dvs7/crc-debug-fch7g" event={"ID":"3b7e9e60-6e19-49be-866a-fba6fd4b0780","Type":"ContainerStarted","Data":"ab84431cd9516b39dc9cfd9729d8b5ee60ad8c495c7dcade7941ee04e47fe49d"} Jan 27 17:50:32 crc kubenswrapper[4772]: I0127 17:50:32.355298 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7dvs7/crc-debug-fch7g" event={"ID":"3b7e9e60-6e19-49be-866a-fba6fd4b0780","Type":"ContainerStarted","Data":"1ea32c15c368715377429d9d913ced250c3cced4bf09ea7a9096c6fa505dcd27"} Jan 27 17:50:32 crc kubenswrapper[4772]: I0127 17:50:32.379981 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-7dvs7/crc-debug-fch7g" podStartSLOduration=1.298720013 podStartE2EDuration="12.379961991s" podCreationTimestamp="2026-01-27 17:50:20 +0000 UTC" firstStartedPulling="2026-01-27 17:50:20.682106669 +0000 UTC m=+9806.662715797" lastFinishedPulling="2026-01-27 17:50:31.763348677 +0000 UTC m=+9817.743957775" observedRunningTime="2026-01-27 17:50:32.367924557 +0000 UTC m=+9818.348533655" watchObservedRunningTime="2026-01-27 17:50:32.379961991 +0000 UTC m=+9818.360571089" Jan 27 17:50:42 crc kubenswrapper[4772]: I0127 17:50:42.057983 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:50:42 crc kubenswrapper[4772]: I0127 17:50:42.058504 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:50:42 crc kubenswrapper[4772]: I0127 17:50:42.058556 4772 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" Jan 27 17:50:42 crc kubenswrapper[4772]: I0127 17:50:42.059362 4772 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025"} pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 27 17:50:42 crc kubenswrapper[4772]: I0127 17:50:42.059423 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" containerID="cri-o://b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" gracePeriod=600 Jan 27 17:50:42 crc kubenswrapper[4772]: E0127 17:50:42.370472 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:50:42 crc kubenswrapper[4772]: I0127 17:50:42.442883 4772 generic.go:334] "Generic (PLEG): container finished" podID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" exitCode=0 Jan 27 17:50:42 crc kubenswrapper[4772]: I0127 17:50:42.442945 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerDied","Data":"b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025"} Jan 27 17:50:42 crc kubenswrapper[4772]: I0127 17:50:42.442988 4772 scope.go:117] "RemoveContainer" containerID="c95b967b4844ba74a00799daa1c360319d22d3be7e94ef067fa431b967e2d966" Jan 27 17:50:42 crc kubenswrapper[4772]: I0127 17:50:42.443840 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:50:42 crc kubenswrapper[4772]: E0127 17:50:42.444230 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:50:44 crc kubenswrapper[4772]: I0127 17:50:44.781563 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wwjcs"] Jan 27 17:50:44 crc kubenswrapper[4772]: I0127 17:50:44.784380 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wwjcs" Jan 27 17:50:44 crc kubenswrapper[4772]: I0127 17:50:44.798466 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wwjcs"] Jan 27 17:50:44 crc kubenswrapper[4772]: I0127 17:50:44.936958 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m8q5\" (UniqueName: \"kubernetes.io/projected/4bec6411-aae2-47f7-a721-a39d6d381c17-kube-api-access-9m8q5\") pod \"redhat-operators-wwjcs\" (UID: \"4bec6411-aae2-47f7-a721-a39d6d381c17\") " pod="openshift-marketplace/redhat-operators-wwjcs" Jan 27 17:50:44 crc kubenswrapper[4772]: I0127 17:50:44.937041 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bec6411-aae2-47f7-a721-a39d6d381c17-utilities\") pod \"redhat-operators-wwjcs\" (UID: \"4bec6411-aae2-47f7-a721-a39d6d381c17\") " pod="openshift-marketplace/redhat-operators-wwjcs" Jan 27 17:50:44 crc kubenswrapper[4772]: I0127 17:50:44.937146 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bec6411-aae2-47f7-a721-a39d6d381c17-catalog-content\") pod \"redhat-operators-wwjcs\" (UID: \"4bec6411-aae2-47f7-a721-a39d6d381c17\") " pod="openshift-marketplace/redhat-operators-wwjcs" Jan 27 17:50:45 crc kubenswrapper[4772]: I0127 17:50:45.038680 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bec6411-aae2-47f7-a721-a39d6d381c17-catalog-content\") pod \"redhat-operators-wwjcs\" (UID: \"4bec6411-aae2-47f7-a721-a39d6d381c17\") " pod="openshift-marketplace/redhat-operators-wwjcs" Jan 27 17:50:45 crc kubenswrapper[4772]: I0127 17:50:45.039211 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m8q5\" (UniqueName: \"kubernetes.io/projected/4bec6411-aae2-47f7-a721-a39d6d381c17-kube-api-access-9m8q5\") pod \"redhat-operators-wwjcs\" (UID: \"4bec6411-aae2-47f7-a721-a39d6d381c17\") " pod="openshift-marketplace/redhat-operators-wwjcs" Jan 27 17:50:45 crc kubenswrapper[4772]: I0127 17:50:45.039261 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bec6411-aae2-47f7-a721-a39d6d381c17-utilities\") pod \"redhat-operators-wwjcs\" (UID: \"4bec6411-aae2-47f7-a721-a39d6d381c17\") " pod="openshift-marketplace/redhat-operators-wwjcs" Jan 27 17:50:45 crc kubenswrapper[4772]: I0127 17:50:45.039647 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bec6411-aae2-47f7-a721-a39d6d381c17-catalog-content\") pod \"redhat-operators-wwjcs\" (UID: \"4bec6411-aae2-47f7-a721-a39d6d381c17\") " pod="openshift-marketplace/redhat-operators-wwjcs" Jan 27 17:50:45 crc kubenswrapper[4772]: I0127 17:50:45.039790 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bec6411-aae2-47f7-a721-a39d6d381c17-utilities\") pod \"redhat-operators-wwjcs\" (UID: \"4bec6411-aae2-47f7-a721-a39d6d381c17\") " pod="openshift-marketplace/redhat-operators-wwjcs" Jan 27 17:50:45 crc kubenswrapper[4772]: I0127 17:50:45.071511 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m8q5\" (UniqueName: \"kubernetes.io/projected/4bec6411-aae2-47f7-a721-a39d6d381c17-kube-api-access-9m8q5\") pod \"redhat-operators-wwjcs\" (UID: \"4bec6411-aae2-47f7-a721-a39d6d381c17\") " pod="openshift-marketplace/redhat-operators-wwjcs" Jan 27 17:50:45 crc kubenswrapper[4772]: I0127 17:50:45.108478 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wwjcs" Jan 27 17:50:45 crc kubenswrapper[4772]: I0127 17:50:45.645001 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wwjcs"] Jan 27 17:50:46 crc kubenswrapper[4772]: I0127 17:50:46.481037 4772 generic.go:334] "Generic (PLEG): container finished" podID="4bec6411-aae2-47f7-a721-a39d6d381c17" containerID="89b9352f77c65a63cab44c13397c73a1d6c97dafb21ef79db13cbb8873b80e0b" exitCode=0 Jan 27 17:50:46 crc kubenswrapper[4772]: I0127 17:50:46.481112 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwjcs" event={"ID":"4bec6411-aae2-47f7-a721-a39d6d381c17","Type":"ContainerDied","Data":"89b9352f77c65a63cab44c13397c73a1d6c97dafb21ef79db13cbb8873b80e0b"} Jan 27 17:50:46 crc kubenswrapper[4772]: I0127 17:50:46.481865 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwjcs" event={"ID":"4bec6411-aae2-47f7-a721-a39d6d381c17","Type":"ContainerStarted","Data":"bc5839b7d468657c30d46f03fedecb8b4e0ef4ce5baaaea4f34f78a6a2e457ca"} Jan 27 17:50:49 crc kubenswrapper[4772]: I0127 17:50:49.505460 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwjcs" event={"ID":"4bec6411-aae2-47f7-a721-a39d6d381c17","Type":"ContainerStarted","Data":"407a6958420e1bfb3d0405a21d3152ae16fe037a897624d55dc2bc173169ab06"} Jan 27 17:50:52 crc kubenswrapper[4772]: I0127 17:50:52.662851 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:50:52 crc kubenswrapper[4772]: E0127 17:50:52.663681 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:50:53 crc kubenswrapper[4772]: I0127 17:50:53.329644 4772 generic.go:334] "Generic (PLEG): container finished" podID="3b7e9e60-6e19-49be-866a-fba6fd4b0780" containerID="1ea32c15c368715377429d9d913ced250c3cced4bf09ea7a9096c6fa505dcd27" exitCode=0 Jan 27 17:50:53 crc kubenswrapper[4772]: I0127 17:50:53.329813 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7dvs7/crc-debug-fch7g" event={"ID":"3b7e9e60-6e19-49be-866a-fba6fd4b0780","Type":"ContainerDied","Data":"1ea32c15c368715377429d9d913ced250c3cced4bf09ea7a9096c6fa505dcd27"} Jan 27 17:50:53 crc kubenswrapper[4772]: I0127 17:50:53.334608 4772 generic.go:334] "Generic (PLEG): container finished" podID="4bec6411-aae2-47f7-a721-a39d6d381c17" containerID="407a6958420e1bfb3d0405a21d3152ae16fe037a897624d55dc2bc173169ab06" exitCode=0 Jan 27 17:50:53 crc kubenswrapper[4772]: I0127 17:50:53.334647 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwjcs" event={"ID":"4bec6411-aae2-47f7-a721-a39d6d381c17","Type":"ContainerDied","Data":"407a6958420e1bfb3d0405a21d3152ae16fe037a897624d55dc2bc173169ab06"} Jan 27 17:50:54 crc kubenswrapper[4772]: I0127 17:50:54.344285 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwjcs" event={"ID":"4bec6411-aae2-47f7-a721-a39d6d381c17","Type":"ContainerStarted","Data":"fc9c3049ded284cad5528fcbc82380287f3e9e60849f185709b1cbf6091758dd"} Jan 27 17:50:54 crc kubenswrapper[4772]: I0127 17:50:54.378141 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wwjcs" podStartSLOduration=2.85193684 podStartE2EDuration="10.378113555s" podCreationTimestamp="2026-01-27 17:50:44 +0000 UTC" firstStartedPulling="2026-01-27 17:50:46.484785628 +0000 UTC m=+9832.465394726" lastFinishedPulling="2026-01-27 17:50:54.010962343 +0000 UTC m=+9839.991571441" observedRunningTime="2026-01-27 17:50:54.371753413 +0000 UTC m=+9840.352362531" watchObservedRunningTime="2026-01-27 17:50:54.378113555 +0000 UTC m=+9840.358722673" Jan 27 17:50:54 crc kubenswrapper[4772]: I0127 17:50:54.456596 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7dvs7/crc-debug-fch7g" Jan 27 17:50:54 crc kubenswrapper[4772]: I0127 17:50:54.506291 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7dvs7/crc-debug-fch7g"] Jan 27 17:50:54 crc kubenswrapper[4772]: I0127 17:50:54.510432 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7dvs7/crc-debug-fch7g"] Jan 27 17:50:54 crc kubenswrapper[4772]: I0127 17:50:54.649550 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3b7e9e60-6e19-49be-866a-fba6fd4b0780-host\") pod \"3b7e9e60-6e19-49be-866a-fba6fd4b0780\" (UID: \"3b7e9e60-6e19-49be-866a-fba6fd4b0780\") " Jan 27 17:50:54 crc kubenswrapper[4772]: I0127 17:50:54.649696 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b7e9e60-6e19-49be-866a-fba6fd4b0780-host" (OuterVolumeSpecName: "host") pod "3b7e9e60-6e19-49be-866a-fba6fd4b0780" (UID: "3b7e9e60-6e19-49be-866a-fba6fd4b0780"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 17:50:54 crc kubenswrapper[4772]: I0127 17:50:54.650116 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpdvc\" (UniqueName: \"kubernetes.io/projected/3b7e9e60-6e19-49be-866a-fba6fd4b0780-kube-api-access-xpdvc\") pod \"3b7e9e60-6e19-49be-866a-fba6fd4b0780\" (UID: \"3b7e9e60-6e19-49be-866a-fba6fd4b0780\") " Jan 27 17:50:54 crc kubenswrapper[4772]: I0127 17:50:54.650778 4772 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3b7e9e60-6e19-49be-866a-fba6fd4b0780-host\") on node \"crc\" DevicePath \"\"" Jan 27 17:50:54 crc kubenswrapper[4772]: I0127 17:50:54.658420 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b7e9e60-6e19-49be-866a-fba6fd4b0780-kube-api-access-xpdvc" (OuterVolumeSpecName: "kube-api-access-xpdvc") pod "3b7e9e60-6e19-49be-866a-fba6fd4b0780" (UID: "3b7e9e60-6e19-49be-866a-fba6fd4b0780"). InnerVolumeSpecName "kube-api-access-xpdvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:50:54 crc kubenswrapper[4772]: I0127 17:50:54.677154 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b7e9e60-6e19-49be-866a-fba6fd4b0780" path="/var/lib/kubelet/pods/3b7e9e60-6e19-49be-866a-fba6fd4b0780/volumes" Jan 27 17:50:54 crc kubenswrapper[4772]: I0127 17:50:54.753488 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpdvc\" (UniqueName: \"kubernetes.io/projected/3b7e9e60-6e19-49be-866a-fba6fd4b0780-kube-api-access-xpdvc\") on node \"crc\" DevicePath \"\"" Jan 27 17:50:55 crc kubenswrapper[4772]: I0127 17:50:55.109334 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wwjcs" Jan 27 17:50:55 crc kubenswrapper[4772]: I0127 17:50:55.109404 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wwjcs" Jan 27 17:50:55 crc kubenswrapper[4772]: I0127 17:50:55.353675 4772 scope.go:117] "RemoveContainer" containerID="1ea32c15c368715377429d9d913ced250c3cced4bf09ea7a9096c6fa505dcd27" Jan 27 17:50:55 crc kubenswrapper[4772]: I0127 17:50:55.353801 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7dvs7/crc-debug-fch7g" Jan 27 17:50:55 crc kubenswrapper[4772]: I0127 17:50:55.728787 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7dvs7/crc-debug-5r4qc"] Jan 27 17:50:55 crc kubenswrapper[4772]: E0127 17:50:55.729603 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b7e9e60-6e19-49be-866a-fba6fd4b0780" containerName="container-00" Jan 27 17:50:55 crc kubenswrapper[4772]: I0127 17:50:55.729625 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b7e9e60-6e19-49be-866a-fba6fd4b0780" containerName="container-00" Jan 27 17:50:55 crc kubenswrapper[4772]: I0127 17:50:55.729867 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b7e9e60-6e19-49be-866a-fba6fd4b0780" containerName="container-00" Jan 27 17:50:55 crc kubenswrapper[4772]: I0127 17:50:55.730662 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7dvs7/crc-debug-5r4qc" Jan 27 17:50:55 crc kubenswrapper[4772]: I0127 17:50:55.873025 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jf8sr\" (UniqueName: \"kubernetes.io/projected/a2ac0d9e-ba08-48c0-9a6c-307c3438b86e-kube-api-access-jf8sr\") pod \"crc-debug-5r4qc\" (UID: \"a2ac0d9e-ba08-48c0-9a6c-307c3438b86e\") " pod="openshift-must-gather-7dvs7/crc-debug-5r4qc" Jan 27 17:50:55 crc kubenswrapper[4772]: I0127 17:50:55.873222 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a2ac0d9e-ba08-48c0-9a6c-307c3438b86e-host\") pod \"crc-debug-5r4qc\" (UID: \"a2ac0d9e-ba08-48c0-9a6c-307c3438b86e\") " pod="openshift-must-gather-7dvs7/crc-debug-5r4qc" Jan 27 17:50:55 crc kubenswrapper[4772]: I0127 17:50:55.974801 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jf8sr\" (UniqueName: \"kubernetes.io/projected/a2ac0d9e-ba08-48c0-9a6c-307c3438b86e-kube-api-access-jf8sr\") pod \"crc-debug-5r4qc\" (UID: \"a2ac0d9e-ba08-48c0-9a6c-307c3438b86e\") " pod="openshift-must-gather-7dvs7/crc-debug-5r4qc" Jan 27 17:50:55 crc kubenswrapper[4772]: I0127 17:50:55.974874 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a2ac0d9e-ba08-48c0-9a6c-307c3438b86e-host\") pod \"crc-debug-5r4qc\" (UID: \"a2ac0d9e-ba08-48c0-9a6c-307c3438b86e\") " pod="openshift-must-gather-7dvs7/crc-debug-5r4qc" Jan 27 17:50:55 crc kubenswrapper[4772]: I0127 17:50:55.975041 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a2ac0d9e-ba08-48c0-9a6c-307c3438b86e-host\") pod \"crc-debug-5r4qc\" (UID: \"a2ac0d9e-ba08-48c0-9a6c-307c3438b86e\") " pod="openshift-must-gather-7dvs7/crc-debug-5r4qc" Jan 27 17:50:55 crc kubenswrapper[4772]: I0127 17:50:55.996813 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jf8sr\" (UniqueName: \"kubernetes.io/projected/a2ac0d9e-ba08-48c0-9a6c-307c3438b86e-kube-api-access-jf8sr\") pod \"crc-debug-5r4qc\" (UID: \"a2ac0d9e-ba08-48c0-9a6c-307c3438b86e\") " pod="openshift-must-gather-7dvs7/crc-debug-5r4qc" Jan 27 17:50:56 crc kubenswrapper[4772]: I0127 17:50:56.051647 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7dvs7/crc-debug-5r4qc" Jan 27 17:50:56 crc kubenswrapper[4772]: W0127 17:50:56.102328 4772 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda2ac0d9e_ba08_48c0_9a6c_307c3438b86e.slice/crio-2567decd5ebaaa4f4f481700f74363f342d3f4cbc7cd8ddc7aefddfba8d1986a WatchSource:0}: Error finding container 2567decd5ebaaa4f4f481700f74363f342d3f4cbc7cd8ddc7aefddfba8d1986a: Status 404 returned error can't find the container with id 2567decd5ebaaa4f4f481700f74363f342d3f4cbc7cd8ddc7aefddfba8d1986a Jan 27 17:50:56 crc kubenswrapper[4772]: I0127 17:50:56.163290 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wwjcs" podUID="4bec6411-aae2-47f7-a721-a39d6d381c17" containerName="registry-server" probeResult="failure" output=< Jan 27 17:50:56 crc kubenswrapper[4772]: timeout: failed to connect service ":50051" within 1s Jan 27 17:50:56 crc kubenswrapper[4772]: > Jan 27 17:50:56 crc kubenswrapper[4772]: I0127 17:50:56.369108 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7dvs7/crc-debug-5r4qc" event={"ID":"a2ac0d9e-ba08-48c0-9a6c-307c3438b86e","Type":"ContainerStarted","Data":"1323a4c6a49cb722cb00e0194b58825dbab94b8aff7bf910a6baddc9ff7aedd8"} Jan 27 17:50:56 crc kubenswrapper[4772]: I0127 17:50:56.369146 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7dvs7/crc-debug-5r4qc" event={"ID":"a2ac0d9e-ba08-48c0-9a6c-307c3438b86e","Type":"ContainerStarted","Data":"2567decd5ebaaa4f4f481700f74363f342d3f4cbc7cd8ddc7aefddfba8d1986a"} Jan 27 17:50:56 crc kubenswrapper[4772]: I0127 17:50:56.390494 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-7dvs7/crc-debug-5r4qc" podStartSLOduration=1.390460085 podStartE2EDuration="1.390460085s" podCreationTimestamp="2026-01-27 17:50:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-27 17:50:56.381288233 +0000 UTC m=+9842.361897361" watchObservedRunningTime="2026-01-27 17:50:56.390460085 +0000 UTC m=+9842.371069183" Jan 27 17:50:56 crc kubenswrapper[4772]: E0127 17:50:56.587818 4772 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda2ac0d9e_ba08_48c0_9a6c_307c3438b86e.slice/crio-conmon-1323a4c6a49cb722cb00e0194b58825dbab94b8aff7bf910a6baddc9ff7aedd8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda2ac0d9e_ba08_48c0_9a6c_307c3438b86e.slice/crio-1323a4c6a49cb722cb00e0194b58825dbab94b8aff7bf910a6baddc9ff7aedd8.scope\": RecentStats: unable to find data in memory cache]" Jan 27 17:50:57 crc kubenswrapper[4772]: I0127 17:50:57.395588 4772 generic.go:334] "Generic (PLEG): container finished" podID="a2ac0d9e-ba08-48c0-9a6c-307c3438b86e" containerID="1323a4c6a49cb722cb00e0194b58825dbab94b8aff7bf910a6baddc9ff7aedd8" exitCode=1 Jan 27 17:50:57 crc kubenswrapper[4772]: I0127 17:50:57.395840 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7dvs7/crc-debug-5r4qc" event={"ID":"a2ac0d9e-ba08-48c0-9a6c-307c3438b86e","Type":"ContainerDied","Data":"1323a4c6a49cb722cb00e0194b58825dbab94b8aff7bf910a6baddc9ff7aedd8"} Jan 27 17:50:58 crc kubenswrapper[4772]: I0127 17:50:58.511537 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7dvs7/crc-debug-5r4qc" Jan 27 17:50:58 crc kubenswrapper[4772]: I0127 17:50:58.545513 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7dvs7/crc-debug-5r4qc"] Jan 27 17:50:58 crc kubenswrapper[4772]: I0127 17:50:58.554458 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7dvs7/crc-debug-5r4qc"] Jan 27 17:50:58 crc kubenswrapper[4772]: I0127 17:50:58.621637 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a2ac0d9e-ba08-48c0-9a6c-307c3438b86e-host\") pod \"a2ac0d9e-ba08-48c0-9a6c-307c3438b86e\" (UID: \"a2ac0d9e-ba08-48c0-9a6c-307c3438b86e\") " Jan 27 17:50:58 crc kubenswrapper[4772]: I0127 17:50:58.621744 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a2ac0d9e-ba08-48c0-9a6c-307c3438b86e-host" (OuterVolumeSpecName: "host") pod "a2ac0d9e-ba08-48c0-9a6c-307c3438b86e" (UID: "a2ac0d9e-ba08-48c0-9a6c-307c3438b86e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 27 17:50:58 crc kubenswrapper[4772]: I0127 17:50:58.621793 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jf8sr\" (UniqueName: \"kubernetes.io/projected/a2ac0d9e-ba08-48c0-9a6c-307c3438b86e-kube-api-access-jf8sr\") pod \"a2ac0d9e-ba08-48c0-9a6c-307c3438b86e\" (UID: \"a2ac0d9e-ba08-48c0-9a6c-307c3438b86e\") " Jan 27 17:50:58 crc kubenswrapper[4772]: I0127 17:50:58.622364 4772 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a2ac0d9e-ba08-48c0-9a6c-307c3438b86e-host\") on node \"crc\" DevicePath \"\"" Jan 27 17:50:58 crc kubenswrapper[4772]: I0127 17:50:58.629430 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2ac0d9e-ba08-48c0-9a6c-307c3438b86e-kube-api-access-jf8sr" (OuterVolumeSpecName: "kube-api-access-jf8sr") pod "a2ac0d9e-ba08-48c0-9a6c-307c3438b86e" (UID: "a2ac0d9e-ba08-48c0-9a6c-307c3438b86e"). InnerVolumeSpecName "kube-api-access-jf8sr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:50:58 crc kubenswrapper[4772]: I0127 17:50:58.679055 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2ac0d9e-ba08-48c0-9a6c-307c3438b86e" path="/var/lib/kubelet/pods/a2ac0d9e-ba08-48c0-9a6c-307c3438b86e/volumes" Jan 27 17:50:58 crc kubenswrapper[4772]: I0127 17:50:58.723990 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jf8sr\" (UniqueName: \"kubernetes.io/projected/a2ac0d9e-ba08-48c0-9a6c-307c3438b86e-kube-api-access-jf8sr\") on node \"crc\" DevicePath \"\"" Jan 27 17:50:59 crc kubenswrapper[4772]: I0127 17:50:59.415092 4772 scope.go:117] "RemoveContainer" containerID="1323a4c6a49cb722cb00e0194b58825dbab94b8aff7bf910a6baddc9ff7aedd8" Jan 27 17:50:59 crc kubenswrapper[4772]: I0127 17:50:59.415204 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7dvs7/crc-debug-5r4qc" Jan 27 17:51:06 crc kubenswrapper[4772]: I0127 17:51:06.153611 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wwjcs" podUID="4bec6411-aae2-47f7-a721-a39d6d381c17" containerName="registry-server" probeResult="failure" output=< Jan 27 17:51:06 crc kubenswrapper[4772]: timeout: failed to connect service ":50051" within 1s Jan 27 17:51:06 crc kubenswrapper[4772]: > Jan 27 17:51:06 crc kubenswrapper[4772]: I0127 17:51:06.664926 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:51:06 crc kubenswrapper[4772]: E0127 17:51:06.665528 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:51:16 crc kubenswrapper[4772]: I0127 17:51:16.330393 4772 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wwjcs" podUID="4bec6411-aae2-47f7-a721-a39d6d381c17" containerName="registry-server" probeResult="failure" output=< Jan 27 17:51:16 crc kubenswrapper[4772]: timeout: failed to connect service ":50051" within 1s Jan 27 17:51:16 crc kubenswrapper[4772]: > Jan 27 17:51:17 crc kubenswrapper[4772]: I0127 17:51:17.663253 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:51:17 crc kubenswrapper[4772]: E0127 17:51:17.663771 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:51:25 crc kubenswrapper[4772]: I0127 17:51:25.446152 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wwjcs" Jan 27 17:51:25 crc kubenswrapper[4772]: I0127 17:51:25.500489 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wwjcs" Jan 27 17:51:25 crc kubenswrapper[4772]: I0127 17:51:25.688820 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wwjcs"] Jan 27 17:51:26 crc kubenswrapper[4772]: I0127 17:51:26.659578 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wwjcs" podUID="4bec6411-aae2-47f7-a721-a39d6d381c17" containerName="registry-server" containerID="cri-o://fc9c3049ded284cad5528fcbc82380287f3e9e60849f185709b1cbf6091758dd" gracePeriod=2 Jan 27 17:51:27 crc kubenswrapper[4772]: I0127 17:51:27.378687 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wwjcs" Jan 27 17:51:27 crc kubenswrapper[4772]: I0127 17:51:27.511431 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bec6411-aae2-47f7-a721-a39d6d381c17-utilities\") pod \"4bec6411-aae2-47f7-a721-a39d6d381c17\" (UID: \"4bec6411-aae2-47f7-a721-a39d6d381c17\") " Jan 27 17:51:27 crc kubenswrapper[4772]: I0127 17:51:27.511867 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bec6411-aae2-47f7-a721-a39d6d381c17-catalog-content\") pod \"4bec6411-aae2-47f7-a721-a39d6d381c17\" (UID: \"4bec6411-aae2-47f7-a721-a39d6d381c17\") " Jan 27 17:51:27 crc kubenswrapper[4772]: I0127 17:51:27.511997 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9m8q5\" (UniqueName: \"kubernetes.io/projected/4bec6411-aae2-47f7-a721-a39d6d381c17-kube-api-access-9m8q5\") pod \"4bec6411-aae2-47f7-a721-a39d6d381c17\" (UID: \"4bec6411-aae2-47f7-a721-a39d6d381c17\") " Jan 27 17:51:27 crc kubenswrapper[4772]: I0127 17:51:27.512473 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bec6411-aae2-47f7-a721-a39d6d381c17-utilities" (OuterVolumeSpecName: "utilities") pod "4bec6411-aae2-47f7-a721-a39d6d381c17" (UID: "4bec6411-aae2-47f7-a721-a39d6d381c17"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:51:27 crc kubenswrapper[4772]: I0127 17:51:27.512658 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bec6411-aae2-47f7-a721-a39d6d381c17-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 17:51:27 crc kubenswrapper[4772]: I0127 17:51:27.529402 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bec6411-aae2-47f7-a721-a39d6d381c17-kube-api-access-9m8q5" (OuterVolumeSpecName: "kube-api-access-9m8q5") pod "4bec6411-aae2-47f7-a721-a39d6d381c17" (UID: "4bec6411-aae2-47f7-a721-a39d6d381c17"). InnerVolumeSpecName "kube-api-access-9m8q5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:51:27 crc kubenswrapper[4772]: I0127 17:51:27.614741 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9m8q5\" (UniqueName: \"kubernetes.io/projected/4bec6411-aae2-47f7-a721-a39d6d381c17-kube-api-access-9m8q5\") on node \"crc\" DevicePath \"\"" Jan 27 17:51:27 crc kubenswrapper[4772]: I0127 17:51:27.670106 4772 generic.go:334] "Generic (PLEG): container finished" podID="4bec6411-aae2-47f7-a721-a39d6d381c17" containerID="fc9c3049ded284cad5528fcbc82380287f3e9e60849f185709b1cbf6091758dd" exitCode=0 Jan 27 17:51:27 crc kubenswrapper[4772]: I0127 17:51:27.670148 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwjcs" event={"ID":"4bec6411-aae2-47f7-a721-a39d6d381c17","Type":"ContainerDied","Data":"fc9c3049ded284cad5528fcbc82380287f3e9e60849f185709b1cbf6091758dd"} Jan 27 17:51:27 crc kubenswrapper[4772]: I0127 17:51:27.670194 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwjcs" event={"ID":"4bec6411-aae2-47f7-a721-a39d6d381c17","Type":"ContainerDied","Data":"bc5839b7d468657c30d46f03fedecb8b4e0ef4ce5baaaea4f34f78a6a2e457ca"} Jan 27 17:51:27 crc kubenswrapper[4772]: I0127 17:51:27.670229 4772 scope.go:117] "RemoveContainer" containerID="fc9c3049ded284cad5528fcbc82380287f3e9e60849f185709b1cbf6091758dd" Jan 27 17:51:27 crc kubenswrapper[4772]: I0127 17:51:27.670292 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wwjcs" Jan 27 17:51:27 crc kubenswrapper[4772]: I0127 17:51:27.676015 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bec6411-aae2-47f7-a721-a39d6d381c17-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4bec6411-aae2-47f7-a721-a39d6d381c17" (UID: "4bec6411-aae2-47f7-a721-a39d6d381c17"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:51:27 crc kubenswrapper[4772]: I0127 17:51:27.691852 4772 scope.go:117] "RemoveContainer" containerID="407a6958420e1bfb3d0405a21d3152ae16fe037a897624d55dc2bc173169ab06" Jan 27 17:51:27 crc kubenswrapper[4772]: I0127 17:51:27.707279 4772 scope.go:117] "RemoveContainer" containerID="89b9352f77c65a63cab44c13397c73a1d6c97dafb21ef79db13cbb8873b80e0b" Jan 27 17:51:27 crc kubenswrapper[4772]: I0127 17:51:27.716206 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bec6411-aae2-47f7-a721-a39d6d381c17-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 17:51:27 crc kubenswrapper[4772]: I0127 17:51:27.758052 4772 scope.go:117] "RemoveContainer" containerID="fc9c3049ded284cad5528fcbc82380287f3e9e60849f185709b1cbf6091758dd" Jan 27 17:51:27 crc kubenswrapper[4772]: E0127 17:51:27.758503 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc9c3049ded284cad5528fcbc82380287f3e9e60849f185709b1cbf6091758dd\": container with ID starting with fc9c3049ded284cad5528fcbc82380287f3e9e60849f185709b1cbf6091758dd not found: ID does not exist" containerID="fc9c3049ded284cad5528fcbc82380287f3e9e60849f185709b1cbf6091758dd" Jan 27 17:51:27 crc kubenswrapper[4772]: I0127 17:51:27.758570 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc9c3049ded284cad5528fcbc82380287f3e9e60849f185709b1cbf6091758dd"} err="failed to get container status \"fc9c3049ded284cad5528fcbc82380287f3e9e60849f185709b1cbf6091758dd\": rpc error: code = NotFound desc = could not find container \"fc9c3049ded284cad5528fcbc82380287f3e9e60849f185709b1cbf6091758dd\": container with ID starting with fc9c3049ded284cad5528fcbc82380287f3e9e60849f185709b1cbf6091758dd not found: ID does not exist" Jan 27 17:51:27 crc kubenswrapper[4772]: I0127 17:51:27.758603 4772 scope.go:117] "RemoveContainer" containerID="407a6958420e1bfb3d0405a21d3152ae16fe037a897624d55dc2bc173169ab06" Jan 27 17:51:27 crc kubenswrapper[4772]: E0127 17:51:27.759022 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"407a6958420e1bfb3d0405a21d3152ae16fe037a897624d55dc2bc173169ab06\": container with ID starting with 407a6958420e1bfb3d0405a21d3152ae16fe037a897624d55dc2bc173169ab06 not found: ID does not exist" containerID="407a6958420e1bfb3d0405a21d3152ae16fe037a897624d55dc2bc173169ab06" Jan 27 17:51:27 crc kubenswrapper[4772]: I0127 17:51:27.759059 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"407a6958420e1bfb3d0405a21d3152ae16fe037a897624d55dc2bc173169ab06"} err="failed to get container status \"407a6958420e1bfb3d0405a21d3152ae16fe037a897624d55dc2bc173169ab06\": rpc error: code = NotFound desc = could not find container \"407a6958420e1bfb3d0405a21d3152ae16fe037a897624d55dc2bc173169ab06\": container with ID starting with 407a6958420e1bfb3d0405a21d3152ae16fe037a897624d55dc2bc173169ab06 not found: ID does not exist" Jan 27 17:51:27 crc kubenswrapper[4772]: I0127 17:51:27.759073 4772 scope.go:117] "RemoveContainer" containerID="89b9352f77c65a63cab44c13397c73a1d6c97dafb21ef79db13cbb8873b80e0b" Jan 27 17:51:27 crc kubenswrapper[4772]: E0127 17:51:27.759461 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89b9352f77c65a63cab44c13397c73a1d6c97dafb21ef79db13cbb8873b80e0b\": container with ID starting with 89b9352f77c65a63cab44c13397c73a1d6c97dafb21ef79db13cbb8873b80e0b not found: ID does not exist" containerID="89b9352f77c65a63cab44c13397c73a1d6c97dafb21ef79db13cbb8873b80e0b" Jan 27 17:51:27 crc kubenswrapper[4772]: I0127 17:51:27.759493 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89b9352f77c65a63cab44c13397c73a1d6c97dafb21ef79db13cbb8873b80e0b"} err="failed to get container status \"89b9352f77c65a63cab44c13397c73a1d6c97dafb21ef79db13cbb8873b80e0b\": rpc error: code = NotFound desc = could not find container \"89b9352f77c65a63cab44c13397c73a1d6c97dafb21ef79db13cbb8873b80e0b\": container with ID starting with 89b9352f77c65a63cab44c13397c73a1d6c97dafb21ef79db13cbb8873b80e0b not found: ID does not exist" Jan 27 17:51:28 crc kubenswrapper[4772]: I0127 17:51:28.028277 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wwjcs"] Jan 27 17:51:28 crc kubenswrapper[4772]: I0127 17:51:28.037858 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wwjcs"] Jan 27 17:51:28 crc kubenswrapper[4772]: I0127 17:51:28.673996 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bec6411-aae2-47f7-a721-a39d6d381c17" path="/var/lib/kubelet/pods/4bec6411-aae2-47f7-a721-a39d6d381c17/volumes" Jan 27 17:51:30 crc kubenswrapper[4772]: I0127 17:51:30.019975 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7dcb7f9846-lrk6t_5c77e7c3-5320-4fa6-810d-bc819a6f7b03/barbican-api/0.log" Jan 27 17:51:30 crc kubenswrapper[4772]: I0127 17:51:30.247516 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7dcb7f9846-lrk6t_5c77e7c3-5320-4fa6-810d-bc819a6f7b03/barbican-api-log/0.log" Jan 27 17:51:30 crc kubenswrapper[4772]: I0127 17:51:30.269935 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-75df4b6d74-xpp9t_2da34b58-6b43-4e25-bdec-39985c344819/barbican-keystone-listener/0.log" Jan 27 17:51:30 crc kubenswrapper[4772]: I0127 17:51:30.434893 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-75df4b6d74-xpp9t_2da34b58-6b43-4e25-bdec-39985c344819/barbican-keystone-listener-log/0.log" Jan 27 17:51:30 crc kubenswrapper[4772]: I0127 17:51:30.500045 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-69b64c5dd7-dj9pw_0d49f4dc-fd69-4e43-9866-87af6da31197/barbican-worker/0.log" Jan 27 17:51:30 crc kubenswrapper[4772]: I0127 17:51:30.578922 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-69b64c5dd7-dj9pw_0d49f4dc-fd69-4e43-9866-87af6da31197/barbican-worker-log/0.log" Jan 27 17:51:30 crc kubenswrapper[4772]: I0127 17:51:30.690350 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_a46febaf-97b6-4ed3-8958-316e2a542a5f/cinder-api-log/0.log" Jan 27 17:51:30 crc kubenswrapper[4772]: I0127 17:51:30.730624 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_a46febaf-97b6-4ed3-8958-316e2a542a5f/cinder-api/0.log" Jan 27 17:51:30 crc kubenswrapper[4772]: I0127 17:51:30.984694 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_37170132-cd9f-44e7-827d-b98486cefb39/probe/0.log" Jan 27 17:51:31 crc kubenswrapper[4772]: I0127 17:51:31.025239 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_37170132-cd9f-44e7-827d-b98486cefb39/cinder-backup/0.log" Jan 27 17:51:31 crc kubenswrapper[4772]: I0127 17:51:31.075894 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_66512484-80ba-4887-b9a9-9cc87a65ad18/cinder-scheduler/0.log" Jan 27 17:51:31 crc kubenswrapper[4772]: I0127 17:51:31.233525 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_66512484-80ba-4887-b9a9-9cc87a65ad18/probe/0.log" Jan 27 17:51:31 crc kubenswrapper[4772]: I0127 17:51:31.284711 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_7b6810b2-bc50-486d-9a87-cf4cd50d33c5/cinder-volume/0.log" Jan 27 17:51:31 crc kubenswrapper[4772]: I0127 17:51:31.327316 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_7b6810b2-bc50-486d-9a87-cf4cd50d33c5/probe/0.log" Jan 27 17:51:31 crc kubenswrapper[4772]: I0127 17:51:31.477111 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8896c5c8c-s6z7x_22e08251-8371-4470-bc3e-d88d673d56f3/init/0.log" Jan 27 17:51:31 crc kubenswrapper[4772]: I0127 17:51:31.644553 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8896c5c8c-s6z7x_22e08251-8371-4470-bc3e-d88d673d56f3/init/0.log" Jan 27 17:51:31 crc kubenswrapper[4772]: I0127 17:51:31.663571 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:51:31 crc kubenswrapper[4772]: E0127 17:51:31.663865 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:51:31 crc kubenswrapper[4772]: I0127 17:51:31.675119 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8896c5c8c-s6z7x_22e08251-8371-4470-bc3e-d88d673d56f3/dnsmasq-dns/0.log" Jan 27 17:51:31 crc kubenswrapper[4772]: I0127 17:51:31.732574 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_a7917063-9e04-41e8-8fb9-e8383f839bd6/glance-httpd/0.log" Jan 27 17:51:31 crc kubenswrapper[4772]: I0127 17:51:31.839386 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_a7917063-9e04-41e8-8fb9-e8383f839bd6/glance-log/0.log" Jan 27 17:51:31 crc kubenswrapper[4772]: I0127 17:51:31.937360 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a1535f57-0540-45ea-b53c-1b4cac461cf3/glance-log/0.log" Jan 27 17:51:31 crc kubenswrapper[4772]: I0127 17:51:31.965273 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a1535f57-0540-45ea-b53c-1b4cac461cf3/glance-httpd/0.log" Jan 27 17:51:32 crc kubenswrapper[4772]: I0127 17:51:32.191772 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29492221-sbxjp_8459d055-62d3-4699-b477-ea15946b982c/keystone-cron/0.log" Jan 27 17:51:32 crc kubenswrapper[4772]: I0127 17:51:32.191947 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-56b5f9d6fc-hmz72_f6255916-357e-42c6-b936-27151f6b2260/keystone-api/0.log" Jan 27 17:51:32 crc kubenswrapper[4772]: I0127 17:51:32.473386 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-copy-data_7db35434-01e2-470d-bb27-8e30189936b3/adoption/0.log" Jan 27 17:51:32 crc kubenswrapper[4772]: I0127 17:51:32.820734 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-8bf5d4b7c-bfg78_b5a89957-107d-449b-b438-2215fd4ed522/neutron-api/0.log" Jan 27 17:51:32 crc kubenswrapper[4772]: I0127 17:51:32.883277 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-8bf5d4b7c-bfg78_b5a89957-107d-449b-b438-2215fd4ed522/neutron-httpd/0.log" Jan 27 17:51:33 crc kubenswrapper[4772]: I0127 17:51:33.177588 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_f178d7e3-af69-4014-8209-5e766a130997/nova-api-api/0.log" Jan 27 17:51:33 crc kubenswrapper[4772]: I0127 17:51:33.266457 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_f178d7e3-af69-4014-8209-5e766a130997/nova-api-log/0.log" Jan 27 17:51:33 crc kubenswrapper[4772]: I0127 17:51:33.489959 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_3331c1dd-ff2d-4a41-9cb3-731297ae0dc3/nova-cell0-conductor-conductor/0.log" Jan 27 17:51:33 crc kubenswrapper[4772]: I0127 17:51:33.617718 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_9549c89c-f55f-484d-80b2-ca1ad19bf758/nova-cell1-conductor-conductor/0.log" Jan 27 17:51:33 crc kubenswrapper[4772]: I0127 17:51:33.853686 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_db0a7ce2-c175-4632-abe5-f35a6b5ce680/nova-cell1-novncproxy-novncproxy/0.log" Jan 27 17:51:34 crc kubenswrapper[4772]: I0127 17:51:34.043259 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_3a983cf0-2c51-4d6a-af53-f115f3a57360/nova-metadata-log/0.log" Jan 27 17:51:34 crc kubenswrapper[4772]: I0127 17:51:34.351077 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_3a983cf0-2c51-4d6a-af53-f115f3a57360/nova-metadata-metadata/0.log" Jan 27 17:51:34 crc kubenswrapper[4772]: I0127 17:51:34.418796 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_c9832ca2-4d35-4533-bdb3-7ac3773e5242/nova-scheduler-scheduler/0.log" Jan 27 17:51:34 crc kubenswrapper[4772]: I0127 17:51:34.550125 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-6d96bf4746-x9c97_9371d269-02b3-4049-aeea-4fd56c648b89/init/0.log" Jan 27 17:51:34 crc kubenswrapper[4772]: I0127 17:51:34.753899 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-6d96bf4746-x9c97_9371d269-02b3-4049-aeea-4fd56c648b89/octavia-api-provider-agent/0.log" Jan 27 17:51:34 crc kubenswrapper[4772]: I0127 17:51:34.789663 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-6d96bf4746-x9c97_9371d269-02b3-4049-aeea-4fd56c648b89/init/0.log" Jan 27 17:51:34 crc kubenswrapper[4772]: I0127 17:51:34.996847 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-6d96bf4746-x9c97_9371d269-02b3-4049-aeea-4fd56c648b89/octavia-api/0.log" Jan 27 17:51:35 crc kubenswrapper[4772]: I0127 17:51:35.365422 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-jmckq_d32d3e24-6f03-46cc-b7ae-61383778b183/init/0.log" Jan 27 17:51:35 crc kubenswrapper[4772]: I0127 17:51:35.581917 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-jmckq_d32d3e24-6f03-46cc-b7ae-61383778b183/init/0.log" Jan 27 17:51:35 crc kubenswrapper[4772]: I0127 17:51:35.636386 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-6v7mr_b6f78da3-da1c-4e27-ab65-581c656f74d9/init/0.log" Jan 27 17:51:35 crc kubenswrapper[4772]: I0127 17:51:35.673837 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-jmckq_d32d3e24-6f03-46cc-b7ae-61383778b183/octavia-healthmanager/0.log" Jan 27 17:51:35 crc kubenswrapper[4772]: I0127 17:51:35.793133 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-6v7mr_b6f78da3-da1c-4e27-ab65-581c656f74d9/init/0.log" Jan 27 17:51:35 crc kubenswrapper[4772]: I0127 17:51:35.834728 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-6v7mr_b6f78da3-da1c-4e27-ab65-581c656f74d9/octavia-housekeeping/0.log" Jan 27 17:51:35 crc kubenswrapper[4772]: I0127 17:51:35.901429 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-4wcv5_c163e8de-ea19-4a1c-8791-8659b9a09ba3/init/0.log" Jan 27 17:51:36 crc kubenswrapper[4772]: I0127 17:51:36.072376 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-4wcv5_c163e8de-ea19-4a1c-8791-8659b9a09ba3/init/0.log" Jan 27 17:51:36 crc kubenswrapper[4772]: I0127 17:51:36.080012 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-4wcv5_c163e8de-ea19-4a1c-8791-8659b9a09ba3/octavia-rsyslog/0.log" Jan 27 17:51:36 crc kubenswrapper[4772]: I0127 17:51:36.163863 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-4q859_ab431622-b724-4ed4-be2b-67ec8b5956db/init/0.log" Jan 27 17:51:36 crc kubenswrapper[4772]: I0127 17:51:36.383331 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-4q859_ab431622-b724-4ed4-be2b-67ec8b5956db/init/0.log" Jan 27 17:51:36 crc kubenswrapper[4772]: I0127 17:51:36.545426 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_be7c27a4-64d2-4581-8271-5aaf74103b04/mysql-bootstrap/0.log" Jan 27 17:51:36 crc kubenswrapper[4772]: I0127 17:51:36.564238 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-4q859_ab431622-b724-4ed4-be2b-67ec8b5956db/octavia-worker/0.log" Jan 27 17:51:36 crc kubenswrapper[4772]: I0127 17:51:36.951021 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_be7c27a4-64d2-4581-8271-5aaf74103b04/mysql-bootstrap/0.log" Jan 27 17:51:37 crc kubenswrapper[4772]: I0127 17:51:37.014048 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_eefd7ff4-5222-45cf-aaad-20ebfd50a2ff/mysql-bootstrap/0.log" Jan 27 17:51:37 crc kubenswrapper[4772]: I0127 17:51:37.022348 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_be7c27a4-64d2-4581-8271-5aaf74103b04/galera/0.log" Jan 27 17:51:37 crc kubenswrapper[4772]: I0127 17:51:37.245394 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_eefd7ff4-5222-45cf-aaad-20ebfd50a2ff/mysql-bootstrap/0.log" Jan 27 17:51:37 crc kubenswrapper[4772]: I0127 17:51:37.289842 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_e05a90a8-dbbb-4e24-ac89-f30360482af9/openstackclient/0.log" Jan 27 17:51:37 crc kubenswrapper[4772]: I0127 17:51:37.315345 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_eefd7ff4-5222-45cf-aaad-20ebfd50a2ff/galera/0.log" Jan 27 17:51:37 crc kubenswrapper[4772]: I0127 17:51:37.392186 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_36e53353-e817-4d3d-878e-2b34f7c9192f/memcached/0.log" Jan 27 17:51:37 crc kubenswrapper[4772]: I0127 17:51:37.509676 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-jv694_febb140e-d26e-43db-9924-0f06739b9a4a/ovn-controller/0.log" Jan 27 17:51:37 crc kubenswrapper[4772]: I0127 17:51:37.545257 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-sx2ff_3d12bbd4-3d0b-444b-a462-b620a7a5d73d/openstack-network-exporter/0.log" Jan 27 17:51:37 crc kubenswrapper[4772]: I0127 17:51:37.639123 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-s5xbr_7e5eabb2-229a-4d75-b62d-65be688f753a/ovsdb-server-init/0.log" Jan 27 17:51:37 crc kubenswrapper[4772]: I0127 17:51:37.873599 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-s5xbr_7e5eabb2-229a-4d75-b62d-65be688f753a/ovsdb-server/0.log" Jan 27 17:51:37 crc kubenswrapper[4772]: I0127 17:51:37.874062 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-s5xbr_7e5eabb2-229a-4d75-b62d-65be688f753a/ovsdb-server-init/0.log" Jan 27 17:51:37 crc kubenswrapper[4772]: I0127 17:51:37.878195 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-s5xbr_7e5eabb2-229a-4d75-b62d-65be688f753a/ovs-vswitchd/0.log" Jan 27 17:51:37 crc kubenswrapper[4772]: I0127 17:51:37.909449 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-copy-data_6d673b09-a15f-48fc-b399-212dc30fce29/adoption/0.log" Jan 27 17:51:38 crc kubenswrapper[4772]: I0127 17:51:38.059369 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_9c221996-e15f-4fe3-bc62-98aac08f546f/ovn-northd/0.log" Jan 27 17:51:38 crc kubenswrapper[4772]: I0127 17:51:38.088362 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_9c221996-e15f-4fe3-bc62-98aac08f546f/openstack-network-exporter/0.log" Jan 27 17:51:38 crc kubenswrapper[4772]: I0127 17:51:38.179836 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c7dba285-1db4-44d8-bdf4-9de6e8d80adb/openstack-network-exporter/0.log" Jan 27 17:51:38 crc kubenswrapper[4772]: I0127 17:51:38.254794 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c7dba285-1db4-44d8-bdf4-9de6e8d80adb/ovsdbserver-nb/0.log" Jan 27 17:51:38 crc kubenswrapper[4772]: I0127 17:51:38.503561 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_d647bfb5-69e6-4b10-96ac-5f7fcd72514f/ovsdbserver-nb/0.log" Jan 27 17:51:38 crc kubenswrapper[4772]: I0127 17:51:38.511570 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_d647bfb5-69e6-4b10-96ac-5f7fcd72514f/openstack-network-exporter/0.log" Jan 27 17:51:38 crc kubenswrapper[4772]: I0127 17:51:38.635581 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_3e19e84f-6d5e-455b-be78-ae3f04c925b7/openstack-network-exporter/0.log" Jan 27 17:51:38 crc kubenswrapper[4772]: I0127 17:51:38.655979 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_3e19e84f-6d5e-455b-be78-ae3f04c925b7/ovsdbserver-nb/0.log" Jan 27 17:51:38 crc kubenswrapper[4772]: I0127 17:51:38.730554 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f70d2878-d629-4772-b2a4-697fe18a3760/openstack-network-exporter/0.log" Jan 27 17:51:38 crc kubenswrapper[4772]: I0127 17:51:38.809872 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f70d2878-d629-4772-b2a4-697fe18a3760/ovsdbserver-sb/0.log" Jan 27 17:51:38 crc kubenswrapper[4772]: I0127 17:51:38.880676 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_e6e7e5c6-90b8-4de9-ae6a-11034616734a/openstack-network-exporter/0.log" Jan 27 17:51:38 crc kubenswrapper[4772]: I0127 17:51:38.919012 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_e6e7e5c6-90b8-4de9-ae6a-11034616734a/ovsdbserver-sb/0.log" Jan 27 17:51:39 crc kubenswrapper[4772]: I0127 17:51:39.003367 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2/openstack-network-exporter/0.log" Jan 27 17:51:39 crc kubenswrapper[4772]: I0127 17:51:39.051794 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_e1fdf1bd-2cdb-4164-84c4-5c780a0a95b2/ovsdbserver-sb/0.log" Jan 27 17:51:39 crc kubenswrapper[4772]: I0127 17:51:39.128179 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6ffd9fc5c6-99g52_c8daf690-d375-4d0a-b763-4b610aaeac45/placement-api/0.log" Jan 27 17:51:39 crc kubenswrapper[4772]: I0127 17:51:39.205499 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6ffd9fc5c6-99g52_c8daf690-d375-4d0a-b763-4b610aaeac45/placement-log/0.log" Jan 27 17:51:39 crc kubenswrapper[4772]: I0127 17:51:39.244682 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88/setup-container/0.log" Jan 27 17:51:39 crc kubenswrapper[4772]: I0127 17:51:39.416054 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88/setup-container/0.log" Jan 27 17:51:39 crc kubenswrapper[4772]: I0127 17:51:39.442891 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f3b5f224-602e-454a-b35e-2e55160016b5/setup-container/0.log" Jan 27 17:51:39 crc kubenswrapper[4772]: I0127 17:51:39.453571 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1c2904f0-5ba8-4bb4-9952-ca1ee06a4d88/rabbitmq/0.log" Jan 27 17:51:39 crc kubenswrapper[4772]: I0127 17:51:39.637466 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f3b5f224-602e-454a-b35e-2e55160016b5/setup-container/0.log" Jan 27 17:51:39 crc kubenswrapper[4772]: I0127 17:51:39.667352 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_f3b5f224-602e-454a-b35e-2e55160016b5/rabbitmq/0.log" Jan 27 17:51:43 crc kubenswrapper[4772]: I0127 17:51:43.663118 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:51:43 crc kubenswrapper[4772]: E0127 17:51:43.663929 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:51:54 crc kubenswrapper[4772]: I0127 17:51:54.673930 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:51:54 crc kubenswrapper[4772]: E0127 17:51:54.676269 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:51:58 crc kubenswrapper[4772]: I0127 17:51:58.473282 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4_0f29ea34-f593-4806-b5f6-2f9976c46a12/util/0.log" Jan 27 17:51:58 crc kubenswrapper[4772]: I0127 17:51:58.712267 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4_0f29ea34-f593-4806-b5f6-2f9976c46a12/pull/0.log" Jan 27 17:51:58 crc kubenswrapper[4772]: I0127 17:51:58.713313 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4_0f29ea34-f593-4806-b5f6-2f9976c46a12/pull/0.log" Jan 27 17:51:58 crc kubenswrapper[4772]: I0127 17:51:58.739136 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4_0f29ea34-f593-4806-b5f6-2f9976c46a12/util/0.log" Jan 27 17:51:58 crc kubenswrapper[4772]: I0127 17:51:58.863525 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4_0f29ea34-f593-4806-b5f6-2f9976c46a12/util/0.log" Jan 27 17:51:58 crc kubenswrapper[4772]: I0127 17:51:58.863709 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4_0f29ea34-f593-4806-b5f6-2f9976c46a12/pull/0.log" Jan 27 17:51:58 crc kubenswrapper[4772]: I0127 17:51:58.897096 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6ebbdeb42ee59bc46cd5a9affeefe7a428e186e004b54bc44478e0857b2c5g4_0f29ea34-f593-4806-b5f6-2f9976c46a12/extract/0.log" Jan 27 17:51:59 crc kubenswrapper[4772]: I0127 17:51:59.154481 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-65ff799cfd-t42n9_674f4da6-f50d-4bab-808d-56ab3b9e2cb4/manager/0.log" Jan 27 17:51:59 crc kubenswrapper[4772]: I0127 17:51:59.234023 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-655bf9cfbb-cgh7j_fde95124-892b-411a-ba05-fa70927c8838/manager/0.log" Jan 27 17:51:59 crc kubenswrapper[4772]: I0127 17:51:59.314702 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-77554cdc5c-tkr6j_d395f105-54f0-4497-a119-57802be313a3/manager/0.log" Jan 27 17:51:59 crc kubenswrapper[4772]: I0127 17:51:59.458607 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-67dd55ff59-hgscb_fb300814-fca7-4419-ac6e-c08b33edd4be/manager/0.log" Jan 27 17:51:59 crc kubenswrapper[4772]: I0127 17:51:59.499654 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-575ffb885b-mtd9d_4c63a702-50b9-42f3-858e-7e27da0a8d8f/manager/0.log" Jan 27 17:51:59 crc kubenswrapper[4772]: I0127 17:51:59.665091 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-77d5c5b54f-jcb4p_e85aef3a-e235-473c-94cc-1f6237798b3e/manager/0.log" Jan 27 17:51:59 crc kubenswrapper[4772]: I0127 17:51:59.881747 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-768b776ffb-sxbjn_2df85221-33ed-49be-949c-516810279e4d/manager/0.log" Jan 27 17:52:00 crc kubenswrapper[4772]: I0127 17:52:00.249381 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-849fcfbb6b-tvrx9_27ec5082-c170-465b-b3a3-1f27a545fd71/manager/0.log" Jan 27 17:52:00 crc kubenswrapper[4772]: I0127 17:52:00.304321 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55f684fd56-wzjrz_783d8159-e67a-4796-83d8-4eff27d79505/manager/0.log" Jan 27 17:52:00 crc kubenswrapper[4772]: I0127 17:52:00.326384 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7d75bc88d5-t54fr_e7465bd0-3b6e-4199-9ee6-28b512198847/manager/0.log" Jan 27 17:52:00 crc kubenswrapper[4772]: I0127 17:52:00.489330 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6b9fb5fdcb-zhd82_b73c175a-e89e-434f-996a-65c1140bb8dd/manager/0.log" Jan 27 17:52:00 crc kubenswrapper[4772]: I0127 17:52:00.601540 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7ffd8d76d4-gcpj4_0a88aa66-b634-44ee-8e5b-bfeacb765e57/manager/0.log" Jan 27 17:52:00 crc kubenswrapper[4772]: I0127 17:52:00.884643 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7875d7675-ktfbt_80584c24-3c75-4624-802f-e608f640eeaa/manager/0.log" Jan 27 17:52:00 crc kubenswrapper[4772]: I0127 17:52:00.918141 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-ddcbfd695-6wltn_e7fc5297-101a-496e-a7c6-e7296e08a5af/manager/0.log" Jan 27 17:52:01 crc kubenswrapper[4772]: I0127 17:52:01.063921 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6b68b8b854q994c_1389813b-42ea-433f-820c-e5b8b41713d7/manager/0.log" Jan 27 17:52:01 crc kubenswrapper[4772]: I0127 17:52:01.169753 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-6fb647f7d4-gkjgt_939a692e-65d1-4be8-b78a-22ae83072d51/operator/0.log" Jan 27 17:52:01 crc kubenswrapper[4772]: I0127 17:52:01.408415 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-vs9rk_96fcf3a5-2584-4590-8057-9c18a9866bd4/registry-server/0.log" Jan 27 17:52:01 crc kubenswrapper[4772]: I0127 17:52:01.737930 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-6f75f45d54-ww79v_e76712a7-ebf6-4f04-a52c-c8d2bacb87f7/manager/0.log" Jan 27 17:52:01 crc kubenswrapper[4772]: I0127 17:52:01.792033 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-79d5ccc684-vwnwk_c3d2f06d-4dd3-49a8-a0a0-54a83cc3f4e8/manager/0.log" Jan 27 17:52:01 crc kubenswrapper[4772]: I0127 17:52:01.946846 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-h9297_abaf1142-1b7c-4987-8a9d-c91e6456c4a5/operator/0.log" Jan 27 17:52:02 crc kubenswrapper[4772]: I0127 17:52:02.188699 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-547cbdb99f-l8d48_c09741c3-6bae-487a-9b4c-7c9f01d8c5bf/manager/0.log" Jan 27 17:52:02 crc kubenswrapper[4772]: I0127 17:52:02.313771 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-799bc87c89-k2l8k_e4a99865-64a7-49e5-bdce-ff929105fc0d/manager/0.log" Jan 27 17:52:02 crc kubenswrapper[4772]: I0127 17:52:02.461743 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-69797bbcbd-ln7xf_6242683c-24ad-4e22-a7b3-8463e07388c2/manager/0.log" Jan 27 17:52:02 crc kubenswrapper[4772]: I0127 17:52:02.572217 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6c9bb4b66c-ws2mh_783285f4-2e9d-4af5-b017-32676e7d1b01/manager/0.log" Jan 27 17:52:02 crc kubenswrapper[4772]: I0127 17:52:02.758904 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-ff554fc88-clt4p_8087d8d3-d2f6-4bca-abec-f5b5335f26fa/manager/0.log" Jan 27 17:52:05 crc kubenswrapper[4772]: I0127 17:52:05.663877 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:52:05 crc kubenswrapper[4772]: E0127 17:52:05.665514 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:52:19 crc kubenswrapper[4772]: I0127 17:52:19.682272 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:52:19 crc kubenswrapper[4772]: E0127 17:52:19.683158 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:52:23 crc kubenswrapper[4772]: I0127 17:52:23.938504 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-5vmlj_db4a3858-5afa-44c8-a435-2010f7e7340d/control-plane-machine-set-operator/0.log" Jan 27 17:52:24 crc kubenswrapper[4772]: I0127 17:52:24.150791 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-mfh29_625f7e2d-0e3f-4c2c-8f49-b09fc3638536/kube-rbac-proxy/0.log" Jan 27 17:52:24 crc kubenswrapper[4772]: I0127 17:52:24.214265 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-mfh29_625f7e2d-0e3f-4c2c-8f49-b09fc3638536/machine-api-operator/0.log" Jan 27 17:52:34 crc kubenswrapper[4772]: I0127 17:52:34.663476 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:52:34 crc kubenswrapper[4772]: E0127 17:52:34.664534 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:52:38 crc kubenswrapper[4772]: I0127 17:52:38.670371 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-bmz8p_7b51171e-5d65-415e-8052-3cc8991f5de4/cert-manager-controller/0.log" Jan 27 17:52:38 crc kubenswrapper[4772]: I0127 17:52:38.741941 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-s7ksj_8d401dfc-33d3-416f-abba-cad4a1e173bd/cert-manager-cainjector/0.log" Jan 27 17:52:38 crc kubenswrapper[4772]: I0127 17:52:38.849453 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-4886f_ab67a3dd-5a79-400f-9b27-294ef256823d/cert-manager-webhook/0.log" Jan 27 17:52:48 crc kubenswrapper[4772]: I0127 17:52:48.663251 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:52:48 crc kubenswrapper[4772]: E0127 17:52:48.663878 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:52:52 crc kubenswrapper[4772]: I0127 17:52:52.727064 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-jlq99_3706a5f9-4370-4cca-abb9-b23e8b9c828f/nmstate-console-plugin/0.log" Jan 27 17:52:52 crc kubenswrapper[4772]: I0127 17:52:52.900436 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-mmdc5_d834ccf6-9b3a-4a3e-8980-7f0a102babd0/nmstate-handler/0.log" Jan 27 17:52:52 crc kubenswrapper[4772]: I0127 17:52:52.963964 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-g7d66_004d59b7-1d3b-41af-8c3d-c6562dd9716a/kube-rbac-proxy/0.log" Jan 27 17:52:52 crc kubenswrapper[4772]: I0127 17:52:52.975233 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-g7d66_004d59b7-1d3b-41af-8c3d-c6562dd9716a/nmstate-metrics/0.log" Jan 27 17:52:53 crc kubenswrapper[4772]: I0127 17:52:53.186195 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-gpfqc_cf21b49c-f01b-4c7c-bdb9-57e115b364d9/nmstate-webhook/0.log" Jan 27 17:52:53 crc kubenswrapper[4772]: I0127 17:52:53.208610 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-mdjph_2914eab0-19c8-464b-a774-d30a492f6763/nmstate-operator/0.log" Jan 27 17:52:59 crc kubenswrapper[4772]: I0127 17:52:59.663087 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:52:59 crc kubenswrapper[4772]: E0127 17:52:59.663831 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:53:13 crc kubenswrapper[4772]: I0127 17:53:13.663516 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:53:13 crc kubenswrapper[4772]: E0127 17:53:13.664465 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:53:24 crc kubenswrapper[4772]: I0127 17:53:24.032327 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-vdg69_d2282b46-452e-402e-b929-23875b572727/kube-rbac-proxy/0.log" Jan 27 17:53:24 crc kubenswrapper[4772]: I0127 17:53:24.211790 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jhpnb_5d41fcac-7044-4f36-b9f8-0b656bb3bcca/cp-frr-files/0.log" Jan 27 17:53:24 crc kubenswrapper[4772]: I0127 17:53:24.298642 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-vdg69_d2282b46-452e-402e-b929-23875b572727/controller/0.log" Jan 27 17:53:24 crc kubenswrapper[4772]: I0127 17:53:24.400020 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jhpnb_5d41fcac-7044-4f36-b9f8-0b656bb3bcca/cp-reloader/0.log" Jan 27 17:53:24 crc kubenswrapper[4772]: I0127 17:53:24.426023 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jhpnb_5d41fcac-7044-4f36-b9f8-0b656bb3bcca/cp-frr-files/0.log" Jan 27 17:53:24 crc kubenswrapper[4772]: I0127 17:53:24.475542 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jhpnb_5d41fcac-7044-4f36-b9f8-0b656bb3bcca/cp-metrics/0.log" Jan 27 17:53:24 crc kubenswrapper[4772]: I0127 17:53:24.485610 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jhpnb_5d41fcac-7044-4f36-b9f8-0b656bb3bcca/cp-reloader/0.log" Jan 27 17:53:24 crc kubenswrapper[4772]: I0127 17:53:24.623184 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jhpnb_5d41fcac-7044-4f36-b9f8-0b656bb3bcca/cp-frr-files/0.log" Jan 27 17:53:24 crc kubenswrapper[4772]: I0127 17:53:24.659031 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jhpnb_5d41fcac-7044-4f36-b9f8-0b656bb3bcca/cp-reloader/0.log" Jan 27 17:53:24 crc kubenswrapper[4772]: I0127 17:53:24.662828 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jhpnb_5d41fcac-7044-4f36-b9f8-0b656bb3bcca/cp-metrics/0.log" Jan 27 17:53:24 crc kubenswrapper[4772]: I0127 17:53:24.662838 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jhpnb_5d41fcac-7044-4f36-b9f8-0b656bb3bcca/cp-metrics/0.log" Jan 27 17:53:24 crc kubenswrapper[4772]: I0127 17:53:24.878454 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jhpnb_5d41fcac-7044-4f36-b9f8-0b656bb3bcca/cp-frr-files/0.log" Jan 27 17:53:24 crc kubenswrapper[4772]: I0127 17:53:24.889741 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jhpnb_5d41fcac-7044-4f36-b9f8-0b656bb3bcca/controller/0.log" Jan 27 17:53:24 crc kubenswrapper[4772]: I0127 17:53:24.915290 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jhpnb_5d41fcac-7044-4f36-b9f8-0b656bb3bcca/cp-metrics/0.log" Jan 27 17:53:24 crc kubenswrapper[4772]: I0127 17:53:24.921726 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jhpnb_5d41fcac-7044-4f36-b9f8-0b656bb3bcca/cp-reloader/0.log" Jan 27 17:53:25 crc kubenswrapper[4772]: I0127 17:53:25.064714 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jhpnb_5d41fcac-7044-4f36-b9f8-0b656bb3bcca/kube-rbac-proxy/0.log" Jan 27 17:53:25 crc kubenswrapper[4772]: I0127 17:53:25.065464 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jhpnb_5d41fcac-7044-4f36-b9f8-0b656bb3bcca/frr-metrics/0.log" Jan 27 17:53:25 crc kubenswrapper[4772]: I0127 17:53:25.087536 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jhpnb_5d41fcac-7044-4f36-b9f8-0b656bb3bcca/kube-rbac-proxy-frr/0.log" Jan 27 17:53:25 crc kubenswrapper[4772]: I0127 17:53:25.229869 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jhpnb_5d41fcac-7044-4f36-b9f8-0b656bb3bcca/reloader/0.log" Jan 27 17:53:25 crc kubenswrapper[4772]: I0127 17:53:25.318643 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-qpxrs_28ed9da3-cd29-4d80-9703-472bdbb3c64b/frr-k8s-webhook-server/0.log" Jan 27 17:53:25 crc kubenswrapper[4772]: I0127 17:53:25.453857 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5c6dd9c74b-84qz4_f72c611d-60d8-4649-a410-38434d01d8e2/manager/0.log" Jan 27 17:53:25 crc kubenswrapper[4772]: I0127 17:53:25.627745 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-66986f9f9f-bmvms_c5ee8d7f-0160-4526-8ae0-45a50a450725/webhook-server/0.log" Jan 27 17:53:25 crc kubenswrapper[4772]: I0127 17:53:25.719818 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-cl54q_24577bed-b34e-4419-9e9e-7068155ba0d1/kube-rbac-proxy/0.log" Jan 27 17:53:26 crc kubenswrapper[4772]: I0127 17:53:26.610611 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-cl54q_24577bed-b34e-4419-9e9e-7068155ba0d1/speaker/0.log" Jan 27 17:53:27 crc kubenswrapper[4772]: I0127 17:53:27.664094 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:53:27 crc kubenswrapper[4772]: E0127 17:53:27.664818 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:53:27 crc kubenswrapper[4772]: I0127 17:53:27.829407 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jhpnb_5d41fcac-7044-4f36-b9f8-0b656bb3bcca/frr/0.log" Jan 27 17:53:41 crc kubenswrapper[4772]: I0127 17:53:41.339769 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg_44b239be-466d-4995-9c33-38d68a00550d/util/0.log" Jan 27 17:53:41 crc kubenswrapper[4772]: I0127 17:53:41.485186 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg_44b239be-466d-4995-9c33-38d68a00550d/util/0.log" Jan 27 17:53:41 crc kubenswrapper[4772]: I0127 17:53:41.521899 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg_44b239be-466d-4995-9c33-38d68a00550d/pull/0.log" Jan 27 17:53:41 crc kubenswrapper[4772]: I0127 17:53:41.563690 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg_44b239be-466d-4995-9c33-38d68a00550d/pull/0.log" Jan 27 17:53:41 crc kubenswrapper[4772]: I0127 17:53:41.662957 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:53:41 crc kubenswrapper[4772]: E0127 17:53:41.663316 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:53:42 crc kubenswrapper[4772]: I0127 17:53:42.195078 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg_44b239be-466d-4995-9c33-38d68a00550d/util/0.log" Jan 27 17:53:42 crc kubenswrapper[4772]: I0127 17:53:42.215365 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg_44b239be-466d-4995-9c33-38d68a00550d/pull/0.log" Jan 27 17:53:42 crc kubenswrapper[4772]: I0127 17:53:42.242097 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a4lwpg_44b239be-466d-4995-9c33-38d68a00550d/extract/0.log" Jan 27 17:53:42 crc kubenswrapper[4772]: I0127 17:53:42.402066 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk_09090577-fdfa-4f36-badf-f32c6ee2ab7d/util/0.log" Jan 27 17:53:42 crc kubenswrapper[4772]: I0127 17:53:42.594131 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk_09090577-fdfa-4f36-badf-f32c6ee2ab7d/pull/0.log" Jan 27 17:53:42 crc kubenswrapper[4772]: I0127 17:53:42.614982 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk_09090577-fdfa-4f36-badf-f32c6ee2ab7d/util/0.log" Jan 27 17:53:42 crc kubenswrapper[4772]: I0127 17:53:42.651215 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk_09090577-fdfa-4f36-badf-f32c6ee2ab7d/pull/0.log" Jan 27 17:53:42 crc kubenswrapper[4772]: I0127 17:53:42.782669 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk_09090577-fdfa-4f36-badf-f32c6ee2ab7d/util/0.log" Jan 27 17:53:42 crc kubenswrapper[4772]: I0127 17:53:42.814282 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk_09090577-fdfa-4f36-badf-f32c6ee2ab7d/pull/0.log" Jan 27 17:53:42 crc kubenswrapper[4772]: I0127 17:53:42.843326 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcbzzsk_09090577-fdfa-4f36-badf-f32c6ee2ab7d/extract/0.log" Jan 27 17:53:42 crc kubenswrapper[4772]: I0127 17:53:42.991222 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf_d583171b-99cd-49da-9a9f-48931806cb45/util/0.log" Jan 27 17:53:43 crc kubenswrapper[4772]: I0127 17:53:43.163728 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf_d583171b-99cd-49da-9a9f-48931806cb45/pull/0.log" Jan 27 17:53:43 crc kubenswrapper[4772]: I0127 17:53:43.180718 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf_d583171b-99cd-49da-9a9f-48931806cb45/util/0.log" Jan 27 17:53:43 crc kubenswrapper[4772]: I0127 17:53:43.214851 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf_d583171b-99cd-49da-9a9f-48931806cb45/pull/0.log" Jan 27 17:53:43 crc kubenswrapper[4772]: I0127 17:53:43.413117 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf_d583171b-99cd-49da-9a9f-48931806cb45/extract/0.log" Jan 27 17:53:43 crc kubenswrapper[4772]: I0127 17:53:43.432474 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf_d583171b-99cd-49da-9a9f-48931806cb45/util/0.log" Jan 27 17:53:43 crc kubenswrapper[4772]: I0127 17:53:43.442908 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713vlmsf_d583171b-99cd-49da-9a9f-48931806cb45/pull/0.log" Jan 27 17:53:43 crc kubenswrapper[4772]: I0127 17:53:43.599989 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dfvcs_881b071c-048c-4f66-96e7-fd1f91ca23f8/extract-utilities/0.log" Jan 27 17:53:43 crc kubenswrapper[4772]: I0127 17:53:43.766688 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dfvcs_881b071c-048c-4f66-96e7-fd1f91ca23f8/extract-utilities/0.log" Jan 27 17:53:43 crc kubenswrapper[4772]: I0127 17:53:43.790340 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dfvcs_881b071c-048c-4f66-96e7-fd1f91ca23f8/extract-content/0.log" Jan 27 17:53:43 crc kubenswrapper[4772]: I0127 17:53:43.802319 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dfvcs_881b071c-048c-4f66-96e7-fd1f91ca23f8/extract-content/0.log" Jan 27 17:53:44 crc kubenswrapper[4772]: I0127 17:53:44.017056 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dfvcs_881b071c-048c-4f66-96e7-fd1f91ca23f8/extract-content/0.log" Jan 27 17:53:44 crc kubenswrapper[4772]: I0127 17:53:44.056780 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dfvcs_881b071c-048c-4f66-96e7-fd1f91ca23f8/extract-utilities/0.log" Jan 27 17:53:44 crc kubenswrapper[4772]: I0127 17:53:44.264778 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kb62q_e840c358-8d41-4381-b643-3bd35f0716a2/extract-utilities/0.log" Jan 27 17:53:44 crc kubenswrapper[4772]: I0127 17:53:44.430612 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kb62q_e840c358-8d41-4381-b643-3bd35f0716a2/extract-utilities/0.log" Jan 27 17:53:44 crc kubenswrapper[4772]: I0127 17:53:44.478597 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kb62q_e840c358-8d41-4381-b643-3bd35f0716a2/extract-content/0.log" Jan 27 17:53:44 crc kubenswrapper[4772]: I0127 17:53:44.482869 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kb62q_e840c358-8d41-4381-b643-3bd35f0716a2/extract-content/0.log" Jan 27 17:53:44 crc kubenswrapper[4772]: I0127 17:53:44.861159 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kb62q_e840c358-8d41-4381-b643-3bd35f0716a2/extract-utilities/0.log" Jan 27 17:53:44 crc kubenswrapper[4772]: I0127 17:53:44.905426 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kb62q_e840c358-8d41-4381-b643-3bd35f0716a2/extract-content/0.log" Jan 27 17:53:45 crc kubenswrapper[4772]: I0127 17:53:45.085944 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-2glnd_d8591d45-25d0-47ea-a856-9cd5334e4a8c/marketplace-operator/2.log" Jan 27 17:53:45 crc kubenswrapper[4772]: I0127 17:53:45.155591 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-2glnd_d8591d45-25d0-47ea-a856-9cd5334e4a8c/marketplace-operator/3.log" Jan 27 17:53:45 crc kubenswrapper[4772]: I0127 17:53:45.322067 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f5shj_72dcc284-e96b-4605-a428-176ca549eeb2/extract-utilities/0.log" Jan 27 17:53:45 crc kubenswrapper[4772]: I0127 17:53:45.436231 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dfvcs_881b071c-048c-4f66-96e7-fd1f91ca23f8/registry-server/0.log" Jan 27 17:53:45 crc kubenswrapper[4772]: I0127 17:53:45.539903 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-kb62q_e840c358-8d41-4381-b643-3bd35f0716a2/registry-server/0.log" Jan 27 17:53:45 crc kubenswrapper[4772]: I0127 17:53:45.565829 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f5shj_72dcc284-e96b-4605-a428-176ca549eeb2/extract-utilities/0.log" Jan 27 17:53:45 crc kubenswrapper[4772]: I0127 17:53:45.578324 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f5shj_72dcc284-e96b-4605-a428-176ca549eeb2/extract-content/0.log" Jan 27 17:53:45 crc kubenswrapper[4772]: I0127 17:53:45.585359 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f5shj_72dcc284-e96b-4605-a428-176ca549eeb2/extract-content/0.log" Jan 27 17:53:45 crc kubenswrapper[4772]: I0127 17:53:45.746575 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f5shj_72dcc284-e96b-4605-a428-176ca549eeb2/extract-content/0.log" Jan 27 17:53:45 crc kubenswrapper[4772]: I0127 17:53:45.788595 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f5shj_72dcc284-e96b-4605-a428-176ca549eeb2/extract-utilities/0.log" Jan 27 17:53:45 crc kubenswrapper[4772]: I0127 17:53:45.839570 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sdbgg_09211095-3894-4db1-bcea-29d1c2064979/extract-utilities/0.log" Jan 27 17:53:46 crc kubenswrapper[4772]: I0127 17:53:46.091421 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f5shj_72dcc284-e96b-4605-a428-176ca549eeb2/registry-server/0.log" Jan 27 17:53:46 crc kubenswrapper[4772]: I0127 17:53:46.097162 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sdbgg_09211095-3894-4db1-bcea-29d1c2064979/extract-utilities/0.log" Jan 27 17:53:46 crc kubenswrapper[4772]: I0127 17:53:46.101760 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sdbgg_09211095-3894-4db1-bcea-29d1c2064979/extract-content/0.log" Jan 27 17:53:46 crc kubenswrapper[4772]: I0127 17:53:46.130801 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sdbgg_09211095-3894-4db1-bcea-29d1c2064979/extract-content/0.log" Jan 27 17:53:46 crc kubenswrapper[4772]: I0127 17:53:46.258270 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sdbgg_09211095-3894-4db1-bcea-29d1c2064979/extract-utilities/0.log" Jan 27 17:53:46 crc kubenswrapper[4772]: I0127 17:53:46.332344 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sdbgg_09211095-3894-4db1-bcea-29d1c2064979/extract-content/0.log" Jan 27 17:53:46 crc kubenswrapper[4772]: I0127 17:53:46.511602 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sdbgg_09211095-3894-4db1-bcea-29d1c2064979/registry-server/0.log" Jan 27 17:53:54 crc kubenswrapper[4772]: I0127 17:53:54.670491 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:53:54 crc kubenswrapper[4772]: E0127 17:53:54.671450 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:54:05 crc kubenswrapper[4772]: I0127 17:54:05.662588 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:54:05 crc kubenswrapper[4772]: E0127 17:54:05.663229 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:54:08 crc kubenswrapper[4772]: I0127 17:54:08.815387 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vnhgt"] Jan 27 17:54:08 crc kubenswrapper[4772]: E0127 17:54:08.816619 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bec6411-aae2-47f7-a721-a39d6d381c17" containerName="extract-content" Jan 27 17:54:08 crc kubenswrapper[4772]: I0127 17:54:08.816633 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bec6411-aae2-47f7-a721-a39d6d381c17" containerName="extract-content" Jan 27 17:54:08 crc kubenswrapper[4772]: E0127 17:54:08.816652 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bec6411-aae2-47f7-a721-a39d6d381c17" containerName="registry-server" Jan 27 17:54:08 crc kubenswrapper[4772]: I0127 17:54:08.816658 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bec6411-aae2-47f7-a721-a39d6d381c17" containerName="registry-server" Jan 27 17:54:08 crc kubenswrapper[4772]: E0127 17:54:08.816680 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bec6411-aae2-47f7-a721-a39d6d381c17" containerName="extract-utilities" Jan 27 17:54:08 crc kubenswrapper[4772]: I0127 17:54:08.816687 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bec6411-aae2-47f7-a721-a39d6d381c17" containerName="extract-utilities" Jan 27 17:54:08 crc kubenswrapper[4772]: E0127 17:54:08.816715 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2ac0d9e-ba08-48c0-9a6c-307c3438b86e" containerName="container-00" Jan 27 17:54:08 crc kubenswrapper[4772]: I0127 17:54:08.816721 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2ac0d9e-ba08-48c0-9a6c-307c3438b86e" containerName="container-00" Jan 27 17:54:08 crc kubenswrapper[4772]: I0127 17:54:08.817015 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2ac0d9e-ba08-48c0-9a6c-307c3438b86e" containerName="container-00" Jan 27 17:54:08 crc kubenswrapper[4772]: I0127 17:54:08.817039 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bec6411-aae2-47f7-a721-a39d6d381c17" containerName="registry-server" Jan 27 17:54:08 crc kubenswrapper[4772]: I0127 17:54:08.826341 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vnhgt" Jan 27 17:54:08 crc kubenswrapper[4772]: I0127 17:54:08.858125 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vnhgt"] Jan 27 17:54:08 crc kubenswrapper[4772]: I0127 17:54:08.938529 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41b8d2e5-185e-4766-82dc-475963959026-catalog-content\") pod \"community-operators-vnhgt\" (UID: \"41b8d2e5-185e-4766-82dc-475963959026\") " pod="openshift-marketplace/community-operators-vnhgt" Jan 27 17:54:08 crc kubenswrapper[4772]: I0127 17:54:08.938602 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41b8d2e5-185e-4766-82dc-475963959026-utilities\") pod \"community-operators-vnhgt\" (UID: \"41b8d2e5-185e-4766-82dc-475963959026\") " pod="openshift-marketplace/community-operators-vnhgt" Jan 27 17:54:08 crc kubenswrapper[4772]: I0127 17:54:08.938643 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-226hp\" (UniqueName: \"kubernetes.io/projected/41b8d2e5-185e-4766-82dc-475963959026-kube-api-access-226hp\") pod \"community-operators-vnhgt\" (UID: \"41b8d2e5-185e-4766-82dc-475963959026\") " pod="openshift-marketplace/community-operators-vnhgt" Jan 27 17:54:09 crc kubenswrapper[4772]: I0127 17:54:09.040336 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41b8d2e5-185e-4766-82dc-475963959026-catalog-content\") pod \"community-operators-vnhgt\" (UID: \"41b8d2e5-185e-4766-82dc-475963959026\") " pod="openshift-marketplace/community-operators-vnhgt" Jan 27 17:54:09 crc kubenswrapper[4772]: I0127 17:54:09.040407 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41b8d2e5-185e-4766-82dc-475963959026-utilities\") pod \"community-operators-vnhgt\" (UID: \"41b8d2e5-185e-4766-82dc-475963959026\") " pod="openshift-marketplace/community-operators-vnhgt" Jan 27 17:54:09 crc kubenswrapper[4772]: I0127 17:54:09.040446 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-226hp\" (UniqueName: \"kubernetes.io/projected/41b8d2e5-185e-4766-82dc-475963959026-kube-api-access-226hp\") pod \"community-operators-vnhgt\" (UID: \"41b8d2e5-185e-4766-82dc-475963959026\") " pod="openshift-marketplace/community-operators-vnhgt" Jan 27 17:54:09 crc kubenswrapper[4772]: I0127 17:54:09.040957 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41b8d2e5-185e-4766-82dc-475963959026-utilities\") pod \"community-operators-vnhgt\" (UID: \"41b8d2e5-185e-4766-82dc-475963959026\") " pod="openshift-marketplace/community-operators-vnhgt" Jan 27 17:54:09 crc kubenswrapper[4772]: I0127 17:54:09.041031 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41b8d2e5-185e-4766-82dc-475963959026-catalog-content\") pod \"community-operators-vnhgt\" (UID: \"41b8d2e5-185e-4766-82dc-475963959026\") " pod="openshift-marketplace/community-operators-vnhgt" Jan 27 17:54:09 crc kubenswrapper[4772]: I0127 17:54:09.062830 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-226hp\" (UniqueName: \"kubernetes.io/projected/41b8d2e5-185e-4766-82dc-475963959026-kube-api-access-226hp\") pod \"community-operators-vnhgt\" (UID: \"41b8d2e5-185e-4766-82dc-475963959026\") " pod="openshift-marketplace/community-operators-vnhgt" Jan 27 17:54:09 crc kubenswrapper[4772]: I0127 17:54:09.153849 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vnhgt" Jan 27 17:54:09 crc kubenswrapper[4772]: I0127 17:54:09.973852 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vnhgt"] Jan 27 17:54:10 crc kubenswrapper[4772]: I0127 17:54:10.370296 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vnhgt" event={"ID":"41b8d2e5-185e-4766-82dc-475963959026","Type":"ContainerStarted","Data":"199265b153397fccb7e5a85666b63dd6b5b8fed3e925fc3bb5d59089f4acf311"} Jan 27 17:54:11 crc kubenswrapper[4772]: I0127 17:54:11.380500 4772 generic.go:334] "Generic (PLEG): container finished" podID="41b8d2e5-185e-4766-82dc-475963959026" containerID="f41b1e666393e5a2661e4c583dc4dc3e931e50487a86a7a14b7e7e6d2d45c005" exitCode=0 Jan 27 17:54:11 crc kubenswrapper[4772]: I0127 17:54:11.380779 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vnhgt" event={"ID":"41b8d2e5-185e-4766-82dc-475963959026","Type":"ContainerDied","Data":"f41b1e666393e5a2661e4c583dc4dc3e931e50487a86a7a14b7e7e6d2d45c005"} Jan 27 17:54:11 crc kubenswrapper[4772]: I0127 17:54:11.382935 4772 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 27 17:54:13 crc kubenswrapper[4772]: I0127 17:54:13.402139 4772 generic.go:334] "Generic (PLEG): container finished" podID="41b8d2e5-185e-4766-82dc-475963959026" containerID="e380eca5b91487ee3d86108869a3d28c49ca1c5d2415c5ccd5dcff2e75cd190a" exitCode=0 Jan 27 17:54:13 crc kubenswrapper[4772]: I0127 17:54:13.402253 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vnhgt" event={"ID":"41b8d2e5-185e-4766-82dc-475963959026","Type":"ContainerDied","Data":"e380eca5b91487ee3d86108869a3d28c49ca1c5d2415c5ccd5dcff2e75cd190a"} Jan 27 17:54:14 crc kubenswrapper[4772]: I0127 17:54:14.413487 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vnhgt" event={"ID":"41b8d2e5-185e-4766-82dc-475963959026","Type":"ContainerStarted","Data":"aac79acc3e134a8a805aefd2c496fbc70975b6548e84461983c168b510e44115"} Jan 27 17:54:14 crc kubenswrapper[4772]: I0127 17:54:14.432452 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vnhgt" podStartSLOduration=3.985287484 podStartE2EDuration="6.432427148s" podCreationTimestamp="2026-01-27 17:54:08 +0000 UTC" firstStartedPulling="2026-01-27 17:54:11.38267879 +0000 UTC m=+10037.363287898" lastFinishedPulling="2026-01-27 17:54:13.829818464 +0000 UTC m=+10039.810427562" observedRunningTime="2026-01-27 17:54:14.42830048 +0000 UTC m=+10040.408909578" watchObservedRunningTime="2026-01-27 17:54:14.432427148 +0000 UTC m=+10040.413036246" Jan 27 17:54:19 crc kubenswrapper[4772]: I0127 17:54:19.154985 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vnhgt" Jan 27 17:54:19 crc kubenswrapper[4772]: I0127 17:54:19.156620 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vnhgt" Jan 27 17:54:19 crc kubenswrapper[4772]: I0127 17:54:19.233329 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vnhgt" Jan 27 17:54:19 crc kubenswrapper[4772]: I0127 17:54:19.509566 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vnhgt" Jan 27 17:54:19 crc kubenswrapper[4772]: I0127 17:54:19.564098 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vnhgt"] Jan 27 17:54:19 crc kubenswrapper[4772]: I0127 17:54:19.662890 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:54:19 crc kubenswrapper[4772]: E0127 17:54:19.663156 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:54:21 crc kubenswrapper[4772]: I0127 17:54:21.492624 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vnhgt" podUID="41b8d2e5-185e-4766-82dc-475963959026" containerName="registry-server" containerID="cri-o://aac79acc3e134a8a805aefd2c496fbc70975b6548e84461983c168b510e44115" gracePeriod=2 Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.051838 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vnhgt" Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.108675 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-226hp\" (UniqueName: \"kubernetes.io/projected/41b8d2e5-185e-4766-82dc-475963959026-kube-api-access-226hp\") pod \"41b8d2e5-185e-4766-82dc-475963959026\" (UID: \"41b8d2e5-185e-4766-82dc-475963959026\") " Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.108801 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41b8d2e5-185e-4766-82dc-475963959026-catalog-content\") pod \"41b8d2e5-185e-4766-82dc-475963959026\" (UID: \"41b8d2e5-185e-4766-82dc-475963959026\") " Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.108820 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41b8d2e5-185e-4766-82dc-475963959026-utilities\") pod \"41b8d2e5-185e-4766-82dc-475963959026\" (UID: \"41b8d2e5-185e-4766-82dc-475963959026\") " Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.109691 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41b8d2e5-185e-4766-82dc-475963959026-utilities" (OuterVolumeSpecName: "utilities") pod "41b8d2e5-185e-4766-82dc-475963959026" (UID: "41b8d2e5-185e-4766-82dc-475963959026"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.116533 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41b8d2e5-185e-4766-82dc-475963959026-kube-api-access-226hp" (OuterVolumeSpecName: "kube-api-access-226hp") pod "41b8d2e5-185e-4766-82dc-475963959026" (UID: "41b8d2e5-185e-4766-82dc-475963959026"). InnerVolumeSpecName "kube-api-access-226hp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.181160 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41b8d2e5-185e-4766-82dc-475963959026-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "41b8d2e5-185e-4766-82dc-475963959026" (UID: "41b8d2e5-185e-4766-82dc-475963959026"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.210408 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41b8d2e5-185e-4766-82dc-475963959026-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.210449 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41b8d2e5-185e-4766-82dc-475963959026-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.210459 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-226hp\" (UniqueName: \"kubernetes.io/projected/41b8d2e5-185e-4766-82dc-475963959026-kube-api-access-226hp\") on node \"crc\" DevicePath \"\"" Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.502750 4772 generic.go:334] "Generic (PLEG): container finished" podID="41b8d2e5-185e-4766-82dc-475963959026" containerID="aac79acc3e134a8a805aefd2c496fbc70975b6548e84461983c168b510e44115" exitCode=0 Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.502794 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vnhgt" event={"ID":"41b8d2e5-185e-4766-82dc-475963959026","Type":"ContainerDied","Data":"aac79acc3e134a8a805aefd2c496fbc70975b6548e84461983c168b510e44115"} Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.502821 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vnhgt" event={"ID":"41b8d2e5-185e-4766-82dc-475963959026","Type":"ContainerDied","Data":"199265b153397fccb7e5a85666b63dd6b5b8fed3e925fc3bb5d59089f4acf311"} Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.502837 4772 scope.go:117] "RemoveContainer" containerID="aac79acc3e134a8a805aefd2c496fbc70975b6548e84461983c168b510e44115" Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.502839 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vnhgt" Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.522019 4772 scope.go:117] "RemoveContainer" containerID="e380eca5b91487ee3d86108869a3d28c49ca1c5d2415c5ccd5dcff2e75cd190a" Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.547401 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vnhgt"] Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.557936 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vnhgt"] Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.560127 4772 scope.go:117] "RemoveContainer" containerID="f41b1e666393e5a2661e4c583dc4dc3e931e50487a86a7a14b7e7e6d2d45c005" Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.594460 4772 scope.go:117] "RemoveContainer" containerID="aac79acc3e134a8a805aefd2c496fbc70975b6548e84461983c168b510e44115" Jan 27 17:54:22 crc kubenswrapper[4772]: E0127 17:54:22.595892 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aac79acc3e134a8a805aefd2c496fbc70975b6548e84461983c168b510e44115\": container with ID starting with aac79acc3e134a8a805aefd2c496fbc70975b6548e84461983c168b510e44115 not found: ID does not exist" containerID="aac79acc3e134a8a805aefd2c496fbc70975b6548e84461983c168b510e44115" Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.595936 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aac79acc3e134a8a805aefd2c496fbc70975b6548e84461983c168b510e44115"} err="failed to get container status \"aac79acc3e134a8a805aefd2c496fbc70975b6548e84461983c168b510e44115\": rpc error: code = NotFound desc = could not find container \"aac79acc3e134a8a805aefd2c496fbc70975b6548e84461983c168b510e44115\": container with ID starting with aac79acc3e134a8a805aefd2c496fbc70975b6548e84461983c168b510e44115 not found: ID does not exist" Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.595964 4772 scope.go:117] "RemoveContainer" containerID="e380eca5b91487ee3d86108869a3d28c49ca1c5d2415c5ccd5dcff2e75cd190a" Jan 27 17:54:22 crc kubenswrapper[4772]: E0127 17:54:22.596294 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e380eca5b91487ee3d86108869a3d28c49ca1c5d2415c5ccd5dcff2e75cd190a\": container with ID starting with e380eca5b91487ee3d86108869a3d28c49ca1c5d2415c5ccd5dcff2e75cd190a not found: ID does not exist" containerID="e380eca5b91487ee3d86108869a3d28c49ca1c5d2415c5ccd5dcff2e75cd190a" Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.596331 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e380eca5b91487ee3d86108869a3d28c49ca1c5d2415c5ccd5dcff2e75cd190a"} err="failed to get container status \"e380eca5b91487ee3d86108869a3d28c49ca1c5d2415c5ccd5dcff2e75cd190a\": rpc error: code = NotFound desc = could not find container \"e380eca5b91487ee3d86108869a3d28c49ca1c5d2415c5ccd5dcff2e75cd190a\": container with ID starting with e380eca5b91487ee3d86108869a3d28c49ca1c5d2415c5ccd5dcff2e75cd190a not found: ID does not exist" Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.596371 4772 scope.go:117] "RemoveContainer" containerID="f41b1e666393e5a2661e4c583dc4dc3e931e50487a86a7a14b7e7e6d2d45c005" Jan 27 17:54:22 crc kubenswrapper[4772]: E0127 17:54:22.596607 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f41b1e666393e5a2661e4c583dc4dc3e931e50487a86a7a14b7e7e6d2d45c005\": container with ID starting with f41b1e666393e5a2661e4c583dc4dc3e931e50487a86a7a14b7e7e6d2d45c005 not found: ID does not exist" containerID="f41b1e666393e5a2661e4c583dc4dc3e931e50487a86a7a14b7e7e6d2d45c005" Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.596825 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f41b1e666393e5a2661e4c583dc4dc3e931e50487a86a7a14b7e7e6d2d45c005"} err="failed to get container status \"f41b1e666393e5a2661e4c583dc4dc3e931e50487a86a7a14b7e7e6d2d45c005\": rpc error: code = NotFound desc = could not find container \"f41b1e666393e5a2661e4c583dc4dc3e931e50487a86a7a14b7e7e6d2d45c005\": container with ID starting with f41b1e666393e5a2661e4c583dc4dc3e931e50487a86a7a14b7e7e6d2d45c005 not found: ID does not exist" Jan 27 17:54:22 crc kubenswrapper[4772]: I0127 17:54:22.676663 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41b8d2e5-185e-4766-82dc-475963959026" path="/var/lib/kubelet/pods/41b8d2e5-185e-4766-82dc-475963959026/volumes" Jan 27 17:54:30 crc kubenswrapper[4772]: I0127 17:54:30.663922 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:54:30 crc kubenswrapper[4772]: E0127 17:54:30.665032 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:54:45 crc kubenswrapper[4772]: I0127 17:54:45.667890 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:54:45 crc kubenswrapper[4772]: E0127 17:54:45.669743 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:54:58 crc kubenswrapper[4772]: I0127 17:54:58.663944 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:54:58 crc kubenswrapper[4772]: E0127 17:54:58.664959 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:55:10 crc kubenswrapper[4772]: I0127 17:55:10.663996 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:55:10 crc kubenswrapper[4772]: E0127 17:55:10.664955 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:55:13 crc kubenswrapper[4772]: I0127 17:55:13.035802 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5kr2z"] Jan 27 17:55:13 crc kubenswrapper[4772]: E0127 17:55:13.036723 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41b8d2e5-185e-4766-82dc-475963959026" containerName="registry-server" Jan 27 17:55:13 crc kubenswrapper[4772]: I0127 17:55:13.036739 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="41b8d2e5-185e-4766-82dc-475963959026" containerName="registry-server" Jan 27 17:55:13 crc kubenswrapper[4772]: E0127 17:55:13.036769 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41b8d2e5-185e-4766-82dc-475963959026" containerName="extract-content" Jan 27 17:55:13 crc kubenswrapper[4772]: I0127 17:55:13.036775 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="41b8d2e5-185e-4766-82dc-475963959026" containerName="extract-content" Jan 27 17:55:13 crc kubenswrapper[4772]: E0127 17:55:13.036791 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41b8d2e5-185e-4766-82dc-475963959026" containerName="extract-utilities" Jan 27 17:55:13 crc kubenswrapper[4772]: I0127 17:55:13.036798 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="41b8d2e5-185e-4766-82dc-475963959026" containerName="extract-utilities" Jan 27 17:55:13 crc kubenswrapper[4772]: I0127 17:55:13.037027 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="41b8d2e5-185e-4766-82dc-475963959026" containerName="registry-server" Jan 27 17:55:13 crc kubenswrapper[4772]: I0127 17:55:13.039728 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5kr2z" Jan 27 17:55:13 crc kubenswrapper[4772]: I0127 17:55:13.047645 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5kr2z"] Jan 27 17:55:13 crc kubenswrapper[4772]: I0127 17:55:13.197659 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tdd8\" (UniqueName: \"kubernetes.io/projected/84e9f92a-7805-4587-9826-afda667e7ee1-kube-api-access-7tdd8\") pod \"redhat-marketplace-5kr2z\" (UID: \"84e9f92a-7805-4587-9826-afda667e7ee1\") " pod="openshift-marketplace/redhat-marketplace-5kr2z" Jan 27 17:55:13 crc kubenswrapper[4772]: I0127 17:55:13.197740 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84e9f92a-7805-4587-9826-afda667e7ee1-utilities\") pod \"redhat-marketplace-5kr2z\" (UID: \"84e9f92a-7805-4587-9826-afda667e7ee1\") " pod="openshift-marketplace/redhat-marketplace-5kr2z" Jan 27 17:55:13 crc kubenswrapper[4772]: I0127 17:55:13.197763 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84e9f92a-7805-4587-9826-afda667e7ee1-catalog-content\") pod \"redhat-marketplace-5kr2z\" (UID: \"84e9f92a-7805-4587-9826-afda667e7ee1\") " pod="openshift-marketplace/redhat-marketplace-5kr2z" Jan 27 17:55:13 crc kubenswrapper[4772]: I0127 17:55:13.299716 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tdd8\" (UniqueName: \"kubernetes.io/projected/84e9f92a-7805-4587-9826-afda667e7ee1-kube-api-access-7tdd8\") pod \"redhat-marketplace-5kr2z\" (UID: \"84e9f92a-7805-4587-9826-afda667e7ee1\") " pod="openshift-marketplace/redhat-marketplace-5kr2z" Jan 27 17:55:13 crc kubenswrapper[4772]: I0127 17:55:13.299790 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84e9f92a-7805-4587-9826-afda667e7ee1-utilities\") pod \"redhat-marketplace-5kr2z\" (UID: \"84e9f92a-7805-4587-9826-afda667e7ee1\") " pod="openshift-marketplace/redhat-marketplace-5kr2z" Jan 27 17:55:13 crc kubenswrapper[4772]: I0127 17:55:13.299819 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84e9f92a-7805-4587-9826-afda667e7ee1-catalog-content\") pod \"redhat-marketplace-5kr2z\" (UID: \"84e9f92a-7805-4587-9826-afda667e7ee1\") " pod="openshift-marketplace/redhat-marketplace-5kr2z" Jan 27 17:55:13 crc kubenswrapper[4772]: I0127 17:55:13.300420 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84e9f92a-7805-4587-9826-afda667e7ee1-catalog-content\") pod \"redhat-marketplace-5kr2z\" (UID: \"84e9f92a-7805-4587-9826-afda667e7ee1\") " pod="openshift-marketplace/redhat-marketplace-5kr2z" Jan 27 17:55:13 crc kubenswrapper[4772]: I0127 17:55:13.300421 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84e9f92a-7805-4587-9826-afda667e7ee1-utilities\") pod \"redhat-marketplace-5kr2z\" (UID: \"84e9f92a-7805-4587-9826-afda667e7ee1\") " pod="openshift-marketplace/redhat-marketplace-5kr2z" Jan 27 17:55:13 crc kubenswrapper[4772]: I0127 17:55:13.333433 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tdd8\" (UniqueName: \"kubernetes.io/projected/84e9f92a-7805-4587-9826-afda667e7ee1-kube-api-access-7tdd8\") pod \"redhat-marketplace-5kr2z\" (UID: \"84e9f92a-7805-4587-9826-afda667e7ee1\") " pod="openshift-marketplace/redhat-marketplace-5kr2z" Jan 27 17:55:13 crc kubenswrapper[4772]: I0127 17:55:13.414154 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5kr2z" Jan 27 17:55:13 crc kubenswrapper[4772]: I0127 17:55:13.938416 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5kr2z"] Jan 27 17:55:15 crc kubenswrapper[4772]: I0127 17:55:15.012296 4772 generic.go:334] "Generic (PLEG): container finished" podID="84e9f92a-7805-4587-9826-afda667e7ee1" containerID="24934acfd4ff03828952a4d233d786064c58ec45adfd8f4996766292937bef9c" exitCode=0 Jan 27 17:55:15 crc kubenswrapper[4772]: I0127 17:55:15.012354 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5kr2z" event={"ID":"84e9f92a-7805-4587-9826-afda667e7ee1","Type":"ContainerDied","Data":"24934acfd4ff03828952a4d233d786064c58ec45adfd8f4996766292937bef9c"} Jan 27 17:55:15 crc kubenswrapper[4772]: I0127 17:55:15.012661 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5kr2z" event={"ID":"84e9f92a-7805-4587-9826-afda667e7ee1","Type":"ContainerStarted","Data":"37b015df7b3b2fcbad3ebd639dc3d4bb0302697c171f372c59ec684cae6a5d2f"} Jan 27 17:55:16 crc kubenswrapper[4772]: I0127 17:55:16.024401 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5kr2z" event={"ID":"84e9f92a-7805-4587-9826-afda667e7ee1","Type":"ContainerStarted","Data":"d4b51a7cc091f99cda607405d2e8f9684e13bdb11c929a5b8c5b1ac513e7eabb"} Jan 27 17:55:17 crc kubenswrapper[4772]: I0127 17:55:17.037232 4772 generic.go:334] "Generic (PLEG): container finished" podID="84e9f92a-7805-4587-9826-afda667e7ee1" containerID="d4b51a7cc091f99cda607405d2e8f9684e13bdb11c929a5b8c5b1ac513e7eabb" exitCode=0 Jan 27 17:55:17 crc kubenswrapper[4772]: I0127 17:55:17.037641 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5kr2z" event={"ID":"84e9f92a-7805-4587-9826-afda667e7ee1","Type":"ContainerDied","Data":"d4b51a7cc091f99cda607405d2e8f9684e13bdb11c929a5b8c5b1ac513e7eabb"} Jan 27 17:55:18 crc kubenswrapper[4772]: I0127 17:55:18.049507 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5kr2z" event={"ID":"84e9f92a-7805-4587-9826-afda667e7ee1","Type":"ContainerStarted","Data":"d7125b1b6edd7114a279766eaf0708862a3658f1e482bfc3610c0145352fb0fa"} Jan 27 17:55:18 crc kubenswrapper[4772]: I0127 17:55:18.080295 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5kr2z" podStartSLOduration=2.511108062 podStartE2EDuration="5.080275438s" podCreationTimestamp="2026-01-27 17:55:13 +0000 UTC" firstStartedPulling="2026-01-27 17:55:15.014573336 +0000 UTC m=+10100.995182434" lastFinishedPulling="2026-01-27 17:55:17.583740712 +0000 UTC m=+10103.564349810" observedRunningTime="2026-01-27 17:55:18.073253458 +0000 UTC m=+10104.053862566" watchObservedRunningTime="2026-01-27 17:55:18.080275438 +0000 UTC m=+10104.060884546" Jan 27 17:55:21 crc kubenswrapper[4772]: I0127 17:55:21.663489 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:55:21 crc kubenswrapper[4772]: E0127 17:55:21.664137 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:55:23 crc kubenswrapper[4772]: I0127 17:55:23.414942 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5kr2z" Jan 27 17:55:23 crc kubenswrapper[4772]: I0127 17:55:23.415262 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5kr2z" Jan 27 17:55:23 crc kubenswrapper[4772]: I0127 17:55:23.467390 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5kr2z" Jan 27 17:55:24 crc kubenswrapper[4772]: I0127 17:55:24.413124 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5kr2z" Jan 27 17:55:25 crc kubenswrapper[4772]: I0127 17:55:25.019737 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5kr2z"] Jan 27 17:55:26 crc kubenswrapper[4772]: I0127 17:55:26.127995 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5kr2z" podUID="84e9f92a-7805-4587-9826-afda667e7ee1" containerName="registry-server" containerID="cri-o://d7125b1b6edd7114a279766eaf0708862a3658f1e482bfc3610c0145352fb0fa" gracePeriod=2 Jan 27 17:55:26 crc kubenswrapper[4772]: I0127 17:55:26.662227 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5kr2z" Jan 27 17:55:26 crc kubenswrapper[4772]: I0127 17:55:26.774155 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84e9f92a-7805-4587-9826-afda667e7ee1-utilities\") pod \"84e9f92a-7805-4587-9826-afda667e7ee1\" (UID: \"84e9f92a-7805-4587-9826-afda667e7ee1\") " Jan 27 17:55:26 crc kubenswrapper[4772]: I0127 17:55:26.774297 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84e9f92a-7805-4587-9826-afda667e7ee1-catalog-content\") pod \"84e9f92a-7805-4587-9826-afda667e7ee1\" (UID: \"84e9f92a-7805-4587-9826-afda667e7ee1\") " Jan 27 17:55:26 crc kubenswrapper[4772]: I0127 17:55:26.774479 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tdd8\" (UniqueName: \"kubernetes.io/projected/84e9f92a-7805-4587-9826-afda667e7ee1-kube-api-access-7tdd8\") pod \"84e9f92a-7805-4587-9826-afda667e7ee1\" (UID: \"84e9f92a-7805-4587-9826-afda667e7ee1\") " Jan 27 17:55:26 crc kubenswrapper[4772]: I0127 17:55:26.785390 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84e9f92a-7805-4587-9826-afda667e7ee1-kube-api-access-7tdd8" (OuterVolumeSpecName: "kube-api-access-7tdd8") pod "84e9f92a-7805-4587-9826-afda667e7ee1" (UID: "84e9f92a-7805-4587-9826-afda667e7ee1"). InnerVolumeSpecName "kube-api-access-7tdd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:55:26 crc kubenswrapper[4772]: I0127 17:55:26.788488 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84e9f92a-7805-4587-9826-afda667e7ee1-utilities" (OuterVolumeSpecName: "utilities") pod "84e9f92a-7805-4587-9826-afda667e7ee1" (UID: "84e9f92a-7805-4587-9826-afda667e7ee1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:55:26 crc kubenswrapper[4772]: I0127 17:55:26.877525 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84e9f92a-7805-4587-9826-afda667e7ee1-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 17:55:26 crc kubenswrapper[4772]: I0127 17:55:26.877743 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tdd8\" (UniqueName: \"kubernetes.io/projected/84e9f92a-7805-4587-9826-afda667e7ee1-kube-api-access-7tdd8\") on node \"crc\" DevicePath \"\"" Jan 27 17:55:26 crc kubenswrapper[4772]: I0127 17:55:26.908837 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84e9f92a-7805-4587-9826-afda667e7ee1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "84e9f92a-7805-4587-9826-afda667e7ee1" (UID: "84e9f92a-7805-4587-9826-afda667e7ee1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:55:26 crc kubenswrapper[4772]: I0127 17:55:26.980419 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84e9f92a-7805-4587-9826-afda667e7ee1-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 17:55:27 crc kubenswrapper[4772]: I0127 17:55:27.139479 4772 generic.go:334] "Generic (PLEG): container finished" podID="84e9f92a-7805-4587-9826-afda667e7ee1" containerID="d7125b1b6edd7114a279766eaf0708862a3658f1e482bfc3610c0145352fb0fa" exitCode=0 Jan 27 17:55:27 crc kubenswrapper[4772]: I0127 17:55:27.139538 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5kr2z" event={"ID":"84e9f92a-7805-4587-9826-afda667e7ee1","Type":"ContainerDied","Data":"d7125b1b6edd7114a279766eaf0708862a3658f1e482bfc3610c0145352fb0fa"} Jan 27 17:55:27 crc kubenswrapper[4772]: I0127 17:55:27.139577 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5kr2z" event={"ID":"84e9f92a-7805-4587-9826-afda667e7ee1","Type":"ContainerDied","Data":"37b015df7b3b2fcbad3ebd639dc3d4bb0302697c171f372c59ec684cae6a5d2f"} Jan 27 17:55:27 crc kubenswrapper[4772]: I0127 17:55:27.139606 4772 scope.go:117] "RemoveContainer" containerID="d7125b1b6edd7114a279766eaf0708862a3658f1e482bfc3610c0145352fb0fa" Jan 27 17:55:27 crc kubenswrapper[4772]: I0127 17:55:27.139779 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5kr2z" Jan 27 17:55:27 crc kubenswrapper[4772]: I0127 17:55:27.184091 4772 scope.go:117] "RemoveContainer" containerID="d4b51a7cc091f99cda607405d2e8f9684e13bdb11c929a5b8c5b1ac513e7eabb" Jan 27 17:55:27 crc kubenswrapper[4772]: I0127 17:55:27.195685 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5kr2z"] Jan 27 17:55:27 crc kubenswrapper[4772]: I0127 17:55:27.210216 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5kr2z"] Jan 27 17:55:27 crc kubenswrapper[4772]: I0127 17:55:27.217397 4772 scope.go:117] "RemoveContainer" containerID="24934acfd4ff03828952a4d233d786064c58ec45adfd8f4996766292937bef9c" Jan 27 17:55:27 crc kubenswrapper[4772]: I0127 17:55:27.254232 4772 scope.go:117] "RemoveContainer" containerID="d7125b1b6edd7114a279766eaf0708862a3658f1e482bfc3610c0145352fb0fa" Jan 27 17:55:27 crc kubenswrapper[4772]: E0127 17:55:27.254961 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7125b1b6edd7114a279766eaf0708862a3658f1e482bfc3610c0145352fb0fa\": container with ID starting with d7125b1b6edd7114a279766eaf0708862a3658f1e482bfc3610c0145352fb0fa not found: ID does not exist" containerID="d7125b1b6edd7114a279766eaf0708862a3658f1e482bfc3610c0145352fb0fa" Jan 27 17:55:27 crc kubenswrapper[4772]: I0127 17:55:27.255003 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7125b1b6edd7114a279766eaf0708862a3658f1e482bfc3610c0145352fb0fa"} err="failed to get container status \"d7125b1b6edd7114a279766eaf0708862a3658f1e482bfc3610c0145352fb0fa\": rpc error: code = NotFound desc = could not find container \"d7125b1b6edd7114a279766eaf0708862a3658f1e482bfc3610c0145352fb0fa\": container with ID starting with d7125b1b6edd7114a279766eaf0708862a3658f1e482bfc3610c0145352fb0fa not found: ID does not exist" Jan 27 17:55:27 crc kubenswrapper[4772]: I0127 17:55:27.255030 4772 scope.go:117] "RemoveContainer" containerID="d4b51a7cc091f99cda607405d2e8f9684e13bdb11c929a5b8c5b1ac513e7eabb" Jan 27 17:55:27 crc kubenswrapper[4772]: E0127 17:55:27.255423 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4b51a7cc091f99cda607405d2e8f9684e13bdb11c929a5b8c5b1ac513e7eabb\": container with ID starting with d4b51a7cc091f99cda607405d2e8f9684e13bdb11c929a5b8c5b1ac513e7eabb not found: ID does not exist" containerID="d4b51a7cc091f99cda607405d2e8f9684e13bdb11c929a5b8c5b1ac513e7eabb" Jan 27 17:55:27 crc kubenswrapper[4772]: I0127 17:55:27.255465 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4b51a7cc091f99cda607405d2e8f9684e13bdb11c929a5b8c5b1ac513e7eabb"} err="failed to get container status \"d4b51a7cc091f99cda607405d2e8f9684e13bdb11c929a5b8c5b1ac513e7eabb\": rpc error: code = NotFound desc = could not find container \"d4b51a7cc091f99cda607405d2e8f9684e13bdb11c929a5b8c5b1ac513e7eabb\": container with ID starting with d4b51a7cc091f99cda607405d2e8f9684e13bdb11c929a5b8c5b1ac513e7eabb not found: ID does not exist" Jan 27 17:55:27 crc kubenswrapper[4772]: I0127 17:55:27.255560 4772 scope.go:117] "RemoveContainer" containerID="24934acfd4ff03828952a4d233d786064c58ec45adfd8f4996766292937bef9c" Jan 27 17:55:27 crc kubenswrapper[4772]: E0127 17:55:27.255946 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24934acfd4ff03828952a4d233d786064c58ec45adfd8f4996766292937bef9c\": container with ID starting with 24934acfd4ff03828952a4d233d786064c58ec45adfd8f4996766292937bef9c not found: ID does not exist" containerID="24934acfd4ff03828952a4d233d786064c58ec45adfd8f4996766292937bef9c" Jan 27 17:55:27 crc kubenswrapper[4772]: I0127 17:55:27.255989 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24934acfd4ff03828952a4d233d786064c58ec45adfd8f4996766292937bef9c"} err="failed to get container status \"24934acfd4ff03828952a4d233d786064c58ec45adfd8f4996766292937bef9c\": rpc error: code = NotFound desc = could not find container \"24934acfd4ff03828952a4d233d786064c58ec45adfd8f4996766292937bef9c\": container with ID starting with 24934acfd4ff03828952a4d233d786064c58ec45adfd8f4996766292937bef9c not found: ID does not exist" Jan 27 17:55:28 crc kubenswrapper[4772]: I0127 17:55:28.683243 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84e9f92a-7805-4587-9826-afda667e7ee1" path="/var/lib/kubelet/pods/84e9f92a-7805-4587-9826-afda667e7ee1/volumes" Jan 27 17:55:32 crc kubenswrapper[4772]: I0127 17:55:32.202690 4772 generic.go:334] "Generic (PLEG): container finished" podID="62afdf43-7cc5-4d53-aff8-2fd18fbfd493" containerID="daea92a47f41c2b8eac7ad3a9eb9829829dd28a915856ca1714eacc250cb2602" exitCode=0 Jan 27 17:55:32 crc kubenswrapper[4772]: I0127 17:55:32.202827 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7dvs7/must-gather-vr6mx" event={"ID":"62afdf43-7cc5-4d53-aff8-2fd18fbfd493","Type":"ContainerDied","Data":"daea92a47f41c2b8eac7ad3a9eb9829829dd28a915856ca1714eacc250cb2602"} Jan 27 17:55:32 crc kubenswrapper[4772]: I0127 17:55:32.204365 4772 scope.go:117] "RemoveContainer" containerID="daea92a47f41c2b8eac7ad3a9eb9829829dd28a915856ca1714eacc250cb2602" Jan 27 17:55:32 crc kubenswrapper[4772]: I0127 17:55:32.489368 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7dvs7_must-gather-vr6mx_62afdf43-7cc5-4d53-aff8-2fd18fbfd493/gather/0.log" Jan 27 17:55:36 crc kubenswrapper[4772]: I0127 17:55:36.664076 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:55:36 crc kubenswrapper[4772]: E0127 17:55:36.665626 4772 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4hwxn_openshift-machine-config-operator(67794a44-d793-4fd7-9e54-e40437f67c0b)\"" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" Jan 27 17:55:40 crc kubenswrapper[4772]: I0127 17:55:40.035519 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7dvs7/must-gather-vr6mx"] Jan 27 17:55:40 crc kubenswrapper[4772]: I0127 17:55:40.036038 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-7dvs7/must-gather-vr6mx" podUID="62afdf43-7cc5-4d53-aff8-2fd18fbfd493" containerName="copy" containerID="cri-o://7d2e1870707eb567a3ae7d6fc16a9352a0ad811fd3af2c67ad9db194e0056b53" gracePeriod=2 Jan 27 17:55:40 crc kubenswrapper[4772]: I0127 17:55:40.043516 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7dvs7/must-gather-vr6mx"] Jan 27 17:55:40 crc kubenswrapper[4772]: I0127 17:55:40.290945 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7dvs7_must-gather-vr6mx_62afdf43-7cc5-4d53-aff8-2fd18fbfd493/copy/0.log" Jan 27 17:55:40 crc kubenswrapper[4772]: I0127 17:55:40.291745 4772 generic.go:334] "Generic (PLEG): container finished" podID="62afdf43-7cc5-4d53-aff8-2fd18fbfd493" containerID="7d2e1870707eb567a3ae7d6fc16a9352a0ad811fd3af2c67ad9db194e0056b53" exitCode=143 Jan 27 17:55:40 crc kubenswrapper[4772]: I0127 17:55:40.467800 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7dvs7_must-gather-vr6mx_62afdf43-7cc5-4d53-aff8-2fd18fbfd493/copy/0.log" Jan 27 17:55:40 crc kubenswrapper[4772]: I0127 17:55:40.468286 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7dvs7/must-gather-vr6mx" Jan 27 17:55:40 crc kubenswrapper[4772]: I0127 17:55:40.586039 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/62afdf43-7cc5-4d53-aff8-2fd18fbfd493-must-gather-output\") pod \"62afdf43-7cc5-4d53-aff8-2fd18fbfd493\" (UID: \"62afdf43-7cc5-4d53-aff8-2fd18fbfd493\") " Jan 27 17:55:40 crc kubenswrapper[4772]: I0127 17:55:40.586300 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwwmf\" (UniqueName: \"kubernetes.io/projected/62afdf43-7cc5-4d53-aff8-2fd18fbfd493-kube-api-access-vwwmf\") pod \"62afdf43-7cc5-4d53-aff8-2fd18fbfd493\" (UID: \"62afdf43-7cc5-4d53-aff8-2fd18fbfd493\") " Jan 27 17:55:40 crc kubenswrapper[4772]: I0127 17:55:40.600469 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62afdf43-7cc5-4d53-aff8-2fd18fbfd493-kube-api-access-vwwmf" (OuterVolumeSpecName: "kube-api-access-vwwmf") pod "62afdf43-7cc5-4d53-aff8-2fd18fbfd493" (UID: "62afdf43-7cc5-4d53-aff8-2fd18fbfd493"). InnerVolumeSpecName "kube-api-access-vwwmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:55:40 crc kubenswrapper[4772]: I0127 17:55:40.691665 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwwmf\" (UniqueName: \"kubernetes.io/projected/62afdf43-7cc5-4d53-aff8-2fd18fbfd493-kube-api-access-vwwmf\") on node \"crc\" DevicePath \"\"" Jan 27 17:55:40 crc kubenswrapper[4772]: I0127 17:55:40.754134 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62afdf43-7cc5-4d53-aff8-2fd18fbfd493-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "62afdf43-7cc5-4d53-aff8-2fd18fbfd493" (UID: "62afdf43-7cc5-4d53-aff8-2fd18fbfd493"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:55:40 crc kubenswrapper[4772]: I0127 17:55:40.795331 4772 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/62afdf43-7cc5-4d53-aff8-2fd18fbfd493-must-gather-output\") on node \"crc\" DevicePath \"\"" Jan 27 17:55:41 crc kubenswrapper[4772]: I0127 17:55:41.313068 4772 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7dvs7_must-gather-vr6mx_62afdf43-7cc5-4d53-aff8-2fd18fbfd493/copy/0.log" Jan 27 17:55:41 crc kubenswrapper[4772]: I0127 17:55:41.313803 4772 scope.go:117] "RemoveContainer" containerID="7d2e1870707eb567a3ae7d6fc16a9352a0ad811fd3af2c67ad9db194e0056b53" Jan 27 17:55:41 crc kubenswrapper[4772]: I0127 17:55:41.313982 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7dvs7/must-gather-vr6mx" Jan 27 17:55:41 crc kubenswrapper[4772]: I0127 17:55:41.339797 4772 scope.go:117] "RemoveContainer" containerID="daea92a47f41c2b8eac7ad3a9eb9829829dd28a915856ca1714eacc250cb2602" Jan 27 17:55:42 crc kubenswrapper[4772]: I0127 17:55:42.683069 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62afdf43-7cc5-4d53-aff8-2fd18fbfd493" path="/var/lib/kubelet/pods/62afdf43-7cc5-4d53-aff8-2fd18fbfd493/volumes" Jan 27 17:55:51 crc kubenswrapper[4772]: I0127 17:55:51.663196 4772 scope.go:117] "RemoveContainer" containerID="b406a942c3fb69b474f2dc48f4fd84a565681c5c0e15f723d9fd971770b5e025" Jan 27 17:55:52 crc kubenswrapper[4772]: I0127 17:55:52.422952 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" event={"ID":"67794a44-d793-4fd7-9e54-e40437f67c0b","Type":"ContainerStarted","Data":"27602b2436534c6f3f068e3bb084a37c973a59402b8f43ca60850070bae7c8e1"} Jan 27 17:57:35 crc kubenswrapper[4772]: I0127 17:57:35.791972 4772 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hqpxg"] Jan 27 17:57:35 crc kubenswrapper[4772]: E0127 17:57:35.793257 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62afdf43-7cc5-4d53-aff8-2fd18fbfd493" containerName="gather" Jan 27 17:57:35 crc kubenswrapper[4772]: I0127 17:57:35.793283 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="62afdf43-7cc5-4d53-aff8-2fd18fbfd493" containerName="gather" Jan 27 17:57:35 crc kubenswrapper[4772]: E0127 17:57:35.793317 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84e9f92a-7805-4587-9826-afda667e7ee1" containerName="extract-utilities" Jan 27 17:57:35 crc kubenswrapper[4772]: I0127 17:57:35.793330 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="84e9f92a-7805-4587-9826-afda667e7ee1" containerName="extract-utilities" Jan 27 17:57:35 crc kubenswrapper[4772]: E0127 17:57:35.793367 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84e9f92a-7805-4587-9826-afda667e7ee1" containerName="extract-content" Jan 27 17:57:35 crc kubenswrapper[4772]: I0127 17:57:35.793380 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="84e9f92a-7805-4587-9826-afda667e7ee1" containerName="extract-content" Jan 27 17:57:35 crc kubenswrapper[4772]: E0127 17:57:35.793410 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84e9f92a-7805-4587-9826-afda667e7ee1" containerName="registry-server" Jan 27 17:57:35 crc kubenswrapper[4772]: I0127 17:57:35.793422 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="84e9f92a-7805-4587-9826-afda667e7ee1" containerName="registry-server" Jan 27 17:57:35 crc kubenswrapper[4772]: E0127 17:57:35.793451 4772 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62afdf43-7cc5-4d53-aff8-2fd18fbfd493" containerName="copy" Jan 27 17:57:35 crc kubenswrapper[4772]: I0127 17:57:35.793463 4772 state_mem.go:107] "Deleted CPUSet assignment" podUID="62afdf43-7cc5-4d53-aff8-2fd18fbfd493" containerName="copy" Jan 27 17:57:35 crc kubenswrapper[4772]: I0127 17:57:35.793774 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="84e9f92a-7805-4587-9826-afda667e7ee1" containerName="registry-server" Jan 27 17:57:35 crc kubenswrapper[4772]: I0127 17:57:35.793811 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="62afdf43-7cc5-4d53-aff8-2fd18fbfd493" containerName="copy" Jan 27 17:57:35 crc kubenswrapper[4772]: I0127 17:57:35.793849 4772 memory_manager.go:354] "RemoveStaleState removing state" podUID="62afdf43-7cc5-4d53-aff8-2fd18fbfd493" containerName="gather" Jan 27 17:57:35 crc kubenswrapper[4772]: I0127 17:57:35.796439 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hqpxg" Jan 27 17:57:35 crc kubenswrapper[4772]: I0127 17:57:35.824540 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hqpxg"] Jan 27 17:57:35 crc kubenswrapper[4772]: I0127 17:57:35.965913 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27mzs\" (UniqueName: \"kubernetes.io/projected/a29fc4ba-71f0-482a-8d16-888aff51e921-kube-api-access-27mzs\") pod \"certified-operators-hqpxg\" (UID: \"a29fc4ba-71f0-482a-8d16-888aff51e921\") " pod="openshift-marketplace/certified-operators-hqpxg" Jan 27 17:57:35 crc kubenswrapper[4772]: I0127 17:57:35.965994 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a29fc4ba-71f0-482a-8d16-888aff51e921-catalog-content\") pod \"certified-operators-hqpxg\" (UID: \"a29fc4ba-71f0-482a-8d16-888aff51e921\") " pod="openshift-marketplace/certified-operators-hqpxg" Jan 27 17:57:35 crc kubenswrapper[4772]: I0127 17:57:35.966012 4772 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a29fc4ba-71f0-482a-8d16-888aff51e921-utilities\") pod \"certified-operators-hqpxg\" (UID: \"a29fc4ba-71f0-482a-8d16-888aff51e921\") " pod="openshift-marketplace/certified-operators-hqpxg" Jan 27 17:57:36 crc kubenswrapper[4772]: I0127 17:57:36.067246 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27mzs\" (UniqueName: \"kubernetes.io/projected/a29fc4ba-71f0-482a-8d16-888aff51e921-kube-api-access-27mzs\") pod \"certified-operators-hqpxg\" (UID: \"a29fc4ba-71f0-482a-8d16-888aff51e921\") " pod="openshift-marketplace/certified-operators-hqpxg" Jan 27 17:57:36 crc kubenswrapper[4772]: I0127 17:57:36.067332 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a29fc4ba-71f0-482a-8d16-888aff51e921-catalog-content\") pod \"certified-operators-hqpxg\" (UID: \"a29fc4ba-71f0-482a-8d16-888aff51e921\") " pod="openshift-marketplace/certified-operators-hqpxg" Jan 27 17:57:36 crc kubenswrapper[4772]: I0127 17:57:36.067360 4772 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a29fc4ba-71f0-482a-8d16-888aff51e921-utilities\") pod \"certified-operators-hqpxg\" (UID: \"a29fc4ba-71f0-482a-8d16-888aff51e921\") " pod="openshift-marketplace/certified-operators-hqpxg" Jan 27 17:57:36 crc kubenswrapper[4772]: I0127 17:57:36.067919 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a29fc4ba-71f0-482a-8d16-888aff51e921-utilities\") pod \"certified-operators-hqpxg\" (UID: \"a29fc4ba-71f0-482a-8d16-888aff51e921\") " pod="openshift-marketplace/certified-operators-hqpxg" Jan 27 17:57:36 crc kubenswrapper[4772]: I0127 17:57:36.068470 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a29fc4ba-71f0-482a-8d16-888aff51e921-catalog-content\") pod \"certified-operators-hqpxg\" (UID: \"a29fc4ba-71f0-482a-8d16-888aff51e921\") " pod="openshift-marketplace/certified-operators-hqpxg" Jan 27 17:57:36 crc kubenswrapper[4772]: I0127 17:57:36.098062 4772 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27mzs\" (UniqueName: \"kubernetes.io/projected/a29fc4ba-71f0-482a-8d16-888aff51e921-kube-api-access-27mzs\") pod \"certified-operators-hqpxg\" (UID: \"a29fc4ba-71f0-482a-8d16-888aff51e921\") " pod="openshift-marketplace/certified-operators-hqpxg" Jan 27 17:57:36 crc kubenswrapper[4772]: I0127 17:57:36.166736 4772 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hqpxg" Jan 27 17:57:36 crc kubenswrapper[4772]: I0127 17:57:36.771506 4772 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hqpxg"] Jan 27 17:57:37 crc kubenswrapper[4772]: I0127 17:57:37.518405 4772 generic.go:334] "Generic (PLEG): container finished" podID="a29fc4ba-71f0-482a-8d16-888aff51e921" containerID="3cf46495e6c295b8311bd4db838af9d233d54cd2e679f422d69d5dc99e36eaea" exitCode=0 Jan 27 17:57:37 crc kubenswrapper[4772]: I0127 17:57:37.518827 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqpxg" event={"ID":"a29fc4ba-71f0-482a-8d16-888aff51e921","Type":"ContainerDied","Data":"3cf46495e6c295b8311bd4db838af9d233d54cd2e679f422d69d5dc99e36eaea"} Jan 27 17:57:37 crc kubenswrapper[4772]: I0127 17:57:37.518861 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqpxg" event={"ID":"a29fc4ba-71f0-482a-8d16-888aff51e921","Type":"ContainerStarted","Data":"d0d879401429a76bebba81c1b4334ca4f4e0c68e48881234b455c1831c80224e"} Jan 27 17:57:38 crc kubenswrapper[4772]: I0127 17:57:38.528724 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqpxg" event={"ID":"a29fc4ba-71f0-482a-8d16-888aff51e921","Type":"ContainerStarted","Data":"1dc2f6290a8ba6dcfac8a909eaee7377853a16025b23d62b1fd7d2d2249b6d7f"} Jan 27 17:57:39 crc kubenswrapper[4772]: I0127 17:57:39.553668 4772 generic.go:334] "Generic (PLEG): container finished" podID="a29fc4ba-71f0-482a-8d16-888aff51e921" containerID="1dc2f6290a8ba6dcfac8a909eaee7377853a16025b23d62b1fd7d2d2249b6d7f" exitCode=0 Jan 27 17:57:39 crc kubenswrapper[4772]: I0127 17:57:39.554151 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqpxg" event={"ID":"a29fc4ba-71f0-482a-8d16-888aff51e921","Type":"ContainerDied","Data":"1dc2f6290a8ba6dcfac8a909eaee7377853a16025b23d62b1fd7d2d2249b6d7f"} Jan 27 17:57:40 crc kubenswrapper[4772]: I0127 17:57:40.567071 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqpxg" event={"ID":"a29fc4ba-71f0-482a-8d16-888aff51e921","Type":"ContainerStarted","Data":"74a5bdd2c327de4c131d638a5bfc55206bf72f2cb957ad4b1edb4a47508787d7"} Jan 27 17:57:40 crc kubenswrapper[4772]: I0127 17:57:40.589302 4772 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hqpxg" podStartSLOduration=3.109391427 podStartE2EDuration="5.589281316s" podCreationTimestamp="2026-01-27 17:57:35 +0000 UTC" firstStartedPulling="2026-01-27 17:57:37.520957007 +0000 UTC m=+10243.501566105" lastFinishedPulling="2026-01-27 17:57:40.000846856 +0000 UTC m=+10245.981455994" observedRunningTime="2026-01-27 17:57:40.587553086 +0000 UTC m=+10246.568162234" watchObservedRunningTime="2026-01-27 17:57:40.589281316 +0000 UTC m=+10246.569890414" Jan 27 17:57:46 crc kubenswrapper[4772]: I0127 17:57:46.167524 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hqpxg" Jan 27 17:57:46 crc kubenswrapper[4772]: I0127 17:57:46.168344 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hqpxg" Jan 27 17:57:46 crc kubenswrapper[4772]: I0127 17:57:46.222991 4772 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hqpxg" Jan 27 17:57:46 crc kubenswrapper[4772]: I0127 17:57:46.693931 4772 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hqpxg" Jan 27 17:57:46 crc kubenswrapper[4772]: I0127 17:57:46.747996 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hqpxg"] Jan 27 17:57:48 crc kubenswrapper[4772]: I0127 17:57:48.646722 4772 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hqpxg" podUID="a29fc4ba-71f0-482a-8d16-888aff51e921" containerName="registry-server" containerID="cri-o://74a5bdd2c327de4c131d638a5bfc55206bf72f2cb957ad4b1edb4a47508787d7" gracePeriod=2 Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.147869 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hqpxg" Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.255589 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a29fc4ba-71f0-482a-8d16-888aff51e921-catalog-content\") pod \"a29fc4ba-71f0-482a-8d16-888aff51e921\" (UID: \"a29fc4ba-71f0-482a-8d16-888aff51e921\") " Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.255805 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27mzs\" (UniqueName: \"kubernetes.io/projected/a29fc4ba-71f0-482a-8d16-888aff51e921-kube-api-access-27mzs\") pod \"a29fc4ba-71f0-482a-8d16-888aff51e921\" (UID: \"a29fc4ba-71f0-482a-8d16-888aff51e921\") " Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.255971 4772 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a29fc4ba-71f0-482a-8d16-888aff51e921-utilities\") pod \"a29fc4ba-71f0-482a-8d16-888aff51e921\" (UID: \"a29fc4ba-71f0-482a-8d16-888aff51e921\") " Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.257466 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a29fc4ba-71f0-482a-8d16-888aff51e921-utilities" (OuterVolumeSpecName: "utilities") pod "a29fc4ba-71f0-482a-8d16-888aff51e921" (UID: "a29fc4ba-71f0-482a-8d16-888aff51e921"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.258439 4772 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a29fc4ba-71f0-482a-8d16-888aff51e921-utilities\") on node \"crc\" DevicePath \"\"" Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.261872 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a29fc4ba-71f0-482a-8d16-888aff51e921-kube-api-access-27mzs" (OuterVolumeSpecName: "kube-api-access-27mzs") pod "a29fc4ba-71f0-482a-8d16-888aff51e921" (UID: "a29fc4ba-71f0-482a-8d16-888aff51e921"). InnerVolumeSpecName "kube-api-access-27mzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.311282 4772 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a29fc4ba-71f0-482a-8d16-888aff51e921-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a29fc4ba-71f0-482a-8d16-888aff51e921" (UID: "a29fc4ba-71f0-482a-8d16-888aff51e921"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.360700 4772 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a29fc4ba-71f0-482a-8d16-888aff51e921-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.360733 4772 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27mzs\" (UniqueName: \"kubernetes.io/projected/a29fc4ba-71f0-482a-8d16-888aff51e921-kube-api-access-27mzs\") on node \"crc\" DevicePath \"\"" Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.666804 4772 generic.go:334] "Generic (PLEG): container finished" podID="a29fc4ba-71f0-482a-8d16-888aff51e921" containerID="74a5bdd2c327de4c131d638a5bfc55206bf72f2cb957ad4b1edb4a47508787d7" exitCode=0 Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.666840 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqpxg" event={"ID":"a29fc4ba-71f0-482a-8d16-888aff51e921","Type":"ContainerDied","Data":"74a5bdd2c327de4c131d638a5bfc55206bf72f2cb957ad4b1edb4a47508787d7"} Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.666856 4772 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hqpxg" Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.666871 4772 scope.go:117] "RemoveContainer" containerID="74a5bdd2c327de4c131d638a5bfc55206bf72f2cb957ad4b1edb4a47508787d7" Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.666861 4772 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqpxg" event={"ID":"a29fc4ba-71f0-482a-8d16-888aff51e921","Type":"ContainerDied","Data":"d0d879401429a76bebba81c1b4334ca4f4e0c68e48881234b455c1831c80224e"} Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.696488 4772 scope.go:117] "RemoveContainer" containerID="1dc2f6290a8ba6dcfac8a909eaee7377853a16025b23d62b1fd7d2d2249b6d7f" Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.705940 4772 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hqpxg"] Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.715810 4772 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hqpxg"] Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.726600 4772 scope.go:117] "RemoveContainer" containerID="3cf46495e6c295b8311bd4db838af9d233d54cd2e679f422d69d5dc99e36eaea" Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.786673 4772 scope.go:117] "RemoveContainer" containerID="74a5bdd2c327de4c131d638a5bfc55206bf72f2cb957ad4b1edb4a47508787d7" Jan 27 17:57:49 crc kubenswrapper[4772]: E0127 17:57:49.786972 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74a5bdd2c327de4c131d638a5bfc55206bf72f2cb957ad4b1edb4a47508787d7\": container with ID starting with 74a5bdd2c327de4c131d638a5bfc55206bf72f2cb957ad4b1edb4a47508787d7 not found: ID does not exist" containerID="74a5bdd2c327de4c131d638a5bfc55206bf72f2cb957ad4b1edb4a47508787d7" Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.787012 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74a5bdd2c327de4c131d638a5bfc55206bf72f2cb957ad4b1edb4a47508787d7"} err="failed to get container status \"74a5bdd2c327de4c131d638a5bfc55206bf72f2cb957ad4b1edb4a47508787d7\": rpc error: code = NotFound desc = could not find container \"74a5bdd2c327de4c131d638a5bfc55206bf72f2cb957ad4b1edb4a47508787d7\": container with ID starting with 74a5bdd2c327de4c131d638a5bfc55206bf72f2cb957ad4b1edb4a47508787d7 not found: ID does not exist" Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.787033 4772 scope.go:117] "RemoveContainer" containerID="1dc2f6290a8ba6dcfac8a909eaee7377853a16025b23d62b1fd7d2d2249b6d7f" Jan 27 17:57:49 crc kubenswrapper[4772]: E0127 17:57:49.787217 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dc2f6290a8ba6dcfac8a909eaee7377853a16025b23d62b1fd7d2d2249b6d7f\": container with ID starting with 1dc2f6290a8ba6dcfac8a909eaee7377853a16025b23d62b1fd7d2d2249b6d7f not found: ID does not exist" containerID="1dc2f6290a8ba6dcfac8a909eaee7377853a16025b23d62b1fd7d2d2249b6d7f" Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.787242 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dc2f6290a8ba6dcfac8a909eaee7377853a16025b23d62b1fd7d2d2249b6d7f"} err="failed to get container status \"1dc2f6290a8ba6dcfac8a909eaee7377853a16025b23d62b1fd7d2d2249b6d7f\": rpc error: code = NotFound desc = could not find container \"1dc2f6290a8ba6dcfac8a909eaee7377853a16025b23d62b1fd7d2d2249b6d7f\": container with ID starting with 1dc2f6290a8ba6dcfac8a909eaee7377853a16025b23d62b1fd7d2d2249b6d7f not found: ID does not exist" Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.787257 4772 scope.go:117] "RemoveContainer" containerID="3cf46495e6c295b8311bd4db838af9d233d54cd2e679f422d69d5dc99e36eaea" Jan 27 17:57:49 crc kubenswrapper[4772]: E0127 17:57:49.787435 4772 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cf46495e6c295b8311bd4db838af9d233d54cd2e679f422d69d5dc99e36eaea\": container with ID starting with 3cf46495e6c295b8311bd4db838af9d233d54cd2e679f422d69d5dc99e36eaea not found: ID does not exist" containerID="3cf46495e6c295b8311bd4db838af9d233d54cd2e679f422d69d5dc99e36eaea" Jan 27 17:57:49 crc kubenswrapper[4772]: I0127 17:57:49.787455 4772 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cf46495e6c295b8311bd4db838af9d233d54cd2e679f422d69d5dc99e36eaea"} err="failed to get container status \"3cf46495e6c295b8311bd4db838af9d233d54cd2e679f422d69d5dc99e36eaea\": rpc error: code = NotFound desc = could not find container \"3cf46495e6c295b8311bd4db838af9d233d54cd2e679f422d69d5dc99e36eaea\": container with ID starting with 3cf46495e6c295b8311bd4db838af9d233d54cd2e679f422d69d5dc99e36eaea not found: ID does not exist" Jan 27 17:57:50 crc kubenswrapper[4772]: I0127 17:57:50.683578 4772 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a29fc4ba-71f0-482a-8d16-888aff51e921" path="/var/lib/kubelet/pods/a29fc4ba-71f0-482a-8d16-888aff51e921/volumes" Jan 27 17:58:12 crc kubenswrapper[4772]: I0127 17:58:12.058833 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:58:12 crc kubenswrapper[4772]: I0127 17:58:12.059676 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 27 17:58:42 crc kubenswrapper[4772]: I0127 17:58:42.058657 4772 patch_prober.go:28] interesting pod/machine-config-daemon-4hwxn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 27 17:58:42 crc kubenswrapper[4772]: I0127 17:58:42.075500 4772 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4hwxn" podUID="67794a44-d793-4fd7-9e54-e40437f67c0b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515136176347024462 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015136176350017371 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015136151671016513 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015136151671015463 5ustar corecore